Given n samples from a normally-distributed variable X, we estimate variance as $s^2=\frac{1}{n-1}\sum{(x_i - \bar{x})^2}$. We can also get a confidence interval for such a variance estimate as: $$CI_\alpha = [\frac{(n-1)s^2}{\chi^2_{\frac{\alpha}{2},n-1}},\frac{(n-1)s^2}{\chi^2_{1-\frac{\alpha}{2},n-1}}]$$
Now suppose we have two variables $X \sim N(0,\sigma_x), Y \sim N(0,\sigma_y)$, and we want to test whether $\sigma_x = \sigma_y$ for some level of significance $\alpha$. In this case it is common to use the F-distribution to perform an F-test on $\frac{s_x^2}{s_y^2}$.
However, we can also look at the confidence intervals for the two variance estimates.
If have found that it is often the case that if the confidence intervals overlap then the F-test (for the same $\alpha$) does not reject the null hypothesis (which is that $\sigma_x = \sigma_y$), and if the confidence intervals do not overlap then the F-test does reject the null hypothesis. But this is not always the case. Here is a sample set where, for $\alpha$=0.1, the confidence intervals overlap but the F-test rejects the null hypothesis:
X Y
13 -24
-10 -24
-12 7
10 -51
2 28
-10 -4
-6 2
0 5
-1 -32
-21 -10
Variance X = 108, 90% CI = [57, 292]
Variance Y = 522, 90% CI = [277, 1412]
F-test = .028
Squinting at the math, it looks like the F-test is considering a ratio where the confidence intervals are considering differences. It's unclear to me why one is better than the other for purposes of this test.
Question: For the purposes of testing the hypothesis $\sigma_x = \sigma_y$, is it better to use the F-test than to use confidence intervals as described above? Why?