My question is, if you take a random variable $X$ with an arbitrary distribution, which has a known and well defined mean & variance, then why does the following interval contain $\approx 95\% $ of the distribution?
$$ \left[\mu - 1.959 \times \sqrt{\sigma^2}, \quad \mu + 1.959 \times \sqrt{\sigma^2}\right] $$
Like I understand why this is true for normally distributed variables as $\pm 1.959$ is just the magic number that maps to the 2.5th and 97.5th percentile of the standard normal distribution, but why does this approximately still hold for non-normally distributed variables?
Some R-code examples:
##### Beta Distribution
a <- 3
b <- 40
# Formulas from wiki https://en.wikipedia.org/wiki/Beta_distribution
mu <- a / (a+b)
sigma2 <- a*b / (((a+b)^2) * ( a + b + 1))
bounds <- c(
mu - 1.959 * sqrt(sigma2),
mu + 1.959 * sqrt(sigma2)
)
pbeta(bounds[[2]], a , b) - pbeta(bounds[[1]], a , b)
#> 0.9543413
##### Weibull Distribution
k <- 0.9
lambda <- 10
# Formulas from wiki https://en.wikipedia.org/wiki/Weibull_distribution
mu <- lambda * gamma(1 + 1/k)
sigma2 <- lambda^2 * (gamma(1 + 2/k) - gamma(1 + 1/k)^2)
bounds <- c(
mu - 1.959 * sqrt(sigma2),
mu + 1.959 * sqrt(sigma2)
)
pweibull(bounds[[2]], k , lambda) - pweibull(bounds[[1]], k , lambda)
#> 0.9484728
My guess is that this is just a property of the definition of the variance function itself however nothing obvious standards out to me as to why this holds, particularly on heavily skewed distributions.