I'm studying from a Deep Learning book (Ian Goodfellow et al). At page 256 the text explains that, considering a set of $k$ regression models, each produces an error $ϵ_i$ for every example, drawn from a zero-mean multivariate normal distribution with variances $E[ϵ_i^2]=v$ and covariances $E[ϵ_iϵ_j]=c$. Then the error made by the average prediction of all the ensemble models is $\frac{1}{k} \sum_{i=1}^k \epsilon_i$. The expected squared error of the ensemble predictor is:
$ \begin{align*} E\left[\left(\frac{1}{k} \sum_{i=1}^k \epsilon_i\right)^2\right] &= \frac{1}{k^2} E\left[\sum_{i=1}^k (\epsilon_i)^2 + \sum_{i \neq j} \epsilon_i \epsilon_j\right] \ &= \frac{1}{k} v + \frac{k-1}{k} c. \end{align*} $$
So, the question is: How do we obtain the first quadratic decomposition? I'm having trouble with it. Could someone explain?
Thank you so much!