7
$\begingroup$

I am studying Quantum Information now, and I need to understand the entropy of a quantum system. But before I go there, I need to understand Shannon Entropy which is defined as :

$H(X) = -\sum_{i=1}^{n} {p(x_i) \log_2{p(x_i)}} $

where $X$ is a discrete random variable with possible outcomes $x_{1},...,x_{n}$ which occur with probability $p_{1},...,p_{n}$. This is entropy that works in information theory, but we know that entropy is already defined back way in thermodynamics by Clausius as :

$$dS = \frac{\delta Q}{T}$$

Then, in statistical physics, entropy is defined by Boltzmann as :

$S=k_B \ln{\Omega}$

where $\Omega$ is the number of microstates of a system. How can I derive the Shannon entropy from these thermodynamics and statistical physics entropy definitions?

$\endgroup$
2

5 Answers 5

5
$\begingroup$

These are not the same.

Shannon entropy (Information entropy), $H_\alpha=-\sum_i p_i\log_\alpha p_i$ applies to any system with specified probabilities $p_i$.

Boltzmann entropy, defined via the famous $S=k\log\Omega$ implies that the system occupies all the accessible states with equal probability, $p_i=1/\Omega$ (this is a particular case of the Information entropy, as can be seen by plugging $p_i=1/\Omega$ into the Shannon formula, taking natural logarithm base, and discarding customary dimensional coefficient).

Gibbs entropy, defined via the Clausius inequality, $dS\geq \delta Q/T_{env}$, is defined empirically, as a quantity that always monotonuously increases, and thus makes thermodynamic processes irreversible.

Furthermore, Boltzmann entropy and Gibbs entropy can be shown to be equivalent, reflecting the equivalence between the microscopic statistical physics and the phenomenological thermodynamics.

Finally, Let me first point out that entropy may mean different things. As Jaynes, in his article The minimum entropy production principle claims that there are six different types of entropy with somewhat different meaning.

Remark:
There is some disagreement about what is called Gibbs entropy, as Gibbs actually introduced two - one along the lines of Clausius, and another more similar to Boltzmann entropy. These are sometimes referred to as Gibbs I and Gibbs II. For more ways to introduce entropy see this answer to Is information entropy the same as thermodynamic entropy?.

$\endgroup$
6
  • $\begingroup$ I do not understand why you call Gibbs entropy the Thermodynamic expression. Usually, people use the name Gibbs entropy for the same formula as Shannon's entropy but limited to the case of the equilibrium ensemble probability distribution. $\endgroup$ Commented Dec 7, 2022 at 6:36
  • $\begingroup$ @GiorgioP sometimes these are referred to as Gibbs I and Gibbs II. There is now a more complete summary here. Anyhow the point is that there is entropy defined phenomenologically (in thermodynamics), there is entropy in statistical physics, and there is information entropy. Was your downvote only due to the disagreement in terminology or do you sincerely think that this answer is not useful as per SE policies? $\endgroup$
    – Roger V.
    Commented Dec 7, 2022 at 8:29
  • $\begingroup$ I never downvoted an answer for terminology reasons. However, I think that your answer does not provide a clear answer to the question (derivation of Sh. entropy from Clausius). Moreover, it contains some inaccuracy (the equivalence between Boltzmann entropy and Clausius entropy is not unconditionally valid). $\endgroup$ Commented Dec 7, 2022 at 10:54
  • $\begingroup$ @GiorgioP the point of the answer is that Shannon entropy cannot be derived from Clausius. $\endgroup$
    – Roger V.
    Commented Dec 7, 2022 at 10:55
  • $\begingroup$ I think it should be stated more clearly than saying they are not the same. Also Shannon and Boltzmann's entropies are not the same, but one can obtain the latter from the first. $\endgroup$ Commented Dec 7, 2022 at 11:51
5
$\begingroup$

Yes, you can.

I'm going to show that the expression of the Shannon entropy can be deduced from statistical and thermodynamic relations.

We know that the entropy defined by $\mathrm{d}S=\delta{Q}/T$ can be related to thermodynamic potentials. The formula we will use here is $$ F=U-TS, $$ where $U$ is the energy and $F$ is the Helmholtz function (sometimes maybe called the free engergy?).

And in statistical mechanics, $F$ is related to the partition function $Q$ for the canonical ensemble: $$ F=-kT\ln Q, $$ where $Q=\sum_{r}\mathrm{e}^{-\beta E_r}$ is the partition function (of course you know that $\beta = \frac{1}{kT}$).

The probability of any microstate $r$ is given by $$ P_r=\frac{\mathrm{e}^{-\beta E_r}}{Q}. $$

OK. Here it comes. Take its logarithm and average it, and we have Eq. (3.3.13) in (Pathria & Beale 2021) $$ \langle \ln P_r \rangle=-\ln Q-\beta \langle E_r\rangle=\beta(F-U) = -\frac{S}{k}. $$ The average formula is just $$ \langle \ln P_r \rangle = \sum_{r} P_r \ln P_r. $$ At last, we obtain the Shannon entropy (forget about the Boltzmann constant $k$) $$ S = -\sum_{r} P_r \ln P_r. $$


Ref.

R. K. Pathria, Paul D. Beale. Statistical Mechanics 4th ed., Academic Press, 2021.

$\endgroup$
3
  • $\begingroup$ I don't understand why this is not the chosen answer. This is an excellent derivation, thank you $\endgroup$ Commented Jul 8, 2022 at 18:38
  • 1
    $\begingroup$ @bananenheld This answer is possibly not chosen because OP seems to be asking about the microcanonical ensemble, but this answer works in the canonical ensemble. I.e., OP asks about why $S=k\log(\Omega),$ which only holds when the microstate probabilities are all equal (all equal to $1/\Omega$), but in the canonical ensemble the probabilities are given by $e^{-\beta E_i}/Q$. $\endgroup$
    – hft
    Commented Dec 7, 2022 at 1:19
  • $\begingroup$ @bananenheld The problem with this answer is that it allows deriving a formula looking like Shannon's formula but is limited to the case of the probability of equilibrium ensembles in Statistical Mechanics. The derivation does not allow the use of the formula in other contexts, for example, the case of signal transmission that was Shannon's starting point. $\endgroup$ Commented Dec 7, 2022 at 6:32
3
$\begingroup$

You can't. It is not possible to derive a more general formula from a less general one. Of course, one can find hints for the generalization, but the validity of the generalization has to be proved independently.

The relationship between the formulas is the following: Shannon's formula is more general (it applies to every probability distribution, even non-equilibrium ones and even if there is no energy underlying the probabilities).

Statistical mechanics entropies (different in different ensembles) are the special case of Shannon's formula for the case where the probabilities have the correct values for each ensemble.

Clausius formula has a connection with statistical mechanics entropies only at the so-called thermodynamic limit. I.e. in the limit of a very large system.

$\endgroup$
9
  • 1
    $\begingroup$ It is not a more general formula, it is a formula for different thing. Shannon entropy, or information-theoretic entropy, is a different concept altogether from the Clausius entropy. Clausius entropy can be sometimes calculated as value of the Shannon entropy; but this does not make the latter a more general formula for the same thing. $\endgroup$ Commented May 23, 2022 at 19:33
  • $\begingroup$ @JánLalinský I do not agree with your point of view. Shannon entropy can be defined for every probability distribution, not only the one corresponding to an equilibrium ensemble. But if used with the equilibrium probability distribution of a Statistical Mechanics ensemble it provides exactly the Statistical Mechanics expression for the entropy. I would call this situation a particularization of a more general formula to a more specific case. In a similar way, the Statistical Mechanics entropy, in general, does not coincide with the thermodynamics entropy. However, with additional conditions... $\endgroup$ Commented May 23, 2022 at 21:57
  • $\begingroup$ ... i.e. with the thermodynamic limit, it coincides with the thermodynamic entropy. What else, to be allowed to speak about a transition from more general to more specific concepts? $\endgroup$ Commented May 23, 2022 at 21:59
  • 1
    $\begingroup$ > "...Statistical Mechanics...I would call this situation a particularization of a more general formula to a more specific case." -- that is true, but that generalization consists of going from Shannon entropy of special probability distribution to Shannon entropy of any probability distribution. Not from going from Clausius entropy to Shannon entropy. Clausius entropy isn't logically dependent on statistical mechanics or probability distributions. $\endgroup$ Commented May 23, 2022 at 22:07
  • 1
    $\begingroup$ If we defined entropy for a body out of equilibrium as function of macroscopic description quantities such as energy density field and mass density field, using SM or not, then that would be generalization of the Clausius entropy, producing a concept of the same kind for more general class of systems. But Shannon entropy is not a generalization in this sense; it is a different concept that turned out to be relevant to Clausius entropy, but it is not the same kind of concept. $\endgroup$ Commented May 23, 2022 at 22:42
1
$\begingroup$

A better approach would be to use the Shannon Entropy to derive Gibbs entropy: $S=−k\cdot∑p_n \cdot \ln(p_n)$. The two equations are very similar and therefore it is much easier understand. From there it is easy to arrive at Boltzman entropy and finally Clausius.

$\endgroup$
0
$\begingroup$

You can derive the Gibbs entropy, which is simply the Shannon entropy multiplied by the Boltzmann constant, from the Boltzmann entropy and a multinomial distribution. This video proceeds with the derivation:

https://www.youtube.com/watch?v=h3xVAVcYfjk

The Boltzmann entropy is not a particular case of the Gibbs entropy in the demonstration. It's quite the other way around. Sure, the Boltzmann entropy has equal probabilities while they may differ in the Gibbs entropy. However, the probabilities refer to microstates of a canonical ensemble in the Boltzmann entropy, but to individual particle energy states in the Gibbs entropy.

$\endgroup$

Not the answer you're looking for? Browse other questions tagged or ask your own question.