Download Gibbs Distribution in Quantum Statistics

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Quantum state wikipedia , lookup

Molecular Hamiltonian wikipedia , lookup

Particle in a box wikipedia , lookup

Perturbation theory (quantum mechanics) wikipedia , lookup

Canonical quantization wikipedia , lookup

Theoretical and experimental justification for the Schrödinger equation wikipedia , lookup

Probability amplitude wikipedia , lookup

Renormalization group wikipedia , lookup

T-symmetry wikipedia , lookup

Transcript
Gibbs Distribution in Quantum Statistics
Quantum Mechanics is much more complicated than the Classical one. To fully characterize a state
of one particle in Classical Mechanics we just need to specify its radius vector and velocity, while
in Quantum Mechanics the state of the same system corresponds to a complex-valued function of
coordinates. One may expect thus that the Quantum Statistics is much more sophisticated than the
Classical one. And this is indeed the case if we mean the complexity of statistical calculations for some
strongly interacting systems. It is amazing, however, that the general statistical considerations leading
to Gibbs distribution—and subsequently to the derivation of all generic thermodynamic relations—are
much simpler in the quantum case.
The derivation of the quantum Gibbs distribution is not compulsory in this course. Below I
present it for those who really like mathematical rigor of theoretical physics. The others may safely
skip the proof and go directly to the final result, Eq. (33).
Derivation of quantum Gibbs distribution. We start with considering the simplest quantummechanical system—two-level system—and then generalize the analysis.
Two-level system. By definition, a two-level system is a system which wavefunction is a two-component
complex vector and, correspondingly, the Hamiltonian is a hermitian matrix 2×2. Each wavefunction
|ψi can be represented as a linear combination of two orthogonal eigenstates of the Hamiltonian
|ψi = C1 |1i + C2 |2i ,
H|1i = E1 |1i ,
H|2i = E2 |2i ,
(1)
(2)
where E1 and E2 are the energies of the two eigenstates. If the system is isolated and |ψi is its initial
state, then the time evolution is given by the formula
|ψ(t)i = C1 e−iE1 t/h̄ |1i + C2 e−iE2 t/h̄ |2i .
(3)
Suppose now that our system interacts with an equilibrium heat bath. Equilibrium heat bath is a
crucial notion of Statistical Mechanics. Basically, it is a sort of intuitively clear axiom saying that a
macroscopic system with an extremely large excited degrees of freedom admits statistical description
in the sense that in the limit of t → ∞—in practice, when t is much larger than the relaxation
time—the observables of the system can be treated as random numbers with some distributions.
By equilibrium heat bath we understand such a macroscopic system interacting with a particular
system of interest (two-level system in our case). We assume that the interaction with a heat bath is
infinitesimally weak. Otherwise the properties of our system would be significantly distorted by this
interaction. The effect of a weak interaction with the heat bath can be represented as some weak
time-dependent noise in the parameters C1 , C2 , E1 , and E2 . Let us trace the consequences of this
noise. To this end we consider some observable, A, which in our case is a hermitian matrix 2 × 2.
The quantum-mechanical expectation value of the observable A is given by
hψ(t)|A|ψ(t)i = |C1 |2 A11 + |C2 |2 A22 + C1 C2∗ ei(E2 −E1 )t/h̄ A21 + C2 C1∗ ei(E1 −E2 )t/h̄ A12 ,
(4)
A11 = h1|A|1i , A22 = h2|A|2i , A12 = h1|A|2i , A21 = h2|A|1i .
(5)
where
1
Whatever weak, the interaction with heat bath qualitatively changes the problem at large times: the
system forgets about its initial state. Hence, the only meaningful quantity is the average of the
expectation value (4) over large times. Using symbol hh. . .ii to denote this average we write
hhAii = hh|C1 |2 ii A11 + hh|C2 |2 ii A22 + hhC1 C2∗ ei(E2 −E1 )t/h̄ ii A12 + hhC2 C1∗ ei(E1 −E2 )t/h̄ ii A21 .
(6)
There is a crucial difference between the first two terms in (6) and the last two ones. The first two
are averages of non-negative numbers and at least one of them is nonzero. Indeed, if
w1 = hh|C1 |2 ii ,
w2 = hh|C2 |2 ii ,
(7)
then
w1 + w2 = 1 ,
(8)
since |C1 |2 + |C2 |2 ≡ 1 (wavefunction is always normalized to unity). The two last terms in (6) are
complex numbers which phases, ϕ, can take on any value in the interval [0, 2π] with equal probability
densities: The only possible reason for one phase to be more probable than another would be the
interaction with the heat bath, in which case the difference of probability densities for two phases,
W (ϕ1 )−W (ϕ2 ), would depend on the strength of the interaction. But we assume that the interaction
is arbitrarily weak and thus the distribution of the phases should be flat: W (ϕ) ≡ 1/2π. The averaging
over time will thus include averaging over all possible and equally probable phases. With this averaging
we have
Z
Z
heiϕ i =
2π
0
eiϕ W (ϕ) dϕ = (1/2π)
2π
0
eiϕ dϕ = 0 ,
(9)
and conclude that
hhC1 C2∗ ei(E2 −E1 )t/h̄ ii = 0 ,
hhC2 C1∗ ei(E1 −E2 )t/h̄ ii = 0 .
(10)
That is only diagonal terms survive time averaging:
hhAii = w1 A11 + w2 A22 .
(11)
In effect, the system lives only in its eigenstates, with probability wj (j = 1, 2) of finding it in the
j-th eigenstate. The question now is: What are these probabilities? We get the clue to the answer by
analyzing the degenerate case: E1 = E2 . In a degenerate two-level system, each state is an eigenstate.
This means that instead of the basis { |1i, |2i } we can use, say, { |ai, |bi }, where
√
√
|ai = (|1i + |2i)/ 2 ,
|bi = (|1i − |2i)/ 2 .
(12)
In terms of this new basis, Eq. (11) reads
hhAii = wa Aaa + wb Abb .
(13)
It is supposed to yield exactly the same value for hhAii, and we are going to check how it can be the
case. With (12) we find
Aaa ≡ ha|A|ai = (1/2)A11 + (1/2)A22 + (1/2)A12 + (1/2)A21 ,
(14)
Abb ≡ ha|A|ai = (1/2)A11 + (1/2)A22 − (1/2)A12 − (1/2)A21 .
(15)
Substituting it into (13) leads to
hhAii = (1/2)(wa + wb ) A11 + (1/2)(wa + wb ) A22 + (1/2)(wa − wb ) (A12 + A21 ) .
2
(16)
And this has to be identically equal to (11) for any matrix A. The only way to satisfy this requirement
is to have
wa = wb = w1 = w2 ,
(17)
and we arrive at the fundamental statement that the probabilities for the system to be found in
each of its eigenstates of the same energy are equal. This statement as well as the whole analysis
presented above apply to any system weakly interacting with a heat bath. One just need to expand
the wavefunction in terms of the eigenstates of the Hamiltonian and repeat all the derivation, from
(3) to (10). Eq. (10) now reads
∗ i(Em −En )t/h̄
hhCn Cm
e
ii = 0 ,
Eqs. (11) and (8) generalize to
hhAii =
X
X
if n 6= m .
wn Ann ,
(18)
(19)
n
wn = 1 ,
(20)
n
where the sum is over all the eigenstates. To prove that the probabilities of two states with equal
energies are equal we can actually utilize the derivation for two-level system, because for the two
given states of interest it is enough to consider operators A of a special form—having non-zero matrix
elements with only these two states.
Now we are ready to derive the explicit form of wn . We already know that wn can depend only
on the energy: wn ≡ f (En ), where f is some function.—This is nothing else than a mathematical
expression of the requirement that wn = wm , if En = Em . Next, we take into account the following
consideration. If we have two independent (non-interacting with each other, but interacting with the
heat bath) systems, A and B, we can formally treat them as a united system AB. The Hamiltonian
of this system is a sum of Hamiltonians:
HAB = HA + HB ,
(21)
and, correspondingly, the eigenstates of such systems are direct products of the eigenstates of HA
and HB :
|AB : n, mi = |A : ni |B : mi,
(22)
HA |A : ni = En(A) |A : ni ,
(23)
(B)
HB |B : mi = Em
|B : mi ,
(24)
(AB)
HAB |AB : n, mi = Enm
|AB : n, mi ,
(25)
(AB)
(B)
Enm
= En(A) + Em
.
(26)
where
The distribution function for the system AB should have the form
(AB)
(B)
wnm
= fAB (En(A) + Em
).
(AB)
On the other hand, the two systems are independent, and wnm
individual distributions for the two systems:
(27)
should be equal to the product of
(AB)
(B)
wnm
= wn(A) wm
.
(28)
Recollecting that
wn(A) = fA (En(A) ) ,
(B)
(B)
wm
= fB (Em
),
3
(29)
we obtain the requirement
(B)
(B)
fAB (En(A) + Em
) = fA (En(A) ) fB (Em
)
(30)
that should be valid for any two systems weakly coupled to one and the same heat bath, but not to
each other. Mathematically, there is only one function that satisfies Eq. (30). It is the exponential:
fA (E) ∝ fB (E) ∝ fAB (E) ∝ e−βE .
(31)
The only free parameter here is β, since the proportionality coefficients are fixed by the normalization
conditions.
We arrive at the following formulation of quantum Gibbs distribution. As we have demonstrated,
without loss of generality we can believe that any quantum mechanical system in equilibrium with a
heat bath is in one of the eigenstates of its Hamiltonian. If we label these state by n and introduce
corresponding energies En :
H |ni = En |ni ,
(32)
then the probability of finding the system in the n-th eigenstate is
wn = e−βEn /Z ,
where
Z=
X
e−βEn
(33)
(34)
n
is a normalization constant; it is called partition function. The particular value of parameter β is
related to (and characteristic of) the heat bath. The inverse of β is called temperature, T = 1/β.
The physical meaning of the parameter T becomes clear from further analysis of generic and specific
properties of Gibbs distributions. The statistical expectation value of any observable represented by
an operator A—all the observables in Quantum Mechanics are operators—is given by
hAi =
X
Ann wn ,
(35)
n
where
Ann = hn|A|ni
(36)
is the quantum mechanical expectation value of A in the state |ni. Hence, we have a two-fold averaging of each observable: (i) with respect to each energy eigenstate, and (ii) with respect to Gibbs
distribution of these eigenstates.
Simplest example: Two-level system. There are only two eigenvalues of energy: E1 and E2 . Physically,
only the energy difference ε = E2 − E1 is relevant, and without loss of generality, we can set E1 = 0,
E2 = ε. Then the partition function is
Z = 1 + e−ε/T ,
and the probabilities are
(37)
1
1
,
w2 =
.
(38)
−ε/T
1+e
1 + eε/T
We introduce a simple observable—polarization, P . Its operator has the following matrix elements:
P11 = 1, P22 = −1, P12 = P21 = 0. The expectation value of P —average polarization, or simply
w1 =
4
polarization (for briefness)—yields the difference between probabilities of finding the system in the
state 1 and 2, respectively. If we are dealing with an ensemble of many identical two-level systems,
then polarization times the total number of systems yields the difference between the number of
systems in the state 1 and the number of systems in the state 2. Calculating P̄ ≡ hP i, we find
P̄ =
sinh(ε/T )
.
1 + cosh(ε/T )
(39)
Polarization of the two-level system provides us with a simplest model of thermometer. Given P̄ from
experiment, we restore the temperature T = T (P̄ ) from Eq. (39).
Problem 16. Draw the plot of the function T (P̄ ). Note that for the systems with bounded spectrum a
negative temperature is also meaningful.
One of the most important observables is the energy. Its operator is the Hamiltonian of the
system, Hnn = En . For the mean value of energy, E ≡ hHi, we have
E =
X
En wn = Z −1
n
X
En e−βEn .
(40)
n
Another useful quantity is the heat capacity
dE
.
dT
C =
(41)
Heat capacity shows how fast the mean energy of the system changes per unit variation of temperature.
By directly differentiating Gibbs distribution, we can show that
C =
1 X
(En − E)2 wn .
T2 n
(42)
Hence, the heat capacity is always equal to the square of the dispersion of energy divided by T 2 . Note
that C ≥ 0.
Problem 17. Derive Eq. (42). [Do not forget to differentiate Z.]
Problem 18. For the two-level system, calculate E by Eq. (40) and C by both (41) and (42). Draw the plots
E(T ) and C(T ). [Do not forget about negative temperatures.]
Entropy. One of the central thermodynamic variables is entropy. In the probability theory, the
entropy of a distribution is defined as
S=−
X
wn ln wn .
(43)
n
[The entropy defined by Eq. (43) is dimensionless. If one works in units kB 6= 0, then for thermodynamical purposes it proves convenient to include the factor kB in the definition of entropy.] What
does the entropy of a distribution tell us about? It characterize the complexity of the distribution. If
a system consists of a number of independent parts—typical situation in Statistical Physics, then the
entropy of the distribution is equal to the sum of entropies of the distributions of each independent
subsystem. To prove this statement, consider a system AB which is a combination of two independent
systems, A and B. With the distribution (28) we have
S (AB) = −
X
n,m
(AB)
(AB)
wnm
ln wnm
=−
X
(B)
(B)
wn(A) wm
(ln wn(A) + ln wm
) = S (A) + S (B) .
n,m
5
(44)
Problem 19. Calculate the entropy of the two-level system using Eq. (43). Draw the plot S(T ). What
happens when T → +0 and T → −0 (note that these are different limits) ?
Simplest Thermodynamic Relations
Once the partition function Z ≡ Z(T ) is found for a given system from a direct calculation, the rest
of the thermodynamic quantities can be obtained from generic relations which we establish below.
We start with introducing one more quantity—Helmholtz free energy
F = −T ln Z .
(45)
wn = e(F −En )/T .
(46)
S = (E − F )/T ,
(47)
E = F + TS .
(48)
dZ
d X −En /T X
=
e
=
(En /T 2 ) e−En /T ,
dT
dT n
n
(49)
In terms of F , Gibbs distribution reads
Plugging this into (43), we see that
or
Consider now the derivative
and note that
T 2 dZ
=E.
Z dT
Expressing Z through F as Z = exp(−F/T ), we rewrite this relation as
F −T
dF
=E.
dT
(50)
(51)
Comparing this to (48), we see that
dF
= −S .
(52)
dT
Finally, by differentiating (48) with respect to temperature and taking into account (52), we get
C = T
dS
.
dT
(53)
Conclusion: To establish all basic thermodynamic characteristics of a system it is enough to directly
evaluate from Gibbs distribution only its partition function. The rest is done by the following algorithm:
• Find F in accordance with (45).
• Find S in accordance with (52).
• Find E in accordance with (48).
6
• Find C in accordance with either (53), or the definition (41) [or both, to make sure that the two
answers coincide, and thus your algebra is free of errors which, unfortunately, happen too often.]
Problem 20. Apply this algorithm to the two-level system.
Macroscopic System: Heat Bath for Itself
By macroscopic system we mean a system which can be viewed as a set of a very large number of
weakly coupled (and thus almost independent) subsystems. Typical example: Consider a liquid, gas,
or solid containing a large number of molecules. Divide the whole thing into cells, each cell much
smaller than the whole system, but still large enough to contain many particles. Statistically, the
cells are almost independent, because they interact with each other only through the surface.—The
number of close-to-surface particles is much smaller than the number of the particles in the bulk.
The key idea behind the notion of macroscopic system is that for each its small quasi-independent
cell the rest of the system behaves like a heat bath. A macroscopic system thus forms a heat bath
for itself, and one may apply statistical description even if the system is isolated from any other heat
baths and its energy is exactly conserved. This conclusion is known as the equivalence of the canonical
and microcanonical distributions. Here by canonical distribution one means Gibbs distribution, while
microcanonical distribution corresponds to the distribution over the states within limitingly small
energy interval—each state having the same probability—and models an isolated macroscopic system
where the energy is conserved. The above conclusion is also clear from the Central Limiting Theorem,
because the energy of a macroscopic system can be represented as a sum a large number of statistically
independent energies of its cells. Hence, the relative fluctuations of energy in a macroscopic system
are negligibly small and cannot be relevant to any observable quantity.
Sometimes instead of canonical and microcanonical distributions one speaks of canonical and
microcanonical ensembles representing these distribution. Speaking the ensemble language, an equilibrium macroscopic system can be viewed as an ensemble of its smaller subsystems (cells).
The natural thermodynamic variable for an isolated system is the energy. It is of crucial importance, however, that the equivalence of the microcanonical and canonical descriptions allows one to
use the notion of temperature, entropy, and other statistical characteristics for isolated macroscopic
systems. By definition, these characteristics are the same as if the system is in contact with some
heat bath such that the mean energy of the Gibbs distribution equals to the energy of the isolated
system of interest. Moreover, with a macroscopic system even the microcanonical distribution is not
necessary to speak of temperature, entropy, etc. Just one typical state of the system will do, because
dividing the system into cells we still have the statistical description in terms of the ensemble of cells.
Energy as a thermodynamic parameter. Now we derive thermodynamic relations in which the
(average) energy (rather than temperature) is the main thermodynamic parameter. These relations
are especially useful for macroscopic systems, while being also valid for microscopic systems. For a
macroscopic system, mean energy is a very good thermodynamic variable because the fluctuations of
energy from its mean value are negligibly small. Suppose we have expressed entropy as a function of
(mean) energy: S = S(E). What about the first derivative of this function?
dS
≡
dE
µ
dS
dT
With (53) we get the famous result
¶µ
dT
dE
¶
µ
≡
dS
dT
¶µ
dE
dT
dS
1
= =β.
dE
T
7
¶−1
≡ C −1
dS
.
dT
(54)
(55)
What about the second derivative? Differentiating (55) with respect to E, we find
d2 S
1 dT
1
=− 2
=− 2
2
dE
T dE
T
µ
dE
dT
¶−1
,
(56)
and see that
d2 S
C −1
=
−
≤ 0,
(57)
dE 2
T2
since, in accordance with (42), we have C ≥ 0. Below we will use this inequality to prove the increase
of entropy.
Problem 21. A macroscopic system is formed by N À 1 two-level systems with interlevel separation ε.
(Interactions between the systems are negligible.) Plot the function S(E), where E is the total energy and
S(E) is the entropy of corresponding equilibrium state.
Increase of Entropy
Heat exchange. Suppose we have two isolated macroscopic systems, A and B, each of which is
in equilibrium and is characterized by its own temperature. We then allow the systems to weakly
interact (exchange energy). As a result of this energy exchange, the systems will relax to some
common temperature. In such a process, the entropy always increases. Let us prove this fundamental
fact. It is convenient to characterize the thermodynamic states of our systems with energy rather
than temperature, because the total energy, E, is conserved and we can consider the energy EA of
the system A as the only free parameter describing the state of the two systems. For the entropy,
which is an additive function, we then have
SAB (EA ) = SA (EA ) + SB (EB = E − EA ) .
(58)
We want to understand the properties of the total entropy as a function of EA . To this end we
calculate the first and the second derivatives:
dSA
dSB
dSAB
=
−
,
EB = E − EA ,
(59)
dEA
dEA dEB
d2 SAB
d2 SA d2 SB
=
EB = E − EA .
(60)
2
2 + dE 2 ,
dEA
dEA
B
First derivative tells us about the extremum points, and the second derivative allows us to distinguish
the maxima from minima. Now we note that all the derivatives in the right-hand sides of (59) and
(60) correspond to equilibrium states, and we can use Eqs. (55), (57). From (55) we see that
dSAB
= βA − βB ,
(61)
dEA
that is the extremal point of the total entropy corresponds to the case when two temperatures are
equal. From (57) we have
d2 SAB
(62)
2 ≤0,
dEA
which means that our extremal point is the point of maximum. Hence, while the system evolves
towards its equilibrium, βA = βB , the entropy always increases. And the energy always goes from
the body with the smallest β to the body with the largest β. Indeed, from (61) we have
∆SAB = (βA − βB ) ∆EA .
(63)
The increase of entropy means that the l.h.s. is positive. Hence, the sign of ∆EA should be equal to
the sign of (βA − βB ).
8