Dec 15, 2010 ... 1.2 Fundamental Assumption of Statistical Mechanics . . . . . . . . . . . . . . . . . . . . . . .
.... Textbook: Statistical Mechanics by Pathria. 1.2 Fundamental ...
Lecture Notes for Statistical Mechanics Fall 2010 Lecturer: Professor Malvin Ruderman Transcriber: Alexander Chen December 15, 2010
Contents 1 Lecture 1 1.1 Basic Information . . . . . . . . . . . . . . . . . . . . . . . . 1.2 Fundamental Assumption of Statistical Mechanics . . . . . 1.3 Energy and Expectation Values . . . . . . . . . . . . . . . . 1.4 Justification of the Assumption that P Only Depends on E 1.5 Deriving the Fundamental Theorem . . . . . . . . . . . . . 1.6 An Example . . . . . . . . . . . . . . . . . . . . . . . . . . .
. . . . . .
3 3 3 3 4 4 5
2 Lecture 2 2.1 Some Remarks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.2 Harmonic Oscillator Example, Continued . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.3 Some Note about Counting . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
6 6 6 7
3 Lecture 3
9
. . . . . .
. . . . . .
. . . . . .
. . . . . .
. . . . . .
. . . . . .
. . . . . .
. . . . . .
. . . . . .
. . . . . .
. . . . . .
. . . . . .
. . . . . .
. . . . . .
. . . . . .
. . . . . .
. . . . . .
4 Lecture 4 12 4.1 A Homework Problem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12 4.2 About the definiton of Entropy . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12 4.3 About Thermodynamics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13 5 Lecture 5 14 5.1 Some Statements on Entropy . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14 5.2 The Chemical Potential . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14 6 Lecture 6 15 6.1 Review . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15 6.2 Reactions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15 7 Lecture 7 18 7.1 The Entropy of Mixing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18 7.2 The Chemical Potential for Photons . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19 1
8 Lecture 8
21
9 Lecture 9
22
10 Lecture 10 24 10.1 Digestion on Distributions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24 10.2 Historical Notes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25 11 Lecture 11 26 11.1 Classical Systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26 12 Lecture 12 28 12.1 Canonical Ensemble . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 29 13 Lecture 13 30 13.1 Continue on Canonical Ensembles . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 30 13.2 Grand Canonical Ensemble . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31 14 Lecture 14
33
15 Lecture 15
34
16 Lecture 16
37
17 Lecture 17 40 17.1 Problem of White Dwarf Star . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40 17.2 Heavy Atoms . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41 18 Lecture 18 43 18.1 Paramagnetism . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43 18.2 Diamagnetism . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 44 19 Lecture 18
45
20 Lecture 20 47 20.1 Superfluids . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 47 21 Lecture 21
49
22 Lecture 22
52
2
Statistical Mechanics
1
Lecture 1
Lecture 1
1.1
Basic Information
Course Instructor: Malvin Ruderman Office Hour: Tue 1:10 - 2:00 pm, Wed 12:00 - 1:30 pm Textbook: Statistical Mechanics by Pathria
1.2
Fundamental Assumption of Statistical Mechanics
If we take a system of N particles in a container of volume V , in order to know the basic properties of the system, we need to construct a general Hamiltonian of N coordinates and N momentum of these particles, and solve for its eigenvalues, which are the energy levels achievable by this system. In general this will be a very difficult problem, as the Hamiltonian will usually involve complicated interactions between the particles and it will become a messy manybody problem. But we don’t do that in statistical mechanics. What we do is to assume that the energy eigenvalues are known, and we want to know the behavior of the system in finite temperature. Take a system of N particles in a heat bath of temperature T , wait until equilibrium, then we measure the energy of the system. Note that by wait until equilibrium we mean that we wait until every state is accessible, after sufficiently long time. The fundamental theorem of statistical mechanics states that the probability of finding the system in a state with energy Ei is: Pi = P
all
e−Ei /kB T −Ei /kB T states e
(1.1)
The quantity in the denominator is called the Partition Function and is usually denoted by Q (not Z?). This theorem tells us that the probability is only dependent on the energy of that state, and that the state with lowest energy, i.e. the ground state, is the most probable state.
1.3
Energy and Expectation Values
Now that we know the probability of states, we can calculate the average energy of the system, and the expectation of other observables, too. The average energy, by definition, is just hEi =
X
Pi Ei
(1.2)
i
P However, the expectation of the square of the energy, E 2 = Pi Ei2 is not, in general, equal to the square of the average energy. So there usually exists fluctuations:
(∆E)2 = E 2 − hEi > 0
In general we can estimate the magnitude of the fluctuation:
(∆E 2 ) 1 ∼ O( ) 2 N hEi
(1.3)
(1.4)
so the fluctuation is negligible for a large enough system. In such a system, we can talk about “the energy” of the system, but in smaller systems we need to distinguish between the expectation value of energy, the median value of energy, or the most probable energy. 3
Statistical Mechanics
Lecture 1
The expectation value of any observable can be calculated in the same way, i.e. hAi =
X
Pi hi|A|ii
(1.5)
i
1.4
Justification of the Assumption that P Only Depends on E
Let us consider a 2-state problem where states φ1 and φ2 have the same energy. This can be due to physical separation by a potential barrier. Now we can form symmetric and antisymmetric linear combinations of these states which will be the ground state and the excited state. No matter what is the initial state, the time evolution of that state will look like the following: |φ(t)i = cos ωt |φ1 i + sin ωt |φ2 i
(1.6)
The state just oscillates between the two states. So if we do a time average, the probability of finding the particle in either state is 1/2. We can consider a particle with n states using Fermi’s Golden Rule: 1 τi→j
=
2π 2 dni |V | ~ ij dE
(1.7)
If a particle is in state i and can decay into some states j, and particle in states j can also jump back, then we can calculate the probability change of the particle to be found in state i: X X X dPi =− | hj|V |ii |2 Pi + | hi|V |ji |2 Pj = | hi|V |ji |2 (Pj − Pi ) dt j
j
(1.8)
j
so if the system is in equilibrium and probabilities become stable, then the probability of the particle to be in state i must be the same as in state j.
1.5
Deriving the Fundamental Theorem
Our fundamental assumption is that the probabilities only depend on the temperature and the energy of the state Ei . So we can write the probability of finding the system in state i as a function Pi (T, Ei ). Now we want to study it using some invariance. Because energy scale is usually arbitrary, and we should have the freedom to choose the reference point of energy without modifying the relative probabilities, we have the following invariance: Pi (T, Ei + α) Pi (T, Ei ) = Pj (T, Ej ) Pj (T, Ej + α)
(1.9)
This is a functional equation and the only solution is: e−Ei β(T ) Pi = P −E β(T ) i ie
(1.10)
where β(T ) is an (yet) unknown function of T . Now we ask the question whether β(T ) depends on the detail properties of the system we consider. Suppose we have two systems 1 and 2, and they have β1 (T ) and β2 (T ) in the probability formula. Now suppose they are each in equilibrium at temperature T , and brought together to form a new system 3, 4
Statistical Mechanics
Lecture 1
which has β(T ). The probability of system to be in state ij (system 1 is in state i and system 2 in state j) is: e−(Ei +Ej )β(T ) Pij = P −(E +E )β(T ) i j ij e
(1.11)
But we can also think of the system as a composite one and the same probability can be expressed as the product of individual probabilities: e−Ei β1 (T ) e−Ej β2 (T ) Pij = Pi Pj = P −E β (T ) P −E β (T ) i 1 j 2 ie je
(1.12)
the two ways of calculation are the same if and only if β(T ) = β1 (T ) = β2 (T ). So we can conclude that the β(T ) function is independent of the details of the system. So the only thing we need to do is to determine β(T ) for one simple system. Consider one particle in a 1D box (infinite well). We know from elementary quantum mechanics that ~2 nπ 2 En = 2m . Plug this into the above formula for probability and expectation value, we get: l hEi =
1 2β(T )
(1.13)
Generalizing this calculation into 3D and N particles, we get hEi = classical theory of ideal gases, that hEi = 32 N kB T . So we know that: β(T ) =
1.6
3N 2β(T ) .
1 kB T
But we know from the
(1.14)
An Example
Consider the harmonic oscillator potential. We know that the energy level is proportional to n. Let us omit the zero point energy of 12 ~ω for now and assume that En = n~ω. We can easily calculate the partition function for one particle since it is just a geometric series: Q=
X i
− kn~ωT
e
B
=
1 1−
e−~ω/kB T
,
1 = 1 − e−~ω/kB T Q
(1.15)
So we can see that the probability of finding the particle in state n goes like e−αn . Now suppose we have more particles. If there are two particle, then the energy becomes E = (n1 + n2 )~ω and there is degeneracy. The number of states corresponding to the same E is proportional to E. Generalizing this result, for N particles, the number of states with the same energy is proportional to E n−1 . To be continued. . .
5
Statistical Mechanics
2
Lecture 2
Lecture 2
2.1
Some Remarks
Recall that last time we argued the form of the fundamental theorem of Statistical Mechanics. The argument was based on first principles, but there is one assumption which is not first principle, i.e. we assumed that the perturbation of energy is hermitian: | hi|V |ji |2 = | hj|V |ii |2
(2.1)
This assumption may not be true if the transition is not time-reversible.
2.2
Harmonic Oscillator Example, Continued
Consider again the harmonic oscillator example. We want to study how physics change from 1 particle to N particles. We have already known the form of probability of the system taking energy level n for 1 and 2 particles, we can anticipate that for N particles it is: e−n~ω/kT = 1 − e−~ω/kT e−n~ω/kT Q ne−n~ω/kT N =2: Pn = Q ... ... nN −1 e−n~ω/kT N =N : Pn ∼ Q P the Q is there to ensure proper normalization, so that i Pn = 1. We will show that the final line is true using some approximations. Because the energy level of the whole N particle system is just the sum of energy levels of individual particles, we have: N =1:
Pn =
n = n1 + n2 + · · · + nN
(2.2)
E = E1 + E2 + · · · + EN
(2.3)
In the case where kT is much greater than ~ω, we can think of the energy levels being continuous, because the little steps of energy will not introduce much error in our approximation. Therefore we can replace summation by integrals and the result will look much nicer. In order to calculate the total probability, we basically integrate over all the possible configurations: Z Z Z P (n) = . . . P (n − n1 − n2 − · · · − nN −1 )P1 (n1 )P2 (n2 ) . . . dn1 dn2 . . . dnN −1 (2.4) This integral looks quite difficult, but we can do it in Fourier space. After Fourier transform, we have: Z
∞ ikn
e −∞
Z
N
∞
P (n)dn =
ikn1
e
P1 (n1 )dn1
−∞
We can take the P1 (n) = 0 for all n < 0 and do the integral on P1 : 6
(2.5)
Statistical Mechanics
Lecture 2
P˜1 (k) =
Z
∞
eikn−~ωβn dn =
0
1 ik − ~ωβ
And doing a Fourier transform back, we can find Z ∞ 1 e−ikn P˜i (k) dk P (n) = 2π −∞
(2.6)
(2.7)
Because P˜1 (k) ≤ 1, the N power will damp everywhere except for the highest point, which will turn to somewhat like a gaussian. Therefore the P (n) should also resemble a gaussian in the large N limit. Now let’s evaluate the above integral analytically. We substitute z = k − (−iβ~ω) and convert the integral to a contour integral: I dz inz P (n) = e−n~ωβ e (2.8) zN where we close the countour from below, because z has a pole in the lower halfplane. In the end we have eβE E N −1 ∂ PN (E) = . The most probable energy can be found by solving the equation ∂E PN (E) = 0, and (N − 1)! the result is: E = (N − 1)kT
(2.9)
We can also find the width of this near-gaussian distribution, by looking at the two points where the ∂2 second derivatives change sign. These can be found by solving ∂E 2 PN (E) = 0 and the answer is: 2 δE = ± (N − 1)1/2 β
2.3
(2.10)
Some Note about Counting
Consider N particles in a volume V . We have a smaller volume v inside V and we want to ask questions like “what is the probability of finding n particles inside the volume v?”. First let’s recall the binomial expansion N X N n N −n (p + q) = p q (2.11) n n=0 N N! where the binomial coefficients are defined to be = . If we let p + q = 1 then we have n n!(N − n)! N
1=
N X n=1
N! pn (1 − p)N −n n!(N − n)!
(2.12)
for any p < 1. Suppose we are considering classical particles, finding a particle inside a volume v inside a box V is just v/V , and if we substitute p = v/V into the above equation, then the summand is just the probability of finding n out of N particles inside volume v at any given time. If we consider quantum mechanics, then we need to integrate the wavefunction square, but just need to replace p. Consider some extremes. For n = N , i.e. all the particles squeezed inside v, the probability is:
7
Statistical Mechanics
Lecture 2
P =
v N V
v
= e−N log V
(2.13)
For n = 0, however, we have: P =
V −v V
N
1 vN N N →∞ −N v = 1− −−−−→ e V N V
8
(2.14)
Statistical Mechanics
3
Lecture 3
Lecture 3
What we have discussed is how statistical mechanics might have developed if we had the knowledge of quantum levels and stuff. We started out with a fundamental assumption that all states with the same energy are equally probable in equilibrium. The key is that we have a hamiltonian with a little perturbation which permits time-reversible transitions. Recall last time we did some counting exercise, and found that for N particles in volume V , the probability of finding n particles in smaller volume v is: v n N! P (n) = n!(N − n)! V
V −v V
N −n (3.1)
We will often use some approximations: ∞
Z n! =
e−x xn dx
(3.2)
0
log n! = n log n − n +
1 log n 2
(3.3)
where the latter equation is called the Stirling approximation. Which means that for n! = n(n − 1)(n − n 2) · · · ∼ nen = nn e−n . If we take these expressions and use them on the previous expression of P (n) to calculate log P (n), then we √ can see the graph is like √ a thin Gaussian, which peaks at around N/2. The peak probability is about 1/ N , and width is about N . Recall if we want to measure a quantum mechanical variable A, the expectation value we get is hAi =
X
Pi hi|A|ii
(3.4)
i
and usually we have A2 6= hAi2 . If we go to N → ∞ limit, then the system has well-defined energy, but there are some variables of particular interest. This is called the thermodynamics limit. The thermodynamic variables fall into two categories: • Extensive variables: U , CV , S, N , these variables increase as the size of the system increases. • Intensive variables: T , P , these variables do not increase as the size of the system increases. Now let’s turn to the special variable “entropy”, which is the missing link between thermodynamics and statistical mechanics. We want to define entropy and find the way to calculate it. Consider two adjacent boxes labeled 1 and 2, they are connected to form system 3. Label the states in 1 by i and states in 2 by j, we have: (3)
(1)
(2)
Pij = Pi Pj
(3.5)
P To form an extensive quantity, i.e. S (3) = S (2) + S (1) , it is appealing to define S = i log Pi , but because probability for some high P energy states goes to zero, this definition will lead easily to infinities. What about define S = −k i Pi log Pi ? We can calculate the entropy of system 3 as:
9
Statistical Mechanics
S (3) = −k
XX i
Lecture 3
(Pi Pj ) log(Pi Pj ) = −k
j
X X X (Pi Pj log Pi + Pi Pj log Pj ) = −k( Pi log Pi + Pj log Pj ) ij
i
j
(3.6) this quantity is manifestly extensive and this is what we want. This definition is good even if the system is not in thermal equilibrium, because whatever is the stucture of Pi , this expression still works. 1 If we look at distribution of Pi of states with almost exactly same energy, then Pi = where Ni is Ni the number of that state. The entropy in this case is: S = −k
X
Pi log Pi = k log Ni
(3.7)
i
This definition of entropy is such that when you depart from the thermodynamic equilibrium, the entropy lowers. In other words, the entropy takes maximum at thermodynamic equilibrium. This is how we interpret entropy physically. Remember in thermodynamics we defined entropy as a function of P and V and T . When you integrate along the change of the system from A to B: Z
B
S= A
dQ T
(3.8)
the integral is independent of path and is defined to be the entropy change. The second law of thermodynamics says this function always exists. We introduced two potentials in thermodynamics: 1. Helmholtz Free Energy: F = U − ST 2. Gibbs Free Energy: G = U − ST + P V These quantities are useful in special circumstances when we want to isolate some thermodynamic effect. Say if we study a system with fixed temperature, then the system tends to minimize its Helmholtz free energy. If we consider pressure fixed instead of the volume, then we use Gibbs free energy, which also tends to be minimized. The name “Entropy” is the greek word for “transformation”. In some sense, it measures how much transformation was done on the system. What about “free energy”? Let’s consider the change of Helmholtz free energy: dF = dU − T dS − SdT
(3.9)
Because we have 1. ∆Q = dU + P dV 2. ∆Q = T dS so we can rewrite the change of free energy to dF = −P dV − SdT . On an isotherm, then dF = −P dV which is the work done to the outside world if we allow the volume to change. In this sense, it is “free”. Let’s look back at the definition of S. We can plug in the canonical probability distribution: 10
Statistical Mechanics
Lecture 3
S = −k
X e−Ei /kT Q
log
X Ei X e−Ei /kT U =k Pi + k log Q Pi = + k log Q Q kT T
(3.10)
Compare the expression with the definition of the Helmholtz energy, we see that: F = −kT log Q,
Q = e−F/kT
(3.11)
which relates the Helmholtz free energy directly to the partition function. Let’s consider an example. Combining the first and second law of thermodynamics, we have T dS = dU + P dV , so we know that: 1 ∂U 1 ∂U P = dV + dT + dV (3.12) T ∂V T T ∂T V T ∂S ∂S dT + dV (3.13) = ∂T V ∂V T ∂P ∂S = , we get P = Equating corresponding terms, and using the Maxwell relation ∂V T ∂T V ∂U ∂P − +T . ∂V T ∂T V In microscopic view, we can calculate the pressure by calculating the expectation of energy change when changing volume: dU P dS = + dV T T
P =
X ∂Ei e−Ei /kT i
∂V
which is much harder.
11
Q
(3.14)
Statistical Mechanics
4
Lecture 4
Lecture 4
4.1
A Homework Problem
Remember in our Hilbert space the P basis we use is the energy eigenbasis H |ii = Ei |ii. So the expectation value of variable A is just hAi = Pi hi|A|ii. Now what if we change into the eigenbasis of operator A, i.e. A |ji = Aj |ji, then is the following expression equivalent to our previous expectation value? X
4.2
Aj e−hj|H|jiβ(T ) = hAi
(4.1)
About the definiton of Entropy
There are two points about the entropy we introduced by the formula S = −k
P
Pi log Pi :
• It is the same entropy as we introduced in thermodynamics. • It is still a good definition even when not in equilibrium. The second point is easier to see because the definition is robust, as it only uses the concept of probability. In order to see the first point, let’s calculate the change in entropy: X X ∆S = −k ∆(Pi log Pi ) = −k (∆Pi log Pi − ∆Pi ) (4.2) i
i
And by the fundamental theorem, we have log Pi = −βEi − log Q, we now have: X X X ∆S = kβ (∆Pi )Ei + k ∆Pi log Q = kβ (∆Pi )Ei i
i
= kβ
i
X
∆(Pi Ei ) − kβ
i
X
(4.3) Pi ∆Ei
i
The first term is just ∆U , where U = hEi. The second term can be interpreted by noting that dEi = ∂Ei /∂V dV , which is just −P dV . Therefore we have: T ∆S = ∆U + P ∆V
(4.4)
which is just the statistical mechanics version of the second law of thermodynamics. We saw that in a system with N states of equal energy, the probability of finding the system in any state is the same: Pi = 1/N . This can be accomplished by introducing a small perturbation which allows reversible transition between different states. If we adopt the above definition of entropy, we can calculate the entropy of the system: S = −khlog P i = −k
X 1 1 log = k log N N N
(4.5)
i
The same definition leads to the fact that at zero temperature, all states have probability of unity, so the entropy is zero. This is the third law of thermodynamics.
12
Statistical Mechanics
4.3
Lecture 4
About Thermodynamics
Remember our introduction of thermodynamic potentials. In addition to what we have introduced, we want to study situations where the number of particles can vary, like that in a chemical reaction. Similar problems include e− + e+ −→ γ where we are interested in the equilibrium temperature and absolute value of entropy. Remember last time we derived the expression of pressure: ∂U ∂P P =− +T (4.6) ∂V T ∂T V this was done by using the thermodynamic identities, which was much easier than using first principles from statistical mechanics. When there is a large number of particles, we can do things easier by using thermodynamics. Remember we introduced the Helmholtz free energy F = U − ST , which satisfies an elegant relation F = −kT log Q. And the change in free energy is: dF = dU − T dS − SdT = −P dV − SdT (4.7) ∂U ∂U By the way of taking differentials, we know that P = − and S = − . Now suppose ∂V T ∂T V we change the number of particles in the system, we introduce the change in the free energy by: dF = −P dV − SdT + µdN
(4.8)
∂F . It is similar to a binding ∂N V,T energy, but in a more general form. Now if we look at the system energy instead, we add the same term ∂U to the change of internal energy, and we get another expression of µ as µ = . Now what if we ∂N V,S consider the Gibbs free energy? where N is the number of particles and µ can be identified as µ =
dG = dU − SdT − T dS + P dV + V dP + µdN = −SdT + V dP + µdN
(4.9)
Consider if we fix the temperature and pressure of the system, and we remove one particle, the system adjusts its volume automatically to accomodate the change, and the other particles do not feel the change. Another way to say this is that both P and T are intensive quantities, so they don’t depend on the particle number. Therefore dG = µ(P, T ), G = µ(P, T )N (4.10) dN P,T Next time we will consider things like vapor pressure and osmotic pressure.
13
Statistical Mechanics
5
Lecture 5
Lecture 5
5.1
Some Statements on Entropy
Remember we defined the entropy in the statistical mechanics sense: X S = −k Pi log Pi
(5.1)
and we derived some properties such as −kT log Q = U − T S = F . If we are looking at an isolated system, it tends to maximize its entropy. If we are looking S and treat it classically, then it is always infinity. In classical situation, the distance between states is going to zero, and instead of discrete states we get a contiuum of states. So all the entropies go to infinity. However, sometimes knowing the value of S will be helpful in some situations when we want to ask all kinds of questions about the system. The second statement about entropy is that when temperature goes to zero, S goes to 0. This has the implication that all specific heat should approach zero at zero temperature. In order to see this, let’s suppose that S = κT α where α is a positive number. We have dS 1 DQ 1 = καT α−1 = = CV =⇒ CV ∝ T α dT T dT T
5.2
(5.2)
The Chemical Potential
Remember that we derived the various expressions for the chemical potential ∂F ∂G ∂U ∂S µ= = = = −T ∂N T,V ∂N P,T ∂N S,V ∂N U,V
(5.3)
And we also know that G = N µ(P, T ). The change in chemical potential is dµ =
dG S V = − dT − dP N N N
14
(5.4)
Statistical Mechanics
6
Lecture 6
Lecture 6
6.1
Review
Remember we introduced the fundamental assumption of statistical mechanics, and defined the entropy as X S = −hk log Pi i = −k Pi log Pi (6.1) i
and we have the relation between the Helmholtz free energy and partition function F = U − ST = −kT log Q From the definition of Helmholtz free energy we have some relations ∂F S=− ∂T N,V ∂F P =− ∂V N,T U = hEi = −
∂ log Q ∂β
(6.2)
(6.3) (6.4) (6.5)
and we have several expressions for chemical potential ∂G ∂U ∂S ∂F = = = −T µ= ∂N T,V ∂N P,T ∂N S,V ∂N U,V
(6.6)
In order to find the absolute value of µ, the easiest way is to fix temperature and volume and calculate the Helmholtz free energy of the system, then differentiate it with respect to particle number N and we get µ. We are going to do this in a moment.
6.2
Reactions
Consider a chemical (or particle) reaction A + B ←→ C
(6.7)
When equilibrium is achieved, the Gibbs free energy is at its minimum, and the particle numbers of different species should be equal, so the chemical potentials should satisfy µA + µ B − µ C = 0
(6.8)
But does the method that we achieve equilibrium matter? What if we release a photon during this chemical process? Do we need to calculate the chemical potential for photons? We will address these questions in this lecture. We will consider processes like e− + e+ −→ γ + γ and H + + e− −→ H which are elementary processes of the universe. Consider a 1 dimensional harmonic oscillator, we know the energy levels to be En = (n + 1/2)~ω0 , so we can calculate the partition function 1
Q1 =
X
1
e−(n+ 2 )~ω0 β =
n
15
e− 2 ~ω0 β 1 − e−~ω0 β
(6.9)
Statistical Mechanics
Lecture 6
Suppose now we have 3 dimensions and N particles, then the total partition function is N Q = (Q1 )3
(6.10)
So the Helmholtz free energy can be calculated as F = −kT log(Q1 )3N = −kT 3N log Q1
(6.11)
Now we need to think a little harder. The way we thought about the system was treating the particles as distinguishable. But in reality we have indistinguishable particles, and we need to incorporate that effect. However, if the probability for finding more than one particle in any single excited state is negligible, i.e. the states are diluted enough, then we can just divide our Q by N ! which is just the factor of overcounting. In fact we should have used the factor N! (6.12) n0 !n1 ! . . . n∞ ! where ni is the number of particles in state i. Our assumption of dilutedness is equivalent to taking all ni to be either 0 or 1. So under our assumption of diluted density of particles, our free energy for indistinguishable particles becomes (in the limit when N is large) Findistinguishable = −kT log
Q = Fdistinguishable + kT (N log N − N + . . . ) N!
(6.13)
Remember our sum for the 1 dimensional harmonic oscillator. We can treat the sum in contiuum approximation and treat the sum as an integral: Z ∞ 1 kT dne−~ω0 βn dn = = (6.14) ~ω0 β ~ω0 0 This is just the first term if we expand the exponential in the exact partition function in equation (6.9). This discourse justifies our use of continuum limit. So when the sum is hard to do, we will try to do the integral instead. A word for the Gibbs Paradox. Consider a chamber partitioned into two equal parts, with some particles in part 1 and some in part 2. We can calculate the entropy for each partition and they will work out. Now we want to remove the partition and join the two parts. If the particles were distinguishable, then we would run into some paradoxical situation as in history. We wouldn’t have the paradox because we put in the N ! which accounts for overcounting. We can carry out the above calculations for other potentials as well, say the 1 dimensional box, i.e. infinite potential well. For 1 dimensional box the allowed energies are nπ~ 2 (6.15) En = 2m 2mL and the partition function is Q1 =
X
π 2 ~2
2
contiuum limit
e− 2mL2 βn −−−−−−−−−−→
Z 0
n
The integral is just a Gaussian integral and easy to evaluate. 16
∞
π 2 ~2
dne− 2mL2 βn
2
(6.16)
Statistical Mechanics
Lecture 6
The Helmholtz free energy for N identical particles in a box of volume V = L3 , then we have F = −kT log
Q3N 1 N!
(6.17)
Carrying out the previous integral and plug into the expression for free energy, we get " # 3/2 h2 N F = N kT log −1 V 2πmkT
(6.18)
and we can calculate the chemical potential by differentiating N µ = kT log V
h2 2πmkT
3/2 =
G N
(6.19)
What is the term in the bracket? Because we have p = ~k = h/λ. If we take the characteristic kinetic energy p2 /2m ∼ kT then the term in the bracket is just (apart from a factor of π) λ2 which is the DeBroglie wavelength squared. Our dilutedness condition here translated into N 3 λ 1 V
(6.20)
Questions left to consider: When we have photons in the game, the chemical potential will become zero. Why is this true? Another question is that when the particles are relativistic, the energy and momentum relation becomes different, and the integral becomes hard to evaluate. What to do?
17
Statistical Mechanics
7
Lecture 7
Lecture 7
Remember last time we studied the N particle partition function when they are indistinguishable X Q= e−Ei /kT = (Q1 )N
(7.1)
And the free energy is proportional to F ∝ −kT log
V λ3
N (7.2)
Now if we have identical particles, i.e. indistinguishable. If we continue to use the above expression then we are overcounting the states. Remember the approximation we used last time we to assume that states are sufficiently diluted and the occupation of any state is no more than 1. Under this approximation Q=
(Q1 )N N!
therefore the Helmholtz free energy becomes N N 1 V F ∝ −kT log ≈ −kT N log 3 λ N! N λ3
7.1
(7.3)
(7.4)
The Entropy of Mixing
Consider a box of volume V being partitioned into two equal halves, both with volume V /2. Particles are distributed equally inside two partitions. We know the entropy of the whole system is the sum of the entropies for the two partitions: N V N S = S1 + S2 = 2 −k log + N log (7.5) 3 2 2(N/2)λ 2 Now if we remove the lid in between, then the entropy of the whole system becomes V S = −kN log + N log N N λ3
(7.6)
and the entropy has increased by ∆S = kN log 2. This is a reversible process. This means that, if we had done this process carefully, we can extract some work from it, like the Carnot engine. Indeed we can do this by imagining that we insert a semi-permeable membrane in between which only blocks one specific particle from going through. When we move the membrane to one end of the box, the system is doing work by this particular particle. Imagine if we have a huge number of this kind of membrane, we can extract work from every particle in this box, and this is the work we could have extracted from the system. The amount of energy is gained from the heat absorbed from the reservoir which is kept at constant T . Now suppose the particles inside this box are indistinguishable. In the case of partitioned box the total partition function is QN 1 QN = (7.7) (N/2)!(N/2)!
18
Statistical Mechanics
Lecture 7
so the initial entropy will be different from what we calculated above. The final entropy will also be different. But the difference in entropy will be the same as what we had above. This is reasonable because we have essentially labeled the particles on the left differently from the particles on the right. If we insert two semi-permeable membrane and pushing them to either end, then we can extract the same amount of work as the above case. This change of entropy is called the entropy of mixing. Now suppose all the particles in the box are the same. Then it doesn’t matter if there is a partition in the middle. So the entropy of mixing has disappeared. But the final entropy is not the same as the final entropy above, as the total partition function has been changed into QN =
QN 1 N!
(7.8)
The above discussion is to differentiate the three cases of distinguishable/indistinguishable particles and explain how to quantify the entropy change.
7.2
The Chemical Potential for Photons
Remember we had an explicit expression for the chemical potential of an ideal gas from last time N µ = kT log V
h2 2πmkT
3/2 =
G N
(7.9)
Now consider the case when there is photon involved in the reaction like e− + p −→ N + γ. We would have trouble if the chemical potential of photon were not zero. So we are going to show that this is true. Suppose we have a box of photons with volume V at temperature T . We know that we can calculate the energy using blackbody radiation U = u(T )V . The pressure on the wall is just P = u(T )/3
(7.10)
Now suppose we take away a photon in the box, while fixing the temperature and volume, the change in free energy must be ∂F =0 (7.11) ∂N T,V because F is totally decided by T and V . Let’s take another way of looking at this. In 3 dimensions we have P = u/3 where u(T ) = κT 4
(7.12)
where κ is a constant. Let’s suppose there is some heat coming into the system ∆Q = dU + P dV = V u0 dT + udV + Therefore we have that
dS =
V u0 T
u dV = T dS 3
dT +
19
4u 3T
(7.13)
dV
(7.14)
Statistical Mechanics
Lecture 7
Using the rule for partial derivatives we get 0 Vu ∂ 4u ∂ = =⇒ u = κT 4 ∂V T ∂T 3T
(7.15)
Therefore thermodynamics requires that u ∝ T 4 . Dimensional analysis will show that u = kT
kT ~c
3
k0
(7.16)
where k 0 is a constant that we must introduce. Now we can calculate the chemical potential G U + PV − TS = N N Z T uV uV T CV 0 = + − dT N 3N N 0 T0
µ=
Because CV = u0 V , we can plug in the u = κT 4 then we can see that µ = 0 for photons.
20
(7.17)
Statistical Mechanics
8
Lecture 8
Lecture 8
Consider a box of fixed volume V kept at fixed temperature T . The system insided tends to minimize its Helmholtz free energy F . Remember X X dF = dU − T dS − SdT + µi dNi = −SdT − P dV + µi dNi (8.1) i
i
where the sum is over all species of particles inside the box. If we fix T and V , then at equilibrium dF = 0, we have X µi dNi = 0 for all allowed changes dNi (8.2) i
Consider weP have a chemical reaction involving particles Ai with number coefficients νi , then we write our reaction as i νi Ai = 0. For example, we can consider H2 + Cl2 ↔ 2HCl. Then νH2 = 1, νCl2 = 1 and νHCl = −2. In this scenario, then we have X νi µi = 0 (8.3) i
We want to make an assumption that the existence of other kinds of particles will not affect the chemical potential of any one species of particles, in the limit of dilute gas. This makes sense because here interaction is not so strong, so that the sum of entropies of two species of particles in two boxes of V is the same as the entropy if the two species of particles in the same box. Any particle is not aware of the presence of other species of particles. Remember for photons µ = 0. For non-relativistic particles we have Ni λ3i µi = kT log + mi c2 (8.4) V Note because of different energies in binding, we have different mass for, e.g. HCl and H + Cl. That is why we include the mi c2 term. If we plug the above equation into equation (8.3) then we can get X kT log ni λ3i + mi c2 νi = 0 (8.5) i
[n1 λ31 ][n2 λ32 ] = e−∆E/kT [n3 λ33 ]
(8.6)
where ni is the density of state of state in V . Remember this is non-relativistic. Consider now the process e− + e+ ←→ γ + γ. Because there is degeneracy associated with spin s, we replace V −→ V (2s + 1). We write n− λ3− + me c2 2s + 1 n+ λ3+ = kT log + me c2 2s + 1
µe− = kT log
(8.7)
µe +
(8.8)
Therefore we can evaluate
2s + 1 2 −2mc2 n+ n− = e β (8.9) λ3 But this equation does not tell us either n− or n+ . We have to put in some extra condition such as n+ = n− to get the exact value of either number densities.
21
Statistical Mechanics
9
Lecture 9
Lecture 9
Up to now we have been considering problems where many particles share a box, but for each state there is only zero or one particle at that particular state. We derived the chemical potential N 3 (9.1) µ = kT log λ + mc2 V Here λ is about the effective wavelength of the particle. The assumption of diluteness is equivalent to N 3 λ 1 V
(9.2)
∂ µ= (U − T S) ∂N T,V
(9.3)
Remember also that we have the relation
So µ can also be thought of something like the binding energy. We can also write N=
V µ/kT e λ3
(9.4)
Consider the reaction e− + H + ←→ H. We will assume the system is neutral, so that the number of electrons is the same as number of ions. We had the result [λ3e ne ][nH + λ3H + ] = e−b (H)/kT [λ3H nH ]
(9.5)
But because we know that the number density of e is the same as H + , while λH ∼ λH + , so we have ne =
nH −b /2kT e λ3e
(9.6)
Because the above process is essentially reversible, the rate of both directions should be the same ne n+ |H|2 ρHγ = nγ nH |H|2 ρe+
(9.7)
In principle we can invert this equation to get the same ratio as the above. But in nγ there is a factor of c coming into play, while there is no speed of light in the above equation. So there must be a delicate cancelation among the terms to eliminate the speed of light. We know the form of the product ne nH + on the left hand is approximately ne nH + ∼ nH nγ e−|b |/kT (9.8) Because b ≈ 13.6 eV, the corresponding temperature for equilibrium is k(105 K) ∼ 10 eV. Suppose we have a box with a magnetic field B(x) pointing down which is a function of position. The particles inside have spin 1/2. We want to know how the distribution of the spin of the particle. Let’s define µ↑ = µB=0 − mB, µ↓ = µB=0 + mB (9.9) We know from the fundamental assumption of statistical mechanics that in equilibrium we have n↑ = n(0)e−mB/kT ,
n↓ = n(0)e+mB/kT 22
(9.10)
Statistical Mechanics
Lecture 9
Therefore if we take the sum of these two equations, we have mB/kT
n↑ + n↓ ∝ e
+e
−mB/kT
mB mB − + =2+ kT kT
mB kT
2 + ...
(9.11)
Which means that there is an accumulation of particles at places with larger B field. Consider a system with available energy levels. We denote the number of particles at energy level r as nr . We have been dealing with the partition function, which can now be written as the sum of states P X Q= e− r nr r /kT (9.12) all distributions of states
subject to the condition that
P
r
nr = N . If we ignore the constraint for now, we can write Q=
YX r
e−nr r /kT =
r
nr
1
Y 1−
e−r /kT
(9.13)
This looks like the Einstein-Bose distribution, so if we calculate the Helmholtz free energy we get X F = −kT log Q = kT log 1 − e−r /kT (9.14) r
Remember this result we have is right for photons. But note that F does not depend on N explicitly, so the partial derivative is zero, which means this applies to particles with µ = 0. So we reached another result where the number of particles can vary freely. Let’s think about this problem in another way. Remember from the beginning we have thought of µ as something resulting from the properties of the system. However, we can treat µ as given a priori, and see what can we get. Consider a box connected with an reservoir with fixed entropy and infinite number of particles. The chemical potential of the reservoir is just ∂F = µ = −|b | (9.15) ∂N T,V So in equilibrium the µ of the system is the same as that of the reservoir, and we have a way to control the chemical potential. If we call the energy of the reservoir zero, then the energy of the system is shifted by −µ. We want to calculate the average number of particles in a certain state by summing over the possible “state configurations” P −(r −µ)nr β n r nr e P hnr i = (9.16) −(r −µ)nr β nr e Remember the chemical potential of a system is often negative, so the inclusion of µ effectively raises the energy and suppressed the probability of every state. We can continue playing this game and find an expression for the expectation value hnr i =
1 ∂ log β ∂µ
∞ X nr =0 fix r
e−(r −µ)nr β = −kT
∂ 1 log 1 − e−(r −µ)/kT ∝ ( −µ)/kT ∂µ e r −1
So we have derived the Bose-Einstein distribution from first principles. 23
(9.17)
Statistical Mechanics
10 10.1
Lecture 10
Lecture 10 Digestion on Distributions
Remember given a system, if we know the partition function of the system, then we can basically answer any questions about the system. However the partition function may be very hard to calculate. Now consider we have a system with N particles and some energy levels, last time we employed a trick to calculate the expectation of occupation number of each state. We shifted the system energy by −µN by adding a reservoir of fixed chemical potential, but now the number of particles in the system is no longer N , but can be varied. Now let’s see how we can do it physically without using the trick. The results we get from last time is P −(r −µ)nr /kT nr 1 re = ( −µ)/kT hnr i = P (10.1) −( −µ)n /kT r r r e −1 re This result applies to bosons. What if we have fermions? Remember we can treat the sum as fixing r and summing up nr , but in the case of fermions nr = 0 or 1, so P −(n −µ)nr /kT r nr e−(r −µ)/kT 1 re = hnr if = P = ( −µ)/kT (10.2) −( −µ)n /kT −( −µ)/kT r r r r 1+e e +1 nr e This is the Fermi-Dirac distribution. Now we want to hold the number of particles in the box fixed. Then we can use the above formula to compute the chemical potential. We can define the probability with some chemical potential e−(Ei −N µ)/kT −(Ei −N µ)/kT ) all states e
P (µ) = P
(10.3)
Let’s look at an example. Suppose we have an atmosphere of O2 , and some Hemoglobin (proteins that can have 1 binding site that can bind to oxygen) with binding energy . How does the amount of binding depend on the pressure of oxygen? We know the chemical potential of oxygen is µO2 = kT log nO2 λ3
(10.4)
This formula applies to perfect dilute gas of indistinguishable paricles. So we can solve for the number density of O2 1 nO2 = 3 eµ/kT (10.5) λ So the chemical potential tells us how dilute the gas is compared to the wavelength of the gas particle. The partition function of binding is simple, as it only has two different configurations Q = 1 + e−/kT eµ/kT
(10.6)
So the average number of binding hnr i =
1 e(−µ)/kT
+1
=
1 e/kT /n
3 O2 λ
+1
(10.7)
The physical meaning of this expectation is the average occupation number on one Hemoglobin site, which is between 0 and 1. When temperature goes to 0, the exponential goes to 0 and there is no problem in binding. When the density of oxygen is infinite then 1/n goes to 0, and this occupation goes to 1 so there is no objection to binding either. 24
Statistical Mechanics
10.2
Lecture 10
Historical Notes
Suppose we have a hamiltonian H(q1 , . . . , q3N , p1 , . . . , p3N ) which is a function of 6N variables in the phase space. A point in the phase space specifies the system completely. There is an interesting theorem attributed to Liouville, which says the following. Let us assume the hamiltonian to be independent of time, and suppose we have a large number of systems moving around in the phase space with ρ(q1 , . . . , p3N ) denoting the number density of different systems in the phase space, and the total number of systems is conserved, then we have the conservation equation ∂ρ ∂ρ + ∇ · (ρv) = 0 = + ρ∇ · v + v · ∇ρ ∂t ∂t
(10.8)
The comoving density of systems, which is defined to be the total time derivative of ρ is ∂ρ dρ = + (v · ∇)ρ = −ρ∇ · v dt ∂t
(10.9)
Remember we have the Hamilton’s equations q˙i = +
∂H , ∂pi
p˙j = −
∂H ∂qj
So the above equation for derivative for ρ becomes X ∂ dρ ∂ = −ρ q˙i + p˙j = 0 dt ∂qi ∂pj
(10.10)
(10.11)
So wherever we sit, the density of system at my neighborhood remains constant, regardless of the details of the dynamics of the system. So there is no way you can use a static hamiltonian to focus a density distribution in the phase space. If we want to establish a density distribution which does not change with time, i.e. ∂ρ/∂t = 0, then the only solution is that dρ/dt = 0 and ρ is a constant. For a classical system we have the expectation of a physical quantity R dq1 . . . dp3N A(q1 , . . . , p3N )e−H/kT R hAi = (10.12) dq1 . . . dp3N e−H/kT Note that this is purely classical and we don’t need to know the density of states. But there is no way we can calculate the entropy in this classical way, because we defined it to be X S = −k Pi log Pi (10.13) which assumes the knowledge of probabilities of micro states.
25
Statistical Mechanics
11
Lecture 11
Lecture 11
We discussed last time that because what we derived, there is no way to focus a parallel beam to a smaller area of concentration. That was wrong. Only when the incoming beams are diverging that we can’t use a static device to focus it to a concentrated parallel beam. Another way of interpreting the Liouville’s theorem is that the comoving density ρ, which is equal to ρ=
d2 n dpdq
(11.1)
is constant wherever we move to in phase space. Suppose the distribution n is constant, then this amounts to the constancy of any volume element dpdq that we are moving. So if we have an incoming parallel beam, then the volume is zero because there is no p dispersion in the y direction, so it is fine to transport it to another zero volume with smaller q dispersion in the y direction. But we can’t do this for diverging incoming beam, because the phase space volume should be constant. Another way to look at this is that suppose we have a focusing device as that. Suppose we have a body on the left that radiates at temperature T and another body at the right of the device to absorb that radiation and radiate back with a smaller surface area but the same temperature, then because they have the same power of radiation, the left body will radiate more energy and will get colder after a period of time, while the right body will gain energy and get hotter, and we did not do any work! So that is forbidden by the second law of thermal dynamics.
11.1
Classical Systems
Let’s consider the expectation value of a variable with respect to a classical system R dq1 . . . dp3N e−βH(p,q) A hAi = R dq1 . . . dp3N e−βH(p,q)
(11.2)
Note that there should be a factor of (2π)k and some factorial, but the coefficients in the numerator and denominator cancel. This expression translates into our similar expression in the quantum case where the integral is replaced by a summing over accessible states. The following relation is true regardless of we are dealing with classical or quantum physics DX E DX E q˙i pi + qi p˙i = 0 (11.3) To see this, note that
d dt pq
= pq ˙ + pq, ˙ and that d hpqi = pq ˙ + pq˙ = hpqi ˙ + hpqi ˙ dt
(11.4)
But in a confined system the time average of the derivative is zero, so the sum is zero, whether in classical or quantum systems. Note that in the Hamiltonian formulation, if our Hamiltonian is of the form H = P p2i 2m + V (q1 . . . qN ) then the kinetic energy is just DX E DX E DX E 2 hKEi = pi q˙i = − qi p˙i = − r·F (11.5)
26
Statistical Mechanics
Lecture 11
P The force on particle i can be written as Fi = j6=i Fij where Fij is the force of the j-th particle on the i-th one. Then we can write X X X r i · Fi = ri · Fij (11.6) j6=i
If suppose the force between particles can be written as the gradient of a potential U (r) = Arn , then we have F · r = nU (r) so we can get the virial theorem: 2hKEi = nhU i
(11.7)
So if the potential is gravity or electrostatic, then we know that n = −1 and Etotal = −hKEi
(11.8)
3 So if in a dilute system we know that P E ∼ − 2 N khT i. In an extreme relativistic case p · v → KE, which says that rest mass is ignorable to the energy, then hKEi = −hU i so Etotal → 0. So it is very difficult to form a bound system in this extreme case. In a classical system we can evaluate the expectation value as R dq1 . . . dp3N pi q˙j e−βH(p,q) hpi q˙j i = R (11.9) dq1 . . . dp3N e−βH(p,q)
∂H , so the numerator can be written as ∂pi Z 1 ∂ −βH ∂pi −βH =− e dq1 . . . dp3N pi e − β ∂pj ∂pj
But by the Hamilton’s equations we can write q˙i = Z pi
∂H −βH e dq1 . . . dp3N ∂pj
(11.10)
The former term is zero because it is a boundary term, therefore hpi q˙j i =
1 δij β
(11.11)
and hpqi ˙ = 3N kT = 2hKEi And this is what we have for ideal gas.
27
(11.12)
Statistical Mechanics
12
Lecture 12
Lecture 12
Remember last time we derived that for a confined system DX E DX E q˙i pi + qi p˙i = 0
(12.1)
Remember the first quantity is equal to 2hKEi in nonrelativistic case while it is equal to hKEi in relativistic case. Remember we have the relation DX E 3N kT = qi Fi (12.2) Now if we have a box of volume V of noninteracting particles. The force that the box is exerting on the particle averaged is just I Z r · P dS = P ∇ · rdV = 3P V = 3N kT (12.3) ∂V
V
So we recovered the ideal gas law. Suppose the gas particles have finite size and they are hard balls that can’t penetrate each other. We can rederive the Van der Waals equation N kT = P (V − b)
(12.4)
where b is proportional to the space occupied by a gas particle. Now suppose that the particles have nontrivial interactions. We have already seen this kind of problem, approximating the interaction to be harmonic oscillator potential or gravitational potential. We will write down the result first Z N (N − 1) ∂ 4πr2 3N kT = 3P V + r u(r) g(r) dr (12.5) 2 ∂r V where u(r) is the two particle potential, and g(r) is defined to be the probability of finding two particles with distance r. If there is no correlation in the positions of the particles then g(r) = 1. We may approximate g(r) = e−u(r)/kT which is a sensible way of approximation. Now we can write the integral out and integrate by parts to get Z ∞ PV = 1 − 2πN e−u(r)/kT − 1 r2 dr (12.6) N kT 0 If u(r)/kT 1 for r > 2σ, and u(r)/kT = ∞ inside r < 2σ, which is just the hard core repulsion, then Z ∞ PV 4π N u(r)4πr2 2 NV =1+ (2π) + dr + . . . (12.7) N kT 3 2 2V 2σ kT We can regroup the terms so that it looks like the Van der Waals equation N kT = P (1 − a)(V − b), where a is the integral above and b is half the excluded volume.
28
Statistical Mechanics
12.1
Lecture 12
Canonical Ensemble
Suppose we have n boxes, and the total energy of the whole system is the sum of energies of individual systems. We are dealing with indistinguishable particles, so if we take exchange the particles in one box with the particles in another box we can’t tell the difference. Every box has N particles. Now let’s use nr to denote the number of boxes in which the system has energy Er . This is the energy of the N particle interacting system. We can write X X nr (12.8) nr Er , n= Etotal = r
r
These two quantities are conserved throughtout the process in question. The collection {nr } = (n1 , n2 , . . . ) specifies a configuration of energy states of all boxes. The number of ways to find the whole system to be in a specific configuration is n! W ({nr }) = (12.9) n1 !n2 ! . . .
29
Statistical Mechanics
13 13.1
Lecture 13
Lecture 13 Continue on Canonical Ensembles
Remember last time we introduced the canonical ensemble as n boxes of volume V with particle number N . We used nr to denote the number of boxes in which the system has energy Er , so the total energy is X nr Er (13.1) Etotal = r
P with the constraint that nr = n. A distribution is a set of numbers {n1 , . . . , nr , . . . } which characterizes the energy state of the whole system. It is derived in the book that * + ∆nr 2 1 1 = (13.2) −o hnr i hnr i n This says that for very large n the gaussian distribution becomes very narrow and well defined. Remember we know that the number of ways to arrange the boxes for the same energy level is W ({nr }) =
n! n1 !n2 ! . . . nr ! . . .
(13.3)
The maximum of W corresponds to the most probable state. It is equivalent to finding the maximum of the logarithm: log W ({nr }) = n log n − n − n1 log n1 + n1 − n2 log n2 + n2 − · · · = n log n − n1 log n1 − n2 log n2 − . . . (13.4) In order to find the maximum, we differentiate it with respect to nr and we can get X δ log W = (log nr − 1) δnr
(13.5)
r
subject to the constraint that equation
P
δnr = 0 and X
P
δnr Er = 0. So we can add these zeroes to the above
(log nr − 1 + α + βEr ) δnr = 0
(13.6)
r
This is satisfied when nr = n∗r which corresponds to the most probable distribution, and since δnr is arbitrary apart from the constraints, it must be the terms in the bracket that vanish for all values of r. The α and β parameters are added to ensure this. So we know that the most probable distribution is n∗r = e−(α−1)−βEr
(13.7)
But if we divide by n then the factor of e−(α−1) cancels so we get e−βEr n∗r = P −βEr = Pr n e which is the distribution of a canonical ensemble.
30
(13.8)
Statistical Mechanics
Lecture 13
As long as the number of particles in each box N is finite, the above probability will fluctuate, as it is essentially an expectation value. For each box we have a volume V and particle number N , so its energy is P X Er e−βEr (13.9) Pr Er = P −βEr U = hEi = e r We can differentiate this expression with respect to β and we get ∂ hEi = −hE 2 i + hEi2 ∂β
(13.10)
So we get the expression which is useful 2
h(∆E) i = kT
2
∂U ∂T
= kT 2 CV
If we further divide by the square of expectation value then we get kT 2 CV kT h(∆E)2 i = ∼O hEi2 hEi2 hEi
13.2
(13.11)
V,N
(13.12)
Grand Canonical Ensemble
Now we want to generalize and consider the grand canonical ensemble. On top of the canonical ensemble, we add many boxes with different number of particles from N . We add the constraint that the sum of numbers of particles in all boxes is fixed. We follow the same procedure as above, but will arrive at the result X [log n∗rs − 1 + α + βErs + γNs ] δnrs = 0 (13.13) r,s
where Ns is the number of particles in the box s and the energy of a box s is Er (Ns , V ). So energy has two subscripts r, s. We can turn the crank and obtain the probability of a box to have N particles at energy level r(N ) is e−βEr (N,V )−γN PN,r(N ) = P −βE (N,V )−γN (13.14) r e This is the distribution for the grand canonical ensemble. Note that when γ is a positive number the most probable state is with 0 particle in the box. We want to make connections between the quantities we introduce here and the old familiar thermodynamic quantities. We call the sum in the denominator the grand partition function X e−βEr −γN = Q (13.15) r,N
Remember in canonical ensemble we have the relations X Q= e−βEr , hEi − T S = −kT log Q
(13.16)
We can also make connection to the grand canonical ensemble by defining the quantity q = log Q,
kT q = T S − hEi + µhN i 31
(13.17)
Statistical Mechanics
Lecture 13
where γ = −βµ. Bet remember that N µ = G = U − T S + P V . So we get q=
hP iV kT
Remember our microscopic definition of entropy X S= PN,r log PN,r
(13.18)
(13.19)
N,r
We also have the expressions for U and P which actually look similar X X ∂EN,r (V ) U= EN,r PN,r , P = PN,r − ∂V N,r
N,r
32
(13.20)
Statistical Mechanics
14
Lecture 14
Lecture 14
Remember last time we introduced the quantity q which is the logarithm of the grand partition function Q X q = log e−γN e−βEr (N,V ) (14.1) N,r
Its differential is dq = −hN idγ − hEidβ + hP idV
(14.2)
d (q + γhN i − hP iV ) + hEiβ = γdN − V dP + βdhEi
(14.3)
We can do some massaging to get
Remember the Gibbs free energy G = hN iµ = hEi − T S + P V
(14.4)
dhEi = T dS − P dV + µdN
(14.5)
Therefore we know that We can compare this expression with equation (14.3) and end up with q=
T S + G − hEi , kT
q=
PV , kT
γ = −βµ
(14.6)
From the same logic we can get that hN i =
1 ∂q β ∂µ
We can get this by simply look at the definition of q. We can work out the fluctuation of the number density of the particles h(∆n)2 i kT ∂V =− hni2 V N ∂P T where ∂V /∂P is the compressability of the system.
33
(14.7)
(14.8)
Statistical Mechanics
15
Lecture 15
Lecture 15
Remember we introduced the grand canonical ensemble last time, and we defined the grand partition function X Q= QN eβµN (15.1) N
where QN is the partition function for the canonical ensemble with N particles. Remember we have log QN = −A/kT
(15.2)
where A is the Helmholtz free energy of the canonical ensemble. However with grand canonical ensembles we have hP iV log Q = (15.3) kT which we also derived last time. Now let’s consider an Einstein-Bose gas of noninteractive and indistinguishable particles. The number of particles in one particular single particle state s has no limit. The expectation value of the number of particles in state s is just P∞ −Ns (s −µ)β 1 Ns =0 Ns e hns i = P = (15.4) ∞ −µ −Ns (s −µ)β e s −1 Ns =0 e In the book they took the limit where particles are far away from each other and there is little chance for overlap of wavefunctions, then the above distribution reduces to the so-called Maxwell-Boltzmann distribution. But that’s not interesting as in condense matter physics there are few scenarios where this assumption is really applicable. The probability for a single particle state s to have ns number of particles is ns e−ns (s −µ)β hns i 1 P P (ns ) = ∞ = (15.5) −ns (s −µ)β 1 + hns i 1 + hns i ns =0 e This was accomplished summing the geometric series. It can be checked that The book introduced a new variable, the fugacity, which is just z = eµβ
P
P (n) = 1.
(15.6)
For dilute gas µ is usually a very large negative number, so z will be very small. Because µ is always negative for a Bose-Einstein gas, so z is always less than 1. We can write the grand partition function as X Y 1 −β1 n1 −β2 n2 −βs ns Q= (ze ) (ze ) . . . (ze ) ··· = (15.7) 1 − ze−βi n n ...n ... 1 2
i
s
We introduced the logarithm of the grand partition function q, and we can compute it and find it to be q=
1X PV = log 1 + αze−β kT α
where α = −1 for Bose-Einstein gas and α = 1 for Fermi-Dirac gas.
34
(15.8)
Statistical Mechanics
Lecture 15
Remember for indistinguishable particles we need to account for overcounting by dividing the N -particle partition function by N ! QN QN = 1 (15.9) N! but this is true as long as we know that ns 1 for all states s. The factor needs to be changed if the gas is not dilute. Consider a Bose gas in a potential well. The ground state is the state where all particles are in the ground state. This statement also applies to distinguishable particles. In fact the Bose-Einstein gas has the same ground state as when the particles are distinguishable, as long as the Hamiltonian is symmetric among the particles and the ground state is unique, regardless of there being interaction or not. So it is not the ground state that differentiates the Bose gas and a gas of distinguishable particles. However there is profound difference in the excited states. For example, if we have N distinguishable particles. The first excited state of the system has N different possible configurations, corresponding to the excited states of the N individual particles. However for the indistinguishable case we only have one first excited state, i.e. a state where exactly one particle is in its first excited state. This is an enormous suppression of the excited states when the number of particles is large, which is usually the case in condense matter systems. Let’s consider a quantum mechanical system of 2 particles of Bose statistics. One excited state can be 1 ψ(x1 , x2 ) = √ eik1 x1 eik2 x2 + eik1 x2 eik2 x1 (15.10) 2 The probability distribution is |ψ(x1 , x2 )|2 = 1 + cos (k1 − k2 ) (x1 − x2 )
(15.11)
So when then particles are on top of one another then the probability is actually larger. This means that for excited particles they tend to come close to each other. So if we add a compulsive potential to the particle interaction, then the excited state acquires an energy gap because they tend to be closer to get repulsed. This effectively introduces an energy gap between the ground state and the first excited state. By the reasoning above, for a Bose gas the particles prefer the stay in the ground state compare to the Maxwell-Boltzmann gas, firstly because there are much less excited states in Bose gases, and secondly because there is an energy gap between the ground state and the excited state. We can evaluate the expectation value of the total number of particles X X 1 (15.12) hN i = hn i = −1 β z e −1 We know from this expression already that z must be less than 1, because for the lowest energy level = 0 then if z were greater than 1 then we have a negative number of particles, which is not physical. We can replace the above sum with an integral Z ∞ X −→ dg() (15.13)
0
where g() is the density of states(DOS) of the system, and for Bose gas we know that Z ∞ 4πV p2 dp 1 z hN i = + 3 −1 εβ (2π~) 1−z z e −1 0 35
(15.14)
Statistical Mechanics
Lecture 15
where the last term is added to give credit to the ground state where = 0. Thus the first term is explicitly the number of particles in excited states. If we look at the first term alone, we want to make it as large as possible, that is to make z as large as possible, which is at most 1. If z approaches 1 then the term becomes Z ∞ 2 4πV p dp (15.15) 3 β (2π~) 0 e − 1 which can be evaluated by change of variables, assuming that ∼ p2 , and it turns out to be finite. So there is a maximum value of number of particles in the excited states. If we put more particles in, then these particles have nowhere to go except into the ground state. If we do the same thing in 1-dim, then what happens? The above integral diverges when µ → 1, so there is no problem for unlimited particles into the excited states. How about 2-dim? We need to work harder and it needs special studies. So if we want to talk about Bose-Einstein condensation, we need to specify the space dimension, as well as how does depend on p.
36
Statistical Mechanics
16
Lecture 16
Lecture 16
Recall that we have the following results for the Bose-Einstein gas in 3D 1 hni = (−µ)β e −1 Z ∞ V x1/2 N = ¯3 λ 0 z −1 ex − 1
(16.1) (16.2)
where z = eµβ is the fugacity of the system, and it has to be less than 1, otherwise the whole machinery breaks. The wavelength in the second formula is ¯=~=√ ~ λ p 2mkT
(16.3)
The integral is usually hard to evaluate analytically, and we usually need to look up a table. The z factor effectively tunes the relevance of the integral. Note that the x 1 part is always irrelevant because they get suppressed by e−x . When z = 1 the integral converges, so if we put more particles into the system there is nowhere to go for the extra particles except for the ground state. This is the cause for Bose-Einstein condensation. We can write out the ground state contribution explicitly Z ∞ V x1/2 1 N = ¯3 + (16.4) −1 x λ 0 z e −1 1−z Now the integral means the number of particles in the excited state. If we divide both sides by V , then the number density is a fixed number plus the ground state contribution. The ground state is suppressed by 1/V , and is relevant only when z is very close to 1. The temperature dependence of the total number of particles is N = T 3/2 κ + NGS
(16.5)
The temperature for Bose-Einstein condensation Tc is when the temperature is so low that the particles are just able to be in the excited state. NE /N
T /Tc
Figure 16.1: Illustration of Bose-Einstein Condensation Note when T < Tc then NE /N ∝ (T /Tc )3/2 . The critical temperature is estimated to be (2π~)2 kTc = 2πm
37
N 2.612V
3/2 (16.6)
Statistical Mechanics
Lecture 16
If we compute the pressure of the system, we should get 2U (16.7) 3V This result can also be obtained without doing the integrals, as long as we have non-relativistic particles, by just arguing about the velocity and momentum transfer at collisions. For a relativistic particle we should have U P = (16.8) 3V The specific heat of the system looks like this. The specific heat has a discontinuity at T = Tc . P =
CV N
1
T /Tc
Figure 16.2: Specific Heat for a Bose-Einstein Gas ¯3N . We can also plot the fugacity of the system vs the quantity V /λ
z
1/2.612
¯3N V /λ
Figure 16.3: Fugacity for a Bose-Einstein Gas Now we want to study the phonons in solids. How much of the result here can be translated to condense matter systems? We want to ask when can we treat it as a continuum. The answer is just when the wave length of the phonon is larger than the interparticle spacing we can treat the background as a continuum. That is c (16.9) kTD ∼ ~ω > ~ a where a is the interparticle spacing. In this regime we can copy our results for photons directly to phonons, except for that the velocity is changed to the speed of sound in the solid. Remember the energy density for black body radiation is 3 kT U/V = kT (16.10) ~c 38
Statistical Mechanics
Lecture 16
There are only 2 polarizations for photons, but 3 for phonons. Plus the energy density is suppressed by 1/c3 , so the energy density of phonons in solid is much higher than that of photons, so we usually neglect the effect of black body radiation inside the solid. If we do the calculation more carefully we can get U = V
Z 0
kmax
~ω V 4πk 2 dk e~ωβ − 1 (2π)3
(16.11)
where kmax is defined as ~cs kmax = kTD
(16.12)
where TD is where our approximation breaks, and is the Debye temperature. Now we move to Chapter 8 of the book. We already know the occupation number ns of a single particle state of energy s for a fermion is 1 hns i = −1 βs (16.13) z e +1 Now there is no constraint on z. It can be larger than 1 or less than 1. The total number of particles is N=
X X hns i = s
s
1 z −1 eβs
+1
(16.14)
The quantity q can be calculated as X VP = log Q = log 1 + ze−βs kT
(16.15)
Remember when we deal with Bose-Einstein particles we can sum the logarithm and get a geometric series which is easy to express in a closed form. But now we only have two terms in the logarithm corresponding to occupation number 1 or 0. We want to see what happens for very low temperatures in the case of Fermi-Dirac gas. For very low temperatures, the term eβ will become very large, so we have problems unless we also have very large z. The largest energy in the fermi system is the fermi energy. It makes sense because for below fermi energy then the occupation number goes to 1, whereas when > F then at β → ∞ limit the occupation number goes to 0. Now because of Pauli exclusion principle, there are far more particles in the excited states than when we don’t have Fermi-Dirac statistics. So now P V /N kT ∼ 1 + . . . instead of 1 − . . . which is the case for Bose-Einstein gas. The average energy per particle is # " U 3 5π 2 kT 2 = F 1 + + ... (16.16) N 5 12 F When T = 0 we have the usual 3F /5. The first correction depends on (kT )2 because the energy is proportional to kT and the number of fermions that can play the game is also proportional to kT .
39
Statistical Mechanics
17
Lecture 17
Lecture 17
We want to study the fermions in greater detail. Usually we quantify our energy or kT by electron volts eV. 1 eV is about 104 K. We define the work function w to be the energy required it takes an electron at ground state to become a free electron. Because for typical systems the energy is actually much larger than the room temperature, so the system is essentially glowing. We can calculate the thermal emission Z Z Z dpy 1 dpz dpx 4πme R=2 uz (−µ)β = (kT )2 e−(w−f )β (17.1) 1/2 2π~ 2π~ e (2π~)3 (2mw) (2π~) +1 This result is not very trivial because the number density of electrons disappear in the final expression. We need to think about it deeper. Ref. the book.
17.1
Problem of White Dwarf Star
A white dwarf is a dense star where the electrons are highly degenerate. The internal energy is mainly from the kinetic energy of the electrons, because the ions have the energy around kT , but electrons because of their degeneracy have much higher energy at F . We will try to write down an answer by physics arguments. The potential of the star is described by the Newton’s equation ∇2 ϕ = 4πGρmass
(17.2)
∇P = −ρ∇ϕ
(17.3)
2 P = ne k 5
(17.4)
The gradient of the pressure is At T = 0 the pressure would be
The kinetic energy of the electron is k = p2 /2me for non-relativistic motion and k = pc for relativistic. But when electrons are relativistic the white dwarf has a maximum mass. The questions we want to ask are: What is the radius of the star? How does the radius depend on the mass of the star? What is the maximum mass of the star? Recall the Virial Theorem for inverse square law forces. We have a relation between the kinetic energy and potential energy ( K = − 12 U, for non-relativistic case (17.5) K = −U, for relativistic case The potential energy of the star itself can be written as U =−
GM 2 R
(17.6)
This is sloppy, but we are allowed to make approximations and this is accurate to a factor of order unity. The kinetic energy is the sum of all electron energies 3 K = Ne F , 5
where
40
F =
~2 2/3 n 2me e
(17.7)
Statistical Mechanics
Lecture 17
Combining the expression of K and U and the virial theorem we can compute the mass of the white dwarf and get Msun 1/3 ~2 9 ∼ 10 cm (17.8) R∼ 5/3 M M 1/3 Gme mHe For a neutron star, the idea is similar but the particles providing the degeneracy kinetic energy are neutrons. So we should have RNS me MWD 1/3 ∼ (17.9) RWD mn MNS Note that the fermi energy of the electrons in the white dwarf goes like 2/3 M 2/3 ∼ M 4/3 (17.10) F ∼ ne ∼ R3 So when the mass is large enough the electrons will go relativistic. In extreme relativistic case we have the following 1/3 Ne GM ∼ Ne ~c (17.11) R R3 Note for extreme relativistic electrons K = −V the total energy would be zero. So the whole system is not stable. From the above equation we can find an expression for the mass r ~c ~c 2/3 M ∼ Ne ∼ mH (17.12) G Gm2H We can call the quantity αG = Gm2H /~c the gravitational fine structure constant because it is the attraction between two protons. This is the Chandrasekar limit for the white dwarf mass. The mass can be smaller because then it will be in the non-relativistic regime, but can’t be larger. There is a mass limit for neutron stars, too, but it is of a different origin. This mass is also a characteristic mass for astronomical objects which are heavy enough to be stars.
17.2
Heavy Atoms
Let’s also look at heavy atoms. It is very similar to a star, except that the positive charge are all lumped in a heavy nucleus. Otherwise it is very similar to what we have done because the electrons also form a degenerate fermi sea. This is called a Thomas-Fermi atom. The potential here is the Coulomb potential Ze2 (17.13) r We take the radius of the atom to be the average radius of electron orbits. Then kinetic energy of the electrons can be approximated as 2/3 2 N ~ (Ze)2 N ∼ (17.14) r3 2me r ϕ=
Note we got the kinetic energy by p2 /2me and p is about ~ divided by the characteristic length of the orbit. From this expression we can get the so-called Thomas-Fermi radius rTF ∼
1 ~2 me e2 Z 1/3 41
(17.15)
Statistical Mechanics
Lecture 17
We can ask what is the maximum mass of an atom. Of course it can’t be too big because we have fission. But we can also do something like the white dwarf one. What would that value of Z be? If the electrons become extremely relativistic we have 1 ~ ~2 → me e2 Z 1/3 me c
(17.16)
This places a limit on Z just like the case of white dwarf, which is Z 1/3 ∼
~c ∼ 106 e2
(17.17)
which is ridiculously big. But similar things happen in the Dirac equation where things scale differently than here. There when Z is larger than 137 the electrons become relativistic and there is no solution.
42
Statistical Mechanics
18
Lecture 18
Lecture 18
18.1
Paramagnetism
Consider a particle in a static magnetic field B. The energy is just −µ · B. The partition function is Q1 =
J X
exp (βmgµB B)
(18.1)
m=−J
And for dilute N particles we have QN 1 N! The classical approximation is to replace the sum with an integral Q=
J X m=−J
Z
J
−→
Z
1
dm = −J
(18.2)
d cos θ
(18.3)
−1
Carrying out the integral we can find the magnetization for N particles to be 1 = N µL(x) M = N µ coth(βµB) − βµB
(18.4)
where L(x) is called the Langevin function and x = βµB. For small x the magnetization is N µ2 B M= (18.5) 3kT This is a paramagnetic response. However we can also have a spontaneous magnetization if the temperature is low enough. In that case B = M/V is the field generated by the material itself. This is the mean field approximation where B is taken as a local average of the field generated by nearby atoms. This spontaneous magnetization disappears at high enough temperature, when 1 = N µ2 /3V kT . This temperature is called Curie temperature and we write N µ2 kTc = (18.6) V 3 For a large solid if we lower the temperature below Curie temperature we will see the material is selfmagnetized in small patches, which are called domains. The magnetizations in different domains arrange themselves so that as little magnetic field stick out of the material as possible. In general this is very complicated. Consider a Fermi sea of electrons in a potential well. The energies fill up to F . At zero temperature all the states below F are filled while all the states above it are empty. If we put a magnetic field into this potential well, depending on the spin orientation of the electrons inside, the energy will be raised or lowered. Ref. the textbook. We could use the chemical potential for this problem and write µ↑ + µB B = µ↓ − µB B
(18.7)
in equilibrium. If we find the chemical potential for a degenerate fermi gas then we can find that the induced magnetization divided by B, which forms a dimensionless quantity, is equal to M N µ2B N λ3 = 1− + ... (18.8) VB V kB T V 23/2 43
Statistical Mechanics
Lecture 18
This is for kB T F . For the other end of the spectrum kB T F we have " # kB T 2 π 2 N µ2B M 1− = + ... VB V F F 12
18.2
(18.9)
Diamagnetism
For a closed shell atom like a Helium, if we apply a magnetic field, there will not be any net paramagnetism. Recall in elementary quantum mechanics we deal with the problem by replacing e p −→ p − A c
(18.10)
in the Hamiltonian and solve for the eigenvalues. For a constant homogeneous magnetic field we choose A to be 1 (18.11) A= B×r 2 The expectation value for energy is 2
2 e2 p + hp · A + A · pi + r⊥ B2 (18.12) hHi = 2m 8mc2 The second term is zero, and the last term is the correction ∆. Its sign is positive which tells us that this is diamagnetism. Now let’s take the atom and smear it out into a larger volume. If we take it large the system goes to the limit of a box of free electrons. In this case the above energy correction gets larger and larger because 2 i gets larger and eventually it will be larger than the binding energy itself. hr⊥ Don’t want to take notes. . . Ref. Pathria Problem 3.43 and page 208. The Bohr-van Leeuwen Theorem says that with a magnetic field does not change the velocity distribution for classical free electron gas. There will be no diamagnetic response for classical electron gas. Consider a box of electrons under magnetic field. The classical electron diamagnetism will give rise to small current loops that repels the magnetic field. If we add up all the current loops we get a big current loop around the edge of the box. This seems to a diamagnetic response, but if we also consider edge electrons which only respond with a partial current loop the contributions cancel. So this is another way to see classical diamagnetic response is zero.
44
Statistical Mechanics
19
Lecture 19
Lecture 18
Given a box of particles, we want to ask what is the magnetic susceptibility. Remember the Hamiltonian is 1 e 2 H= p− A (19.1) 2m c The quantum diamagnetic contribution to the energy is
e2 H = r2 B2 8mc2
(19.2)
Now we want to know what is r2 . We can have a guess of it using the uncertainty principle r∼
~ , mv
r2 ∼
~2 ~2 = m2 v 2 mK
(19.3)
where K is the kinetic energy of the electrons. Again we must make a choice what is K . It could be kT if the electron gas is dilute, or could be the Fermi energy F if the electron gas is dense. Then the diamagnetic susceptibility is M e~ 2 1 = −ne (19.4) VB mc kT or replace kT with F . Now we use quantum mechanic ways to do this problem. We need to choose a gauge to write down our vector potential. The naive choice is A = B · r/2. But we choose another gauge which is called Landau gauge, where ˆ Bx A=y (19.5) Then the Hamiltonian is
(py − eBx/c)2 p2x p2 + + z 2m 2m 2m We can readily write down the eigenfunctions H=
Ψ (kx , ky , kz ) = eikz z eiky y ψ(x − x0 )
(19.6)
(19.7)
The motion around x0 will look like a harmonic oscillator, and x0 is about x0 ∼
~c kx eB
(19.8)
The final eigenvalues of the energy are 1 p2 j = ~ωe j + + z 2 2m
(19.9)
Now that we know the energies of the single particle states, we want to know how many possible configurations correspond to the same energy. We can write the density of states as dx dy dpx dpy dn = d (2π~)2 45
(19.10)
Statistical Mechanics
Lecture 19
Now we can carry out the integral over x and y and combine the dpx dpy into 2πp dp. We can recognize this later expression is just m d, and we can write ∆ = ~ω, so we have Lx Ly dn = 2πm~ω d (2π~)2
(19.11)
where ω = eB/mc. We can think about the problem in another way. We can think about the total flux of the magnetic field through the region as Lx Ly B, and we want to put in the electron orbits and ask how many electron orbits can we put into the region. The flux quanta is eΦ0 = 2π~ c
(19.12)
and we know n = BLx Ly /Φ0 . We argue that for higher and higher values of j, the extra flux introduced by increasing j by 1 is fixed, so that we can count the number using this unit flux. Now with all the knowledge, we want to compute the grand partition function of the system. Recall X log Q = log 1 + ze−β (19.13)
We can calculate hN i and hM i from the grand partition function. Results ref. Pathria Chapter 8. The result is not very different from our heuristic result from the above, apart from a constant factor.
46
Statistical Mechanics
20 20.1
Lecture 20
Lecture 20 Superfluids
Recall the fluid equation ∂v ∇P + (v · ∇)v − + η∇2 v = 0 (20.1) ∂t ρ For superfluid η = 0 and there is no viscosity. Superfluidity is found in He at very low temperature. If the lattice spacing for the fluid is a, then the kinetic energy for localization is E∼
(~/a)2 2m
(20.2)
In order to understand superfluidity, we first need to understand what are the forces between particles in the fluid. When the forces are zero, we have the familiar result from Einstein-Bose statistics 2/3 N (2π~)2 ∼ 3.13 K (20.3) kB Tc = 2πm V · 2.612 Note for He superfluidity happens for T < 2.2 K. So we know superfluidity is related to but does not come solely from the Bose condensation effect. Recall the Van der Waals interaction between particles. Consider an atom in a uniform electric field. In equilibrium the atom will be stationary and the electron cloud will be polarized. The electrons will form a dipole such that the nucleus in the middle will experience no net force. The dipole moment is about ed ∼ a3 E0
(20.4)
where a is the size of the atom. Now suppose we have two atoms 1 and 2 separated by distance r. The Van der Waals potential between them due to polarization is V12 = −
e2 a32 a21 a6 e2 ∼ − r6 r6 a
(20.5)
The first ratio (a/r)6 is about 10−6 in order of magnitude, and the second ratio is about e2 /a ∼ 102 eV. So the attraction by Van der Waals force is about 10−4 K in strength. Now the potential is attractive for larger distance but increase rapidly for smaller distances. It is hard to see what to neglect in our approximation. Consider a box of volume V and N bosons inside. Suppose the interaction between particles is a weak attraction of range approximately a, with potential energy about −u0 . We first put all particles at the ground state so the total energy is 0. Now we want to put every particle to squeeze into a small box of size a. The energy corresponding to this configuration is EN =
~2 N (N − 1) N− u0 2ma 2
(20.6)
This says that if there is no repulsion, the energy will be lowered by the particles all coming together. We we need to add a repulsion to prevent the particles to all fall together. We introduce a pseudopotential to represent repulsion. Suppose we add a weak repulsion u(r) > 0. The ground state again has zero kinetic energy, but because there is repulsion, the energy is actually EGS = 0 +
N (N − 1) w 2 V 47
(20.7)
Statistical Mechanics
Lecture 20
R where w = u(r)dV is the integrated interaction. Now what about the excited state? Consider a 1-particle excited state which has momentum p. The energy is now Ee1 =
wk p2 + N (N − 1) 2m V
(20.8)
R where wk = 4π r2 dr cos(k · r)u(r), and k = p/~. The effect of repulsion is almost twice as big as the repulsion in ground state. So excited states cost energy. If there are only two particles we can write out the wave function explicitly. Assuming periodic boundary condition, the ground state wave function is ΨGS =
1
1
L3/2
L3/2
(20.9)
whereas the 1-particle excited state is eik·r1 1 1 eik·r2 = (20.10) L3/2 L3/2 L3/2 L3/2 Now if the particles are bosons then there is only one excited state which is the symmetric combination of the above two states ik·r1 1 e 1 1 eik·r2 ΨBE = √ + (20.11) 2 L3/2 L3/2 L3/2 L3/2 The probability for finding the particles at distance r12 is, in the case of Einstein-Bose particles Ψe1 =
1 (1 + cos k · r12 ) (20.12) L3 So there is twice more probability for the particles to be close to each other. This is another factor that prevents the formation of excited states when there is repulsion. So the one-particle excited state has the spectrum p2 E= +∆ (20.13) 2m where ∆ is a kind of gap between excited state and the ground state. It has a lowest point at some p0 which shows the tendency of forming a lattice of that spacing. However the phonon spectrum is gapless, and the dispersion relation is linear Ephonon = cs k (20.14) |ΨBE |2 =
Now consider a particle going through the liquid at velocity V0 . Its energy is E = V0 · P0 /2, so the energy change is δE = V0 · δP0 . Now if the dispersion of the particle (which is just a straight line) lies lower than the lowest excitation states in the liquid, including both phonon and electrons, then the particle will not see any dissipation and the liquid will turn out as if it is a superfluid. Suppose we can describe the electrons in the fluid by a wave function ψ0 = |ψ0 | eiϕ . The density of particles is just ρ = N |ψ0 |2 (20.15) We define the velocity as V=
~ψ∇ψ ∗ − ~ψ ∗ ∇ψ 2µψ ∗ ψ
(20.16)
~ ∇ϕ m
(20.17)
and we have V=
48
Statistical Mechanics
21
Lecture 21
Lecture 21
Now consider the many-electron wavefunction of the form ψ = |ψ| eiϕ
(21.1)
We can define the “velocity” of a particle inside the many body system as v=
~ ∇ϕ m
(21.2)
Note that the velocity is a gradient, so its curl vanishes. Now because ϕ is a phase factor, the integral of v over a closed loop must be quantized if there is some matter distribution inside the loop. I 2π~ (21.3) v · ds = m Now recall the fluid equation ∂v ∇P + (v · ∇)v − + η∇2 v = 0 ∂t ρ
(21.4)
We define the following quantity Z π=
dP ρ
(21.5)
When there is no dissapation, we have ∂v ∇P Dv = + (v · ∇)v = Dt ∂t ρ
(21.6)
There are many interesting things for superfluids with ∇ × v = 0. We want to study the vortex lines in superfluids. A vortex line must start and end on the boundary of the fluid, or just exist as a loop. First we introduce a theorem I D v · ds = 0 (21.7) Dt We show this is true by evaluating I I Z D D Ds v · ds = v · ds + v · d Dt Dt Dt
(21.8)
The first term is just the integral of the gradient of π, so it is zero. Now we consider the second term Dr ∂r = + (v · ∇)r = 2v Dt ∂t Now the second term becomes
I
I v · d(2v) =
49
∇(v2 ) · ds = 0
(21.9)
(21.10)
Statistical Mechanics
Lecture 21
Suppose we have a vortex line along the z direction and we consider the motion on the x-y plane. If the velocity is around the vortex line, then it must be v=
~ ϕ mr
(21.11)
However if there is a vortex line and the flow is going pass it at velocity v, we want to know how the vortex line itself moves. We still have the quantization condition (21.3). The force on the vortex line is Fvortex = 2π
~ ˆ ρ` × vδ 2 (r) m
(21.12)
where ˆ` is the direction of the vortex line. This says that when there is a flow past the vortex line, there will be a force perpendicular to the vortex line and the flow direction. This is called the Magnus force. If we have a vortex line, there will be some tension on it, because we need to do work to stretch it longer. The tension is Z Z rmax v2 r dr 2π~2 rmax T = 2πr dr ρ = ρ ∼ log (21.13) 2 2m2 r2 rmin rmin The tension is related to the Magnus force ~ 2T = F = 2π ρv R m
(21.14)
so the velocity of the vortex line is v∼
~ mR
(21.15)
except for a logarithm factor of pure number. Consider a large container with a small hole connecting to the outside. Now the phonon excitation of long wavelengthes can’t go through the hole, because it can’t even see it. Everything flowing out of the small hole must be excitationless, so it is a 0 temperature fluid with no entropy. The entropy in the bulk is unchanged and volume decreased. But there is one kind of excitation that can occur when the fluid is coming out, which is vortex excitation. Now if the vortex loop is moving at a velocity comparable to the velocity of the fluid, we can get an estimation of the radius of the vortex loop, and we know Eloop ∼ ρR3
~2 , m2 R2
p ∼ ρR3
~ mR
(21.16)
ˆ = 0. We can create an Now consider a vortex near a solid boundary with boundary condition v · n image vortex at the other side of the boundary with the same magnitude and opposite direction, so that the normal component of velocity at the wall is zero. Thus we can calculate the velocity of fluid near the vortex. We usually think about vortex lines of minimum excitation, because for higher excitations say 2, then the kinetic energy will be 4 times as large, and it pays energetically for the line to become two separate vortex lines. So the vortex lines tend to repel each other when they are very close. Consider a superfluid in a container, and we spin the container at angular velocity Ω0 . If it were a normal fluid we know we will have a curved surface in the shape of a parabola, with ∇ × v = 2Ω0 50
(21.17)
Statistical Mechanics
Lecture 21
For classical fluids we have a lot of ways to communicate the motion of the container to the fluid, but it is not the case for superfluids. It is convenient to change into the comoving frame and minimize the energy there, and then transform back to the lab frame. The quantity E − Ω0 · L =
IΩ2 − Ω0 ΩI 2
(21.18)
is what we want to minimize. The minimum is at Ω = Ω0 with magnitude −Ω20 I/2. For a superfluid what happens is that the motion of the container creates vortex lines in the fluid, and the number density of the vortex lines is Ω0 m nV = (21.19) ~π It mimics classical fluid rotation, with the same parabolic surface, but it achieves it by creating a large array of vortices which comoves with the rotation of the bucket. If we don’t have a magnifying glass, the motion won’t be very different from classical motion, but microscopically it will be quite different.
51
Statistical Mechanics
22
Lecture 22
Lecture 22
Last time we did the superfluid in a rotating bucket. The way we think about this formally is to minimize H − L · Ω0
(22.1)
and it is equivalent to minimizing the following Ω2 I − Ω0 ΩI 2
(22.2)
The physics happening in the rotating bucket is as follows. When you start to rotate the bucket, the superfluid does not respond until a critical velocity is reached. At that time a vortex line is created somewhere and in the steady state will correspond to the vortex line stay in the middle of the fluid. The kinetic energy of the vortex line is Z
R
K= ∼0
1 2π ρN 2 L2πr⊥ dr⊥ = ρL 2 2
~2 m
2 Z 0
R
1 r
2 r dr
(22.3)
And we know that Ω0 L = Ω0 (~N )
(22.4)
When these two quantities are equal we have a critical point and the vortex line is formed. Then after a while we have another vortex line, and after a while we have another. When there are more vortex lines, the behavior of the fluid mimics that of the classical fluid. The critical angular velocity for creating one vortex line is ~ 1 R Ωc = log (22.5) 2 mR a where a is about the atomic separation. We want to know the number density of vortex line lattice in the bucket when it is spinning at Ω0 . The total circulation is I 2π~ v · ds = 2πR2 Ω = NV (22.6) m So if we divide by the total area we can get nV =
Ωm NV = 2 πR π~
(22.7)
Note that this number density does not depend on the density of the underlying material, only the fundamental mass of the atoms. Vortices in the lattice talk to each other even at long distances, because the velocity around a vortex line drops at 1/r which is slow. So the whole lattice co-rotates as a body. But there is some subtlety because there will be some variation in speed of the vortex lines. Another question to ask is what kind of lattice will the vortices form. The answer is that they will form a hexagonal lattice. Then we can introduce perturbation onto the lattice and produce sound waves just as we have sound waves in solids. We can calculate the velocity in the lattice vT ∼
~ masep
52
(22.8)
Statistical Mechanics
Lecture 22
where asep is the separation of the vortex lattice. This velocity is extremely small, so all the waves are at very large wavelengthes, they don’t contribute much to the heat capacity of the substance. Now suppose we have an irregular container with superfluid in it. If it is spinning at Ω0 then we will have some vortex lines in the fluid. Now if we are spinning down, the vortex lines will move out and disappear at the edge, so there is no problem. But suppose the container is not smooth and there is some place where the separation between upper and lower edge is small, the vortex line will be pinned there because it takes energy for it to move elsewhere. Remember the tension is K ≈ 10−9 L
Tv =
(22.9)
Now if T ∼ 1 ◦ K then the ∆s for pinning the vortex line should satisfy ∆sTv kT
(22.10)
If this is satisfied then termal excitation will not have enough energy to move the vortex line, so the line will be stuck when the container is spinning down. Remember in a material we have the dispersion for sound waves ω = cs k
(22.11)
The energy of the excitation looks like black body radiation kT 3 E ∼ kT ~cs
(22.12)
The number density of phonons goes like 1 c3s
ns ∼
(22.13)
Because cs c the number density of phonons is enormous compared to the number density of photons. You don’t need to worry about the blackbody radiation contributing to the heat capacity of the solid. The scattering cross section between photons is neglegible compared to the scattering cross section between phonons. So the blackbody radiation just bounces off the wall and don’t scatter off each other, and its mean free path is just the dimension of the box. But the mean free path of phonons is much much shorter, because of the cross section and much larger density. Now in superfluid the only excitation is the phonon excitation and the phonon collides with nothing but other phonons. There is no other degree of freedom except the number density of phonons. We can think of it as an interacting gas and disturbance will travel out at a speed different from cs , which we call the speed of second sound. The pressure in a phonon gas is related to energy density by u =P 3
(22.14)
u c22
(22.15)
and we have the mass distribution ρ=
where c2 is the speed of second sound. Combining equations we can get c22 =
dP c2 = s dρ 3
=⇒ 53
cs c2 = √ 3
(22.16)
Statistical Mechanics
Lecture 22
and this is the relation between speed of second sound and speed of sound. Suppose we have a small container in a big one with a impermeable membrane connecting the fluids inside. The membrane allows superfluid to go through but not phonons. Now in equilibrium the temperature in the smaller container T1 is greater than the temperature T2 in the bigger container. This is reminiscent of the osmotic pressure problem we introduced earlier in the course. We have P (T1 ) − P (T2 ) = ρSF gh
(22.17)
where h is the height difference of the two fluids. Now if we seal off the membrane, the superfluid can move from one container to another along the wall of the smaller container, while classical liquid cannot. When superfluid is exchanged this way, the total entropy and number of phonons are not changed in either containers. Now if T1 > T2 , energy can only flow from T1 to T2 , but here what is flowing is not energy, so we can have some “paradox-like” scenarios where fluid flow from T2 to T1 . Now what happens if we elevate the smaller container from the large one? The superfluid will still come out because of capillary motion. This is not because of He is small and permeates the container, but rather it moves along the wall of the container. The final scenario is a container with a small pipe at the bottom and we know phonon can’t escape while superfluid can. The temperature of the fluid inside the container will rise because no entropy is changed, and the container at the other end of the pipe will be lowered in temperature. This will lead to hot object getting hotter and cold object getting colder. Now we shift our attention to neutral Einstein-Bose gas to charged fermion gas. We want to study superconductivity treating it like a charged superfluid. In superconductors we know that no magnetic field will penetrate it. We want to study why. Recall the Maxwell equations. In a perfect conductor the force on a particle is eE = mv˙
(22.18)
j = env
(22.19)
and the current inside is just In a perfect conducter there is no damping and current will not dissipate. We combine the above two equations and Maxwell equations and so some algebra, then will get mc ˙ ∇ × 2 j˙ = −B (22.20) e n Discarding displacement current, we can replace j˙ with curl of B and we get 2 ˙ = 4πne B ˙ ∇2 B mc2
(22.21)
The coefficient in front is just the plasma frequency squared. Note that there are dots above B. Now suppose we are considering charged Einstein bosons and they are in a big ground-state wavefunction. The current has the form ∗ ∗ p − eA/c j = en(ψ0 vψ0 ) = en ψ0 ψ0 (22.22) m Remember quantum mechanics tells us that p = ~∇φ, and the curl of this quantity is zero, so we have ∇×j=− 54
e2 nB mc
(22.23)
Statistical Mechanics
Lecture 22
This lead to the same equation as above except we don’t have dots above B. This is the peculiarity of the charged quantum boson system. It is like integrating the above equation and select the one solution without any time independent extra term.
55