Äîêóìåíò âçÿò èç êýøà ïîèñêîâîé ìàøèíû. Àäðåñ îðèãèíàëüíîãî äîêóìåíòà : http://www.mccme.ru/dubna/2015/notes/simon-notes.pdf
Äàòà èçìåíåíèÿ: Tue Jul 21 18:55:00 2015
Äàòà èíäåêñèðîâàíèÿ: Sun Apr 10 12:41:08 2016
Êîäèðîâêà:
First Principles of Statistical Mechanics
M.Simon July 21, 2015
Abstract The theory of thermodynamics aims at understanding how two types of energy - mechanical and thermal - can be converted into each other. At the end of the XIXth century, the Austrian physicist Ludwig Boltzmann stated that the laws of thermodynamics should be derived from Newton's mechanical first principles, on the basis of the atomistic theory of matter. A gas, for instance, can be represented as a collection of atoms - or point particles - moving under Newton's laws. Mechanical systems are mainly characterized by geometric quantities, such as the positions and the velocities of its mass points. However, this description completely fails for gases, which have properties that may interfere with their mechanical characteristics, in particular their temperature. The kinetic theory of gases has turned into what we know as statistical mechanics through the work of the American mathematical physicist J. Willard Gibbs. In this context, finding the equilibrium value of macroscopic variables (like temperature) amounts to computing a probability distribution. Such a probability law depends on a finite number of parameters, and describes the possible states of the system composed of a huge amount of particles. This introductory course will take a mathematical look at the first principles of statistical mechanics, and will aim at: · defining the notion of a thermodynamic system of particles; understanding the natural probability distributions that characterize such systems, namely the micro-canonical ensemble and the Boltzmann-Gibbs distribution; · proving (at least in simpler cases) the theorem of equivalence of ensembles, which justifies the use of the Boltzmann-Gibbs distribution to compute thermodynamics quantities; · asking some mathematical questions (some of them still unsolved) that naturally arise in this area.

1 Introduction
Classical physics mainly relies on the following assumption: if the state of the system is precisely known in one instance, the laws of physics determine its future states completely. Therefore, why and how to introduce probabilities into classical physics? Here are listed a few examples of physical statements whose precise formulation requires probabilistic concepts: 1. Water freezers at 0 degree Celsius. 2. When two bodies are in contact with one another, the heat flows from the warmer body to the colder one. 3. When the heat of a container full of gas increases, the pressure on the walls of the contained increases too.

1


1

INTRODUCTION
The formulation of these statements should involve probability, for the following reason: one can indeed construct special scenarios that are consistent with the laws of physics where each of the above statements is false. These scenarios are physically possible but extremely unlikely! The most famous example of such a physically possible but highly improbable scenario was given by Maxwell: let A be an insulated contained that is divided into two boxes, B and C. Both boxes are filled with gas, and separated by a sliding door. Assume that the temperature of B is higher than the temperature of C. When the door is open, molecules may flow from one compartment to the other; when it is closed, B and C are completely insulated. Assume now that the door is intelligent enough to select the most improbable situations: · when the door sees a very "fast" molecule (with high temperature) approaching from C toward the door, it quickly opens and lets the molecule pass from C to B; · when the door sees a very "slow" molecule (with low temperature) approaching from B toward the door, it quickly opens and lets the molecule pass from B to C. This will cause a heat transfer from the colder body to the warmer one! Now, take a second container that is an exact replica of the first. This time, the door cannot influence the passage of molecules any more, but instead there is a random device that opens the door at irregular intervals. It may happen (very unlikely) that this device opens the door exactly at the same times as before! And the same conclusion follows: a heat flow is created from C to B. This second case is consistent with the laws of physics.

2

1.1

Thermodynamics

The purpose of thermodynamics is to describe the properties of various macroscopic systems at (or near) equilibrium. This is done with the help of a few macroscopic variables such as the internal energy E, the volume V, the pressure P, the number of particles N, the temperature T, the entropy S, and others. These variables are not all independent, and thermodynamics provides general relations among them. Some of these relations are not specific to particular systems, but are valid in a very general setting. However, because thermodynamics makes no link with the microscopic physics that is involved at the level of particles, it cannot give a complete picture. This is the role of statistical mechanics. In fact, thermodynamics could stand very well on its own. Statistical mechanics aims at extending and complementing its framework. We recall here the most famous laws of thermodynamics: 1. Thermodynamics variables can be divided into two groups, depending on their behavious under a scaling of the system: extensive variables scale with the system size (like E, S, V, and N), while intensive variables stay invariant (like T, P). 2. An isolated system is in thermal equilibrium if T is uniform throughout the system; it is in mechanical equilibrium if P is uniform. 3. The first law of thermodynamics states that the internal energy E of a system can be increased by either adding heat or doing work. Mathematically, dE



¯ dQ



¯ d W,

¯ ¯ where d Q and d W are inexact differentials, meaning that the integrals »B »B ¯Q and ¯ d dW
A A

depend not only on the end points, but also on the path taken to go from state A to state B. However, we have »B dE E B EA,
A

p q¡ p q

independently of the path.


1

INTRODUCTION
4. The second law of thermodynamics states that there exists a function S called the entropy, such that, ¯ for thermally isolated systems (such that d Q 0), it satisfies dS 0.

3



¥

1.2

A few notions of probability theory

The concept of probability is one of the foundations of general statistics. The likelihood of a particular outcome among the set of possible outcomes is expressed by a number from 0 to 1 (0 representing the impossibility and 1 representing the absolute certainty). Dealing with basic probability as a discrete counting process is satisfactory if we work with reasonably small numbers, like throwing dice or picking cards: in these cases the probability for a given event can be thought of as the ratio "the number of ways that event can happen" divided by "the number of ways that any possible outcome could happen". If the number of events is very large, as in the distribution of energy among the molecules of a gas, then the probability has to be approximated by a continuous variable. In the discrete case we denote by x i the probability that x i is observed. The normalization constraint reads as ¸ xi 1,

pq

p q
x
i

i

and the mean value of x is

xxy

¸
i

pxiq

.

If the outcome x takes a continuous range of real values, then the probability x takes a different character, and the sum above takes the form of an integral: the probability of finding outcomes x between a and b can be expressed as »b »b d f x dx dx. a a dx

pq

pq

The function f is called the probability density of the random variable. The normalization condition is then » f x d x 1,

pq
»

and the mean value equals

xxy pq r p

x f x dx.

pq

In other words, if d x is an infinitely small number, the probability that x belongs to x 0 , x to f x 0 d x : x x0, x0 d x f x0 d x .

p

0

dxq

is equal

qs p q

EX A M P L E 1.1 (Probability distributions). We give here the main probability densities that are used in statistical mechanics. 1. Taking a number x uniformly at random in the interval a, b means that its probability density is 5 1b a if x a, b , Ux 0 otherwise.

pq

p q {p ¡ q
¡x

r s

2. The normal distribution is widely used in various domains, and its density is defined on N 1 pxq c exp

as (1)

p ¡ µq
2
2

2

©

2

,

µ



,

¡

0.


1

INTRODUCTION
The normal distribution is remarkably useful because of the Central Limit Theorem1 . Physical quantities that are expected to be the sum of many independent processes (such as measurement errors) often have distributions that are nearly normal. When µ 0 and 1, N is called the standard normal distribution.

4





3. Perhaps the most widely used distribution function in classical physics is the Boltzmann-Gibbs distribution function, which describes the probability of finding particles with an amount of energy E 0 at a given temperature T: 1 ¡E{pkB Tq fE e , (2) Z where kB is the Boltzmann constant, and Z is the normalization constant which has to satisfy:

¥

p q
V
e

1 Z

»

¡E{pkB Tq

dE

0



1.

In other words, the probability that any one molecule will be found with energy E decreases exponentially with the energy.

1.3

Foundations of statistical mechanics

In what follows, a thermodynamic system should be thought as a gas enclosed in a container of a given (but possibly variable) volume V. This container provides the mean to couple the system to an external mechanical system. A general definition will be give in Section 3. D
EFINITION

1.1. 1. The macro-state of a thermodynamic system depends on a relatively small number of thermodynamic coordinates (for instance: pressure and temperature).

2. A micro-state corresponds to a microscopic description (of the degrees of freedom that compose the system, for instance: the position and velocity of each molecule of gas). The description of each microstate requires an enormous amount of information, and the corresponding time evolution is usually quite complicated. 3. Rather than following the evolution of an individual (pure) micro-state, statistical mechanics examines an ensemble of micro-states corresponding to a given (mixed) macro-state. In other words, macro-states can be characterized by a probability distribution of possible states across a certain statistical ensemble of all micro-states. Each thermodynamic state, say V, T, N for instance, corresponds precisely to one probability distribution fpV,T,Nq on the micro-states space.

p

q

p

EX A M P L E 1.2 (Tossed coin). A tossed coin can land in two positions: head up, or tail up. Considering the coin as a particle, one can say that this particle has two states: 0 or 1. If N coins are tossed, this can be considered as a system of N particles with two states each. The micro-states of the system are specified by the states occupied by each point: they are elements of 0, 1 N , there are 2N possible micro-states. The macro-states of this system are defined by the number of particles in each state, N0 and N1 , which have to satisfy the condition N0 N1 N. The number of micro-states associated to one macro-state N0 , N1 is given by ¢ N! N . N0 N0 ! N N0 !

tu

q







p¡ q

The main postulate of statistical physics is that the equilibrium macro-state is the one realized by the greatest number of micro-states, i.e. the most probable macro-state. For instance, a macro-state of an ideal gas with all molecules in one half of the container is much less probable than the macro-state with
1

Roughly speaking, it states that large averages of independent random variables are normally distributed.


1

INTRODUCTION
the molecules equally distributed over the whole container: this exactly corresponds to N0 N1 N 2 in the example above. More precisely, if the initial state is all molecules in one half of the container, then in the course of evolution the system will come to the most probably state with the molecules equally distributed over the whole container, and will stay in this state forever. One of the law of thermodynamics states that: if an isolated system is initially in a non-equilibrium state, then it will evolve to the equilibrium state characterized by the maximal entropy. More precisely, the entropy S and the number of micro-states associated to a macro-state should be related, one being a monotonic function of the other. If one chooses2 S

5



{



kB log ,

then the entropy has the advantage to be additive: if a system consists of two subsystems that weakly interact with each other then S S1 S2 while 1 2 .





1.4

Maximisation of entropy and Boltzmann distribution

We consider now a system of N particles with p states each (as a generalization of Example 1.2). Assume that, if a particle is in the state i 1, ..., p , then it has an energy i 0. Let N1 , ..., N p be a macro-state, namely: Ni is the number of particles in state i . The number of possible micro-states is given by N! . N1 ! N p !

t

u

¡

p

q



¤¤¤

Suppose that there is no energy exchange with the system: in particular, not only the total number of particles N is fixed, but also the total energy E of the system. We will define the notion of temperature by looking for the equilibrium macro-state. In that case, the equilibrium is achieved for the macro-state that contains the maximum number of micro-states. Therefore, the task is to find the maximum of 3 with respect to all Ni that satisfy
p ¸ i1



N

i



N,

p ¸ i1



N

ii



NE,

(3)

where i is the energy of the particle in state i . Practically it is more convenient to maximise log than itself. Consequently our problem reads
4
N1 ,...,N

pq

¡

max

log
p

N1 !

¤¤¤

N!

©

Np !

; such that

p ¸ i1



N

i



N,

p ¸ i1



N

ii



B

NE .

Using the method of Lagrange multipliers (see Appendix B), one searches for the maximum of the target function p p ¸ ¸ N1 , ..., N p log Ni Ni i .

p

q p q

i1





i1



We are interesting in the behaviour of the macroscopic system with N (and therefore N1 , ..., N p are very large). Using Stirling's formula4 , we only keep the first order in the asymptotic expansion of
2 3 4

ÿV

This is an idea from L.Boltzmann; kB is the Boltzmann constant. The letter comes from the German word Wahrscheinlichkeit which means probability. Stirling's formula states that ¢ N N N! 2N, as N . e



c

ÿV


1

INTRODUCTION
which reads as N1 , ..., N

6

p

p

q

log N!

¡

p ¸ i1



Ni log N

p i q p 1q


p ¸ i1



N

i





p ¸ i1



N

ii

.

Computing the derivatives with respect to Ni and solving its zeros, it yields N
i



e



i

.

The Lagrange multipliers can be found from (3). Finally, the probability that one particle is found in i th state is p ¸ Ni 1 ¡ i pi e where Z e ¡ i , N Z i 1



pq

p q

which is the discrete version of the Boltzmann-Gibbs distribution (2), and is implicitly given by d log Z d

p

qp q ¡

E.

(4)

Notice that pi is actually the density of particles which are in i th state. The macro-state of the infinite system is defined as p1 , ..., p p . We also easily check the relations

p

q

p ¸ i1



p

i



1,

p ¸ i1



p

ii



E.

In the following we denote by

DE F I N I T I O N 1.2. The parameter T kB ¡1 , where is solution to (4), is called temperature of the thermodynamical system. In particular, the temperature is only defined at equilibrium and in the thermodynamic limit N .

eq

the number of possible micro-states associated to N

p

q

i



Ne

¡



i

{Zp q

.

ÿV

D

EFINITION

1.3. The quantity S p1 , ..., p

p

p

q : ¡

kB

p ¸ i1



pi log p

p iq

is called thermodynamic entropy5 . P
ROPOSITION

1.1. The thermodynamic entropy can be written as ¨ S p1 , ..., p p S E, kB E log Z .

p

q p

q



pq

kB log eq S E, . N In particular, the equilibrium macro-state is the one that maximises the finite dimensional entropy SN N1 , ..., N p : kB log (among all that satisfy the conditions (3)), and that is what Boltzmann called entropy. This formula is now written on his gravestone. Indeed, when N is very large, one can use Stirling's formula to write
N

We also have

lim

ÿV

p q p

q

p

q

pq

SN kB Finally, one can easily compute

N

ÿV



N log N

p q¡ fS p fE

N



E



log Z

p q

E.

E,

q

1 . T

This is a standard formula of thermodynamics.
5

That entropy takes the same form than the Shannon entropy which is used in information theory.


2

THE IDEAL GAS IN ONE DIMENSION

7

2 The ideal gas in one dimension
2.1 The micro-canonical distribution
We give another simple (and highly unrealistic) example. Let us consider N particles, all of mass 1, that move on a one-dimensional interval 0, L and do not interact with each other. For each particle i , we denote by pi its momentum and by qi 0, L its position. When reaching the boundary of the interval, particles are reflected. Assume the right boundary of the interval (the piston) to be movable: a constant force f is acting, and the energy due to this external force is given by Eext f L, if the piston's position is L. At initial time, the interior energy of the system of particles is given by



rs

r s



N ¸ p2 0 i i1

pq



2



Eint .

When time evolves, we assume that the dynamics of the system is such that it runs uniformly over all configurations that are compatible with the constraint that the total energy is kept constant, equal to Et

Since Eint 0, the maximal value for the position of the piston is Lmax H f . Therefore, the state space for the whole system L, q1 , ..., qN , p1 , ... pN is ¡ $1{2 © . : 0, Lmax 0, L N S N 2 H f L

¥

p qH

E

int



f L.

p

r

q s¢r s ¢

{

p¡ q

AS S U M P T I O N 2.1. We assume that the initial configuration of the process is chosen uniformly at random in the set , and that when time evolves, the configuration remains uniformly distributed on this set. The probability distribution on is called the micro-canonical distribution of the system (see Section 3 below for the general definition). Assume that L is known, in that case we can compute the normalization constant for the uniform probability on 0, L N S N 2 H f L as

r s ¢ p p ¡ qq
L
N

Z

N

p

L, H, f :

q

2N{2 2H N2

p { q p ¡ f Lq

¨p

N12

¡ q{

.

This is indeed an easy consequence of Exercise 3.

2.2

Finding the equilibrium position

Thermodynamics is concerned with macroscopic variables (or macro-states): in that case, we are looking for the distribution of the piston position L. Precisely, the probability that L belongs to a small interval L0 , L0 dL equals

p

q

rL p

L0 , L

0

dLqs

q ¢ dL ZN pL, H, f qdL 0 2 @ exp N logpL0 q pN ¡ 1q logp2pH ¡ f L0 qq{2q dL ¨L 2 @ exp N logpLq pN ¡ 1q logp2pH ¡ f Lqq{2q dL 0
Z ¨L
N

p

L0 , H, f

max

max

(5)

When N is very large6 we can estimate the behaviour of the integral above thanks to Laplace's method (see Appendix A). For that purpose, let us look for the maximal value of the function fL
6

p q logpLq

1 log 2 H 2

p p ¡ f Lqq

.

N should be thought as the number of atoms, which is of order 1023 (Avogadro number).


2

THE IDEAL GAS IN ONE DIMENSION
The maximum of f is achieved at L

8

2H . 3f Therefore, with very high probability (close to 1), the piston stays at its equilibrium position L .





2.3

Computing the pressure on the piston

The pressure on the piston is defined as the average force that the gas molecules exercise when they are reflected on it. Each time a molecule i is reflected, its momentum changes from pi to pi . The time between two hits is 2L pi . Therefore, during a time interval 0, t , the molecule i hits the piston around t pi 2L times. The average force exerted during a time interval 0, t equals: »¢ ¸ N d pi t pi 2Eint 2H 2 f L 1t 1¸ 1¸ pt ds 2 pi f. 2 pi s t 0 i t1,...,Nu d s t i t1,...,Nu t i 1 2L L L

{p q

p{q

rs rs ¢

¡

p q



p p qq





¡



qi s

p qL



qi s L s 0, t

p q r s

In other words, in equilibrium, the thermodynamic pressure p force f acting on the piston.



lim

t

ÿV

p t is equal to the external

pq

2.4

Concepts of entropy and temperature

The pressure is the first intensive quantity that we met: it is independent of the system size (or the amount of material in the system). The other quantities (mass, internal energy) are classified as extensive. Let us now introduce two new intensive quantities: the temperature and the entropy. For that purpose, we define the rescaled variables Eint L H , e , h . N N N The probability distribution (5) rewrites as 2 @ exp N log 0 N 1 log 2 h f 0 2 d d 2 @. ¨ max 0, 0 exp N log N 1 log 2 h f 2d 0







r p

qs p

p q p ¡ q p p ¡ qq{ q p q p ¡ q p p ¡ qq{ q
1 log 2 h 2

Define Then, up to negligible terms,

s h, f ,

q : logp q
0

p p ¡ f qq p p q

.

r p p

0

,

d qs

The functions s is called entropy function. When computed at the equilibrium value it is called entropy. In our case it reads 1 2h 2h s h, f : s h, f , log log . 2 3 3f The temperature is defined as the coefficient that relates the change of entropy to the change of internal energy e. Since, at equilibrium, h, f completely determine , e , we can express s in the variables , e as s ,e log 2e .

2 @ exp Ns h, f , 0 d 2 @. ¨ max exp Ns h, f , d 0

q

qp

q



p

q

p q p
¡1

c

pq q

pq

As a result, the temperature T can be defined as T

fs f e p , eq

2e.


3

RIGOROUS DEFINITIONS OF MICRO -CANONICAL OBJECTS

9

2.5

Conclusion

In this simple example, we understand how some famous equations of thermodynamics arise. The equilibrium state of the system is governed by the external force f , and the intrinsic probability of the system to find itself in a state with a given value of the macroscopic parameter ( ). The properties of this probability distribution give rise to some effective force ­ the pressure p 2e ­ that has to be equilibrated against the macroscopic force. We also understand that: in the constant volume ensemble (L fixed), the conserved energy is the kinetic energy Eint ; while in the constant pressure ensemble (Eint L fixed), the conserved energy is the sum of the mechanical and kinetic energy Eint pL.

{



{

3 Rigorous definitions of micro-canonical objects
With the two examples above in mind, we now define what we understand by a thermodynamic system and give precise definitions. D
EFINITION

3.1 (Thermodynamic system). A thermodynamic system involves:

1. a parameter N



, called the particle number,
N

2. a state space X and its product space XN , 3. a Hamilton function HN : X

ÿ

,

4. a few constraints depending on macroscopic parameters. RE M A R K 3.1. In the context of a gas, the space X is the phase space of a single molecule, HN is the interaction between molecules, and the constraint is the indicator function that the position of all molecules should be within the container of volume V. In the following, we will denote this constraint by x V.



DE F I N I T I O N 3.2 (Micro-canonical ensemble). The micro-canonical ensemble of a thermodynamic system is the collection of all uniform probability distributions fE,V,N on the sets 2 @ E,V,N : x XN ; HN x E and x V .





p q



The micro-canonical partition function is the function » 1 ZE,V,N : E H N!



p ¡ N p x qq

1

xV



dx,

(6)

where denotes the Dirac delta-function7 on
7

and 1A is the characteristic function8 of the set A.

The delta function is defined as follows: for all smooth functions , and a

»



.

yy

p q p ¡ aq

dy

paq
»
X
N

.

It follows that, if is a function on XN , then

»»
X
N

y

p ¡ HN p x qqp x q p y q
5

d ydx

xH

p q p N p x qq

dx.

8

The characteristic function 1A is defined as

1A p x

q

1 0

if x A, otherwise.


4

THE THEOREM OF EQUIVALENCE OF ENSEMBLES
The thermodynamic entropy is defined as S E, V, N :

10

Notice that the factor 1 N! is introduced to take into account that particles are indistinguishable and ensures that S is proportional to N (and additive). Finally, the micro-canonical probability density of the microstates x XN associated to the macro-state E, V, N is

{

p

q logp q

Z

E,V,N

q

.



p

f

E,V,N

p x q :

1 Z
E,V,N

¢ q

1 E N!

p ¡ HN p x qq

1

xV



.

In many physics textbooks, this last measure is replaced with the Lebesgue measure (thought as an equivalent of the volume in N ) of the set x XN ; HN x E , which gives the same practical results. From now on we assume that S E, V, N is a differentiable function and we set Boltzmann's constant kB equal to 1.

p

t

| p q¡ | ¤ u

D

EFINITION

3.3. The pressure p, temperature T and chemical potential µ are defined as

fS p fE

E, V, N

1 q T

,

fS p fV

E, V, N

q

p , T

fS p fN

E, V, N

q

1 . µ

In particular, if we want to couple the thermodynamic system to a mechanical source of energy of strength f , it amounts to assume that the quantity H E f V is conserved (meaning that HN x H fV becomes the new constraint, and E can vary). Then, the distribution of V is given by ¨ exp S V0 , H f V0 , N dV ¨ ¨ . V V0 , V0 dV exp S V0 , H f V0 , N dV

p

p q ¡

r p

qs p

¡ p¡ q

q¢ q

From Laplace's method, the equilibrium position of V is determined by the maximal entropy, and in particular is the solution of the equation dS V, H f V, N dV with H and f fixed.

¡

0,

(7)

4 The theorem of equivalence of ensembles
The main difficulty in the computations involving the micro-canonical ensemble comes from the constrained integral (6) over some space HN x E in very high dimension, which turns out to be a pretty difficult geometric problem. We are able to do it for the sphere (Exercise 3), but in general this is hopeless. That is the reason why it is more convenient to change ensembles: we are going to consider a system where the energy is no longer fixed, but allowed to vary, while the temperature is fixed.

t p q u

DE F I N I T given by

ION

4.1. Let

¡

0 be a parameter. The canonical partition function associated to , V, N is
»

p

q

» 1 ,V,N e ZE,V,N dE e dE e ¡ H N p x q d x , N! and is associated to the canonical probability density (or canonical ensemble or Gibbs measure) ¨ ¡ E e ZE,V,N fE,V,N x dE e¡ HN p x q ¨ G ,V,N x . ,V,N e ¡ H N p x q d x

¡

E



»

¡ E SpE,V,Nq



p q

pq
¡



The function F defined as is called the free energy functional.

F , E, V, N :

p

q E¡



1

S E, V, N

p

q


4

THE THEOREM OF EQUIVALENCE OF ENSEMBLES
We are now ready to state the main result of that section: T
HEOREM

11

4.1 (Equivalence of ensembles). Assume that the micro-canonical entropy satisfies lim

N

ÿV

S E, V, N N

p

q spe, v q

,

(8)

where

V v and N NÿV Assume that s is strictly concave, continuous and satisfies, » exp N s e, v Define the function f , v by

ùù ùÿ

E e. N NÿV for all 0,

ùù ùÿ
¨

¥

p

q

te ; spe,vq¡ e¤au

r p q ¡ es
2 min e
e

de

¤
@

C exp N a .

pq

(9)

f , v

p

q



¡ spe, v q
1.

.

(10)

Finally, assume that the convergence in (8) is such that, uniformly in e, S E, V, N N s e, v Then, for any

¥

p q ¡ E ù ù ÿV r p q ¡ es Nù ÿ
log , V, N N

(11)

0,
N

lim

RE M A R K 4.1. The function f is called the Legendre transform of s. One can easily show that, if s is differentiable and strictly concave, then f , v

pq

ÿV

p

q ¡f p
¡

, v .

q

(12)

p

q e p

, v





1

s e , v , v

pp

qq

(13)

where e is the unique solution of the equation Proof. Let us denote D :

fspfee, v q

. e , v

e

2

; s e, v

p q ¡ e ¡ sp



e



¡

@ ,

c and D is the complement of D . First of all, the assumption on uniform convergence (11) ensures that, for large N, we can replace the integrand in » e¡ E SpE,V,Nq dE (14)

by its limit, since, for any § § §N

¡
¡
1

0, there exists N0





S E, V, N

p



E

$

¡ p q¡

such that, for all N N0 , § $§ § § s e, v e § §s e, v

More precisely, for such N, we make the change of variables E » » ¡ eN SpeN,V,Nq de e eNrs¡ es rS¡ E¡Nps¡
c D

¥ ¤ p q¡ eN in (14)
e

for all e § § e§.



,

and we bound

¤

qs

¤ ¤

»

c D

de

e
c D

N s e

rp ¡ q |s¡ e|s

de

C exp N s e , v

rp



e



|sp

e , v



e



| ¡ s

¨ .


4

THE THEOREM OF EQUIVALENCE OF ENSEMBLES
On the other hand, for N large enough, the integral on D can be bounded from (9) as follows: » ¨ e¡ eN SpeN,V,Nq de C exp N s e , v e 1 .
D

12

¤

rp



sp q

Moreover, since the function s has bounded derivatives, on a set of size N¡1 the integrand cannot vary by more than a constant, for some c 0, so that we have » ¨ e¡ eN SpeN,V,Nq de c N¡1 exp N s e , v e 1 .

¡

D

¥

rp



sp q
q

Taking the logarithm, dividing by N, for any

¡ f p
and

, v 1

qp ¡ q ¤

0 we obtain that ¢» 1 lim inf log N e¡ NÿV N
eN S eN,V,N

¡

eN S eN,V,N

p



de

¢» 1 lim sup log N e NÿV N

¡

p

q



de

¤ ¡ f p

, v 1

qp q

.

This implies (12).

Conclusion. The theorem of equivalence of ensembles justifies the use of the canonical ensemble
to compute thermodynamic quantities, from the canonical rather than the micro-canonical partition function. We can define the free energy in terms of the logarithm of the partition function and derive all thermodynamic quantities (including the entropy) from it via Legendre transformations. It is important to note that this equivalence holds in the limit of infinite particle number (and, in consequence, infinite volume, energy, ...). Once again, this illustrates the fact that statistical mechanics in interested in understanding what happens when the size of the system tends to infinity. The beauty of the equivalence of ensembles is that, computationnally, it is much easier (even though still hard enough) to work with the Gibbs distribution than with the micro-canonical distribution: working with constraints is always hard, and the canonical ensemble allows us to get rid of one annoying constraint. The theorem tells us that not fixing the energy is fine.


4

THE THEOREM OF EQUIVALENCE OF ENSEMBLES

13

Statistical Mechanics: Exercices
EX E R C I S E 1 (Exact and inexact differentials). Consider the infinitesimal quantity Ad x Bd y where A and B are functions of x and y . It this quantity is to be an exact differential, then there must exist a function F x , y such that dF Ad x Bd y .



p

q





1. Show that in this case, A and B must satisfy

fA f fy f
2. Is x 3. 4. E

B . x

is an exact differential, then its integral from x 1 , y1 to x 2 , y2 depends on the endpoints only, and not on the path taken to go from one point to the other. 2 (Normal distribution). 1. Prove9 that » 2 e¡ x d x

p 2 ¡ y qd x x d y an Is p1 ¡ y { x 2 qd x p1{ x q Prove that if Ad x Bd y

exact differential? d y an exact differential?

p

qp

q

XERCISE



c

.

2. Check that the function N (given in (1)) defines a probability distribution, in the sense that » N x d x 1,

pq

and compute its mean value. EX E i.e.
RCISE

3 (The surface of a sphere). Let SN r denote the N S
N

pq

p ¡ 1q
N ¸ i1

­dimensional sphere of radius r
A r.

¡

0,

p r q : p
N

3

x 1 , ..., x

N

q

N

;

Let AN r be the surface area of S 1. Compute

pq

pr q



x

2 i



.
»
2 2 ¡p x1 ¤¤¤ xN q

e
N

dx

1

¤¤¤

d xN .
»

2. Deduce that
AN r

p q p { q

2N{2 r N2

N1

¡

,

where s :

pq

V
0

x

s1

¡ e¡

x

dx.

Remind (or prove) that the function satisfies x
9

p 1q c p x q p x q x p1{2q , p n 1q n! p n q .
»
e
2

,

Hint: First of all, use the polar coordinates to compute

¡x ¡
2

y

2

d x d y.


4

THE THEOREM OF EQUIVALENCE OF ENSEMBLES
As a result, the uniform probability density on the sphere SN r is the function that constantly equals the inverse of the partition function Z
N

14

pq
{

pr q

»
SN

pr q

dx

1

¤¤¤

dx

N

2p {2q N
N2

r

N1

¡

on SN r and vanishes elsewhere. EX E R C I S E 4 ( ). Now we consider a random vector We choose to generate each coordinate by using a each coordinate independently ­ i.e., the value of one coordinate. Show that the probability distribution of U is invariant under any rotation in the sphere SN r .

pq



N X that has random coordinates X 1 , ..., X N . standard normal distribution. We are generating coordinate has no effect on the value of any other the random vector U given by



p

q

} }
N

X , X

where X

}}
2

N ¸ i1



X2 , i

pq

. This means that U has the uniform distribution on the surface of

EX E R C I S E 5. Generalize Example 1.2 for a system of N particles that can have p denoted by 1, 2, ..., p : more precisely,

t

u

q

different states,

· show that the number of possible micro-states equals pN ; · what is the most probable macro-state? E
XERCISE

· compute the number of micro-states associated to one macro-state N1 , ..., N p ; 6. Write explicitely the list of elements given in Definition 3.1 for

p

· the placement of N particles in p boxes (Subsection 1.4 and Exercise 5); · the ideal gas in one dimension (Section 2). E
XERCISE

7. Using Definition 3.3, show that dS V, H f V, N dV

p

¡

q

0,

is equivalent to p



f.

EX E R C I S E 8 (Ideal lattice gas). Let be a finite subset of the discrete lattice d . We denote by V the number of vertices of this subset. We consider of fixed number N of particles, for which we are only interested in their positions x i . We assume that two particles can occupy the same vertex, and the total energy of the system, which is conserved, is proportional to the number of particles: E µN. Therefore, the micro-canonical partition function is simply the number of ways to arrange N particles on V vertices of the lattice.

| |





1. Compute Z

E,V,N

and show that S E, V, N is equivalent as N

p

2. Prove that the pressure is given by p

E{pµVq

q

ÿV

to N log 1

p V {N q

.

.

3. Consider the lattice gas with the additional constraint that no more than one particle can occupy the same site. What will be changed? E
XERCISE

9 (Micro-canonical entropy). Let 0 be the function on 0 x :

E,V,N

defined as

pq

1 E N!

p ¡ HN p x qq

1

xV



.


4

THE THEOREM OF EQUIVALENCE OF ENSEMBLES
Notice that 0 is not a probability density since its integral does not equal 1. Let be any probability density whose support is E,V,N . The relative entropy of with respect to 0 is defined as H , 0 :

15

p

q

»

¢

log

d x d0 x

pq pq



d x

pq
E,V,N

Show that the unique minimizer of this function is the uniform probability density on denoted by fE,V,N in the lecture notes, and that Hf

, which is

p

E,V,N

, 0

q ¡Sp ¡ pq

E, V, N .

q

EX E R C I S E 10 (Legendre transform). Let I its Legendre transform is the function f : I f





p yq

be an interval, and f : I defined by 2 @ sup x y f x , y I .



ÿ

ÿ

be a convex function. Then,

xI

1. Show that f is well-defined on I if f is convex.





2. Note that if f is defined on and everywhere differentiablee, then f p can be interpreted as the negative of the y -axis coordinate of the intersection between the tangent line to the graph of f that has slope p and the y -axis.

pq
0

3. Prove that, if f is twice differentiable such that f solution to f I x 0 p.

p q

P

¡

0, then f



p pq

px

¡ f p x0q

where x 0 is the

E

XERCISE

11 (Classical ideal gas in one-dimension). Here the Hamiltonian is H

N

p

p1 , ..., pN , q1 , ..., qN

q

N ¸p i1

2 i



2

.

1. Show that the canonical partition function is 2. Show that f , v
,V,N



1 V N!

¢
N

2



N2

{

.
2 .

p

q NÿV ¡ p



¡

¢d
1

log v

3. Compute the thermodynamic entropy using f , v . 4. Compute the entropy directly from the micro-canonical partition function and compare.

q


A

LAPLACE'S METHOD

16

A Laplace's method
Laplace's method is a technique used to investigate the behaviour as N goes to infinity of integrals of the form »b exp N f x d x ,
a

p p qq

where f is a twice-differentiable function: assume that the function f has a unique global maximum at x 0 . Then, significant contributions to the integral of this function will come only from points x in a of x 0 , which can then be estimated thanks to Taylor expansions. PR O P O S I T I O N A.1. Assume that f is twice differentiable on a, b , and that x 0 a, b is the unique point where f achieves its maximal value on a, b . Assume additionally that f P x 0 0. Then, d »b 2 exp N f x d x exp N f x 0 NÿV N f P x0 a

rs

rs

r s p q pq

p p qq



p p qq ¢ ¡ q

B

Lagrange multipliers

Suppose we want to extremize the function f x , y where the coordinates x , y the constraint g x , y 0. If f is to be an extremum, then

p

q

p

p

q

2

are subject to

df

f f

f dx x

f f

f dy y



0.

If the diplacements were all independent, then we would conclude that all the partial derivatives must be zero: f x f y 0. But the displacements are not independent, more precisely g 0 implies that g g dg dx d y 0, x y (since the displacements must lie on the surface g 0). This means that d y can be expressed in terms of d x . Consider the linear combination d f d g , where is called the Lagrange multiplier. This combination must obviously vanish. Suppose that we can choose such that

f {f f {f



f f

f f





f f fg fy fy f f fg fx fx

0,

(15)

and this implies then

0.

(16)

Maximisation has been achieved. The method of Lagrange multipliers consists in: solving (15) and (16) considering as a parameter ( x , y are the unknowns); then, determining the value of which is consistent with g x , y 0.

p

q

References
[1] A. Bovier, Statistical mechanics of disordered systems, Cambridge University Press (2006), ISBN 0521-84991-3. [2] D. Garanin, Statistical Thermodynamics, Lecture notes (Fall 2009). [3] Y.M. Guttmann, The concept of probability in statistical physics, Cambridge University Press (1999), ISBN 0-521-62128-3. [4] E. Poisson, Statistical Physics II (PHYS*4240), Lecture Notes (Fall 2000).