Docsity
Docsity

Prepare for your exams
Prepare for your exams

Study with the several resources on Docsity


Earn points to download
Earn points to download

Earn points by helping other students or get them with a premium plan


Guidelines and tips
Guidelines and tips

Lecture Notes on Kinetic Theory, Lecture notes of Chemistry

The Boltzmann Equation and The Quantum Boltzmann Equation, Transport with Collisions.

Typology: Lecture notes

2020/2021

Uploaded on 06/21/2021

ekaant
ekaant 🇺🇸

4.6

(34)

270 documents

1 / 39

Toggle sidebar

This page cannot be seen from the preview

Don't miss anything!

bg1
2. Kinetic Theory
The purpose of this section is to lay down the foundations of kinetic theory, starting
from the Hamiltonian description of 1023 particles, and ending with the Navier-Stokes
equation of fluid dynamics. Our main tool in this task will be the Boltzmann equation.
This will allow us to provide derivations of the transport properties that we sketched in
the previous section, but without the more egregious inconsistencies that crept into our
previous attempt. But, perhaps more importantly, the Boltzmann equation will also
shed light on the deep issue of how irreversibility arises from time-reversible classical
mechanics.
2.1 From Liouville to BBGKY
Our starting point is simply the Hamiltonian dynamics for Nidentical point particles.
Of course, as usual in statistical mechanics, here is Nridiculously large: NO(1023 )
or something similar. We will take the Hamiltonian to be of the form
H=1
2m
N
X
i=1
~p2
i+
N
X
i=1
V(~ri)+X
i<j
U(~ri~r j) (2.1)
The Hamiltonian contains an external force ~
F=rVthat acts equally on all parti-
cles. There are also two-body interactions between particles, captured by the potential
energy U(~r i~rj). At some point in our analysis (around Section 2.2.3)wewillneed
to assume that this potential is short-ranged, meaning that U(r)0forrdwhere,
as in the last Section, dis the atomic distance scale.
Hamilton’s equations are
@~pi
@t=@H
@~ri
,@~ri
@t=@H
@~pi
(2.2)
Our interest in this section will be in the evolution of a probability distribution,
f(~ri,~p
i;t)overthe6Ndimensional phase space. This function tells us the proba-
bility that the system will be found in the vicinity of the point (~ri,~p
i). As with all
probabilities, the function is normalized as
ZdV f(~ri,~p
i;t)=1 with dV =
N
Y
i=1
d3rid3pi
Furthermore, because probability is locally conserved, it must obey a continuity equa-
tion: any change of probability in one part of phase space must be compensated by
–14–
pf3
pf4
pf5
pf8
pf9
pfa
pfd
pfe
pff
pf12
pf13
pf14
pf15
pf16
pf17
pf18
pf19
pf1a
pf1b
pf1c
pf1d
pf1e
pf1f
pf20
pf21
pf22
pf23
pf24
pf25
pf26
pf27

Partial preview of the text

Download Lecture Notes on Kinetic Theory and more Lecture notes Chemistry in PDF only on Docsity!

2. Kinetic Theory

The purpose of this section is to lay down the foundations of kinetic theory, starting from the Hamiltonian description of 10 23 particles, and ending with the Navier-Stokes equation of fluid dynamics. Our main tool in this task will be the Boltzmann equation. This will allow us to provide derivations of the transport properties that we sketched in the previous section, but without the more egregious inconsistencies that crept into our previous attempt. But, perhaps more importantly, the Boltzmann equation will also shed light on the deep issue of how irreversibility arises from time-reversible classical mechanics.

2.1 From Liouville to BBGKY

Our starting point is simply the Hamiltonian dynamics for N identical point particles. Of course, as usual in statistical mechanics, here is N ridiculously large: N ⇠ O(10 23 ) or something similar. We will take the Hamiltonian to be of the form

H =

2 m

X^ N

i=

~p (^) i^2 +

X^ N

i=

V (~r (^) i ) +

X

i<j

U (~r (^) i ~rj ) (2.1)

The Hamiltonian contains an external force F~ = rV that acts equally on all parti- cles. There are also two-body interactions between particles, captured by the potential energy U (~ri ~r (^) j ). At some point in our analysis (around Section 2.2.3) we will need to assume that this potential is short-ranged, meaning that U (r) ⇡ 0 for r d where, as in the last Section, d is the atomic distance scale.

Hamilton’s equations are @~pi @t

@H

@~ri

@~ri @t

@H

@~pi

Our interest in this section will be in the evolution of a probability distribution, f (~r (^) i , ~p (^) i ; t) over the 6N dimensional phase space. This function tells us the proba- bility that the system will be found in the vicinity of the point (~r (^) i , ~p (^) i ). As with all probabilities, the function is normalized as

Z dV f (~r (^) i , ~p (^) i ; t) = 1 with dV =

Y^ N

i=

d 3 r (^) i d 3 p (^) i

Furthermore, because probability is locally conserved, it must obey a continuity equa- tion: any change of probability in one part of phase space must be compensated by

a flow into neighbouring regions. But now we’re thinking in terms of phase space, the “r” term in the continuity equation includes both @/@~ri and @/@~pi and, corre- spondingly, the velocity vector in phase space is (~r˙ (^) i , ~p˙ (^) i ). The continuity equation of the probability distribution is then

@f @t

@~ri

~r˙ (^) i f

@~pi

~p˙ (^) i f

where we’re using the convention that we sum over the repeated index i = 1,... , N. But, using Hamilton’s equations (2.2), this becomes

@f @t

@~ri

@H

@~pi f

@~pi

@H

@~ri f

@f @t

@f @~ri

@H

@~pi

@f @~pi

@H

@~ri

This final equation is the Liouville’s equation. It is the statement that probability doesn’t change as you follow it along any trajectory in phase space, as is seen by writing the Liouville equation as a total derivative,

df dt

@f @t

@f @~ri · ~r˙ (^) i + @f @~pi · ~p˙ (^) i = 0

To get a feel for how probability distributions evolve, one often evokes the closely related Liouville’s theorem^2. This is the statement that if you follow some region of phase space under Hamiltonian evolution, then its shape can change but its volume remains the same. This means that the probability distribution on phase space acts like an incompressible fluid. Suppose, for example, that it’s a constant, f , over some region of phase space and zero everywhere else. Then the distribution can’t spread out over a larger volume, lowering its value. Instead, it must always be f over some region of phase space. The shape and position of this region can change, but not its volume.

The Liouville equation is often written using the Poisson bracket,

{A, B} ⌘

@A

@~ri

@B

@~pi

@A

@~pi

@B

@~ri

With this notation, Liouville’s equation becomes simply

@f @t = {H, f }

(^2) A fuller discussion of Hamiltonian mechanics and Liouville’s theorem can be found in Section 4 of the classical dynamics notes: http://www.damtp.cam.ac.uk/user/tong/dynamics.html.

to limit our ambition. We’ll focus not on the probability distribution for all N parti- cles but instead on the one-particle distribution function. This captures the expected number of parting lying at some point (~r, ~p). It is defined by

f 1 (~r, p~; t) = N

Z YN

i=

d 3 r (^) i d 3 p (^) i f (~r, ~r 2 ,... , ~rN , ~p, ~p 2 ,... p~ (^) N ; t)

Although we seem to have singled out the first particle for special treatment in the above expression, this isn’t really the case since all N of our particles are identical. This is also reflected in the factor N which sits out front which ensures that f 1 is normalized as Z d 3 rd 3 p f 1 (~r, ~p; t) = N (2.6)

For many purposes, the function f 1 is all we really need to know about a system. In particular, it captures many of the properties that we met in the previous chapter. For example, the average density of particles in real space is simply

n(~r; t) =

Z

d 3 p f 1 (~r, ~p; t) (2.7)

The average velocity of particles is

~u(~r; t) =

Z

d 3 p ~p m f 1 (~r, ~p; t) (2.8)

and the energy flux is

E^ ~(~r; t) =

Z

d 3 p ~p m E(~p)f 1 (~r, p~; t) (2.9)

where we usually take E(~p) = p 2 / 2 m. All of these quantities (or at least close relations) will be discussed in some detail in Section 2.4.

Ideally we’d like to derive an equation governing f 1. To see how it changes with time, we can simply calculate:

@f (^1) @t

= N

Z YN

i=

d 3 r (^) i d 3 p (^) i

@f @t

= N

Z YN

i=

d 3 r (^) i d 3 p (^) i {H, f }

Using the Hamiltonian given in (2.1), this becomes

@f (^1) @t

= N

Z YN

i=

d 3 r (^) i d 3 p (^) i

X^ N

j=

~p (^) j m

@f @~rj

X^ N

j=

@V

@~rj

@f @~pj

X^ N

j=

X

k<l

@U (~r (^) k ~r (^) l ) @~rj

@f @~pj

Now, whenever j = 2,... N , we can always integrate by parts to move the derivatives away from f and onto the other terms. And, in each case, the result is simply zero because when the derivative is with respect to ~r (^) j , the other terms depend only on ~p (^) i and vice-versa. We’re left only with the terms that involve derivatives with respect to ~r 1 and ~p 1 because we can’t integrate these by parts. Let’s revert to our previous notation and call ~r 1 ⌘ ~r and ~p 1 ⌘ p~. We have

@f (^1) @t

= N

Z YN

i=

d 3 r (^) i d 3 p (^) i

~p m

@f @~r

@V (~r) @~r

@f @~p

X^ N

k=

@U (~r ~rk ) @~r

@f @~p

= {H 1 , f 1 } + N

Z YN

i=

d 3 r (^) i d 3 p (^) i

XN

k=

@U (~r ~rk ) @~r

@f @~p

where we have defined the one-particle Hamiltonian

H 1 = p 2 2 m

  • V (~r) (2.11)

Notice that H 1 includes the external force V acting on the particle, but it knows nothing about the interaction with the other particles. All of that information is included in the last term with U (~r ~r (^) k ). We see that the evolution of the one-particle distribution function is described by a Liouville-like equation, together with an extra term. We write @f (^1) @t

= {H 1 , f 1 } +

@f (^1) @t

coll

The first term is sometimes referred to as the streaming term. It tells you how the particles move in the absence of collisions. The second term, known as the collision integral, is given by the second term in (2.10). In fact, because all particles are the same, each of the (N 1) terms in

P N

k=2 in (2.10) are identical and we can write ✓ @f (^1) @t

coll

= N (N 1)

Z

d 3 r 2 d 3 p (^2) @U (~r ~r 2 ) @~r

@~p

Z YN

i=

d 3 r (^) i d 3 p (^) i f (~r, ~r 2 ,... , ~p, ~p 2 ,... ; t)

But now we’ve got something of a problem. The collision integral can’t be expressed in terms of the one-particle distribution function. And that’s not really surprising. As the name suggests, the collision integral captures the interactions – or collisions – of one particle with another. Yet f 1 contains no information about where any of the other particles are in relation to the first. However some of that information is contained in the two-particle distribution function,

f 2 (~r 1 , ~r 2 , ~p 1 , ~p 2 ; t) ⌘ N (N 1)

Z YN

i=

d 3 r (^) i d 3 p (^) i f (~r 1 , ~r 2 ,... , ~p 1 , p~ 2 ,... ; t)

However, there is an advantage is working with the hierarchy of equations (2.14) because they isolate the interesting, simple variables, namely f 1 and other lower f (^) n. This means that the equations are in a form that is ripe to start implementing various approximations. Given a particular problem, we can decide which terms are important and, ideally, which terms are so small that they can be ignored, truncating the hierarchy to something manageable. Exactly how you do this depends on the problem at hand. Here we explain the simplest, and most useful, of these truncations: the Boltzmann equation.

2.2 The Boltzmann Equation

“Elegance is for tailors” Ludwig Boltzmann

In this section, we explain how to write down a closed equation for f 1 alone. This will be the famous Boltzmann equation. The main idea that we will use is that there are two time scales in the problem. One is the time between collisions, ⌧ , known as the scattering time or relaxation time. The second is the collision time, ⌧ (^) coll , which is roughly the time it takes for the process of collision between particles to occur. In situations where

⌧ ⌧ (^) coll (2.15)

we should expect that, for much of the time, f 1 simply follows its Hamiltonian evolution with occasional perturbations by the collisions. This, for example, is what happens for the dilute gas. And this is the regime we will work in from now on.

At this stage, there is a right way and a less-right way to proceed. The right way is to derive the Boltzmann equation starting from the BBGKY hierarchy. And we will do this in Section 2.2.3. However, as we shall see, it’s a little fiddly. So instead we’ll start by taking the less-right option which has the advantage of getting the same answer but in a much easier fashion. This option is to simply guess what form the Boltzmann equation has to take.

2.2.1 Motivating the Boltzmann Equation

We’ve already caught our first glimpse of the Boltzmann equation in (2.12),

@f (^1) @t = {H 1 , f 1 } +

@f (^1) @t

coll

But, of course, we don’t yet have an expression for the collision integral in terms of f 1. It’s clear from the definition (2.13) that the second term represents the change in

momenta due to two-particle scattering. When ⌧ ⌧ (^) coll , the collisions occur occasion- ally, but abruptly. The collision integral should reflect the rate at which these collisions occur.

Suppose that our particle sits at (~r, ~p) in phase space and collides with another particle at (~r, ~p 2 ). Note that we’re assuming here that collisions are local in space so that the two particles sit at the same point. These particles can collide and emerge with momenta ~p 10 and ~p 20. We’ll define the rate for this process to occur to be

Rate = !(p,~ ~p 2 |~p 10 , ~p 20 ) f 2 (~r, ~r, ~p, ~p 2 ) d 3 p 2 d 3 p 01 d 3 p 02 (2.17)

(Here we’ve dropped the explicit t dependence of f 2 only to keep the notation down). The scattering function! contains the information about the dynamics of the process. It looks as if this is a new quantity which we’ve introduced into the game. But, using standard classical mechanics techniques, one can compute! for a given inter-atomic potential U (~r). (It is related to the di↵erential cross-section; we will explain how to do this when we do things better in Section 2.2.3). For now, note that the rate is proportional to the two-body distribution function f 2 since this tells us the chance that two particles originally sit in (~r, ~p) and (~r, ~p 2 ).

We’d like to focus on the distribution of particles with some specified momentum ~p. Two particles with momenta ~p and ~p 2 can be transformed in two particles with momenta ~p 10 and ~p 20. Since both momenta and energy are conserved in the collision, we have

~p + ~p 2 = ~p 10 + ~p 20 (2.18) p 2 + p 22 = p 01 2 + p 02 2 (2.19)

There is actually an assumption that is hiding in these equations. In general, we’re considering particles in an external potential V. This provides a force on the particles which, in principle, could mean that the momentum and kinetic energy of the particles is not the same before and after the collision. To eliminate this possibility, we will assume that the potential only varies appreciably over macroscopic distance scales, so that it can be neglected on the scale of atomic collisions. This, of course, is entirely reasonable for most external potentials such as gravity or electric fields. Then (2.18) and (2.19) continue to hold.

While collisions can deflect particles out of a state with momentum ~p and into a di↵erent momentum, they can also deflect particles into a state with momentum ~p.

at what we’ve actually assumed. Looking at (2.21), we can see that we have taken the rate of collisions to be proportional to f 2 (~r, ~r, ~p 1 , p~ 2 ) where p 1 and p 2 are the momenta of the particles before the collision. That means that if we substitute (2.22) into (2.21), we are really assuming that the velocities are uncorrelated before the collision. And that sounds quite reasonable: you could imagine that during the collision process, the velocities between two particles become correlated. But there is then a long time, ⌧ , before one of these particles undergoes another collision. Moreover, this next collision is typically with a completely di↵erent particle and it seems entirely plausible that the velocity of this new particle has nothing to do with the velocity of the first. Nonethe- less, the fact that we’ve assumed that velocities are uncorrelated before the collision rather than after has, rather slyly, introduced an arrow of time into the game. And this has dramatic implications which we will see in Section 2.3 where we derive the H-theorem.

Finally, we may write down a closed expression for the evolution of the one-particle distribution function given by

@f (^1) @t = {H 1 , f 1 } +

@f (^1) @t

coll

with the collision integral ✓ @f (^1) @t

coll

Z

d 3 p 2 d 3 p 01 d 3 p 02 !(~p 10 , ~p 20 |~p, ~p 2 )

h f 1 (~r, ~p 10 )f 1 (~r, ~p 20 ) f 1 (~r, ~p)f 1 (~r, ~p 2 )

i (2.24)

This is the Boltzmann equation. It’s not an easy equation to solve! It’s a di↵erential equation on the left, an integral on the right, and non-linear. You may not be surprised to hear that exact solutions are not that easy to come by. We’ll see what we can do.

2.2.2 Equilibrium and Detailed Balance

Let’s start our exploration of the Boltzmann equation by revisiting the question of the equilibrium distribution obeying @f eq^ /@t = 0. We already know that {f, H 1 } = 0 if f is given by any function of the energy or, indeed any function that Poisson commutes with H. For clarity, let’s restrict to the case with vanishing external force, so V (r) = 0. Then, if we look at the Liouville equation alone, any function of momentum is an equilibrium distribution. But what about the contribution from the collision integral?

One obvious way to make the collision integral vanish is to find a distribution which obeys the detailed balance condition,

f 1 eq (~r, ~p 10 )f 1 eq (~r, ~p 20 ) = f 1 eq (~r, ~p)f 1 eq (~r, ~p 2 ) (2.25)

In fact, it’s more useful to write this as

log(f 1 eq (~r, ~p 10 )) + log(f 1 eq (~r, ~p 20 )) = log(f 1 eq (~r, ~p)) + log(f 1 eq (~r, ~p 2 )) (2.26)

How can we ensure that this is true for all momenta? The momenta on the right are those before the collision; on the left they are those after the collision. From the form of (2.26), it’s clear that the sum of log f 1 eq must be the same before and after the collision: in other words, this sum must be conserved during the collision. But we know what things are conserved during collisions: momentum and energy as shown in (2.18) and (2.19) respectively. This means that we should take

log(f 1 eq (~r, ~p)) = (μ E(~p) + ~u · ~p ) (2.27)

where E(p) = p 2 / 2 m for non-relativistic particles and μ, and ~u are all constants. We’ll adjust the constant μ to ensure that the overall normalization of f 1 obeys (2.6). Then, writing ~p = m~v, we have

f 1 eq (~r, ~p) =

N

V

2 ⇡m

e m(~v~u)^ (^2) / 2 (2.28)

which reproduces the Maxwell-Boltzmann distribution if we identify with the inverse temperature. Here ~u allows for the possibility of an overall drift velocity. We learn that the addition of the collision term to the Liouville equation forces us to sit in the Boltzmann distribution at equilibrium.

There is a comment to make here that will play an important role in Section 2.4. If we forget about the streaming term {H 1 , f 1 } then there is a much larger class of solutions to the requirement of detailed balance (2.25). These solutions are again of the form (2.27), but now with the constants μ, and ~u promoted to functions of space and time. In other words, we can have

f 1 local (~r, ~p; t) = n(~r, t)

(~r, t) 2 ⇡m

exp

(~r, t)

m 2 [(~v ~u(~r, t)] 2

Such a distribution is not quite an equilibrium distribution, for while the collision integral in (2.23) vanishes, the streaming term does not. Nonetheless, distributions of this kind will prove to be important in what follows. They are said to be in local equilibrium, with the particle density, temperature and drift velocity varying over space.

The Quantum Boltzmann Equation

Our discussion above was entirely for classical particles and this will continue to be our focus for the remainder of this section. However, as a small aside let’s look at how

We start by returning to the BBGKY hierarchy of equations. For simplicity, we’ll turn o↵ the external potential V (~r) = 0. We don’t lose very much in doing this because most of the interesting physics is concerned with the scattering of atoms o↵ each other. The first two equations in the hierarchy are ✓ @ @t

~p (^1) m

@~r 1

f 1 =

Z

d 3 r 2 d 3 p (^2) @U (~r 1 ~r 2 ) @~r 1

@f (^2) @~p 1

and ✓ @ @t

~p (^1) m

@~r 1

~p (^2) m

@~r 2

@U (~r 1 ~r 2 ) @~r 1

@~p 1

@~p 2

f 2 = (2.31) Z d 3 r 3 d 3 p (^3)

@U (~r 1 ~r 3 ) @~r 1

@~p 1

@U (~r 2 ~r 3 ) @~r 2

@~p 2

f (^3)

In both of these equations, we’ve gathered the streaming terms on the left, leaving only the higher distribution function on the right. To keep things clean, we’ve sup- pressed the arguments of the distribution functions: they are f 1 = f 1 (~r 1 , ~p 1 ; t) and f 2 = f 2 (~r 1 , ~r 2 , p~ 1 , ~p 2 ; t) and you can guess the arguments for f 3.

Our goal is to better understand the collision integral on the right-hand-side of (2.30). It seems reasonable to assume that when particles are far-separated, their distribution functions are uncorrelated. Here, “far separated” means that the distance between them is much farther than the atomic distance scale d over which the potential U (r) extends. We expect

f 2 (~r 1 , ~r 2 , ~p 1 , ~p 2 ; t)! f 1 (~r 1 , ~p 1 ; t)f 1 (~r 2 , ~p 2 ; t) when |~r 1 ~r 1 | d

But, a glance at the right-hand-side of (2.30) tells us that this isn’t the regime of interest. Instead, f 2 is integrated @U (r)/@r which varies significantly only over a region r  d. This means that we need to understand f 2 when two particles get close to each other.

We’ll start by getting a feel for the order of magnitude of various terms in the hierarchy of equations. Dimensionally, each term in brackets in (2.30) and (2.31) is an inverse time scale. The terms involving the inter-atomic potential U (r) are associated to the collision time ⌧ (^) coll.

1 ⌧ (^) coll

@U

@~r

@~p

This is the time taken for a particle to cross the distance over which the potential U (r) varies which, for short range potentials, is comparable to the atomic distance scale, d,

itself and

⌧ (^) coll ⇠ d v ¯ (^) rel

where ¯v (^) rel is the average relative speed between atoms. Our first approximation will be that this is the shortest time scale in the problem. This means that the terms involving @U/@r are typically the largest terms in the equations above and determine how fast the distribution functions change.

With this in mind, we note that the equation for f 1 is special because it is the only one which does not include any collision terms on the left of the equation (i.e. in the Hamiltonian H (^) n ). This means that the collision integral on the right-hand side of (2.30) will usually dominate the rate of change of f 1. (Note, however, we’ll meet some important exceptions to this statement in Section 2.4). In contrast, the equation that governs f 2 has collision terms on the both the left and the right-hand sides. But, importantly, for dilute gases, the term on the right is much smaller than the term on the left. To see why this is, we need to compare the f 3 term to the f 2 term. If we were to integrate f 3 over all space, we get Z d 3 r 2 d 3 p 3 f 3 = N f (^2)

(where we’ve replaced (N 2) ⇡ N in the above expression). However, the right- hand side of (2.31) is not integrated over all of space. Instead, it picks up a non-zero contribution over an atomic scale ⇠ d 3. This means that the collision term on the right-hand-side of (2.31) is suppressed compared to the one on the left by a factor of N d 3 /V where V is the volume of space. For gases that we live and breath every day, N d 3 /V ⇠ 10 ^3 10 ^4. We make use of this small number to truncate the hierarchy of equations and replace (2.31) with

✓ @ @t

~p (^1) m

@~r 1

~p (^2) m

@~r 2

@U (~r 1 ~r 2 ) @~r 1

@~p 1

@~p 2

f 2 ⇡ 0 (2.32)

This tells us that f 2 typically varies on a time scale of ⌧ (^) coll and a length scale of d. Meanwhile, the variations of f 1 is governed by the right-hand-side of (2.30) which, by the same arguments that we just made, are smaller than the variations of f 2 by a factor of N d 3 /V. In other words, f 1 varies on the larger time scale ⌧.

In fact, we can be a little more careful when we say that f 2 varies on a time scale ⌧ (^) coll. We see that – as we would expect – only the relative position is a↵ected by the

b

δ b

θ

φ

Figure 4: The di↵erential cross section.

Let’s think about the collision between two particles. They start with momenta ~p (^) i = m~v (^) i and end with momenta ~p (^) i^0 = m~v (^) i^0 with i = 1, 2. Now let’s pick a favourite, say particle 1. We’ll sit in its rest frame and consider an onslaught of bombarding particles, each with velocity ~v 2 ~v 1. This beam of incoming particles do not all hit our favourite boy at the same point. Instead, they come in randomly distributed over the plane perpendicular to ~v 2 ~v 1. The flux, I, of these incoming particles is the number hitting this plane per area per second,

I =

N

V

|~v 2 ~v 1 |

Now spend some time staring at Figure 4. There are a number of quantities defined in this picture. First, the impact parameter, b, is the distance from the asymptotic trajectory to the dotted, centre line. We will use b and as polar coordinates to parameterize the plane perpendicular to the incoming particle. Next, the scattering angle, ✓, is the angle by which the incoming particle is deflected. Finally, there are two solid angles, d and d⌦, depicted in the figure. Geometrically, we see that they are given by

d = bdbd and d⌦ = sin ✓d✓d

The number of particles scattered into d⌦ in unit time is Id. We usually write this as

I d d⌦ d⌦ = Ib db d (2.35)

Figure 5: On the left: a point particle scattering o↵ a hard sphere. On the right: a hard sphere scattering o↵ a hard sphere.

where the di↵erential cross section is defined as

d d⌦

=^

b sin ✓

db d✓

d(b 2 ) d cos ✓

You should think of this in the following way: for a fixed (~v 2 ~v 1 ), there is a unique relationship between the impact parameter b and the scattering angle ✓ and, for a given potential U (r), you need to figure this out to get |d/d⌦| as a function of ✓.

Now we can compare this to the notation that we used earlier in (2.17). There we talked about the rate of scattering into a small area d 3 p 01 d 3 p 02 in momentum space. But this is the same thing as the di↵erential cross-section.

!(~p, ~p 2 ; ~p 10 , ~p 20 ) d 3 p 01 d 3 p 02 = |~v ~v 2 |

^ d d⌦

d⌦ (2.37)

(Note, if you’re worried about the fact that d 3 p 01 d 3 p 02 is a six-dimensional area while d⌦ is a two dimensional area, recall that conservation of energy and momenta provide four restrictions on the ability of particles to scatter. These are implicit on the left, but explicit on the right).

An Example: Hard Spheres In Section 1.2, we modelled atoms as hard spheres of diameter d. It’s instructive to figure out the cross-section for such a hard sphere.

In fact, there are two di↵erent calculations that we can do. First, suppose that we throw point-like particles at a sphere of diameter d with an impact parameter b  d/ 2 From the left-hand diagram in Figure 5, we see that the scattering angle is ✓ = ⇡ 2 ↵, where

b = d 2 sin ↵ = d 2 sin

d 2 cos

v 1 − v 2

v 1 ’− v 2 ’

b

b

d

x 1 x 2

φ

Figure 6: Two particle scattering

the impact parameter b. We’ve also shown the collision zone in this figure. Using the definitions (2.35) and (2.37), we have ✓ @f (^1) @t

coll

Z

d 3 p 2 |~v 1 ~v 2 |

Z

d db b

Z (^) x (^2)

x (^1)

@f (^2) @x =

Z

d 3 p 2 d 3 p 01 d 3 p 02 !(~p 10 , ~p 20 |~p, ~p 2 ) [f 2 (x 2 ) f 2 (x 1 )]

It remains only to decide what form the two-particle distribution function f 2 takes just before the collision at x = x 1 and just after the collision at x = x 2. At this point we invoke the assumption of molecular chaos. Just before we enter the collision, we assume that the two particles are uncorrelated. Moreover, we assume that the two particles are once again uncorrelated by the time they leave the collision, albeit now with their new momenta

f 2 (x 1 ) = f 1 (~r, ~p 1 ; t)f 1 (~r, ~p 2 ; t) and f 2 (x 2 ) = f 1 (~r, ~p 10 ; t)f 1 (~r, ~p 20 ; t)

Notice that all functions f 1 are evaluated at the same point ~r in space since we’ve assumed that the single particle distribution function is suitably coarse grained that it doesn’t vary on scales of order d. With this final assumption, we get what we wanted: the collision integral is given by ✓ @f (^1) @t

coll

Z

d 3 p 2 d 3 p 01 d 3 p 02 !(~p 10 , p~ 20 |~p, ~p 2 )

h f 1 (~r, ~p 10 )f 1 (~r, ~p 20 ) f 1 (~r, ~p)f 1 (~r, ~p 2 )

i

in agreement with (2.24).

2.3 The H-Theorem

The topics of thermodynamics and statistical mechanics are all to do with the equi- librium properties of systems. One of the key intuitive ideas that underpins their importance is that if you wait long enough, any system will eventually settle down to equilibrium. But how do we know this? Moreover, it seems that it would be rather tricky to prove: settling down to equilibrium clearly involves an arrow of time that dis- tinguishes the future from the past. Yet the underlying classical mechanics is invariant under time reversal.

The purpose of this section is to demonstrate that, within the framework of the Boltzmann equation, systems do indeed settle down to equilibrium. As we described above, we have introduced an arrow of time into the Boltzmann equation. We didn’t do this in any crude way like adding friction to the system. Instead, we merely assumed that particle velocities were uncorrelated before collisions. That would seem to be a rather minor input but, as we will now show, it’s enough to demonstrate the approach to equilibrium.

Specifically, we will prove the “H-theorem”, named after a quantity H introduced by Boltzmann. (H is not to be confused with the Hamiltonian. Boltzmann originally called this quantity something like a German E, but the letter was somehow lost in translation and the name H stuck). This quantity is

H(t) =

Z

d 3 rd 3 p f 1 (~r, ~p; t) log(f 1 (~r, ~p; t))

This kind of expression is familiar from our first statistical mechanics course where we saw that the entropy S for a probability distribution p is S = k (^) B p log p. In other words, this quantity H is simply

S = k (^) B H

The H-theorem, first proven by Boltzmann in 1872, is the statement that H always decreases with time. The entropy always increases. We will now prove this.

As in the derivation (2.4), when you’re looking at the variation of expectation values you only care about the explicit time dependence, meaning

dH dt

Z

d 3 rd 3 p (log f 1 + 1) @f (^1) @t

Z

d 3 rd 3 p log f (^1) @f (^1) @t