A property of a physical system, such as the entropy of a gas, that stays approximately constant when changes occur slowly is called an adiabatic invariant. By this it is meant that if a system is varied between two end points, as the time for the variation between the end points is increased to infinity, the variation of an adiabatic invariant between the two end points goes to zero.

In thermodynamics, an adiabatic process is a change that occurs without heat flow; it may be slow or fast. A reversible adiabatic process is an adiabatic process that occurs slowly compared to the time to reach equilibrium. In a reversible adiabatic process, the system is in equilibrium at all stages and the entropy is constant. In the 1st half of the 20th century the scientists that worked in quantum physics used the term "adiabatic" for reversible adiabatic processes and later for any gradually changing conditions which allow the system to adapt its configuration. The quantum mechanical definition is closer to the thermodynamical concept of a quasistatic process, and has no direct relation with adiabatic processes in thermodynamics.

In mechanics, an adiabatic change is a slow deformation of the Hamiltonian, where the fractional rate of change of the energy is much slower than the orbital frequency. The area enclosed by the different motions in phase space are the adiabatic invariants.

In quantum mechanics, an adiabatic change is one that occurs at a rate much slower than the difference in frequency between energy eigenstates. In this case, the energy states of the system do not make transitions, so that the quantum number is an adiabatic invariant.

The old quantum theory was formulated by equating the quantum number of a system with its classical adiabatic invariant. This determined the form of the Bohr-Sommerfeld quantization rule: the quantum number is the area in phase space of the classical orbit.

## Thermodynamics

In thermodynamics, adiabatic changes are those that do not increase the entropy. They occur slowly in comparison to the other characteristic timescales of the system of interest,[1] and allow heat flow only between objects at the same temperature. For isolated systems, an adiabatic change allows no heat to flow in or out.

### Adiabatic expansion of an ideal gas

If a container with an ideal gas is expanded instantaneously, the temperature of the gas doesn't change at all, because none of the molecules slow down. The molecules keep their kinetic energy, but now the gas occupies a bigger volume. If the container expands slowly, however, so that the ideal gas pressure law holds at any time, gas molecules lose energy at the rate that they do work on the expanding wall. The amount of work they do is the pressure times the area of the wall times the outward displacement, which is the pressure times the change in the volume of the gas:

${\displaystyle dW=PdV={Nk_{B}T \over V}dV}$

If no heat enters the gas, the energy in the gas molecules is decreasing by the same amount. By definition, a gas is ideal when its temperature is only a function of the internal energy per particle, not the volume. So

${\displaystyle dT={1 \over NC_{v}}dE}$

Where ${\displaystyle C_{v}}$ is the specific heat at constant volume. When the change in energy is entirely due to work done on the wall, the change in temperature is given by:

${\displaystyle NC_{v}dT=-dW=-{N{k_{B}}T \over V}dV}$

This gives a differential relationship between the changes in temperature and volume, which can be integrated to find the invariant. The constant ${\displaystyle k_{B}}$ is just a unit conversion factor, which can be set equal to one:

${\displaystyle \,d(C_{v}N\log T)=-d(N\log V)}$

So

${\displaystyle \,C_{v}N\log T+N\log V}$

is an adiabatic invariant, which is related to the entropy

${\displaystyle \,S=C_{v}N\log T+N\log V-N\log N=N\log(T^{C_{v}}V/N)}$

So entropy is an adiabatic invariant. The N log(N) term makes the entropy additive, so the entropy of two volumes of gas is the sum of the entropies of each one.

In a molecular interpretation, S is the logarithm of the phase space volume of all gas states with energy E(T) and volume V.

For a monatomic ideal gas, this can easily be seen by writing down the energy,

${\displaystyle E={1 \over 2m}\sum _{k}p_{k1}^{2}+p_{k2}^{2}+p_{k3}^{2}}$

The different internal motions of the gas with total energy E define a sphere, the surface of a 3N-dimensional ball with radius ${\displaystyle \scriptstyle {\sqrt {2mE}}}$. The volume of the sphere is

${\displaystyle {2\pi ^{3N/2}(2mE)^{{3N-1} \over 2}} \over {\Gamma (3N/2)}}$,

where ${\displaystyle \Gamma }$ is the Gamma function.

Since each gas molecule can be anywhere within the volume V, the volume in phase space occupied by the gas states with energy E is

${\displaystyle {2\pi ^{3N/2}(2mE)^{{3N-1} \over 2}}V^{N} \over {\Gamma (3N/2)}}$.

Since the N gas molecules are indistinguishable, the phase space volume is divided by ${\displaystyle N!=\Gamma (N+1)}$, the number of permutations of N molecules.

Using Stirling's approximation for the gamma function, and ignoring factors that disappear in the logarithm after taking N large,

${\displaystyle S=N{\big (}3/2\log(E)-3/2\log(3N/2)+\log(V)-\log(N){\big )}}$
${\displaystyle =N{\big (}3/2\log(\scriptstyle {\frac {2}{3}}\displaystyle E/N)+\log(V/N){\big )}}$

Since the specific heat of a monatomic gas is 3/2, this is the same as the thermodynamic formula for the entropy.

### Wien's law - adiabatic expansion of a box of light

For a box of radiation, ignoring quantum mechanics, the energy of a classical field in thermal equilibrium is infinite, since equipartition demands that each field mode has an equal energy on average and there are infinitely many modes. This is physically ridiculous, since it means that all energy leaks into high frequency electromagnetic waves over time.

Still, without quantum mechanics, there are some things that can be said about the equilibrium distribution from thermodynamics alone, because there is still a notion of adiabatic invariance that relates boxes of different size.

When a box is slowly expanded, the frequency of the light recoiling from the wall can be computed from the Doppler shift. If the wall is not moving, the light recoils at the same frequency. If the wall is moving slowly, the recoil frequency is only equal in the frame where the wall is stationary. In the frame where the wall is moving away from the light, the light coming in is bluer than the light coming out by twice the Doppler shift factor v/c.

${\displaystyle \Delta f={2v \over c}f}$

On the other hand, the energy in the light is also decreased when the wall is moving away, because the light is doing work on the wall by radiation pressure. Because the light is reflected, the pressure is equal to twice the momentum carried by light, which is E/c. The rate at which the pressure does work on the wall is found by multiplying by the velocity:

${\displaystyle \,\Delta E=v{2E \over c}}$

This means that the change in frequency of the light is equal to the work done on the wall by the radiation pressure. The light that is reflected is changed both in frequency and in energy by the same amount:

${\displaystyle {\Delta f \over f}={\Delta E \over E}}$

Since moving the wall slowly should keep a thermal distribution fixed, the probability that the light has energy E at frequency f must only be a function of E/f.

This function cannot be determined from thermodynamic reasoning alone, and Wien guessed at the form that was valid at high frequency. He supposed that the average energy in high frequency modes was suppressed by a Boltzmann-like factor. This is not the expected classical energy in the mode, which is ${\displaystyle 1/2\beta }$ by equipartition, but a new and unjustified assumption that fit the high-frequency data.

${\displaystyle \,\langle E_{f}\rangle =e^{-\beta hf}}$

When the expectation value is added over all modes in a cavity, this is Wien's distribution, and it describes the thermodynamic distribution of energy in a classical gas of photons. Wien's Law implicitly assumes that light is statistically composed of packets that change energy and frequency in the same way. The entropy of a Wien gas scales as the volume to the power N, where N is the number of packets. This led Einstein to suggest that light is composed of localizable particles with energy proportional to the frequency. Then the entropy of the Wien gas can be given a statistical interpretation as the number of possible positions that the photons can be in.

## Classical mechanics - action variables

Pendulum with extra small vibration where ${\displaystyle \omega (t)={\sqrt {g \over L(t)}}\approx {\sqrt {g \over L_{0}}}}$ and ${\displaystyle L(t)\approx L_{0}+\varepsilon (t)+...}$

Suppose that a Hamiltonian is slowly time varying, for example, a one-dimensional harmonic oscillator with a changing frequency.

${\displaystyle H_{t}(p,x)={p^{2} \over 2m}+{m\omega (t)^{2}x^{2} \over 2}\,}$

The action J of a classical orbit is the area enclosed by the orbit in phase space.

${\displaystyle J=\int _{0}^{T}p(t){dx \over dt}dt\,}$

Since J is an integral over a full period, it is only a function of the energy. When the Hamiltonian is constant in time and J is constant in time, the canonically conjugate variable ${\displaystyle \theta }$ increases in time at a steady rate.

${\displaystyle {d\theta \over dt}={\partial H \over \partial J}=H\,'(J)\,}$

So the constant ${\displaystyle H\,'}$ can be used to change time derivatives along the orbit to partial derivatives with respect to ${\displaystyle \theta }$ at constant J. Differentiating the integral for J with respect to J gives an identity that fixes ${\displaystyle H\,':}$

${\displaystyle {dJ \over dJ}=1=\int _{0}^{T}{\bigg (}{\partial p \over \partial J}{dx \over dt}+p{\partial \over \partial J}{dx \over dt}{\bigg )}dt=H\,'\int _{0}^{T}{\bigg (}{\partial p \over \partial J}{\partial x \over \partial \theta }-{\partial p \over \partial \theta }{\partial x \over \partial J}{\bigg )}dt\,}$

The integrand is the Poisson bracket of x and p. The Poisson bracket of two canonically conjugate quantities like x and p is equal to 1 in any canonical coordinate system. So

${\displaystyle 1=H\,'\int _{0}^{T}\{x,p\}\,dt=H\,'\,T\,}$

and ${\displaystyle H\,'}$ is the inverse period. The variable ${\displaystyle \theta }$ increases by an equal amount in each period for all values of J - it is an angle-variable.

The Hamiltonian is a function of J only, and in the simple case of the harmonic oscillator.

${\displaystyle \,H=\omega J\,}$

When H has no time dependence, J is constant. When H is slowly time varying, the rate of change of J can be computed by re-expressing the integral for J

${\displaystyle J=\int _{0}^{2\pi }p{\partial x \over \partial \theta }d\theta \,}$

The time derivative of this quantity is

${\displaystyle {dJ \over dt}=\int _{0}^{2\pi }{\bigg (}{dp \over dt}{\partial x \over \partial \theta }+p{d \over dt}{\partial x \over \partial \theta }{\bigg )}d\theta \,}$

Replacing time derivatives with theta derivatives, using ${\displaystyle d\theta =\omega dt\,}$ and setting ${\displaystyle \omega :=1\,}$ without loss of generality (${\displaystyle \omega }$ being a global multiplicative constant in the resulting time derivative of the action), yields

${\displaystyle {dJ \over dt}=\int _{0}^{2\pi }{\bigg (}{\partial p \over \partial \theta }{\partial x \over \partial \theta }+p{\partial \over \partial \theta }{\partial x \over \partial \theta }{\bigg )}d\theta \,}$

So as long as the coordinates J, ${\displaystyle \theta }$ do not change appreciably over one period, this expression can be integrated by parts to give zero. This means that for slow variations, there is no lowest order change in the area enclosed by the orbit. This is the adiabatic invariance theorem - the action variables are adiabatic invariants.

For a harmonic oscillator, the area in phase space of an orbit at energy E is the area of the ellipse of constant energy,

${\displaystyle E={p^{2} \over 2m}+{m\omega ^{2}x^{2} \over 2}\,}$

The x-radius of this ellipse is ${\displaystyle \scriptstyle {\sqrt {2E/\omega ^{2}m}}}$, while the p-radius of the ellipse is ${\displaystyle \scriptstyle {\sqrt {2mE}}}$. Multiplying, the area is ${\displaystyle 2\pi E/\omega }$. So if a pendulum is slowly drawn in, so that the frequency changes, the energy changes by a proportional amount.

### Old quantum theory

After Planck identified that Wien's law can be extended to all frequencies, even very low ones, by interpolating with the classical equipartition law for radiation, physicists wanted to understand the quantum behavior of other systems.

The Planck radiation law quantized the motion of the field oscillators in units of energy proportional to the frequency:

${\displaystyle E=hf=\hbar \omega \,}$

The quantum can only depend on the energy/frequency by adiabatic invariance, and since the energy must be additive when putting boxes end to end, the levels must be equally spaced.

Einstein, followed by Debye, extended the domain of quantum mechanics by considering the sound modes in a solid as quantized oscillators. This model explained why the specific heat of solids approached zero at low temperatures, instead of staying fixed at ${\displaystyle 3k_{B}}$ as predicted by classical equipartition.

At the Solvay conference, the question of quantizing other motions was raised, and Lorentz pointed out a problem, known as Rayleigh-Lorentz pendulum. If you consider a quantum pendulum whose string is shortened very slowly, the quantum number of the pendulum cannot change because at no point is there a high enough frequency to cause a transition between the states. But the frequency of the pendulum changes when the string is shorter, so the quantum states change energy.

Einstein responded that for slow pulling, the frequency and energy of the pendulum both change but the ratio stays fixed. This is analogous to Wien's observation that under slow motion of the wall the energy to frequency ratio of reflected waves is constant. The conclusion was that the quantities to quantize must be adiabatic invariants.

This line of argument was extended by Sommerfeld into a general theory: the quantum number of an arbitrary mechanical system is given by the adiabatic action variable. Since the action variable in the harmonic oscillator is an integer, the general condition is:

${\displaystyle \int pdq=nh\,}$

This condition was the foundation of the old quantum theory, which was able to predict the qualitative behavior of atomic systems. The theory is inexact for small quantum numbers, since it mixes classical and quantum concepts. But it was a useful half-way step to the new quantum theory.

## Plasma physics

In plasma physics there are three adiabatic invariants of charged particle motion.

### The first adiabatic invariant, μ

The magnetic moment of a gyrating particle is

${\displaystyle \mu ={\frac {p_{\perp }^{2}}{2mB}}}$

which respects special relativity.[2] ${\displaystyle p_{\perp }=\gamma mv_{\perp }}$is the relativistic momentum perpendicular to the magnetic field. ${\displaystyle \mu }$ is a constant of the motion to all orders in an expansion in ${\displaystyle \omega /\omega _{c}}$, where ${\displaystyle \omega }$ is the rate of any changes experienced by the particle, e.g., due to collisions or due to temporal or spatial variations in the magnetic field. Consequently, the magnetic moment remains nearly constant even for changes at rates approaching the gyrofrequency. When ? is constant, the perpendicular particle energy is proportional to B, so the particles can be heated by increasing B, but this is a 'one shot' deal because the field cannot be increased indefinitely. It finds applications in magnetic mirrors and magnetic bottles.

There are some important situations in which the magnetic moment is not invariant:

• Magnetic pumping: If the collision frequency is larger than the pump frequency, ? is no longer conserved. In particular, collisions allow net heating by transferring some of the perpendicular energy to parallel energy.
• Cyclotron heating: If B is oscillated at the cyclotron frequency, the condition for adiabatic invariance is violated and heating is possible. In particular, the induced electric field rotates in phase with some of the particles and continuously accelerates them.
• Magnetic cusps: The magnetic field at the center of a cusp vanishes, so the cyclotron frequency is automatically smaller than the rate of any changes. Thus the magnetic moment is not conserved and particles are scattered relatively easily into the loss cone.

### The second adiabatic invariant, J

The longitudinal invariant of a particle trapped in a magnetic mirror,

${\displaystyle J=\int _{a}^{b}p_{\parallel }ds}$

where the integral is between the two turning points, is also an adiabatic invariant. This guarantees, for example, that a particle in the magnetosphere moving around the Earth always returns to the same line of force. The adiabatic condition is violated in transit-time magnetic pumping, where the length of a magnetic mirror is oscillated at the bounce frequency, resulting in net heating.

### The third adiabatic invariant, Φ

The total magnetic flux ${\displaystyle \Phi }$ enclosed by a drift surface is the third adiabatic invariant, associated with the periodic motion of mirror-trapped particles drifting around the axis of the system. Because this drift motion is relatively slow, ${\displaystyle \Phi }$ is often not conserved in practical applications.

## References

1. ^ Anosov, D. V.; Favorskii, A. P. (1988). "Adiabatic invariant". In Hazewinkel, Michiel (ed.). Encyclopedia of Mathematics. Vol. 1 (A-B). Reidel, Dordrecht. pp. 43-44. ISBN 9789401512398.
2. ^ Tao, Xin; Chan, Anthony A.; Brizard, Alain J. (2007-09-01). "Hamiltonian theory of adiabatic motion of relativistic charged particles". Physics of Plasmas. 14 (9): 092107. arXiv:0706.1925. doi:10.1063/1.2773702. ISSN 1070-664X. S2CID 119142268.
• Yourgrau, Wolfgang; Stanley Mandelstam (1979). Variational Principles in Dynamics and Quantum Theory. New York: Dover. ISBN 978-0-486-63773-0. §10
• Pauli, Wolfgang (1973). Charles P. Enz (ed.). Pauli Lectures on Physics. Vol. 4. Cambridge, Mass: MIT Press. ISBN 978-0-262-66035-8. pp. 85-89