PhysicsToday.org
 
access your subscriber profile Log off of Physics Today online
Search
advanced search
 
Table of contents
Past issues
RSS feed What is this?
Links to advertisers
Products advertised
Place an ad
Buyers' guide
About us
Contact us
Submit press release
American Institute of Physics
The Industrial Physicist
Computing in Science & Engineering
Journals
Virtual Journals
Articles

Conversations on Nonequilibrium Physics With an Extraterrestrial

Nonequilibrium systems come in many varieties, and a number of not−yet−reconciled mathematical approaches can be applied to them.

A couple of years ago I was visited by a Slimy Galactic Superoctopus−−a being from outer space who, for the occasion, had taken on a female human appearance. Were it not for her spaceship in my garden and a few other details, I would have said she was just a pleasant young woman, very intelligent and pretty. She told me her name was Pallas. She had been writing a galactic PhD thesis on the nature of human mathematics, and we had long discussions on that and other topics.1 Then Pallas had to pass her exams and she disappeared into the darkness of space. I had little hope of seeing her again. But once more, her tall, white, silent spaceship stood in my garden, and when we met at the garden door, it was as if she had never been away.

While we were strolling along one of the paths we had walked before, she said, "I would like to understand your interests in mathematical physics."

Figure 1
"Currently I am looking at problems in statistical mechanics, nonequilibrium to be specific," I answered. "It's an area of rapid progress in which scientists have finally gained some fundamental understanding of situations far from equilibrium. Many biological processes, by the way, are in that category." (See figure 1.)

Equilibrium and nonequilibrium statistical mechanics

"Statistical mechanics," I continued, "attempts to explain the macroscopic properties of matter in terms of the interactions of its microscopic constituents. To be precise, one starts with a mathematical description of the time evolution of a system of many particles−−molecules for example. The evolution is deterministic, defined by a Hamiltonian H that may be classical or quantum. I shall now squeeze more than a century of work on equilibrium and nonequilibrium statistical mechanics into a few words. Ready?"

"Yes. Please go on."

"Statistical mechanics was developed at the end of the 19th century by, among others, James Clerk Maxwell, Ludwig Boltzmann, and Josiah Willard Gibbs. Equilibrium statistical mechanics is concerned with certain states of matter that appear macroscopically at rest, in equilibrium, and that are microscopically a superposition of states i, with probabilities pi. A simple example is that of a quantum spin system, a type of quantum system with a finite−dimensional Hilbert space. Let E1, . . ., En be the eigenvalues of the Hamiltonian H, and express temperature in units of energy. An equilibrium state at inverse temperature β is then defined by associating with each eigenvalue Ei a probability pi = exp(−βEi)/Σj exp(−βEj), obtained by normalizing the so−called Boltzmann factor exp(−βEi). For more general systems, the set of states will be infinite and sums may be replaced by integrals.

"If the Hamiltonian describes particles that repel each other at short distances and have negligible interaction at large distances, then an asymptotic 'thermodynamic limit' exists in which the volume and number of particles tend to infinity at fixed density and temperature. Or, instead of fixing the temperature, one can fix the energy density: The two types of description are equivalent. In favorable cases, one can study phase transitions, the critical points, and so forth. Difficult problems remain, though, such as proving the existence of crystal phases, but the successes of equilibrium statistical mechanics during the 20th century were stupendous.

"After the foundations were laid in the 19th century, equilibrium statistical mechanics developed slowly at first, then rapidly. The 19th century also saw the development of basic ideas of nonequilibrium statistical mechanics. Boltzmann showed how entropy, which measures the amount of microscopic randomness compatible with a macroscopic description, could statistically only increase with time. The philosophically correct views of Boltzmann, however, did not lead to a simple, general, useful prescription that could be systematically used to attack nonequilibrium problems." (For a modern description of Boltzmann's ideas, see the article by Joel Lebowitz in Physics Today, September 1993, page 32.)

"Note that instead of nonequilibrium," I continued, "people also speak of dissipation or irreversibility. The three terms refer to different aspects of the same thing: When a system is outside of equilibrium, it dissipates energy as heat, and that happens with an irreversible increase of the entropy. Mechanical, chemical, and electrical energy may all be dissipated in irreversible processes."

Dynamics

All the time I had been describing statistical mechanics, Pallas had been smiling and had remained silent. Now she said, "Can you tell me the real reason why nonequilibrium is harder than equilibrium? And can you give me an example of both?"

I answered, "In the description that I just gave, equilibrium theory starts with the Boltzmann factors, exp (−βEi): The dynamics−−that is, time evolution−−is eliminated. All one needs to know are the energies of the various states. In nonequilibrium problems, the dynamics cannot be forgotten. Consider specific heat and heat resistance. For specific heat, one asks how much heat needs to be put in a lump of matter to increase its temperature by one degree. For heat resistance, one asks what temperature difference needs to be put across a lump of matter to force a unit of energy to pass through the lump per unit time. The two problems may not sound very different, but the first is equilibrium, does not involve dynamics, and is relatively easy. The second is nonequilibrium, involves dynamics, and is difficult."

By the time she asked her next question, Pallas's smile had definitely become mischievous.

"Because dynamics is important for questions of nonequilibrium, wouldn't you start with simple dynamics that you can understand in detail?"

"A natural idea, indeed, is to study the heat resistance of a chain of harmonically coupled harmonic oscillators," I replied. "This system is completely integrable, the Hamiltonian is quadratic, and everything can be computed explicitly. But the results are pathological: The resistance is not asymptotically proportional to the length of the chain as one would expect on physical grounds.2 Completely integrable systems may be all right for easy computation, but they are not what we need to study nonequilibrium systems. If you think about it, completely integrable dynamics is exceptional, and we should consider a type of dynamics that is common rather than exceptional.

"Chaotic dynamics is a natural choice. That somewhat loose term describes time evolution with sensitive
dependence on initial conditions. Thus, a small change δx(0) in the initial condition x(0) typically gives δx(t) ≈ δx(0) exp(λt), with the so−called Lyapunov exponent λ > 0. There exist mathematical arguments that most dynamical systems are chaotic in some sense. So it is reasonable to assume chaos for the microscopic dynamics−−both for equilibrium and nonequilibrium situations. But while chaotic dynamics is common, such behavior does not cover all situations of physical interest. In fact, Andrei Kolmogorov, Vladimir Arnold, and Jürgen Moser have described distinctly nonchaotic behavior that is relevant to some situations in statistical mechanics."3

Pallas interrupted. "I see a difficulty. You have to base the study of nonequilibrium systems on a mathematical understanding of dynamics, but you don't know what general dynamical systems look like. For thousands of years, scientists throughout the galaxy have tried to understand general dynamical systems, and it is not an easy problem. To avoid getting mired in mathematical questions beyond human capabilities, perhaps you should stay closer to physics."

I get upset when Pallas implies that the human brain is a rather limited intellectual tool, better suited to hunting rabbits than doing mathematics, and that Slimy Galactic Superoctopuses are mathematically quite superior to us. But keeping the physics in mind is certainly a good point.

Diverse nonequilibrium systems

"The Newtonian gravitational force," I noted, "does not lead to thermodynamic equilibrium because it is attracting and also because it is long range. Water left in a bottle outside the house in winter may remain liquid well below 0°C. It is in a metastable state, but will freeze rapidly when one bangs the bottle. When a piece of iron is placed in an oscillating magnetic field, it produces heat, and the magnetization of the iron depends on previous values of the magnetic field, not just the instantaneous value. That behavior is called hysteresis. A temperature difference imposed between two faces of a lump of matter may lead to a nonequilibrium steady state, or NESS, in which heat is transported from the warm to the cold face. Another kind of NESS can arise when reacting chemicals are pumped into a tank and stirred, and excess mixture is able to flow out, although sometimes for such systems, one gets oscillations. Gravitational collapse, metastability, magnetic hysteresis, transport, and chemical reactions are all examples of nonequilibrium phenomena [see figure 2]. Since the physics of those examples is rather different, I'll concentrate on NESSs, and allow both transport and chemical reactions.

Figure 2
"Spatial structure is important for transport phenomena, in which heat, chemical concentration, or momentum carried by a viscous fluid is transported through matter. Typically, the transport phenomena are locally close to equilibrium. The matter in which transport occurs appears close to an equilibrium state at some temperature, pressure, and so forth when one observes on a small but still macroscopic scale, say 10−2 mm. Note that the macroscopic equations for transport, such as the heat and diffusion equations or the Navier− Stokes equation, usually contain a Laplace operator −2 with an appropriate coefficient, called a transport coefficient. The Laplace operator describes diffusion of heat, concentration, or momentum. One might say that in transport phenomena, dissipation is of diffusive nature. Chemical reactions, which are far from equilibrium, normally have no macroscopic spatial structure." Later on in our stroll, I told Pallas a bit more about transport coefficients and introduced her to some important 20th−century contributors to nonequilibrium statistical mechanics. (See box 1.)

Deterministic thermostats

"Physics," I informed Pallas, "seems to be in a foundational period for nonequilibrium statistical mechanics, with an explosion of very interesting and sophisticated work. Assumptions vary−−classical or quantum theory, deterministic or stochastic dynamics, finite or infinite number of degrees of freedom. The relations between the various approaches have not yet been worked out.

"When a system is out of equilibrium, it dissipates energy into heat. Therefore, for it to reach a steady state, the system needs a thermostat to cool it. One useful idealization of such a system has a finite number of degrees of freedom subject to non−Hamiltonian forces. William Hoover and Dennis Evans modified the Hamiltonian equation

Equ1A(1a)

to obtain

Equ1B(1b)

where ξ is no longer assumed to be a gradient, and α = ξ(q) ⋅ p/p2. For simplicity, I'm expressing quantities as if they were one−dimensional, but my discussion is general and applies to phase space of arbitrary even dimension.

"The term −αp implements the so−called isokinetic thermostat and ensures that the time evolution preserves the kinetic energy p2/2m. Note that in the lab, a thermostat usually cools a system at its boundary. By contrast, the isokinetic thermostat cools the bulk. That is physically somewhat different, but not unreasonable, and very convenient for mathematical discussion."4

"Isn't it distressing," Pallas interrupted, "that you abandon Hamiltonian mechanics? Is it necessary?"

"A possible way to look at it," I replied, "is that one starts with an infinite system because an infinite reservoir acts as a thermostat. Then one integrates away the coordinates of the reservoir and is left with a finite non−Hamiltonian system. Because of the non−Hamiltonian dynamics, the volume in phase space is no longer conserved. As I'll explain soon, that volume contraction can be identified with entropy production."

"Integrate away the reservoir," said Pallas. "That is easier said than done."

"Exactly. It can be done formally, but we are still waiting for a serious treatment. Note that Pierre Gaspard has a purely Hamiltonian approach based on diffusion. His work begs to be connected to other approaches."

A theory of NESSs

"A few years ago," I continued, "Giovanni Gallavotti made me aware that various ideas present in the literature, when put together, amount to a theory of NESSs in which one does not assume closeness to equilibrium. That observation was used by Gallavotti and Eddie Cohen to prove their important fluctuation theorem.5 But before I go into that, let me outline a five−point general theoretical framework for NESSs. It's basically what Gallavotti told me, with the addition of the linear−response formula."

Deterministic dynamics. "I assume that the nonequilibrium time evolution is given by

Equ2 (2)

where I have further abbreviated the notation by writing x instead of (p, q) for a point in phase space with fixed kinetic energy. The symbol Χ denotes a vector field, such as appeared on the right−hand side of equation 1b.

"The solution to equation 2 may be expressed in the form x(t) = ftx(0), which defines a nonlinear evolution operator ft. It is often desirable to impose time−reversal symmetry. In the case of equation 1b for example, that symmetry means the equation remains valid when (t, p) goes to (−t, −p) while q and ξ are unchanged."

Chaos. "Earlier I noted that chaotic time evolution is natural in both equilibrium and nonequilibrium systems. A chaotic time evolution stretches distances in some directions of phase space−−that stretching leads to the sensitive dependence on initial conditions−−and it contracts distances in other directions. Gallavotti and Cohen imposed a chaotic hypothesis in the mathematically precise sense of uniform hyperbolicity.6 They also assume that time is a discrete variable. Uniform hyperbolicity is perhaps too strong physically so, for now, I'll remain vague about just what I mean by chaotic dynamics and make further assumptions as I go."

A measure for a NESS. "Since the time evolution described by equation 2 is not Hamiltonian, it will not, in general, preserve the phase−space volume m, no matter how that volume is defined. (Think of dp dq if you want to have a specific example for m.) In light of the nonconservation of volume, I'll consider more general measures ρ with respect to which any continuous function A has an integral −− ρ(dx) A(x). The measure ρ may be singular, as is the case for the Dirac delta measure.

"For a chaotic system, many measures are invariant under time evolution. Which one is the best choice to describe a NESS? Suppose that the time averages of continuous functions A are given by a measure ρ in the sense that

Equ3(eqn 3)

Despite appearances, it is often the case that the limit in the left−hand side is independent of time t and phase−space coordinate x. Thus, for a large set of initial conditions−−in particular, a set for which the measure m gives a nonzero volume−−measures like ρ implement time averages and are physically natural choices for a NESS. Called SRB measures, they were introduced by Yakov Sinai, Rufus Bowen, and me, then discussed more generally by Jean−Marie Strelcyn, François Ledrappier, and Lai−Sang Young. As early as the 1970s, it appeared that such measures might be appropriate for describing nonequilibrium systems, but the idea could be implemented only after deterministic thermostats were introduced.

Box 2
"One may define a NESS in terms of an SRB measure ρ; in particular, one assumes that such a measure exists. Completely integrable systems do not have SRB measures as just described, but an SRB measure is uniquely determined when the chaotic hypothesis of Gallavotti and Cohen is satisfied. Note that the introduction of the measure breaks time−reversal symmetry because I let the time τ go to −∞, not +∞."

  At this point, Pallas voiced a request: "I wonder if you might draw a picture to illustrate the stretching of phase−space dimensions and the singular nature of the SRB measure."

I found a napkin and sketched the figure that appears in box 2.

Entropy production as volume contraction. "The average rate of entropy production in the NESS compatible with ρ is

Equ4 (4)

where the divergence is computed with respect to the volume element m. Invariance of ρ under time evolution implies that any choice of m gives the same value for e(ρ).

"One can obtain equation 4 by a formal calculation based on the Gibbs expression −∫{m(dx)g(x) ln g(x)} for the entropy corresponding to a probability distribution g(x)m(dx). A key technical element in the calculation is that [as discussed in box 2] the appropriately normalized volume m evolves to the SRB measure ρ as time becomes infinite. The relation between equation 4 and the Gibbs expression was apparently first noted by Ladislav Andrey.7 The general relation between entropy production rate and phase−space volume contraction is still debated, but equation 4 appears to be correct when one has a deterministic thermostat.

"If ρ is an SRB measure, one can show that e(ρ) ≥ 0; for the interesting case of dissipation, e(ρ) > 0. The positivity of e(ρ) is tied to the fact that introducing the SRB measure breaks time−reversal symmetry. In the Hamiltonian case, time evolution preserves phase−space volume and divΧ = 0."

Linear response. "Statistical mechanics aims to compute the response of a system to an external action. The specific heat, which describes how a system reacts to external heat input, is an example. To include an external action, one can modify the time evolution, equation 2, to obtain

Equ5A(5a)

where Xt is a small perturbation of Χ that may depend on the time t. Given such a perturbation, the SRB state ρ would be replaced by ρ + δtρ. A simple formal calculation yields

Equ5B(5b)

where ft is the evolution operator for the unperturbed time evolution. In the linear−response formula, equation 5b, A is assumed to be differentiable rather than just continuous. The formula can be rigorously proved under the assumption of uniform hyperbolicity−−the chaotic hypothesis of Gallavotti and Cohen." For further discussion see references 6 and 8.

Fluctuation theorems

"So," said Pallas, "you claim that the five points you just listed constitute a theory of nonequilibrium? In what sense?"

"In the sense that, given a suitable physical question, one can in principle write an answer. There may be integrals, limits, and so forth that one cannot explicitly perform, but such difficulties also accompanied equilibrium statistical mechanics after Boltzmann and Gibbs, and the 'close to equilibrium' theory developed by Lars Onsager, Mel Green, and Ryogo Kubo around 1950. In fact, the theory I just outlined is an extension of that close−to−equilibrium work, with closeness to equilibrium not assumed. Now I want to discuss briefly two physically significant examples.

"First is the remarkable Gallavotti−Cohen fluctuation theorem.5 Their formula, which has no adjustable parameters, gives the fluctuations in the rate of entropy production for a NESS. It has been rigorously proved, given the chaotic hypothesis, and thus adds a new dimension to earlier, more formal results.9 The chaotic hypothesis is overly restrictive, and it is remarkable that it gives the right answer: Apparently, dynamical systems associated with statistical mechanics often behave as if they were uniformly hyperbolic.

Figure 3
"My second example is also related to fluctuations. Consider the fluctuation−dissipation theorem, which is valid close to equilibrium. The theorem relates the spontaneous fluctuations of a system to the response of the system to an external action. The physical idea is that in order to know how a system reacts to a kick, one might as well wait until the system has a fluctuation equivalent to the result of that kick. Then one can see what happens. A special case of the fluctuation−dissipation theorem is the Green−Kubo formula, which expresses transport coefficients in terms of time correlation functions.

"Efforts to extend the fluctuation−dissipation theorem far from equilibrium would be expected to fail in general. Indeed, in the five−point theory I laid out, the NESS lives on a phase−space attractor that is generally fractal. One could kick the system to a point outside of the attractor, but such a point cannot be visited by spontaneous fluctuations. Therefore, fluctuations cannot reveal how the system reacts to such a kick. One might, though, expect that the theorem could be partially extended to allow for kicks that keep the system on the attractor."6,8

Pallas interrupted me. "I can see a couple of big problems that you have not addressed. One is that statistical mechanics is generally about large systems, systems with many degrees of freedom. How do you take the limit?"

"That indeed may be a formidable problem," I admitted. "After all, in some cases a large system will undergo oscillations in time rather than behave as a steady state! My inclination is to postpone the study of the large−system limit: Since it is feasible to analyze the nonequilibrium properties of finite systems−−as Gibbs did for their equilibrium properties−−it seems a good idea to start there. That may not answer all questions, but it advances nonequilibrium statistical mechanics to the point equilibrium had reached after Gibbs."

But Pallas saw another big problem. She noted that varying the parameters of a physical system amounts to changing the time evolution, equation 2, which typically produces many changes of behavior, so−called bifurcations. The linear−response formula, equation 5b, assumes smooth dependence on parameters, and that's unlikely to hold. So the integration over time in that formula may be divergent.

"In fact," I told her, "the same problem arises for the Green−Kubo formula close to equilibrium. Nico van Kampen has remarked that, in situations in which one typically applies Green−Kubo, one would expect the linear−response formula to be mathematically valid only for unphysically small perturbations of equilibrium.10 How serious are these objections? Consider the following three facts. First, Green−Kubo works. That is, it is consistent with experiments and computer simulations with large systems. Second, the linear−response formula can be proved for uniformly hyperbolic systems, including those considered by Gallavotti and Cohen. Third, bifurcations messing up Green−Kubo are seen in systems that are small and nonuniformly hyperbolic.

"Theorists believe that most physical systems are not uniformly hyperbolic and that they would exhibit many bifurcations. Nonetheless, typical large systems seem to behave as if they were uniformly hyperbolic. One would certainly like to understand why that is the case.

"When we have more time, I can explain a variety of technical results that really convince me that we are on the right track to understanding nonequilibrium physics."11

Coda

With a somewhat ironic smile, Pallas said, "On my next visit to Earth, I will check with you to see if ideas on nonequilibrium statistical mechanics have progressed in the direction you indicated. But for now let us leave this discussion. Would you like to listen to some music?"

"Well . . . there are relations between nonequilibrium and music, but is that the real reason you ask?"

"It is something else. You know how unnatural the tampered scale is−−brutally cut by human musicians into 12 equal half−tones, every interval slightly wrong. Because of your imperfect human hearing, you do not realize it, but the result is truly abominable. And the funny thing is that I have come to enjoy it! Isn't that very perverse?"

"Tempered scale," I said weakly, "not tampered. And Johann Sebastian Bach is neither truly abominable nor very perverse." This time I was really upset.

But Pallas just smiled sweetly. "Ready to listen to some preludes and fugues from the Well−Tempered Clavier?"


This article is a reworking of the Shih−I Pai lecture I gave at the University of Maryland, College Park, in April 2003.


David Ruelle is an emeritus professor at the Institut des Hautes Etudes Scientifiques in Bures−sur−Yvette, France, and a distinguished visiting professor of mathematics at Rutgers University in New Brunswick, New Jersey.


References
1. D. Ruelle, in Mathematics: Frontiers and Perspectives, V. Arnold et al., eds., American Mathematical Society, Providence, RI (2000), p. 251.
2. H. Spohn, J. L. Lebowitz, Commun. Math. Phys. 54, 97 (1977) [INSPEC].
3. See, for example, A. Haro, R. de la Llave, Phys. Rev. Lett. 85, 1859 (2000) [INSPEC].
4. For a brief discussion of the independent introduction by William Hoover and by Dennis Evans of the isokinetic thermostat, see D. J. Evans, G. P. Morriss, Statistical Mechanics of Nonequilibrium Liquids, Academic Press, London (1990), section 5.2.
5. G. Gallavotti, E. G. D. Cohen, Phys. Rev. Lett. 74, 2694 (1995) [INSPEC]; J. Stat. Phys. 80, 931 (1995) [INSPEC].
6. See, for example, D. Ruelle, J. Stat. Phys. 95, 393 (1999) [INSPEC].
9. See, in particular, D. J. Evans, E. G. D. Cohen, G. P. Morriss, Phys. Rev. Lett. 71, 2401 (1993) [INSPEC].
10. N. G. van Kampen, Phys. Norv. 5, 279 (1971).
11. N. I. Chernov, G. L. Eyink, J. L. Lebowitz, Ya. G. Sinai, Phys. Rev. Lett. 70, 2209 (1993) [INSPEC]; Commun. Math. Phys. 154, 569 (1993) [INSPEC]. C. P. Dettmann, G. P. Morriss, Phys. Rev. E 53, R5541 (1996) [INSPEC]; M. P. Wojtkowski, C. Liverani, Commun. Math. Phys. 194, 47 (1998) [INSPEC]; D. Ruelle, Proc. Natl. Acad. Sci. U.S.A. 100, 3054 (2003) [MEDLINE].
12. I. Prigogine, Introduction to Thermodynamics of Irreversible Processes, Interscience/Wiley, New York (1968).
13. G. Nicolis, Introduction to Nonlinear Science, Cambridge U. Press, New York (1995); P. Gaspard, Chaos, Scattering, and Statistical Mechanics, Cambridge U. Press, New York (1998).
14. J. R. Dorfman, An Introduction to Chaos in Nonequilibrium Statistical Mechanics, Cambridge U. Press, New York (1999).


© 2004 American Institute of Physics



Also This Month
Light's Orbital Angular Momentum
Cryogenics on a Chip
Conversations on Nonequilibrium Physics With an Extraterrestrial

COMPANY SPOTLIGHT



Help the physics community displaced by Katrina:

Katrina Affected Physics and Astronomy Community Bulletin Board

U.S. Department of Energy Temporary Educational Matching Program