# 4 The temperature. Heat Transfer. Equilibrium. Equipartition of energy. The Ideal Gas Law

We begin this Chapter with a puzzler. Imagine the following setup, where a rigid, heat-insulated container is divided in two parts by a rigid piston that can freely move left of right. Each part contains a nearly ideal gas for which the following parameters are known: mass , concentration , and average velocity squared . The quantities pertaining to the l.h.s. are labeled with “1”, and to the r.h.s. with “2”.

Clearly the ratio of the concentrations, is in one-to-one correspondence with the position of the piston and thus can be changed at will by moving the piston. Indeed, , , where the total volume remains constant by construction. Thus, , where is fixed, also by construction.

Now the puzzler: Suppose we prepare the system so that the following conditions are satisfied:

(1)

while

(2)

**The puzzler is: Will the piston move or will it remain stationary? Is the system in true equilibrium or not?
**

We have seen in the preceding Chapter that the pressure

(3)

Suppose the area of the piston is . Thus, the net force acting on the piston:

(4)

should vanish by Eqns. (1) and (3). Thus the piston is in a state of mechanical equilibrium and should stay put. Or, should it?

Before we proceed further, a short comment regarding the term “nearly ideal gas”. In a nearly ideal gas, the particles are almost never in contact with one another and collide (interact) just often enough to “thermalize”, or randomize their motions so that the particles sample all possible values of the velocity and coordinate without any bias. As a result, any microscopic configuration will be visited again and again. This unbiased, recurrent sampling of a distribution that remains steady for an arbitrarily long time corresponds to a statistically steady, “equilibrated” system. In formal terms, equilibrium means that one may define a fixed, time-independent distribution for all the coordinates and velocities. Every quantity will fluctuate with in time, but the fluctuations each will be around a steady value:

Only if this steady value remains steady *forever*, can we technically talk of “equilibrium”. In actuality, infinite observation times are unrealistic, and so when we say “infinite time”, we really mean “infinite, in principle”.

One useful property of a nearly ideal gas is that its energy is almost exclusively contained in the motion of the molecules, and almost none of it is stored in the potential energy:

(5)

where by we denote the full set of the coordinates of the particles.

Now back to our *puzzler*. As in the preceding Chapter, we are interested only in the horizontal component of particle’s velocity. The shield can move exclusively horizontally, by construction. Let us again use a formula we derived in Chapter 2 to write down an expression for the velocity of particle that had just collided with a shield. The particle and the shield had velocities and , respectively, before the collision:

(6)

After we square this equation and average the resulting expression over the collisions—and, thus, over molecules—we obtain the following:

(7)

Let us see now that in the approximation where the velocity of an incoming particle and the velocity of the shield just prior to the collision are uncorrelated, the average of the product breaks up into the product of the respective averages of the two factors:

(8)

**Digression on statistics**: Uncorrelated variables. Suppose we have two distributed quantities, call them and . Completely analogously to how we handled a single variable case in the last Chapter, one can compute the average of any combination of two variables using a weighted average. For concreteness, consider the product .

(9)

where we now sample two quantities, and . The quantity is the statistical weight (or proability) of a configuration where the variables and to *simultaneously* fall into a small rectangular sector centered around the point .

Because the variables and define a 2D space, these small rectangular sectors form a grid in that 2D space. Thus it is practical to use *two* indices to label those sectors. Other than this, the meaning of the weight is exactly the same as that of the weight we defined for a single distributed quantity in the last Chapter.

Next, let us see that *the probabilities of uncorrelated events multiply*:

(10)

where the weight is the probability of outcome in a standalone experiment involving the variable and the weight is the probability of outcome in a standalone experiment involving variable . The superscripts and are simply labels we need to distinguish the distribution functions for the two variables, which are generally different. We begin a simple example of two fair coins, each of which, if tossed, will land heads up or tails up with equal probability 1/2. Clearly there are four equally likely outcomes for an experiment where the two coins are tossed at the same time:

coin 1 | coin 2 |

heads | heads |

heads | tails |

tails | heads |

tails | tails |

Because there are 4 outcomes total and they are all equally likely, the probability of each outcome is exactly 1/4. For instance, the probability to get two tails at the same time is 1/4. On the other hand, the probability of getting tails for an individual fair coin is 1/2. Consistent with Eq. (10), . By the same token, the probability to guess a 4-digit ATM code in one try is because there are 10,000 equally likely possibilities, i.e., numbers ranging from 0 to 9,999. Alternatively, one can look at this problem as the probability of guessing four uncorrelated one-digit numbers at the same time, again yielding .

These ideas are not difficult to generalize to arbitrary, “unfair” coins. Suppose you are tossing an unfair coin, call it “coin 1”, which yields heads with probability (40%) and tails with probability (60%). Suppose also that your friend is tossing, entirely independently of you, a different unfair coin, call it “coin 2”, whose outcomes are heads with probability and tails with probability . Suppose you have tossed your coin a times. Let’s say your experiment yielded heads. For each one of those heads, you know your friend had heads with probability , and so the number of outcomes that will count toward the “heads (coin 1) + heads (coin 2)” outcome, the top outcome entry in the Table above, is . Thus the statistical weight of the “heads-heads” outcome is . Likewise, the probability of the “tails (coin 1) + heads (coin 2)” outcome is and so on. We see that the multiplication of probabilities comes about because the success rate of a compound event is conditional on the success of both of the constituent events. When the constituent events are statistically independent, the success rates of the individual events must be simply multiplied in order for us to determine the overall success rate. For this very reason, by the way, the rate of bimolecular reaction is proportional to the product of the activities (or concentrations) of the individual reactants: One needs one of *both* species to create a product molecule.

The converse of the above logic is of equal significance for this class: If it turns out that the compound statistical weight of events 1 and 2 is not determined by simply multiplying their individual statistical weights, we may say that there is a statistical *correlation*, or, in more physical terms, “interaction” between the two degrees of freedom, such as and in an above example. In practice, we may not know beforehand whether there is any interaction. This, then, raises a question: How do we determine the individual statistical weights, in the first place? Here is how to do this. Assuming that we have performed enough experiments to sample all relevant configurations of the system, we can histogram the chances of a specific outcome for the degree of freedom 1 where the degree of freedom 2 was not even monitored but, hopefully, has adequately sampled all of its configurations. The resulting histogram is, then, simply the sum of the combined statistical weight over all configurations of the degree of freedom 2:

(11)

Likewise,

(12)

Note that, by construction, and . This automatically ensures that the combined distribution is normalized, too:

(13)

This is particularly obvious when the two variables are statistically independent:

(14)

The above equations illustrate, among other things, how to manipulate double sums where the summed object is factorizable into a product, where one factor depends exclusively on one summation variable and the other factor depends exclusively on the other summation variable. By the same token, we can now compute the average of the product of two independent variables:

(15)

If you do not feel comfortable with this type of algebraic manipulations, it may be worthwhile to practice by writing out such double sums explicitly for small values of and .

Now, we can show, using the same logic as above, that the compound probability of three or more uncorrelated events also multiply. Indeed, one can relabel the compound experiment concerning, say, coin 2 and coin 3 as experiment 2 and repeat the argument. To summarize, for an arbitrary number of distributed variables, :

(16)

**End of statistics digression**.

Now that we have established the (approximate) validity of Eq. (8), we are ready to formulate conditions for true equilibrium. Indeed, in true equilibrium, the piston is not moving, on average: , which implies, by Eq. (8), that

(17)

Furthermore, in equilibrium, all directions in space must be equivalent or, else, there would be a net flow of particles or other properties. For example, suppose for a moment that two opposite directions are not equivalent in that the particle flux in one direction is not equal to the particle flux in the opposite direction. In that case, there would be a non-vanishing *net* flow along this direction which would, then, result in a continuing piling up of particles in one place and a continuing “drain” of particles in another place. In addition, the concentration should be on average spatially uniform, too, so as to also prevent a net flow. Because the flux is the product of the concentration and velocity, respectively, the velocities of individual particles should be distributed in a way that is independent of the place and direction. For this reason, we must conclude that the characteristics of the incoming and outgoing particles, near the “collision zone”, should be on average the same:

(18)

Eqs. (7), (17), and (18) then yield, after a bit of algebra, a remarkably consequential result:

(19)

where I have added the 1/2 factors for future convenience and the subscript to emphasize that we are dealing with motion along the coordinate axis . What Eq. (19) tells us that the kinetic energy of the motion of each particle and the piston must be mutually equal in equilibrium, on the average! Since this applies to the molecules on *both* sides of the piston, we conclude that in equilibrium,

(20)

Let us refocus on an individual compartment of our original setup. Since the directions , , and are all equivalent in equilibrium, one has it that

(21)

At the same time, the velocity vector can be decomposed into three mutually orthogonal components:

Thus

(22)

By the way, because of this geometrical relation, the kinetic energy of motion in spatial dimensions greater than 1 is simply the sum of the kinetic energies of motion along individual directions:

(23)

In any event, Eq. (22) yields

(24)

Which, then, leads, by Eq. (21), to :

(25)

In view of Eq. (20), this yields that in equilibrium, the average kinetic energy of a molecule on the l.h.s and a molecule on the r.h.s., respectively, must be the same:

(26)

On the basis of Eq. (20) or (26) we then find the answer to the puzzler: The piston will actually move because the assumption of the piston being stationary is inconsistent with the condition (2). What gives?? After all, the piston appears to be in mechanical equilibrium, by Eq. (1)! The answer is that condition (1) only ensures that the net mechanical force on the piston vanishes only on the average, but not necessarily at any given moment. This uncompensated force, which fluctuates around zero, then leads to fluctuations in the position of the piston and a gradual drift either leftward or rightward, the direction depending on the sign of inequality in Eq. (2), to be determined shortly. Suppose, for concreteness, that in the beginning:

(27)

To actually reach equilibrium, the piston should be moving in the direction such that the quantity should be decreasing over time, while the quantity should be increasing over time so as to meet somewhere midway by the time equilibrium is reached:

To understand what is happening physically, we note that the total kinetic energy of the gas on the l.h.s., is decreasing, while the total kinetic energy of the gas on the r.h.s., is increasing as the system is approaching equilibrium. In other words, the energy stored in the motions of the particles is being transferred, while no other property is being transferred at the same time. Indeed, the pressures in the two compartments are mutually equal (no momentum transfer) and there is no particle transfer either (the piston is impermeable by construction).

What do we make out of this type of energy transfer, where just the kinetic energy of the molecules’ motion is moved from one place to another, while no macroscopically observable quantity is transferred and there are no visible changes in the shape or position of the objects involved? This is the familiar **heat transfer**! The above argument also provides a simple criterion for the direction of heat transfer, if any: Heat will flow from places where the quantity is greater to places where the quantity is lower. This means the quantity is a good proxy for what we call the **temperature**. In fact, by convention, we define temperature so that:

(28)

We have included the mass of the molecule inside the average to account for the possibility that more than one species can be present in the system. Recall that Eq. (26) was written with two pure gases separated by a partition, but applies equally well to a *mixture* of two nearly ideal gases. In any event, the above definition of the temperature is clearly not unique. Any other definition where is a monotonically increasing function of would be workable, but the linear dependence in the above definition happens to be quite convenient, as we will have a chance to witness many times in the future.

The Boltzmann constant J/K, is just that: a numerical constant used to convert from temperature to energy. Really, temperature has the meaning of energy, consistent with its definition of being equal, up to a multiplicative constant, to the kinetic energy of molecules.

The numerical factor of in Eq. (28) is quite arbitrary as well, but it is the convention that must be followed consistently. With this convention, we obtain that the average kinetic energy of motion per single degree of freedom is exactly :

(29)

a fact often referred to as the the **Equipartition Theorem**. A “degree of freedom” is something whose configuration can be described by a single number. For, instance, motion in 3D space corresponds with three degrees of freedom. We can extend these notions to motions other than purely translational motions. For instance, a rotation around a fixed axis corresponds with a single degree of freedom. Indeed, one can think of such a rotation as a translation of a bead confined to moving in a circle:

And so, for each rotation, there is a worth of energy. Another classic example is a harmonic oscillator. This system has also a potential energy. We have seen, however, that on the average, the kinetic and potential energy of the harmonic oscillator are equal. Thus the total energy of a harmonic oscillator, on average is just the . Non-withstanding the fact that these statements must be somtimes corrected for quantum-mechanical effects (and will be in due time), it is an incredibly useful notion that the typical thermal energy of a single degree of freedom is typically around .

Next we combine Eq. (28) with the earlier derived relation for pressure: to obtain the **Ideal Gas Law**:

(30)

Note all the quantities in the above equation are *intensive*, i.e., provide no information about the size of the system. The concentration can be easily rewritten in terms of *extensive* variables, which, by definition, scale linearly with the system size. Indeed, , which, then, yields

(31)

or

(32)

Alternative ways to write down the Ideal Gas Law is

(33)

where is the number of moles and J/(mol K) is the universal gas constant. Also common, but unphysical way to write the ideal gas law is in terms of the molar volume :

The Ideal Gas Law is both a definition of the temperature and a physical law. How does this work? We can use, as our “thermometer”, a specific gaseous substance that we think behaves like a nearly ideal gas and then make predictions about *other* gases. For instance, the ideal gas law allows one to predict the temperature of a gas, if its concentration and pressure are known. One, then, can compare this prediction by bringing test “test gas” in mechanical and thermal contact with our “thermometer gas” prepared at the same pressure and see, whether heat exchange will take place. In practice, we use thermometers that work using other physical phenomena. Nearly ideal gases are used to calibrate those more practical thermometers in a temperature range where we expect a gas behave nearly ideally. Making sure that temperature measurements are properly standardized in a broad temperature range requires much sophistication, engineering skill, and knowledge of Thermodynamics and Statistical and Quantum Mechanics.

We finish by returning to our puzzler. Suppose that the system was prepared so that , which we now know corresponds with . Recall that the pressures remain fixed and mutually equal all along, as the mechanical equilibrium is never compromised: . Now, as the *full* equilibrium is approached and the two temperatures converge onto some value midway, the concentration of gas 1 must increase while the concentration of gas 2 must decrease with time, by the ideal gas law: . This means that the piston will be moving to the *left* until equilibrium is achieved. One can further use energy conservation to connect the temperatures of the two gases to the eventual temperature of the system, i.e., after the piston attains its equilibrium position. Indeed, . The quantity is, of course, fixed. Combining this with the requirement that the pressures on the opposite sides of the piston are mutually equal, , and that the total volume stays constant, we obtain that . Simplifying this yields , where is the total particle number, a fixed quantity, of course. Consequently, we obtain that the pressure in the l.h.s. section will stay constant as the piston is drifting toward its eventual, equilibrium position. And so will the pressure one the other side, since .

Final remark: We assumed above that the piston is truly rigid to drive home the point that if one neglected thermal fluctuations, one would miss a very important aspect of equilibration in the form of heat exchange. In reality, a piston would have to be made out of molecules, too, and so it would conduct heat much more efficiently than the idealized piston we had in mind. Depending on how well the piston conducts heat, the temperatures in the respective compartments may not be actually uniform. These minor complications are not show stoppers and would distract us from the meat of the argument. Also, we did not mention that the piston’s coordinate and velocity would be fluctuating not just in one, but in all spatial directions, again because of the equipartition.