- Entropy of mixing
The entropy of mixing is the change in the
configuration entropy , an extensive thermodynamic quantity, when two differentchemical substance s orcomponent s are mixed. This entropy change must be positive since there is more uncertainty about the spatial locations of the different kinds ofmolecule s when they are interspersed. We assume that the mixing process has reachedthermodynamic equilibrium so that the mixture is uniform and homogeneous. If the substances being mixed are initially at different temperatures and pressures, there will, of course, be an additional entropy increase in the mixed substance due to these differences being equilibrated, but if the substances being mixed are initially at the same temperature and pressure, the entropy increase will be entirely due to the entropy of mixing.The entropy of mixing may be calculated by Gibbs' Theorem which states that when two different substances mix, the entropy increase upon mixing is equal to the entropy increase that would occur if the two substances were to expand alone into the mixing volume. (In this sense, then the term "entropy of mixing" is a misnomer, since the entropy increase is not due to any "mixing" effect.) Nevertheless, the two substances must be different for the entropy of mixing to exist. This is the
Gibbs paradox which states that if the two substances are identical, there will be no entropy change, yet the slightest detectable difference between the two will yield a considerable entropy change, and this is just the entropy of mixing. In other words, the entropy of mixing is not a continuous function of the degree of difference between the two substances.The entropy of mixing is given by:
:
where is the
gas constant , the total number of moles and themole fraction of each of the mixed components.Note that the increase in entropy involves no heat flow (just the irreversible process of mixing), so the equation
(see
Entropy ) does not apply [Statistical and Thermal Physics, M. D. Sturge, A.K. Peters, 2003, p181.] . Since there is no change ininternal energy (), volume (), orenthalpy (), there is no exchange of heat or work with the surroundings and so the thermal entropy of the system and the surroundings will not change [Physical Chemistry, S. C. Wallwork and D. J. W. Grant, Longman, 1977, p292.] .Proof
Assume that the molecules of two different substances are approximately the same size, and regard space as subdivided into a whose cells are the size of the molecules. (In fact, any lattice would do, including close packing.) This is a
crystal -like conceptual model to identify the molecular centers of mass. If the two phases areliquid s, there is no spatial uncertainty in each one individually.Ref|1 Everywhere we look in component 1, there is a molecule present, and likewise for component 2. After they are intermingled (assuming they are miscible), the liquid is still dense with molecules, but now there is uncertainty about what kind of molecule is in which location. Of course, any idea of identifying molecules in given locations is athought experiment , not something one could do, but the calculation of the uncertainty is well-defined.We can use Boltzmann's equation for the entropy change as applied to the "mixing" process
:
where is Boltzmann’s constant. We then calculate the number of ways of arranging molecules of component 1 and molecules of component 2 on a lattice, where
:
is the total number of molecules, and therefore the number of lattice sites.Calculating the number of permutations of objects, correcting for the fact that of them are "identical" to one another, and likewise for ,
:
After applying
Stirling's approximation , the result is:
This expression can be generalized to a mixture of components, , with
:
where we have introduced the
mole fraction s, which are also the probabilities of finding any particular component in a given lattice site.::
A more direct and logically transparent derivation, not requiring Stirling's approximation, is to start with the Shannon entropy or compositional uncertaintyRef|2
:
The summation is over the various chemical species, so this is the uncertainty about which kind of molecule is in any one site. It must be multiplied by the number of sites to get the uncertainty for the whole system. The entropy of mixing from above can be rearranged as
:
The equivalence of the two follows immediately.
Reverting to two components, we obtain
:
where is the
gas constant , equal to timesAvogadro's number , and are the numbers of moles of the components, and is the total number of moles.Since the mole fractions are necessarily less than one, the values of thelogarithm s are negative. The minus sign reverses this, giving a positive entropy of mixing, as expected.Gibbs free energy of mixing
In an
ideal gas orideal solution (no enthalpy term) the Gibbs free energy change of mixing is given by: :where is the
Gibbs free energy and theabsolute temperature [Graph as function of temperature, number of moles and composition: [http://www.whfreeman.com/elements/content/livinggraphs/E3012.html Link] .]The Gibbs energy is always negative meaning that mixing as ideal solutions is always spontaneous. The lowest value is when the mole fraction is 0.5 for a mixture of two components or 1/n for a mixture of n components.
olutions
If the
solute is acrystal linesolid , the argument is much the same. A crystal has no spatial uncertainty at all, except forcrystallographic defect s, and a (perfect) crystal allows us to localize the molecules using the crystalsymmetry group . The fact that volumes do not add when dissolving a solid in a liquid is not important for condensed phases. If the solute is not crystalline, we can still use a spatial lattice, as good an approximation for an amorphous solid as it is for a liquid.The
Flory-Huggins solution theory provides the entropy of mixing forpolymer solutions, in which themacromolecule s are huge compared to the solute molecules. In this case, the assumption is made that eachmonomer subunit in the polymer chain occupies a lattice site.Note that solids in contact with each other also slowly interdiffuse, and solid mixtures of two or more components may be made at will (
alloy s,semiconductor s, etc.). Again, the same equations for the entropy of mixing apply, but only for homogeneous, uniform phases.Gases
In gases there is a lot more spatial uncertainty because most of their volume is merely empty space. We can regard the mixing process as simply conjoining the two containers. The two lattices which allow us to conceptually localize molecular centers of mass also join. The total number of empty cells is the sum of the numbers of empty cells in the two components prior to mixing. Consequently, that part of the spatial uncertainty concerning whether "any" molecule is present in a lattice cell is the sum of the initial values, and does not increase upon mixing.
Almost everywhere we look, we find empty lattice cells. But we do find molecules in those few cells which are occupied. For each one, there is a "contingent" uncertainty about which kind of molecule it is. Using conditional probabilities, it turns out that the analytical problem for the small
subset of occupied cells is exactly the same as for mixed liquids, and the "increase" in the entropy, or spatial uncertainty, has exactly the same form as obtained previously. Obviously the subset of occupied cells is not the same at different times. But only when an occupied cell is found do we ask which kind of molecule is there.See also:
Gibbs paradox , in which it would seem that mixing two samples of the "same" gas would produce entropy.In a paper on this topic [Lost Work and the Entropy of Mixing, R. G. Keesing, European Journal of Physics, 1986, Vol.7, pp266-268] , R. G. Keesing notes:
It is often implied that the mixing of two perfect gasesleads to an increase in entropy. However when twoperfect gases, in thermal equilibrium, mix there is noincrease in entropy solely as a result of their mixing.The entropy increase on mixing through mutualdiffusion is solely the result of the two individual gasesoccupying larger volumes and is the same whether ornot the gases have mixed. The increase in entropycomes about because each gas absorbs heat in theprocess of isothermal expansion which is the samewhether or not the gases have mixed. Thus the ideathat increases in entropy, in non-interacting systems, issomehow connected with the mixing process is false.
Notes
* 1. This is, of course, an approximation. Liquids have a “free volume” which is why they are (usually) less dense than
solid s.
* 2. Claude Shannon introduced this expression for use ininformation theory , but similar formulas can be found as far back as the work ofLudwig Boltzmann and J. Willard Gibbs. Shannon uncertainty is completely unrelated to the Heisenberguncertainty principle inquantum mechanics .External links
* [http://www.phys.uri.edu/~gerhard/PHY525/tln25.pdf Online lecture]
* [http://www.msm.cam.ac.uk/phase-trans/mphil/MP4-3.pdf Online lecture]References
Wikimedia Foundation. 2010.