Content deleted Content added
ns |
|||
Line 38:
Thermodynamic entropy bears a close relationship to the concept of [[information entropy]] (''H''). Information entropy is a measure of the "spread" of a probability density or probability mass function. Thermodynamics makes no assumptions about the atomistic nature of matter, but when matter is viewed in this way, as a collection of particles constantly moving and exchanging energy with each other, and which may be described in a probabilistic manner, information theory may be successfully applied to explain the results of thermodynamics. The resulting theory is known as [[statistical mechanics]].
An important concept in statistical mechanics is the idea of the [[macrostate|microstate and the macrostate]] of a system. If we have a container of gas, for example, and we know the position and velocity of every molecule in that system, then we know the microstate of that system. If we only know the thermodynamic description of that system, the pressure, volume, temperature, and/or the entropy, then we know the macrostate of that system.
:<math>S=k_\text{B}\,\ln W</math>
|