| Comments: | (or for this version) |
Abstract: We derive a generalization of the Second Law of Thermodynamics that uses Bayesian updates to explicitly incorporate the effects of a measurement of a system at some point in its evolution. By allowing an experimenter's knowledge to be updated by the measurement process, this formulation resolves a tension between the fact that the entropy of a statistical system can sometimes fluctuate downward and the information-theoretic idea that knowledge of a stochastically-evolving system degrades over time. The Bayesian Second Law can be written as $\Delta H(\rho_m, \rho) + \langle \mathcal{Q}\rangle_{F|m}\geq 0$, where $\Delta H(\rho_m, \rho)$ is the change in the cross entropy between the original phase-space probability distribution $\rho$ and the measurement-updated distribution $\rho_m$, and $\langle \mathcal{Q}\rangle_{F|m}$ is the expectation value of a generalized heat flow out of the system. We also derive refined versions of the Second Law that bound the entropy increase from below by a non-negative number, as well as Bayesian versions of the Jarzynski equality. We demonstrate the formalism using simple analytical and numerical examples.

Tutorial on Second law of Thermodynamics

第29講 Entropy and the Second Law of Thermodynamics I A
|
Entropy - God's Dice Game: The book describes the historical evolution of the understanding of entropy, alongside biographies of the scientists who ... communication theory, economy, and sociology Book (CreateSpace Independent Publishing Platform)
|

John Scales Avery (born in 1933 in Lebanon to American parents) is a theoretical chemist noted for his research publications in quantum chemistry, thermodynamics, evolution, and history of science. Since the early 1990s, Avery has been an active World peace...