A measure of the number of microscopic configurations consistent with a macroscopic state — quantifying disorder and missing information theory.

second law of thermodynamics: entropy of an isolated system never decreases

defines the arrow of time: the direction in which entropy increases

Boltzmann entropy: S = k_B ln W, where W is the number of accessible microstates

Shannon entropy: the information-theoretic analog — measures uncertainty in bits, bridging physics and information theory

maximum entropy at thermal equilibrium — see temperature and thermodynamics

black hole entropy (Bekenstein-Hawking): proportional to horizon area, linking gravity, quantum mechanics, and information theory

free energy = energy minus temperature times entropy — governs spontaneous processes

low-entropy initial conditions of the universe are a central puzzle in cosmology

Local Graph