3.2 Binary KL and Pinsker’s inequality
Definition
3.4
Binary KL
Lemma
3.5
Derivative in \(q\)
Theorem
3.6
Binary Pinsker, sharp constant \(2\)
Theorem
3.7
Pinsker’s inequality
Lean: InformationTheory.pinsker_proof
For probability measures \(P \ll Q\) with \(\mathrm{KL}(P,Q) {\lt} \infty \),
\[ \mathrm{TV}(P, Q) \; \le \; \sqrt{\mathrm{KL}_\mathbb {R}(P, Q) / 2} \]
[ 7 , Thm 11.6.1 ] . The convention matches Definition 2.22 (half-sup TV); with the \(L^1\) convention the bound reads \(\| P - Q \| _1 \le \sqrt{2\, \mathrm{KL}}\).
Theorem
3.8
Binary-to-general coarsening
Lean: InformationTheory.klBin_le_klDivReal
For every measurable \(A\) with \(P(A) = p\), \(Q(A) = q\), \(\mathrm{KL}_\mathrm {bin}(p, q) \le \mathrm{KL}_\mathbb {R}(P, Q)\). This is the two-point data-processing inequality and is the step linking the binary Pinsker to the general measure-theoretic Pinsker [ 7 , Thm 2.8.1 ] .