Entropy Functional Analysis Package #
This file provides the functional analytic framework for entropy functionals needed in the meta-variational principle.
Main API (Section 1-1 of plan2) #
InformationTheory.klDiv
: The relative entropy D(ρ‖μ) with values in ℝ≥0∞probability_klDiv_self
: For probability measure μ, D(μ‖μ) = 0
Lower semicontinuity and chain rules (Section 1-2) #
relativeEntropy_lsc
: Liminf-type LSC under RN density convergencellr_add_of_rnDeriv_mul
: Additivity of log-likelihood ratiosrelativeEntropy_chain_rule_prob_toReal
: Chain rule for KL divergence
Integrability conditions (Section 1-3) #
integrable_llr_of_integrable_klFun_rnDeriv
: Transfer lemma for integrability
Implementation notes #
We use ℝ≥0∞ as the codomain for entropy functionals, converting to ℝ via toReal only when interfacing with PLFA/EVI frameworks.
For a probability measure μ, the relative entropy D(μ‖μ) equals zero. This is the key lemma from task 1-1 that highlights the self-entropy property.
Liminf-type lower semicontinuity of relative entropy under a.e. convergence of RN derivatives.
If (ρₙ)
and ρ
are all absolutely continuous w.r.t. μ
, and the RN derivatives
(ρₙ.rnDeriv μ).toReal → (ρ.rnDeriv μ).toReal
converge μ
-a.e., then
klDiv ρ μ ≤ liminf_n klDiv (ρₙ) μ
holds by Fatou's lemma.
This is a key technical result for establishing LSC in the weak topology.
Relative entropy is non-negative
KL divergence equals zero iff measures are equal (for probability measures)
Core lemma: Additivity of log-likelihood ratios #
Under a multiplicative RN-derivative hypothesis, log-likelihood ratios add a.e. This is the key step towards the chain rule formula for KL divergences. It isolates the purely pointwise identity on log-likelihood ratios.
Chain rule for relative entropy (probability measure version with toReal).
For probability measures with μ ≪ ν ≪ ρ and appropriate integrability conditions,
we have the chain rule:
(klDiv μ ρ).toReal = (klDiv μ ν).toReal + ∫ llr ν ρ dμ
This is the concrete form needed for PLFA/EVI frameworks where we work with real values.
The integrability assumptions h_int1
and h_int2
are explicit as required by plan2.
Section 1-3: Integrability conditions for chain rule #
These lemmas provide sufficient conditions to ensure the integrability assumptions
required by the chain rule relativeEntropy_chain_rule_prob_toReal
.
Transfer lemma: integrability of klFun composed with RN derivative implies integrability of llr. This is the main tool for verifying integrability conditions in the chain rule.
When the RN derivative is bounded, llr is integrable for finite measures. This provides a simple sufficient condition for integrability.
For probability measures with finite KL divergence, llr is integrable.
When RN derivatives are uniformly bounded above and below, llr is integrable.
Data processing inequality: KL divergence decreases under stochastic maps
Entropy has compact sublevel sets (abstract statement)
Structure for entropy functional with functional analytic properties
- Ent : MeasureTheory.Measure X → ℝ
The entropy value for a probability measure
- sublevel_nonempty (c : ℝ) (ρₙ : ℕ → MeasureTheory.Measure X) : (∀ (n : ℕ), self.Ent (ρₙ n) ≤ c) → ∃ (ρ : MeasureTheory.Measure X), self.Ent ρ ≤ c
Non-emptiness of sublevel sets (abstract placeholder for LSC)
- bounded_below : ∃ (c : ℝ), ∀ (ρ : MeasureTheory.Measure X), c ≤ self.Ent ρ
Entropy is bounded below
- compact_sublevels (c : ℝ) (ρₙ : ℕ → MeasureTheory.Measure X) : (∀ (n : ℕ), MeasureTheory.IsProbabilityMeasure (ρₙ n)) → (∀ (n : ℕ), self.Ent (ρₙ n) ≤ c) → ∃ (ρ : MeasureTheory.Measure X) (φ : ℕ → ℕ), StrictMono φ ∧ MeasureTheory.IsProbabilityMeasure ρ ∧ self.Ent ρ ≤ c ∧ ∃ (weakly_converges : Prop), weakly_converges
Entropy has compact sublevel sets
Instances For
Concrete entropy functional
Equations
- Frourio.ConcreteEntropyFunctional μ = { Ent := fun (ρ : MeasureTheory.Measure X) => (InformationTheory.klDiv ρ μ).toReal, sublevel_nonempty := ⋯, bounded_below := ⋯, compact_sublevels := ⋯ }
Instances For
Displacement convexity of entropy along Wasserstein geodesics
Gradient flow structure for entropy functional
- flow : ℝ → MeasureTheory.ProbabilityMeasure X → MeasureTheory.ProbabilityMeasure X
The flow map: time → initial condition → solution
Initial condition
- energy_dissipation (t s : ℝ) : 0 ≤ t → t ≤ s → ∀ (ρ₀ : MeasureTheory.ProbabilityMeasure X), InformationTheory.klDiv (↑(self.flow s ρ₀)) μ ≤ InformationTheory.klDiv (↑(self.flow t ρ₀)) μ
Energy dissipation (entropy decreases along flow)
- time_continuous (_ρ₀ : MeasureTheory.ProbabilityMeasure X) (t s : ℝ) : 0 ≤ t → 0 ≤ s → ∀ ε > 0, ∃ δ > 0, |t - s| < δ → True
Continuity in time (abstract property)
Instances For
Connection to PLFA framework: entropy provides F functional
Equations
- Frourio.entropyToPLFA μ ρ = (InformationTheory.klDiv (↑ρ) μ).toReal
Instances For
Entropy functional satisfies convexity along geodesics (for PLFA)
Entropy satisfies the Energy-Dissipation Equality (EDE)
Entropy satisfies Evolution Variational Inequality (EVI)
JKO scheme for entropy: discrete gradient flow
Equations
- Frourio.entropyJKO μ τ _hτ ρ = ρ
Instances For
JKO iterates converge to gradient flow as τ → 0
Instance: Entropy functional for PLFA/EVI framework integration
Equations
- ⋯ = True.intro