Overview
This post provides a rigorous examination of Moser Iteration, a foundational bootstrap technique in the regularity theory of Partial Differential Equations. Developed by Jürgen Moser (Moser, 1961) to provide an alternative proof of the De Giorgi-Nash theorem, the method utilizes the synergy between Caccioppoli-type energy estimates and Sobolev embeddings. It allows one to “pump” the integrability of a sub-solution from to , ultimately yielding local boundedness and the celebrated Harnack Inequality.
🏷️ The Setting: Rough Coefficients and Divergence Form
The power of the Moser technique lies in its ability to handle second-order elliptic operators with measurable coefficients. We consider:
where are only assumed to be bounded and uniformly elliptic:
The Divergence Mandate
The divergence structure is non-negotiable for Moser iteration. It enables the use of integration by parts to convert second-order information into first-order gradient control. For non-divergence form equations (), energy methods fail, and one must instead employ the Krylov-Safonov theory based on the Alexandroff-Bakelman-Pucci (ABP) estimate.
🏷️ The Analytical Engine: The Energy Bootstrap
The iteration proceeds by transforming a local bound into a local bound for some . This is achieved by testing the equation with a power of the solution, , and a cutoff function .
The Caccioppoli Inequality
For a non-negative sub-solution , testing with yields the energy control:
This inequality shows that the “energy” of a power of is controlled by the norm of that same power on a slightly larger domain.
🏷️ Main Theorem: Local Boundedness
Local Sup-Norm Estimate
Let be a non-negative sub-solution of in . For any ball and any , there exists a constant such that:
Proof: The Iterative Limit ( )
To prove the bound, we construct a sequence of nested balls with and integrability exponents (where is the Sobolev gain).
1. The Single Step: Applying the Sobolev inequality to the Caccioppoli inequality on the -th ball:
The factor arises from the gradient of the cutoff function , which must scale as .
2. The Iterative Product: Iterating this relation times:
3. Convergence of the Constant: Taking the logarithm of the product transforms it into a series:
Since grows geometrically, this series converges. This implies the product of constants remains finite as .
4. Conclusion: In the limit, the norm on the shrinking balls converges to the norm on , yielding the desired bound.
🏷️ The Critical Case: Dimension
In two dimensions, the Sobolev gain is formally infinite, which would seemingly break the discrete steps of the iteration.
Proof: The BMO and Exponential Route
For , we exploit the fact that for all .
1. Variable Gain: We can choose a fixed gain, say . The Sobolev constant in grows as . Plugging this into the log-series:
The convergence is preserved, ensuring the bound holds.
2. The BMO Insight: A more modern approach recognizes that the energy estimate implies . In 2D, this places in the space of Bounded Mean Oscillation (BMO). The John-Nirenberg Theorem then guarantees that is locally integrable to any power, providing the necessary “launchpad” for the iteration.
🏷️ Scope and Constraints
The Harnack Inequality
By applying the iteration to both and , Moser derived the Harnack Inequality: . This implies that positive solutions cannot oscillate wildly and are, in fact, Hölder continuous.
Constraints and Failure Cases
The "Non-Divergence" Barrier
The Moser iteration is fundamentally tethered to the divergence structure. This is due to the nature of the energy space:
- Lack of Integration by Parts: In non-divergence form equations, , there is no natural way to transfer a derivative to a test function. Thus, we cannot obtain the Caccioppoli inequality that powers the bootstrap.
- Low Regularity of Coefficients: If the coefficients are only , a solution to a non-divergence form equation may not even possess a gradient in . This “low integrability” of the second derivatives prevents the use of Sobolev embeddings.
- The Solution (Krylov-Safonov): For rough coefficients in non-divergence form, one must switch from the analytical bootstrap (Moser) to the geometric maximum principle (Krylov-Safonov (Krylov & Safonov, 1980)). This method uses the ABP (Alexandroff-Bakelman-Pucci) Estimate to show that if a solution is large at a point, it must be large on a set of “positive measure” near the boundary of its convex envelope. See the definitive account in Caffarelli & Cabré [@caffarelli1995fully].
- Dimension : The Sobolev gain becomes formally infinite. As detailed in the proof above, this critical case is resolved via BMO theory and the John-Nirenberg inequality.
- Smoothness vs. Structure: If the coefficients are (Lipschitz), any non-divergence operator can be rewritten in divergence form plus a lower-order drift. In that case, Moser iteration is again applicable. The limitation is specifically for rough coefficients in non-divergence form.
📝 Notes
- The “Nested Ball” method is a strategy to handle the trade-off between domain size and integrability gain.
- This technique shares the “growth tracking” logic found in on Grönwall’s inequality, but operates on the scale of exponents rather than time.
🔗 See Also
- on convergence of graph Laplacian to manifold’s Laplacian --- Moser estimates ensure the equicontinuity of discrete eigenfunctions in the limit.
- on Weyl’s asymptotic law --- Eigenvalue growth is governed by the same Sobolev embedding constants that power the Moser bootstrap.