# .

In mathematics, Gronwall's lemma or Grönwall's lemma, also called Gronwall–Bellman inequality, allows one to bound a function that is known to satisfy a certain differential or integral inequality by the solution of the corresponding differential or integral equation. There are two forms of the lemma, a differential form and an integral form. For the latter there are several variants.

Grönwall's lemma is an important tool to obtain various estimates in the theory of ordinary and stochastic differential equations. In particular, it is used to prove uniqueness of a solution to the initial value problem; see the Picard–Lindelöf theorem.

It is named for Thomas Hakon Grönwall (1877–1932). Grönwall is the Swedish spelling of his name, but he spelled his name as Gronwall in his scientific publications after emigrating to the United States.

The differential form was proven by Grönwall in 1919. The integral form was proven by Richard Bellman in 1943.

A nonlinear generalization of the Gronwall–Bellman inequality is known as Bihari's inequality.

Differential form

Let I denote an interval of the real line of the form [a, ∞) or [a, b] or [a, b) with a < b. Let β and u be real-valued continuous functions defined on I. If u is differentiable in the interior I o of I  (the interval I without the end points a and possibly b) and satisfies the differential inequality

$$u'(t) \le \beta(t)\,u(t),\qquad t\in I^\circ,$$

then u is bounded by the solution of the corresponding differential equation y ′(t) = β(t) y(t):

$$u(t) \le u(a) \exp\biggl(\int_a^t \beta(s)\, \mathrm{d} s\biggr)$$

for all t in I.

Remark: There are no assumptions on the signs of the functions β and u.
Proof

Define the function

$$v(t) = \exp\biggl(\int_a^t \beta(s)\, \mathrm{d} s\biggr),\qquad t\in I.$$

Note that v satisfies

$$v'(t) = \beta(t)\,v(t),\qquad t\in I^\circ,$$

with v(a) = 1 and v(t) > 0 for all t in I. By the quotient rule

$$\frac{d}{dt}\frac{u}{v} = \frac{u'v-v'u}{v^2} \le \frac{\beta u v - \beta v u}{v^2} = 0,\qquad t\in I^\circ,$$

so an application of the mean value theorem implies

$$\frac{u(t)}{v(t)}\le \frac{u(a)}{v(a)}=u(a),\qquad t\in I,$$

which is Gronwall's inequality.
Integral form for continuous functions

Let I denote an interval of the real line of the form [a,∞) or [a,b] or [a,b) with a < b. Let α, β and u be real-valued functions defined on I. Assume that β and u are continuous and that the negative part of α is integrable on every closed and bounded subinterval of I.

(a) If β is non-negative and if u satisfies the integral inequality

$$u(t) \le \alpha(t) + \int_a^t \beta(s) u(s)\,\mathrm{d}s,\qquad \forall t\in I,$$

then

$$u(t) \le \alpha(t) + \int_a^t\alpha(s)\beta(s)\exp\biggl(\int_s^t\beta(r)\,\mathrm{d}r\biggr)\mathrm{d}s,\qquad t\in I.$$

(b) If, in addition, the function α is non-decreasing, then

$$u(t) \le \alpha(t)\exp\biggl(\int_a^t\beta(s)\,\mathrm{d}s\biggr),\qquad t\in I.$$

Remarks:

There are no assumptions on the signs of the functions α and u.
Compared to the differential form, differentiability of u is not needed for the integral form.
For a version of Grönwall's inequality which doesn't need continuity of β and u, see the version in the next section.

Proof

(a) Define

$$v(s) = \exp\biggl({-}\int_a^s\beta(r)\,\mathrm{d}r\biggr)\int_a^s\beta(r)u(r)\,\mathrm{d}r,\qquad s\in I.$$

Using the product rule, the chain rule, the derivative of the exponential function and the fundamental theorem of calculus, we obtain for the derivative

$$v'(s) = \biggl(\underbrace{u(s)-\int_a^s\beta(r)u(r)\,\mathrm{d}r}_{\le\,\alpha(s)}\biggr)\beta(s)\exp\biggl({-}\int_a^s\beta(r)\mathrm{d}r\biggr), \qquad s\in I,$$

where we used the assumed integral inequality for the upper estimate. Since β and the exponential are non-negative, this gives an upper estimate for the derivative of v. Since v(a) = 0, integration of this inequality from a to t gives

$$v(t) \le\int_a^t\alpha(s)\beta(s)\exp\biggl({-}\int_a^s\beta(r)\,\mathrm{d}r\biggr)\mathrm{d}s.$$

Using the definition of v(t) for the first step, and then this inequality and the functional equation of the exponential function, we obtain

\begin{align}\int_a^t\beta(s)u(s)\,\mathrm{d}s &=\exp\biggl(\int_a^t\beta(r)\,\mathrm{d}r\biggr)v(t)\\ &\le\int_a^t\alpha(s)\beta(s)\exp\biggl(\underbrace{\int_a^t\beta(r)\,\mathrm{d}r-\int_a^s\beta(r)\,\mathrm{d}r}_{=\,\int_s^t\beta(r)\,\mathrm{d}r}\biggr)\mathrm{d}s. \end{align}

Substituting this result into the assumed integral inequality gives Grönwall's inequality.

(b) If the function α is non-decreasing, then part (a), the fact α(s) ≤ α(t), and the fundamental theorem of calculus imply that

\begin{align}u(t)&\le\alpha(t)+\biggl({-}\alpha(t)\exp\biggl(\int_s^t\beta(r)\,\mathrm{d}r\biggr)\biggr)\biggr|^{s=t}_{s=a}\\ &=\alpha(t)\exp\biggl(\int_a^t\beta(r)\,\mathrm{d}r\biggr),\qquad t\in I.\end{align}

Integral form with locally finite measures

Let I denote an interval of the real line of the form [a,∞) or [a,b] or [a,b) with a < b. Let α and u be measurable functions defined on I and let μ be a locally finite measure on the Borel σ-algebra of I (we need μ([a,t]) < ∞ for all t in I). Assume that u is integrable with respect to μ in the sense that

$$\int_a^t|u(s)|\,\mu(\mathrm{d}s)<\infty,\qquad t\in I,$$

and that u satisfies the integral inequality

$$u(t) \le \alpha(t) + \int_{[a,t)} u(s)\,\mu(\mathrm{d}s),\qquad t\in I.$$

the function α is non-negative or
the function t → μ([a,t]) is continuous for t in I and the function α is integrable with respect to μ in the sense that

$$\int_a^t|\alpha(s)|\,\mu(\mathrm{d}s)<\infty,\qquad t\in I,$$

then u satisfies Grönwall's inequality

$$u(t) \le \alpha(t) + \int_{[a,t)}\alpha(s)\exp\bigl(\mu(I_{s,t})\bigr)\,\mu(\mathrm{d}s)$$

for all t in I, where Is,t denotes to open interval (s, t).
Remarks

There are no continuity assumptions on the functions α and u.
The integral in Grönwall's inequality is allowed to give the value infinity.
If α is the zero function and u is non-negative, then Grönwall's inequality implies that u is the zero function.
The integrability of u with respect to μ is essential for the result. For a counterexample, let μ denote Lebesgue measure on the unit interval [0,1], define u(0) = 0 and u(t) = 1/t for t in (0,1], and let α be the zero function.
The version given in the textbook by S. Ethier and T. Kurtz. makes the stronger assumptions that α is a non-negative constant and u is bounded on bounded intervals, but doesn't assume that the measure μ is locally finite. Compared to the one given below, their proof does not discuss the behaviour of the remainder Rn(t).

Special cases

If the measure μ has a density β with respect to Lebesgue measure, then Grönwall's inequality can be rewritten as

$$u(t) \le \alpha(t) + \int_a^t \alpha(s)\beta(s)\exp\biggl(\int_s^t\beta(r)\,\mathrm{d}r\biggr)\,\mathrm{d}s,\qquad t\in I.$$

If the function α is non-negative and the density β of μ is bounded by a constant c, then

$$u(t) \le \alpha(t) + c\int_a^t \alpha(s)\exp\bigl(c(t-s)\bigr)\,\mathrm{d}s,\qquad t\in I.$$

If, in addition, the non-negative function α is non-decreasing, then

$$u(t) \le \alpha(t) + c\alpha(t)\int_a^t \exp\bigl(c(t-s)\bigr)\,\mathrm{d}s =\alpha(t)\exp(c(t-a)),\qquad t\in I.$$

Outline of proof

The proof is divided into three steps. In idea is to substitute the assumed integral inequality into itself n times. This is done in Claim 1 using mathematical induction. In Claim 2 we rewrite the measure of a simplex in a convenient form, using the permutation invariance of product measures. In the third step we pass to the limit n to infinity to derive the desired variant of Grönwall's inequality.
Detailed proof
Claim 1: Iterating the inequality

For every natural number n including zero,

$$u(t) \le \alpha(t) + \int_{[a,t)} \alpha(s) \sum_{k=0}^{n-1} \mu^{\otimes k}(A_k(s,t))\,\mu(\mathrm{d}s) + R_n(t)$$

with remainder

$$R_n(t) :=\int_{[a,t)}u(s)\mu^{\otimes n}(A_n(s,t))\,\mu(\mathrm{d}s),\qquad t\in I,$$

where

$$A_n(s,t)=\{(s_1,\ldots,s_n)\in I_{s,t}^n\mid s_1<s_2<\cdots<s_n\},\qquad n\ge1,$$

is an n-dimensional simplex and

$$\mu^{\otimes 0}(A_0(s,t)):=1.$$

Proof of Claim 1

We use mathematical induction. For n = 0 this is just the assumed integral inequality, because the empty sum is defined as zero.

Induction step from n to n + 1: Inserting the assumed integral inequality for the function u into the remainder gives

$$R_n(t)\le\int_{[a,t)} \alpha(s) \mu^{\otimes n}(A_n(s,t))\,\mu(\mathrm{d}s) +\tilde R_n(t)$$

with

$$\tilde R_n(t):=\int_{[a,t)} \biggl(\int_{[a,q)} u(s)\,\mu(\mathrm{d}s)\biggr)\mu^{\otimes n}(A_n(q,t))\,\mu(\mathrm{d}q),\qquad t\in I.$$

Using the Fubini-Tonelli theorem to interchange the two integrals, we obtain

$$\tilde R_n(t) =\int_{[a,t)} u(s)\underbrace{\int_{(s,t)} \mu^{\otimes n}(A_n(q,t))\,\mu(\mathrm{d}q)}_{=\,\mu^{\otimes n+1}(A_{n+1}(s,t))}\,\mu(\mathrm{d}s) =R_{n+1}(t),\qquad t\in I.$$

Hence Claim 1 is proved for n + 1.
Claim 2: Measure of the simplex

For every natural number n including zero and all s < t in I

$$\mu^{\otimes n}(A_n(s,t))\le\frac{\bigl(\mu(I_{s,t})\bigr)^n}{n!}$$

with equality in case t → μ([a,t]) is continuous for t in I.
Proof of Claim 2

For n = 0, the claim is true by our definitions. Therefore, consider n ≥ 1 in the following.

Let Sn denote the set of all permutations of the indices in {1,2,...,n}. For every permutation σ in Sn define

$$A_{n,\sigma}(s,t)=\{(s_1,\ldots,s_n)\in I_{s,t}^n\mid s_{\sigma(1)}<s_{\sigma(2)}<\cdots<s_{\sigma(n)}\}.$$

These sets are disjoint for different permutations and

$$\bigcup_{\sigma\in S_n}A_{n,\sigma}(s,t)\subset I_{s,t}^n.$$

Therefore,

$$\sum_{\sigma\in S_n} \mu^{\otimes n}(A_{n,\sigma}(s,t)) \le\mu^{\otimes n}\bigl(I_{s,t}^n\bigr)=\bigl(\mu(I_{s,t})\bigr)^n.$$

Since they all have the same measure with respect to the n-fold product of μ, and since there are n! permutations in Sn, the claimed inequality follows.

Assume now that t → μ([a,t]) is continuous for t in I. Then, for different indices i and j in {1,2,...,n}, the set

$$\{(s_1,\ldots,s_n)\in I_{s,t}^n\mid s_i=s_j\}$$

is contained in a hyperplane, hence by an application of Fubini's theorem its measure with respect to the n-fold product of μ is zero. Since

$$I_{s,t}^n\subset\bigcup_{\sigma\in S_n}A_{n,\sigma}(s,t) \cup \bigcup_{1\le i<j\le n}\{(s_1,\ldots,s_n)\in I_{s,t}^n\mid s_i=s_j\},$$

the claimed equality follows.
Proof of Grönwall's inequality

For every natural number n, Claim 2 implies for the remainder of Claim 1 that

$$|R_n(t)| \le \frac{\bigl(\mu(I_{a,t})\bigr)^n}{n!} \int_{[a,t)} |u(s)|\,\mu(\mathrm{d}s),\qquad t\in I.$$

Since μ is locally finite on I, we have μ(Ia,t) < ∞. Hence, the integrability assumption on u implies that

$$\lim_{n\to\infty}R_n(t)=0,\qquad t\in I.$$

Claim 2 and the series representation of the exponential function imply the estimate

$$\sum_{k=0}^{n-1} \mu^{\otimes k}(A_k(s,t)) \le\sum_{k=0}^{n-1} \frac{\bigl(\mu(I_{s,t})\bigr)^k}{k!} \le\exp\bigl(\mu(I_{s,t})\bigr)$$

for all s < t in I. If the function α is non-negative, then it suffices to insert these results into Claim 1 to derive the above variant of Grönwall's inequality for the function u.

In case t → μ([a,t]) is continuous for t in I, Claim 2 gives

$$\sum_{k=0}^{n-1} \mu^{\otimes k}(A_k(s,t)) =\sum_{k=0}^{n-1} \frac{\bigl(\mu(I_{s,t})\bigr)^k}{k!} \to\exp\bigl(\mu(I_{s,t})\bigr)\qquad\text{as }n\to\infty$$

and the integrability of the function α permits to use the dominated convergence theorem to derive Grönwall's inequality.
References

^ Gronwall, Thomas H. (1919), "Note on the derivative with respect to a parameter of the solutions of a system of differential equations", Ann. of Math. 20 (4): 292–296, JSTOR 1967124, MR1502565
^ Bellman, Richard (1943), "The stability of solutions of linear differential equations", Duke Math. J. 10 (4): 643–647, MR0009408
^ Ethier, Steward N.; Kurtz, Thomas G. (1986), Markov Processes, Characterization and Convergence, New York: John Wiley & Sons, p. 498, ISBN 0-471-08186-8, MR0838085