Variational principle for general diffusion problems

0 downloads 0 Views 238KB Size Report
Jun 28, 2004 - sign preservation) are crucial to the variational approach by techniques based on ..... Remark: The property (f1) is sufficient for giving the right growth of φ in order to infer existence of the ...... stein's Variational Principle, Math.
Variational principle for general diffusion problems June 28, 2004 Luca Petrelli & Adrian Tudorascu Department of Mathematical Sciences Carnegie Mellon University, Pittsburgh, PA 15213 [email protected], [email protected] Abstract We employ the Monge-Kantorovich mass transfer theory to study existence of solutions for a large class of parabolic partial differential equations. We deal with non-homogeneous nonlinear diffusion problems (of Fokker-Planck type) with time dependent coefficients. This work greatly extends the applicability of known techniques based on constructing weak solutions by approximation with time-interpolants of minimizers arising from Wasserstein-type implicit schemes ([1], [11], [13], [14], [15]). It also generalizes the results from [18] and [20], where proofs of convergence in the case of a right hand side in the equation is given by these methods. To prove existence of weak solutions we establish an interesting maximum principle for such equations. This involves comparison with the solution for the corresponding homogeneous, time-independent equation. Keywords: Anomalous diffusion, diffusion equations, optimal mass transportation, Wasserstein distance, discretized gradient flow, implicit schemes, nonhomogeneous, non-autonomous problem, weak solution. AMS subject classification: 35G25, 46N10, 49J99.

1

Introduction

In the present work we study general diffusion problems with drift and diffusion coefficients which may be explicitly time dependent. The boundary conditions considered here are of Neumann type. In the absence of a forcing term, this causes the solution to preserve its total mass at all times. Conservation of total mass is clearly not achieved in the presence of a non-homogeneous term. Under certain conditions the sign of the initial data is preserved by the solution. These features (i.e. conservation of mass and sign preservation) are crucial to the variational approach by techniques based on time-discretized implicit schemes in the Wasserstein distance. More precisely, we consider the following problem (where all gradients are spatial): ⎧ t)) − ∆f (t, u) = g(x, t, u) in Ω × (0, T ), ⎨ ut − ∇ · (u∇Ψ(x,  u∇Ψ + ∇f (t, u) · νx = 0 on ∂Ω × (0, T ), (P ) ⎩ in Ω, u(·, 0) = u0 where u0 is the initial datum and f : R2 → R, g : RN +2 → R and Ψ : RN +1 → R are functions with certain properties. This general problem is connected to several issues of current interest in statistical physics, including anomalous diffusion in many systems. Interesting examples are mentioned in [16]: diffusion in plasmas, surface growth and transport of fluid in porous media, relative diffusion in turbulent media. In special cases even exact solutions can be computed ([6], [16], [17]). For example, [16] produces an exact solution for the equation: ut = −∇ · [F(·, t)u] + a(t)∆uν ,

1

where ν ≥ 1 and F(x, t) := k1 (t) + k2 (t)x falls into our admissible category of drift coefficients (indeed, F(·, t) is conservative for all t). The homogeneous case (g ≡ 0) comprises a large class of Fokker-Planck equations which describe the evolution in time of the probability density functions (p.d.f.’s) of stochastic processes that are solutions for certain stochastic differential equations (see [10]). For the linear, generic case see [11]. For the general case, we found [2] of great help as to its stochastic derivation and interpretation. It is shown that one is “[. . . ] led to accept that the stochastic force depends [. . . ] on the probability distribution”. More precisely, u(·, t) is the p.d.f. of a stochastic process {Xt } which is the solution for the Ito stochastic differential equation: (1.1) dXt = −Ψ(Xt , t)dt + H(Xt , t)dWt , where {Wt } is the standard Wiener process and H is the stochastic force. Here  H(x, t) =

2f (t, u(x, t)) u(x, t)

1/2 .

Techniques inspired by the Monge-Kantorovich optimal mass transfer theory are known to be applicable to solving certain parabolic partial differential equations. Jordan, Kinderlehrer and Otto show in [11] that the solution for the heat equation is the gradient flow of a free energy functional with respect to the Wasserstein distance. In [14] Otto gives a similar result for the porous medium equation and more recently Agueh, in [1], proves existence for more general parabolic equations:  ∂u = div u∇c∗ [∇(F  (u) + V )], ∂t where c is some appropriate cost function, c∗ its Legendre transform, F the internal energy and V a convex potential. Agueh uses more general Monge-Kantorovich distances (induced by the cost function c) than the quadratic cost Wasserstein metric; he is, therefore, able to extend previous results, such as Otto’s [14]. Key to all these results is the construction of a time step approximation, which is then linearly interpolated and shown to converge to some notion of weak solution. This sequence in time is recovered from the variational principle:

1 (1.2) d(u, uk−1 )2 + hS(u) , uk := argmin 2 where d(u, uk−1 ) represents the Wasserstein distance between the measures udx and uk−1 dx, having the same total mass, and S(u) represents the free energy of the system. Conservation of mass allows for comparison between different time steps by means of the Wasserstein distance. In [12], Kinderlehrer and Walkington present a numerical algorithm, which uses the main idea of this variational principle, to solve certain parabolic partial differential equations with forcing terms. Mass is no longer conserved, therefore, in order to be able to compare different time steps, they suggest including a time-averaged correction in the discrete scheme:

1 2 g ) + hS(u) , (1.3) d(u, uk−1 + h˜ uk := argmin 2 g and g˜ is the time where now the minimum has to be taken over u with the same mass as uk−1 + h˜ average over the k-th interval (i. e. [kh, (k + 1)h)) of the non-homogeneous data g. In this paper we prove convergence for the cases studied in [12] and we extend the proof to (P ). Let us remark that the presence of g is not the only novelty of this work. In fact, through some nontrivial changes of standard methods we were able to include a direct time dependence for both the drift coefficient Ψ and the nonlinear diffusion term f . While the concept in itself is quite simple, we now work with time averages for the above quantities, the proofs require some nontrivial induction. The authors have not been able to find an existence result in this generality in the literature. Proofs of convergence, using these methods, in less general cases than the ones considered here may be found in [18] and [20]. The main advantage of such a method over more traditional ones is evident in (1.2): no gradient of the approximating functions enters the variational principle. This allows constructing approximations from discontinuous functions, an appealing situation for problems which exhibit steep gradients. Also, with

2

this method we are able to minimize directly in the weak topology, naturally induced by the Wasserstein distance. Furthermore, our method has the remarkable property of constructing good approximations for the solution which presents interesting numerical advantages (e.g. reduction of oscillations, stability). Another interesting consequence of this new approach is that we do not require convexity of the potential. This makes our result available for problems where the potential might have more than one well (e.g. [5]). The paper is organized as follows; in Section 2 we set up the format and we remind the reader of the main properties of the Wasserstein distance. We also present a maximum principle which will be used extensively in the sequel. We find this maximum principle a remarkable result. It is inspired by the pioneering work of Otto ([14]) . In Section 3 we recover the Euler equations of the variational principle and prove necessary inequalities that will be used to show convergence of the interpolation of the time-step approximations to the weak solution. We also prove the main result in the case of a linear diffusion coefficient where all that is needed is weak convergence. In order to prove the main result in the general nonlinear case we need strong convergence, which we recover through a compactness result for uh in L1 (Ω × [0, T ]). This is taken care of in Section 4, where we use the Riesz-Frechet-Kolmogorov criterion to establish the desired compactness (leading to the proof of existence). We conclude in the last section with a partial uniqueness result. Long time behavior of solutions is to be addressed in future work.

2

The Weak Solution

Let u0 ∈ L1 (Ω) be the initial data and denote Q := Ω × (0, T ). Definition 1. A weak solution for (P ) is a function u(x, t) satisfying (i) u ∈ L∞ ((0, T ); L1 (Ω)); (ii) u∇Ψ + ∇f (·, u) ∈ L1 (Q) and g(·, ·, u) ∈ L1 (Q); ¯ × [0, T )), i.e. smooth functions on Ω ¯ × [0, T ] that vanish near t = T , we have (iii) For every ζ ∈ Cc∞ (Ω    u∂t ζ − u∇Ψ + ∇f (·, u) · ∇ζ + g(·, ·, u)ζ dxdt = − u0 ζ(·, 0)dx. (2.1) Ω

Q

By means of the Wasserstein metric Otto proves (in [14]) the existence of weak solutions for (P ) for the case Ω = RN , f (s) = s2 , Ψ ≡ g ≡ 0, and nonnegative, bounded and compactly supported initial data (here T = ∞). The weak solution u is found to satisfy 0 ≤ u ≤ u0 L∞ (Ω) a.e. in Ω × (0, ∞) and conserve u(·, t)L1 (Ω) = u0 L1 (Ω) for all t > 0. For uniqueness we are referred to [21]. Denote by (f 1), (f 2) the following properties of a function f : R × [0, ∞) → R, (u − v)(f (t, u) − f (t, v)) ≥ c|u − v|ω for all u, v ≥ 0,

(f 1)

for some c > 0 independent of t ≥ 0, ω ≥ 2 and f (·, s) is Lipschitz continuous, uniformly for s in bounded sets.

(f 2)

Remark: Note that all power functions f (s) = s , p ≥ 1 satisfy (f 1) (and, trivially, (f 2)). Indeed, notice that, if, say, x ≥ y ≥ 0, then xp −y p ≥ (x−y)p . This leads to (x−y)(xp −y p ) ≥ |x−y|p+1 for all x, y ≥ 0. p

In order to avoid the technical difficulties raised by working with the Wasserstein metric on unbounded domains we assume Ω to be bounded. Also, for convenience, we assume momentarily that 0 < T < ∞.

Let u1 , u2 : Ω → R be two nonnegative and integrable functions that satisfy Ω u1 dx = Ω u2 dx > 0. Denote: 

 ξ(x)dµ(x, y) = ξ(x)u1 (x)dx P (u1 , u2 ) := nonnegative Borel measure µ on Ω × Ω 

Ω×Ω



and

ξ(y)dµ(x, y) = Ω×Ω



¯ ξ(y)u2 (y)dy for all ξ ∈ C(Ω)

Then we have:

3





Definition 2. The pth power of the Wasserstein distance of order p is defined as inf |x − y|p dµ(x, y). dp (u1 , u2 )p := µ∈P (u1 ,u2 )

Ω×Ω

1

Ψ ≡ g ≡ 0 and f ≡ f (u). Formally, a solution of (P ) Assume u0 ≥ 0 is a nonzero element of L (Ω),

is a gradient flow of the functional S(u) := Ω φ(u)dx with respect to the Wasserstein metric of order 2 (simply denoted by d), i.e. a differentiable function (0, ∞) t → u(t) ∈ M such that u(0) = u0 that satisfies mu(t) (∂t u(t), v) + ∇S(u(t)), v = 0 for all v ∈ Tu(t) Mu0 , where Mu0 :=

  u  u is measurable, nonnegative and udx = u0 dx , Ω



Tu(t) Mu0 is the tangent space to Mu0 at u(t) and φ is a solution of sφ (s) − φ(s) = f (s) in [0, ∞)

(2.2)

(see, for example, [15]). It is well-known that any gradient flow on the manifold (M, m) (m is a metric tensor on T M that induces the Wasserstein metric on M; see [15], [14]) admits a natural time-discretized variational scheme of the form (dm being the metric induced by m on M): Fix h > 0 and let ρ0 ≥ 0 be integrable of positive mass. For every integer k ≥ 1 find ρk ∈ Mρ0 that minimizes 1 dm (ρ, ρk−1 )2 + S(ρ) 2h among all ρ ∈ Mρ0 . This motivates our approach of proving that the scheme (2.5) has a unique solution in our case and we will show how to eventually construct a weak solution for (P ) by standard interpolation techniques. In order to do this we need several results. Some of them we are not going to prove since they are simply direct generalizations of similar results proved in [11], [15], [14] and [13]. When we drop the assumption Ψ ≡ 0 and we let f = f (t, u) we need consider time averages for the quantities Ψ and f . We will still be able to define a sequence uk which will be used later to construct the time interpolation. Let: 1 (k+1)h 1 (k+1)h Ψ(·, t) dt and fkh := f (t, ·) dt. (2.3) Ψhk := h kh h kh  The functional S will be replaced by a sequence of functionals Skh k given by  1 (k+1)h Skh (u) := φ(t, u(x)) + u(x)Ψ(x, t) dxdt, h kh Ω where φ(t, ·) is the continuous solution for (2.2) with f = f (t, ·), i. e. φ(t, ·) is a solution of

s

∂φ (t, s) − φ(t, s) = f (t, s) in (0, ∞) × [0, ∞). ∂s

(2.4)

Remark: Note that, whenever f (t, ·) is continuous and has the right growth, then φ(t, ·) is C 1 and it satisfies φ∗ (t, ·) ◦ φ (t, ·) = f (t, ·), where φ∗ is the Legendre transform in space of φ and φ is its spatial derivative. Assume Ψ : RN × (0, ∞) → R is differentiable and locally Lipschitz in x ∈ RN . ∗

1

(Ψ1)

Lemma 1. Let u ∈ L (Ω) be nonnegative and h > 0 be given. If (f 1) holds, then, for each integer k ≥ 1, the functional (k+1)h  1 φ(t, u(x)) + u(x)Ψ(x, t) dxdt (2.5) Ikh [u∗ ](u) := d(u, u∗ )2 + 2 kh Ω admits a unique minimizer in Mu∗ .

4

Notice that, due to (2.3), we have Skh (u) =

 Ω

φhk (u) + uΨhk dx

and

1 d(u, u∗ )2 + hSkh (u). 2 The proof is already quite standard (see, for example, [11]). We just point out that, due to the boundedness of Ω, we can omit all the technical details used there to obtain bounds on the second moments of the measure densities involved. Also as in [11], we can avoid using the strict convexity of the square of the Wasserstein metric (see [14]) by noting that the uniqueness of the minimizer is due to the convexity of d(·, u∗ )2 (easy to prove) and the strict convexity of the Skh (due to (f 1)). The minimizer is obtained as the weak L1 limit of a minimizing sequence by applying the superlinear growth L1 -weak convergence criterion to φhk which, also as a consequence of (f 1), satisfies for all k and all h > 0,   s h fk (σ) φhk (s) 1 h dσ + C ≥ c log s + f (0) 1 − = + C for all s > 1 k s σ2 s 1

(Ψhk has no adverse effect on the superlinear growth of Ikh [u∗ ] since Ω uΨhk dx ≥ −ΨL∞ (Q) uL1 (Ω) ). A trivial iterative application of this lemma yields: Ikh [u∗ ](u) =

Corollary 1. Given u0 ≥ 0 of finite, positive mass, the following iterative scheme: For any integer k ≥ 1, “ uk minimizes Ikh [uk−1 ] in Mu0 ” has a unique solution {uk }k≥0 .

(2.6)

Remark: The property (f 1) is sufficient for giving the right growth of φ in order to infer existence of the minimizer. However, this is not the reason why (f 1) is used throughout the paper (it is needed to prove compactness in section 4). In fact, weaker assumptions on φ (see [1]) than the ones implied by (f 1) are sufficient. These uk ’s turn out to be exactly what we need in the case g ≡ 0 to construct the solution of (P ) by interpolating time. In [12], Kinderlehrer and Walkington suggest modifying (2.6) as follows: Corollary 2. Given u0 ≥ 0 of finite, positive mass, the following iterative scheme: For any integer k ≥ 1, “ uk minimizes Ikh [vk−1 ] in Mvk−1 ” has a unique solution {uk }k≥0 . At each step k ≥ 0 the function vk : Ω → R is defined as follows: (k+1)h vk := uk + g(·, τ, uk (·))dτ.

(2.7)

(2.8)

kh

We make the following assumptions on g: g(x, ·, ·) is nonnegative in [0, ∞) × [0, ∞) for all x ∈ RN ;

(g1)

g(x, t, u) ≤ C(1 + u) for u ≥ 0 locally uniformly with respect to (x, t), t ≥ 0;

(g2)

g(x, t, ·) is continuous on [0, ∞);

(g3)

{g(x, ·, u)}(x,u) is equicontinuous on [0, ∞) with respect to (x, u).

(g4)

The fact that all the vk ’s and the uk ’s from Corollary 2 are nonnegative and have finite, positive mass (nondecreasing in k) follows trivially from (g1)-(g3). The sequel will use the notation uhk instead of uk when we want to emphasize the dependence of each such function on h, dependence that will become increasingly important. For every positive integer n, following [11], [14] etc., we define uh : Ω × [0, T ) → [0, ∞), uh (x, t) := uhk (x) if kh ≤ t < (k + 1)h, for all k ∈ N, k < n where h := T /n.

5

(2.9)

Remark: Now we can explain the intuition behind (2.8). Formally, one can easily get from (P ) the following

t+h u(x, t + h)dx = g(x, τ, u(x, τ ))dτ dx, u(x, t) + Ω



t

for t ∈ (0, T ) and sufficiently small h > 0. Now let t = (k − 1)h and consider the interpolant uh instead of u.

2.1

A discrete comparison principle

 It is obvious that, for a given h > 0, we only need uhk 0≤k≤n−1 to completely define uh . Denote by (f 3) the property f (t, ·) is differentiable and

∂f is positive and monotone in time. ∂s

(f 3)

We are now ready for the following: Lemma 2. If 0 ≤ u0 ≤ M0 < ∞ a.e. in Ω for large enough M0 , then there exists 0 < M = M (M0 ) < ∞ such that 0 ≤ uh ≤ M a.e. in Ω for all h > 0 if one of the following is true (i) f satisfies (f 3), lims↑∞ φs (t, s) = ∞ uniformly with respect to t > 0 and g ≡ 0; (ii) f satisfies (f 3), lims↑∞ φs (t, s) = ∞ uniformly in t > 0 and for s > 0 large enough we have ηs

∂f ∂f (t, ηs + η − 1) − (ηs + η − 1) (t, s) does not change sign for all t > 0, η > 1, ∂s ∂s

being nonnegative if

∂f (·, s) ∂s

(2.10)

is increasing and nonpositive otherwise.

Remark: Note that all functions f (t, s) = a(t)sp for p ≥ 1 and a monotone and bounded from below away from zero, are covered by the above statement even if g is not identically zero. For p ≥ 1 the expression in (2.10) is nonnegative while for p = 1 it is nonpositive. Also, (f 1) implies lims↑∞ φ (s) = ∞. Remark: Note that it follows from some easy calculations that we have properties similar to (f 3), (2.10)(i) and (2.10)(ii) for φk and fk . Also, in the first part of the following proof we drop the k’s to make it more readable. Proof: Observe that in all cases φ is strictly increasing and continuous on (0, ∞) and φ ((0, ∞)) =  −1 . We are going to adjust the proof in [20] to fit (φ (0), ∞). Set, for convenience, θ := φ and ϑ := φ our purpose. First, we claim that: If u∗ ≤ UCφ,Ψ a.e. in Ω, then u ≤ UCφ,Ψ a.e. in Ω,

(2.11)

:= ϑ ◦ (C − Ψ) (C is a constant function) and u is the minimizer from Lemma 1. Take where µ ∈ P (u , u) to be the optimal transfer plan and define v0 and v1 by ¯ v0 ξdx = ξ(x)dµ(x, y) for all ξ ∈ C(Ω), UCφ,Ψ ∗



{u(y)>U (y)}





v1 ξdx =

ξ(y)dµ(x, y) = {u(y)>U (y)}

uξdy {u(y)>U (y)}

¯ Here U := U φ,Ψ . The equality (valid for all ξ ∈ C(Ω ¯ × Ω)) ¯ for all ξ ∈ C(Ω). C ξ(x, y)dµλ (x, y) = ξ(x, y)dµ(x, y)+ Ω×Ω

{u(y)≤U (y)}



+(1 − λ)

ξ(x, y)dµ(x, y) + λ {u(y)>U (y)}

ξ(x, x)dµ(x, y) {u(y)>U (y)}

6

defines for every λ 1 a plan µλ ∈ P (u∗ , uλ ) with uλ := u − λ(v1 − v0 ) ∈ Mu∗ . Then 1 d(u∗ , uλ )2 + {φ(uλ ) + Ψuλ }dx ≤ I[u∗ ](u) + {φ(uλ ) − φ(u) + Ψuλ − Ψu}dx, 2h Ω Ω

(2.12)

due to the definition of d and µλ . Let β(λ) := [φ(uλ ) − φ(u) + Ψ(uλ − u)]dx = [φ(u − λv) − φ(u) − λvΨ]dx, Ω



where v := v1 − v0 ∈ Tu Mu∗ . We have β  (0) = −

v(θ(u) + Ψ)dx. Ω

Since θ is strictly increasing and v1 = uχ{u>U } > U ≥ v0 on {u > U } the splitting (take into account that v1 = 0 on {u ≤ U }) v[θ(u) − θ(U )]dx + v0 [θ(u) − θ(U )]dx − v(Ψ + θ(U ))dx, β  (0) = − {u>U }

{u≤U }





shows that β (0) < 0. Indeed, Ψ + θ(U ) ≡ C = const and v has null average so the last term of the inequality above vanishes. Therefore, β(0) = 0 > β  (0) and so β(λ) is negative for sufficiently small λ > 0. This, however, contradicts the minimality of I[u∗ ](u) (by (2.12)) unless meas{u > U } = 0, i.e. 0 ≤ u ≤ U a.e. in Ω. Thus, the claim (2.11) is proved. Observe that, if u∗ ∈ L∞ (Ω), then we may choose C := θ(u∗ L∞ (Ω) ) + ΨL∞ (Ω) . Indeed, we have   u∗ (x) ≤ u∗ L∞ (Ω) ≤ ϑ θ(u∗ L∞ (Ω) ) + ΨL∞ (Ω) − Ψ(x) , for a.e. x ∈ Ω. Once we reconsider the right notations in the picture we have just proved the following adjusted claim: If u∗ ≤ UMkk

φ ,Ψk

φ ,Ψk

a.e. in Ω, then u ≤ UMkk

a.e. in Ω,

(2.13)

where the constants Mk will be chosen below. Let b a constant, which we choose later, and define the sequence {mk }1≤k≤n by mk := ϑk (θk (ak ) − b) where ak := (1 + Ch)k (M0 + b) + (1 + Ch)k − 1.

(2.14)

Note that, if we choose M0 large enough, we have no trouble defining all of the above. Let also, for ck , constants to be defined later, Mk := θk (mk ) + ck , for 1 ≤ k ≤ n. Next we claim: φ ,Ψk

uhk ≤ UMkk

a.e. in Ω for all 1 ≤ k ≤ n.

(2.15)

This would bring our proof to its conclusion due to the boundedness of the right hand side uniformly with respect to h (proven later). We prove (2.15) by induction. Note that, due to (2.13), it suffices to φ ,Ψk−1 φ ,Ψ h ≤ UMkk k a.e. in Ω. The induction hypothesis reads uhk−1 ≤ UMk−1 a.e. in Ω. Then show that vk−1 k−1 (2.8) and (g2) imply φ

,Ψk−1

vk−1 ≤ (1 + Ch)uhk−1 + Ch ≤ (1 + Ch)UMk−1 k−1

+ Ch.

Consequently, we wish to show φ

,Ψk−1

(1 + Ch)UMk−1 k−1

φk ,Ψk + Ch ≤ UM . k

This is equivalent to (1 + Ch)ϑk−1 (θk−1 (mk−1 ) + ck−1 − Ψk−1 ) + Ch ≤ ϑk (θk (mk ) + ck − Ψk ).

7

(2.16)

The relations in (2.14) give θk (mk ) = θk ((1 + Ch)ak−1 + Ch) − b = θk ((1 + Ch)ϑk−1 (θk−1 (mk−1 ) + b) + Ch) − b. Therefore, as all the θk , ϑk are strictly increasing, (2.16) is equivalent to θk ((1 + Ch)ϑk−1 (θk−1 (mk−1 ) + ck−1 − Ψk−1 ) + Ch) − (ck − Ψk ) ≤ ≤ θk ((1 + Ch)ϑk−1 (θk−1 (mk−1 ) + b) + Ch) − b. We now choose b = LT + ||Ψ0 ||L∞ (Ω) + ||Ψ||L∞ (Q) , where L is the Lipschitz constant in time for Ψ, i.e. we have: |Ψ(x, t + h) − Ψ(x, t)| ≤ Lh for all x ∈ Ω and t ∈ [0, T ] (2.17) and we define the ck recursively as c0 := ||Ψ0 ||L∞ (Ω) and ck := ck−1 + Lh. Therefore, we obtain 0 ≤ ck−1 − Ψk−1 ≤ ck − Ψk ≤ b and to prove (2.16) it is enough to show that q(s) := θk ((1 + Ch)ϑk−1 (θk−1 (mk−1 ) + s) + Ch) − s is increasing on (0, ∞). This result follows from (2.10) with the positive sign. Indeed, q  (s) = (1 + C h)θk ((1 + C h)ϑk−1 (θk−1 (m) + s) + C h)

d [ϑk−1 (θk−1 (m) + s)] − 1 ds

= η θk (η ϑk−1 (θk−1 (m) + s) + η − 1)ϑk−1 (θk−1 (m) + s) − 1,  where η := 1+Ch. Taking into account that ϑk−1 (x) = 1/[θk−1 (ϑ(x))], and setting y = ϑk−1 (θk−1 (m)+s) we obtain that q is nondecreasing if

η

f  (y) fk (η y + η − 1) ≥ k−1 ηy+η−1 y

nondecreasing. which follows from (f 3) (which implies fk ≥ (fk−1 ) ) and (2.10) with ∂f ∂s Else, we set b = 0 and ck as above and show that q(s) is decreasing on (0, ∞). As ck − Ψk ≥ 0, it follows that q(0) ≥ q(ck − Ψk ) which gives us the bound in this case. φ ,Ψ Of course, if g is null, then C = 0 and q is constant. Thus, (i) is proved if we can show that the UMkk k are bounded. recalling the various definitions involved we have: φk ,Ψk = ϑk ◦ (Mk − Ψk ) = ϑk ◦ (θk (ak ) − b + ck − Ψk ) ≤ ϑk ◦ θk (ak ) = ak UM k

when we have (2.10) with the positive sign. Now it follows easily from their definition that the terms ak are bounded by eC T (M0 + b + 1) − 1. In case we have (2.10) with the negative sign then the following is true φ ,Ψk

UMkk

≤ ϑk ◦ (θk (eC T (M0 + 1)) + C T + ||Ψ0 ||L∞ (Ω) + ||Ψk ||L∞ (Ω) ) ≤ ϑk ◦ θk (C  ) = C 

for such a C  which is bound to exist thanks to our hypothesis on the uniform growth of φs .



Remark: If Ψ ≡ 0 and f = f (u), then we can replace UCφ,Ψ in (2.11) by a constant and the proof remains valid. Also, in this case f need only be increasing (none of (f 1), (f 2), (f 3) or (2.10) are needed). Remark: Existence of a uniform lower bound for all uhk provided that u0 be essentially bounded from below away from zero does not seem to follow once we have the uniform upper bound (as it did in [1], [13], [20]). That is, unless we impose unrealistic conditions on the the potential Ψ and the lower bound of u0 . The existence of such uniform lower bound would make possible proving compactness in Section 4 without using (f 1) but weaker assumptions. We thank the referee for this observation.

8

2.2

The approximate Euler equation

Denote by N the set of all nonnegative integers. Proposition 1. Let u0 ∈ L∞ (Ω) be nonnegative of positive mass and {uk }k∈N be the solution of (2.7). Then uk ∈ L∞ (Ω) and fk (uk ) ∈ W 1,∞ (Ω) for all k ≥ 1. Furthermore 

 kh     1 1  − u )ζ + u ∇Ψ · ∇ζ + ∇f (u ) · ∇ζ − g(·, τ, u )dτ ζ dx ≤ (u k k−1 k k k k k−1  h h (k−1)h Ω ≤

1 ∇2 ζ∞ d(vk−1 , uk )2 , 2h

(2.18)

¯ for every ζ ∈ C ∞ (Ω). We will next give a sketch of the proof following mainly [20]. Proof: Consider a smooth vector field ξ ∈ Cc∞ (Ω; RN ) and the corresponding flux, i.e. a one-parameter family {ϑ(·; τ )}τ ∈R of diffeomorphisms, given by ∂τ ϑ(·; τ ) = ξ ◦ ϑ(·; τ ) for all τ ∈ R, ϑ(·; 0) = idΩ . Then, as mentioned in [11], ϑ(·; τ ) is invertible for all τ ∈ R and ϑ(Ω; τ ) = Ω for |τ | sufficiently small. For such τ we define uk ◦ ϑ−1 (·; τ ) uτ := , detJϑ(·; τ ) ◦ ϑ−1 (·; τ ) which can be shown to lie in Mvk−1 . This means that uτ is admissible in the variational principle (2.7) and we have 1 1 (2.19) d(vk−1 , uk )2 + h {φk (uk ) + uk Ψk }dx ≤ d(vk−1 , uτ )2 + h {φk (uτ ) + uτ Ψk }dx. 2 2 Ω Ω Note that uτ is the so-called push-forward density of uk under the diffeomorphism ϑ(·; τ ), i.e. ¯ uτ (x)ζ(x)dx = uk (x)ζ(ϑ(x; τ ))dx for all ζ ∈ C(Ω). Ω



Therefore, we can use Ψk instead of ζ and, as τ1 (Ψk (ϑ(·; τ )) − Ψk (·)) converges as τ ↓ 0 uniformly in Ω to ∇Ψk · ξ, we have   1 (uτ − uk )Ψk dx = ∇Ψk · ξuk dx. (2.20) lim τ →0 τ Ω Ω Now

d dτ

[detJϑ(·; τ )]|τ =0 = ∇ · ξ easily gives   1 lim φk (uτ )dx − φk (uk )dx = − fk (uk )∇ · ξdx. τ →0 τ Ω Ω Ω

(2.21)

Let µ ∈ P (vk−1 , uk ) be optimal in the definition of d(vk−1 , uk ). Then ¯ × Ω) ¯ ξ(x, y)dµτ (x, y) = ξ(x, ϑ(y; τ ))dµ(x, y) for all ξ ∈ C(Ω Ω×Ω

Ω×Ω

defines a µτ ∈ P (vk−1 , uτ ) and thus we have     1 1 1 1 1 1 2 2 2 2 d(vk−1 , uτ ) − d(vk−1 , uk ) |ϑ(y; τ ) − x| − |x − y| dµ(x, y), ≤ τ 2 2 2 2 Ω×Ω τ which, together with (2.19), (2.20) and (2.21), yields (upon replacing ξ also by −ξ)   ∇Ψk · ξuk − fk (uk )∇ · ξ dx = 0 for all ξ ∈ Cc∞ (Ω; RN ). (y − x) · ξ(y)dµ(x, y) + h Ω×Ω



9

(2.22)

Now consider ζ ∈ Cc∞ (Ω) which gives ∇ζ =: ξ ∈ Cc∞ (Ω; RN ). As the following holds ζ(y) − ζ(x) = (y − x) · ∇ζ(y) + (1/2)(y − x)T (∇2 ζ)(z)(y − x), for some z ∈ Ω (here we think of ζ as being extended by zero outside Ω), we infer      (uk − vk−1 )ζdx −  (y − x) · ∇ζ(y)dµ(x, y)     = 



(2.23)

Ω×Ω

  1 [ζ(y) − ζ(x) − (y − x) · ∇ζ(y)]dµ(x, y) ≤ sup |∇2 ζ| |y − x|2 dµ(x, y) 2 Ω Ω×Ω Ω×Ω

1 ¯ with ∇ζ ∈ Cc∞ (Ω; RN ). ∇2 ζ∞ d(vk−1 , uk )2 for all ζ ∈ C ∞ (Ω) 2 According to Lemma 2, all uk are uniformly bounded in L∞ for 1 ≤ k ≤ n − 1 independently of h. Obviously, so are fk (uk ) for 1 ≤ k ≤ n − 1. We will actually prove that fk (uk ) ∈ W 1,∞ (Ω). For this consider:       ≤ diam(Ω) (y − x) · ξ(y)dµ(x, y) |ξ(y)|dµ(x, y) (2.24)   =

Ω×Ω

Ω×Ω



= diam(Ω) Ω

|ξ(y)|uk (y)dy ≤ diam(Ω)uk Lp (Ω) ξLp∗ (Ω;RN ) ,

and so, by (2.22)       1   p ∞ p ≤ f (u )∇ · ξdx  + Ψ u  diam(Ω)u k k k L (Ω) k L (Ω) ξLp∗ (Ω;RN ) , L (Q)   h Ω for all ξ ∈ Cc∞ (Ω; RN ). This implies ∇fk (uk ) ∈ Lp (Ω) for all 1 ≤ p ≤ ∞. Therefore, {fk (uk )}k∈N ⊂ W 1,∞ (Ω) and, as Ω is bounded, we conclude that {fk (uk )}k∈N ⊂ W 1,p (Ω) for any 1 ≤ p ≤ ∞. It is important to realize that (2.22) can be more refined for p = 2 (we will make use of it later). Indeed    

Ω×Ω

   (y − x) · ξ(y)dµ(x, y)  ≤

1/2  1/2 2 |x − y| dµ(x, y) × uk (y)|ξ(y)| dy 2

Ω×Ω

(2.25)



≤ (uk ∞ )1/2 d(vk−1 , uk )ξL2 (Ω) . Note that (2.22) now entails (y − x) · ξ(y)dµ(x, y) + h (uk ∇Ψk + ∇fk (uk )) · ξdx = 0, Ω×Ω

(2.26)



for all ξ ∈ Cc∞ (Ω; RN ). In view of (2.25) and (2.26) we obtain ∇fk (uk )2L2 (Ω) ≤

C d(vk−1 , uk )2 + C∇Ψ2L∞ (Q) , h2

(2.27)

for C > 0 independent of k, h. The only thing that remains to be proved is that (2.18) holds for all ¯ (because (ii) is an obvious consequence of (i)). We refer the reader to [4], [8], [9] for the ζ ∈ C ∞ (Ω) existence of a convex function Φ : Ω → R which a.e. satisfies ∇Φ ∈ Ω and ¯ × Ω). ¯ ϕ(x, y)dµ(x, y) = uk (y)ϕ(∇Φ(y), y)dy for all ϕ ∈ C(Ω Ω×Ω



Applying this to ϕ(x, y) := x · ξ(y) gives, together with (2.26),  1 ∇Φ − idΩ uk − uk ∇Ψk − ∇fk (uk ) = 0 a.e. in Ω. h

10

(2.28)

¯ RN ) and, according to (2.8), (2.23) and (2.26), the proof is Next we integrate back against any ξ ∈ C ∞ (Ω; ¯ (not necessarily compactly supported complete once we note that (2.23) remains valid for any ζ ∈ C ∞ (Ω) in Ω).  Remark: We have seen that fk (uk ) ∈ W 1,∞ (Ω). Therefore, they are Lipschitz functions. Due to (f 1) we can, in fact, deduce that the minimizers themselves are regular. More specifically, the uk ’s themselves are Lipschitz functions if ω = 2 or H¨ older continuous functions of uniform H¨ older exponent if ω > 2.

3

Convergence to the weak solution

We group in this section a collection of results needed in the rest of the proof. Key to the following is the inequality in the first Lemma, which is an extension of a known result, see for example [11], [7] or [12], to take into consideration the non-homogeneity of the equations we consider. Lemma 3. There exists a constant C > 0 independent of h such that n−1 

h d(vk−1 , uhk )2 ≤ Ch,

(3.1)

k=1

where n ≥ 1 is integer and nh = T . Proof: Corollary 2 yields: 1 h h , uhk )2 + Skh (uhk ) ≤ Skh (vk−1 ) d(vk−1 2h for all k ≥ 1. Consequently



1 h , uhk )2 ≤ d(vk−1 2h



h [φk (vk−1 ) − φk (uhk )]dx +



h (vk−1 − uhk )Ψk dx.

Since the φk are convex we have, due to (2.8):   h h (x)) ≤ φk (uhk−1 (x)) + φk (vk−1 (x)) φk (vk−1 This implies

n−1 



k=1

n−1  Ω

k=1





k=1

g(x, τ, uhk−1 (x))dτ.

(k−1)h

n−1  kh k=1



kh

h [φk (vk−1 ) − φk (uhk )]dx ≤

[φk (uhk−1 ) − φk (uhk )]dx + φk (M )

n−1 



[φk (uhk−1 ) − φk (uhk )]dx + φk (M )



(k−1)h





T −h 0



g(·, ·, uhk−1 )dxdτ ≤

g(·, ·, uh )dxdτ,

where M > 0 is a common upper bound for all uhk ’s and vkh ’s. We also have n−1  k=1

[φk (uhk−1 ) − φk (uhk )]dx =



n−1  k=1



n−1  k=1



1 h



kh

[φk (uhk−1 ) − φk−1 (uhk−1 )]dx +



n−1  k=1

[φk−1 (uhk−1 ) − φk (uhk )]dx ≤



[φ(τ, uhk−1 ) − φ(τ + h, uhk−1 )]dτ dx +

(k−1)h



n−1  k=1



1 h



kh (k−1)h



[φ0 (uh0 ) − φN (uhN )]dx ≤

Chdτ dx + C ≤ C  .

11

This and the previous result give us a bound on the first term. Now we estimate n−1  k=1





n−1  k=1



h (vk−1 − uhk )Ψk dx =

n−1  Ω

k=1

(Ψk − Ψk−1 )uhk−1 dx +





n−1  k=1

n−1  k=1

≤ M LT |Ω| +

(uhk−1 − uhk )Ψk dx +





(Ψk−1 uhk−1 − Ψk uhk ) dx +

kh

g(·, t, uhk−1 )Ψk dtdx ≤

(k−1)h n−1  k=1



Ch(1 + M )||Ψ||∞ dx ≤

(Ψ0 uh0 − Ψn uhn )dx + CT (1 + M )||Ψ||∞ |Ω|.

The last two equations, together with (g1), give by summation n−1 1  h d(vk−1 , uhk )2 ≤ C, 2h k=1



for C > 0 independent of h. Since {uh } is bounded in L∞ (Q) we also deduce that uh  u weakly  in L∞ (Q) as h ↓ 0. Assume now that f and g are affine in u. More precisely, consider the problem: ⎧ ⎨ ut − ∇ · (u∇Ψ) − β∆u = a(x, t)u + b(x, t) in Ω × (0, T ), u ∂Ψ + β ∂u =0 on ∂Ω × (0, T ), ∂ν ∂ν ⎩ in Ω. u(·, 0) = u0

(3.2)

(3.3)

We still need all the conditions on f and g to remain imposed so we assume β > 0, 0 ≤ a(x, t), b(x, t) ≤ M < ∞ for all (x, t) ∈ Q. We next develop some useful tools to prove: Proposition 2. The problem (3.3) admits a weak solution for nonnegative and bounded initial data. Let us start by defining f h : (0, ∞) × R → R by f h (t, s) := fkh (s) whenever kh ≤ t < (k + 1)h.

(3.4)

Now, according to (2.27), (3.1) and Proposition 1, we have T h

=h



|∇f h (t, uh (x, t))|2 dxdt =

k=1

n−1  k=1

n−1  (k+1)h



|∇fkh (uhk )|2 dx ≤

kh



 |∇fkh (uhk )|2 dx dt =

n−1 C  h d(vk−1 , uhk )2 + C(n − 1)hΨ2∞ ≤ C  , h k=1



(as (n − 1)h = T − h < T ) for some C > 0 independent of h. Let now Gh be the extension by zero of ∇f h (·, uh ) (which is defined on Ω × (h, T )) to Q. Then {Gh }h↓0 is bounded in L2 (Q), which implies there exists G ∈ L2 (Q; RN ) such that, up to a subsequence (not relabelled) (3.5) Gh  G weakly in L2 (Q; RN ). The following lemma is also crucial to the proof of the main result of the paper (see Theorem 1; next section!). Lemma 4. Let ε > 0 be given. Then, for sufficiently small h > 0, the following holds:  T −h  1 1 h − uh (x, t) (ζ(x, t + h) − ζ(x, t))dxdt − u0 (x)ζ(x, t + h)dxdt+  h h 0 Ω h Ω T  h 1 T ∇f (t, uh ) + uh ∇Ψ · ∇ζdxdt + uh ζdxdt− + h T −h Ω h Ω

12

(3.6)

− for all ζ ∈

¯ Cc∞ (Ω

T −h 0

 n−1   1 h g(·, ·, uh )ζdxdt ≤ ∇2 ζ∞ d(vk−1 , uhk )2 + O(h) + ε 2 Ω k=1

× [0, T )).

¯ × [0, T )) in (2.18), integrate over Proof: Following [1], [11], [14], we consider ζ = ζ(x, t) ∈ Cc∞ (Ω [kh, (k + 1)h] and sum for k = 1, ..., n − 1. However, as shown in [20], these operations do not lead directly to (3.6) and it takes a couple of steps to get there. We basically need to show how O(h) and ε come into play. For this, note that by doing the operations suggested to obtain (3.6), the terms containing g add up to: n−1  (k+1)h 1 kh g(x, τ, uhk−1 (x))ζ(x, t)dτ dt. − h (k−1)h kh k=1

Exactly as in [20], let us now write each term as: 1 kh g(x, τ, uhk−1 (x))ζ(x, t)dτ = T1,k + T2,k , h (k−1)h where T1,k :=

1 h



kh (k−1)h

{g(x, τ, uhk−1 (x)) − g(x, t − h, uhk−1 (x))}ζ(x, t)dτ

and T2,k := g(x, t − h, uhk−1 (x))ζ(x, t). If h is sufficiently small, we have, due to (g4), |g(x, τ, uhk−1 (x)) − g(x, t − h, uhk−1 (x))| ≤ ε/T, as t ∈ [kh, (k + 1)h] implies |τ − t + h| ≤ h. Finally, integrating the T1,k ’s over [kh, (k + 1)h] and adding we obtain a quantity whose absolute value is less than (n − 1)hε/T < ε. As for the T2,k ’s, note that for all t ∈ [kh, (k + 1)h], we have g(x, t − h, uhk−1 (x)) = g(x, t − h, uh (x, t − h)). Therefore, by integrating over [kh, (k + 1)h] and adding we obtain, upon changing the variable t → t − h: T −h n−1  (k+1)h T2,k dt = g(x, t, uh (x, t))ζ(x, t + h)dt. k=1

0

kh

T −h

g(x, t, uh (x, t))ζ(x, t)dt, within O(h) distance. Note that this is close to what we want, i.e. 0 Also, note that the terms containing Ψ do not simply add up to what is displayed in (3.6). However, the argument for the latter terms is similar to the one above. Thus, (3.6) is proved .    Proof of Proposition 2: Fix t ∈ (0, T ) and notice that uh (·, t) = uhk for k := ht . We know that all uhk ’s (since f (uhk ) is in H 1 and f is affine) lie in H 1 (Ω) for k ≥ 1 and so, if we pick any ξ ∈ Cc∞ (Q), we can write (we have equality for the spatial integrals and then integrate in t): T T uhxi ξdxdt = − uh ξxi dxdt. h



h



It follows, in view of (3.2) and (3.5), that   βuξxi + Gi ξ dxdt = 0, i = 1, ..., N, Q

where Gi are the components of G defined in (3.5). This, along with the arbitrariness of ξ ∈ Cc∞ (Q) imply that G is the distributional spatial gradient of βu and so (3.5) becomes Gh  β∇u weakly in L2 (Q; RN ). We now have all the ingredients for passing to the limit in the left hand side of (3.6) if both f and g are affine in u. Therefore u is a weak solution.  What happens if at least one of f and g is not affine? The next section deals with this issue.

13

4

The main existence result

The weak  convergence result (3.2) that follows from the uniform bound of uh in L∞ (Q) is not enough to show convergence for (3.6) as h ↓ 0. The nonlinearity of the problem requires stronger convergence; for that purpose we show, in Proposition 3, that {uh }h↓0 is relatively compact in L1 (Q). Proposition 3. Assume (f 1)-(f 3), and (g1)-(g3) hold. Then {uh }h↓0 is relatively compact in L1 (Q). Proof: Let U h : RN +1 → [0, ∞) be the extension by 0 of uh outside Q. Let wh (x, := fh (t, uh (x, t)).  t)h We shall apply the Riesz-Fr´ echet-Kolmogorov criterion (see, e.g. [3]) to prove that U h↓0 Q = {uh }h↓0 is relatively compact in the strong topology of L1 (Q). We have   h U (x + e, t + τ ) − U h (x, t)dxdt ≤

(4.1)

Q



T +τ

≤ = τ



τ

T Ω

  h U (x + e, t) − U h (x, t)dxdt +

|U h (x + e, t) − U h (x, t)|dxdt +

T T −τ



  h U (x, t + τ ) − U h (x, t)dxdt = Q





T −τ

uh (x, t)dxdt + 0





|uh (x, t + τ ) − uh (x, t)|dxdt.

The idea is to show that the left hand side of (4.1) is bounded by some quantity that tends to vanish independently of h as τ and |e| tend to vanish. As we can see, there are two different issues that need be addressed; precompactness in space and precompactness in time. We will deal with them separately. Lemma 5. (Precompactness in space) There exists a function l = l(e) such that lim|e|↓0 l(e) = 0 and T Ω

τ

|U h (x + e, t) − U h (x, t)|dxdt ≤ l(e)

for all τ, h. Proof: We have T T |U h (x + e, t) − U h (x, t)|dxdt = τ



τ

|uh (x + e, t) − uh (x, t)|dxdt +

Ω∩(Ω−e)

T τ

uh (x, t)dxdt. Ω\(Ω−e)



 From Lemma (2), u is uniformly essentially bounded. Since meas Ω\(Ω − e) → 0 as |e| → 0 and we infer that the second term of the inequality above tends to 0 as |e| ↓ 0 uniformly in τ . We turn now to the first term. According to (3.5) we have h

T h



|∇wh |2 dxdt ≤

n−1 C2  h d(vk−1 , uhk )2 + nhΨ∞ ≤ C  , h k=1

for some C  > 0 independent of h. If τ ≥ h, we have T |wh (x + e, t) − wh (x, t)|dxdt ≤ τ

≤ |e|

1 T 0

τ

Ω∩(Ω−e)

|∇wh (x + se, t)|dxdtds ≤ |e|

Ω∩(Ω−e)

1 T 0

τ

Ωs

|∇wh (x, t)|dxdtds,

where Ωs := [(Ω + se) ∩ (Ω − (1 − s)e)] ⊂ Ω for sufficiently small |e| and for s ∈ [0, 1] (since Ω is convex). Since Q is bounded (thus the L1 norm is bounded from above by the L2 norm) the previous inequality chains and (3.5) implies T |wh (x + e, t) − wh (x, t)| dxdt ≤ C|e|, (4.2) τ

Ω∩(Ω−e)

14

for some constant C > 0. If τ < h all we need to observe is that T |wh (x + e, t) − wh (x, t)|dxdt = Ω∩(Ω−e)

τ



T = Ω∩(Ω−e)

h

|wh (x + e, t) − wh (x, t)|dxdt + (h − τ )

Ω∩(Ω−e)

|w0 (x + e) − w0 (x)| dx.

(Note w0h = w0 = f (u0 ) does not depend on h.) While we treat the first integral as before and obtain something similar to (4.2), we know that the second integral tends to zero as |e| → 0. We end the proof by noting that, if kh < t ≤ (k + 1)h, we have: |wh (x + e, t) − wh (x, t)| = |fk (uhk (x + e)) − fk (uhk (x))| =    1  (k+1)h  =  f (t, uhk (x + e)) − f (t, uhk (x)) dt = h kh

(4.3)

= |f (tk , uhk (x + e)) − f (tk , uhk (x))| ≥ c|uhk (x + e) − uhk (x)|ω−1 , 

where tk is some t in (kh, (k + 1)h). We now prove Lemma 6. (Precompactness in time) There exists a function q = q(τ ) such that limτ ↓0 q(τ ) = 0 and T −τ |uh (x, t + τ ) − uh (x, t)|dxdt ≤ q(τ ). 0



Proof: In order to do this let us fix an h > 0 and let τ = j h + γ with 0 ≤ γ < h so that: T − τ = nh − jh − γ = (n − j)h − γ < (n − j)h and estimate

T −τ

0



|uh (x, t + τ ) − uh (x, t)|ω dxdt.

Now, thanks to (f 1) and the definitions of uh and f h , in kh ≤ t < (k + 1)h, we have

T −τ 0

n−j−1 Ω

|uh (x, t + τ ) − uh (x, t)|ω dxdt ≤

k=0

n−j−1

≤C n−j−1

=C



k=0





k=0

(h − γ)



(k+1)h







(k+1)h

|uh (x, t + τ ) − uh (x, t)|ω dxdt ≤

kh

(uh (x, t + τ ) − uh (x, t))(f (t, uh (x, t + τ )) − f (t, uh (x, t)))dtdx =

kh



(uhk+j − uhk )(fk (uhk+j ) − fk (uhk ))dx + γ



(uhk+j+1 − uhk )(fk (uhk+j+1 ) − fk (uhk ))dx



(4.4) Also     h      uk+j −uhk fk (uhk+j )−fk (uhk ) = uhk+j −uhk fk+j (uhk+j )−fk (uhk ) + uhk+j −uhk fk (uhk+j )−fk+j (uhk+j ) =    uhk+j − uhk (k+1)h  = uhk+j − uhk fk+j (uhk+j ) − fk (uhk ) + f (t, uhk+j ) − f (t + jh, uhk+j ) dt ≤ h kh    ≤ uhk+j − uhk fk+j (uhk+j ) − fk (uhk ) + 2M jhL(f ; I), where I := [0, M ] and L(f ; I) is (according to (f 2)) the uniform Lipschitz constant of f (·, s) for s ∈ I (M being the uniform upper bound for all uhk ). Next we claim the following:  n+j−1  j h h  h h  (uk+j − uk ) fk+j (uk+j ) − fk (uk ) dx ≤ Cj + C (4.5) h Ω k=0

15

for all j ≥ 1 and a constant C > 0 (independent of k, q and h). To prove (4.5), we know that, for any i ≥ 1, there exists a convex potential Φ with ∇Φ : Ω → Ω onto that satisfies h , uhk+i )2 = uhk+i (y)|y − ∇Φ(y)|2 dy, (4.6) d(vk+i−1 Ω





h vk+i−1 (x)ζ(x)dx =



¯ uhk+i (y)ζ(∇Φ(y))dy for all ζ ∈ C(Ω).

(4.7)

As in [14], let ζ := fk+j (uhk+j ) − fk (uhk ). Due to (4.7) we have  h     h uk+i − vk+i−1 fk+j (uhk+j ) − fk (uhk ) dx = uhk+i (y) ζ(y) − ζ(∇Φ(y)) dy ≤ Ω



1 0



 Ω

1/2  1/2 2    2 uhk+i (y)∇ζ (1 − t)y + t∇Φ(y)  dy dt uhk+i (y)y − ∇Φ(y) dy . Ω

This is where Otto’s lemma 3 ([14]; Appendix) applies to |∇ζ| (recall that the wkh ’s are Lipschitz continuous). Thus, using (2.27) and (4.6), we discover   h (uhk+i − vk+i−1 ) fk+j (uhk+j ) − fk (uhk ) dx ≤ (4.8) Ω

  h C h h , uhk+i ) d(vk−1 , uhk ) + d(vk+j−1 , uhk+j ) + 2hΨ∞ ≤ d(vk+i−1 h for some C > 0 independent of k, i, h. Next we claim there exists C > 0 independent of k, j, h such that  h   uk+j − uhk fk+j (uhk+j ) − fk (uhk ) dx ≤ (4.9) Ω

j  C h h h ≤ ) + d(uhk+j , vk+j−1 ) + 2hΨ∞ d(uhk+i , vk+i−1 ) + Cjh. d(uhk , vk−1 h i=1 h − wkh and note that To prove this claim, recall that ζ = wk+j

(uhk+j − uhk )ζ =

j 

h (uhk+i − vk+i−1 )ζ +

i=1

j 

h (vk+i−1 − uhk+i−1 )ζ.

i=1

h We know, by (2.8), (g1), (g2) and the uniform boundedness of uhk that vk+i−1 − uhk+i−1 = O(h). We also h ∞ take into account that {wk }h is bounded in L . Therefore,  h  h  vk+i−1 − uhk+i−1 wk+j − wkh dx ≤ Ch. Ω

Summing up (4.8) and the inequalities above for i = 1..j we prove (4.9). Now write (4.9) for k = 1..n−j−1 and add to obtain n−j−1    h (uhk+j − uhk ) wk+j − wkh dx ≤ k=1



j  C  h h h d(uhk , vk−1 ) + d(uhk+j , vk+j−1 ) + 2hΨ∞ d(uhk+i , vk+i−1 ) + Cjh(n − j − 1). h i=1

n−j−1



k=1

As h(n − j − 1) ≤ T , all that remains to be proved is  

n−j−1

j  h h h d(uhk , vk−1 ) + d(uhk+j , vk+j−1 ) d(uhk+i , vk+i−1 ) ≤ Cjh i=1

k=1

and 2Ψ∞

j 

h d(uhk+i , vk+i−1 )≤C

i=1

16



jh.

In order to prove the first, rewrite the left hand side as j n−j−1   h h h h [d(uhk+i , vk+i−1 )d(uhk , vk−1 ) + d(uhk+i , vk+i−1 )d(uhk+j , vk+j−1 )] ≤ i=1



k=1 j n−j−1 1  h h h [2d(uhk+i , vk+i−1 )2 + d(uhk+j , vk+j−1 )2 + d(uhk , vk−1 )2 ]. 2 i=1 k=1

It is easy to see that (3.1) gives the second inequality. Applying (4.5), (for j and j + 1) in (4.4) concludes our proof.  Proof of Proposition 3 (conclusion): We finally estimate (see (4.1)) the term ially T uh (x, t) dxdt ≤ M τ meas(Ω). T −τ

T

T −τ Ω

uh (x, t)dxdt. Triv-



This last result, together with Lemmas 5 and 6, is enough to show that the Riesz-Fr´ echet-Kolmogorov criterion (see, e.g. [3]) applies to give the desired L1 precompactness. Thus, Proposition 3 is proved.  Therefore, {uh } is precompact in L1 (Q) and, up to a subsequence, we can pass to the limit in (3.6) to infer existence for (P ). We may thus prove: Theorem 1. Assume (f 1) - (f 3), (g1) - (g4) and (2.10). Then the problem (P ) admits a nonnegative essentially bounded weak solution provided that Ω is bounded and convex and the initial u0 is nonnegative and essentially bounded. Proof: In order to be able to pass to the limit in (3.6) we need to show ∇f (·, u) ∈ L2 (Q; RN ) and Gh  ∇f (·, u) weakly in L2 (Q; RN ), i. e. G = ∇f (·, u) (see (3.5)). Note that this would follow from f h (·, uh ) → f (·, u) strongly in L2 (Q) exactly as in the proof of Proposition 2. Trivially ||f h (t, uh ) − f (t, u)||L2 (Q) ≤ ||f h (t, uh ) − f (t, uh )||L2 (Q) + ||f (t, uh ) − f (t, u)||L2 (Q) . As the second term is bounded and tends pointwise to zero (up to a subsequence not relabelled), all we need to show is |f h (·, uh ) − f (·, uh )|2 dxdt = 0. lim h↓0

Q

Due to the boundedness of the integrand it suffices to show that it tends to zero in L1 (Q). We have |f h (t, uh ) − f (t, uh )|dxdt ≤ Q

1 h

n−1  Ω k=0

(k+1)h kh



 |f (τ, uhk (x)) − f (t, uhk (x))|dτ dtdx.

(k+1)h kh

As f (·, s) is Lipschitz continuous in t with respect to s in bounded domains we infer: |f h (t, uh ) − f (t, uh )|dxdt ≤ C h, Q



which brings the proof to its conclusion.

Remark: It is interesting to see what happens if we drop the “strong monotonicity” requirement (f 1). Following the proof of Proposition 3 we discover that (f 1) was only used to bound from above quantities of the type |uh (t + τ, x + e) − uh (x, t)|ω−1 by |f h (t + τ, uh (t + τ, x + e)) − f h (t, uh (x, t))|. In fact, the rest of the proof shows, without employing (f 1) in any form, that {f h (·, uh )}h is relatively compact in L1 (Q). It follows that, up to a subsequence (not relabelled), f h (·, uh ) → V strongly in L1 (Q)

17

for some nonnegative V ∈ L1 (Q). We have seen in the proof of Theorem 1 that this along with (f 2) imply V h := f (·, uh ) → V in L2 (Q). Due to (f 3), the operator T : L2 (Q) → L2 (Q) given by (T v)(x, t) := f (t, v(x, t)) is maximal monotone. The strong convergence of V h to V together with (3.2) imply uh , V h L2 (Q) → u, V L2 (Q) as h ↓ 0. Therefore, V = f (·, u) (see [19]). This observation allows us to generalize the existence result in [20] to the problem ⎧ ut − ∇ · (u∇Ψ(x, t)) − ∆θ = a(x, t)u + b(x, t, θ) in Ω × (0, T ), ⎪ ⎪ ⎨ u(x, for a. e. (x, t) ∈ Q,   t) ∈ H(t, θ(x, t)) on ∂Ω × (0, T ), u∇Ψ + ∇θ · νx = 0 ⎪ ⎪ ⎩ in Ω. θ(·, 0) = θ0 where θ0 is the initial datum, H is a maximal graph with an inverse f satisfying (f 2), (f 3). Also, a : RN +1 → R, b : RN +2 → R are such that (g1)-(g4) are satisfied by g(x, t, u) := a(x, t)u + b(x, t, f (t, u)).

5

Remarks on uniqueness

Uniqueness can be relatively easily proved under some restrictions. For the general case, see [1]. Proposition 4. Assume now g is Lipschitz in u. Within the hypotheses of Theorem 1 there exists at most one solution for (P ) which satisfies ∂u/∂t ∈ L1 (Q) provided that ∂f /∂s is bounded from below away from zero. Proof: Define ζδ as

ζδ (x, t) := φδ (f (t, u1 (x, t)) − f (t, u2 (x, t))), ⎧ ⎨

where φδ (s) :=

0 s/δ ⎩ 1

if if if

s≤0 0≤s≤δ s≥δ

We can then use, for the difference u1 − u2 of the solutions in (2.1), a smooth approximation of ζδ so that we can write  (u1 −u2 )∇Ψ·∇ζδ +∇(f (·, u1 )−f (·, u2 ))·∇ζδ −(g(·, ·, u1 )−g(·, ·, u2 ))ζδ dxdt ∂t (u1 −u2 ) ζδ dxdt = − Q

Q

Call now Qδ = Q ∩ {0 < f (·, u1 ) − f (·, u2 ) < δ} then, recalling the definition of ζδ , we have  1 (u1 − u2 )∇Ψ · ∇(f (·, u1 ) − f (·, u2 )) + |∇(f (·, u1 ) − f (·, u2 ))|2 dxdt ∂t (u1 − u2 )ζδ dxdt = − δ Q Qδ 1 (g(·, ·, u1 ) − g(·, ·, u2 ))(f (·, u1 ) − f (·, u2 ))dxdt + (g(·, ·, u1 ) − g(·, ·, u2 ))dxdt + δ Qδ Rδ 1/2  1 2 2 (u1 − u2 ) |∇Ψ| dxdt ∇f (·, u1 ) − ∇f (·, u2 )L2 (Qδ ) ≤ δ Qδ f (·, u1 ) − f (·, u2 ) +C(1 + M ) (u1 − u2 )+ dxdt, dxdt + Lipu (g) δ Qδ Q where Rδ := {f (·, u1 ) − f (·, u2 ) ≥ δ}. Here we use (g1), (g2) and (Ψ1). We recall that solutions of (P ) are bounded, 0 ≤ u ≤ M , and that in Qδ we have f (·, u1 ) − f (·, u2 ) < δ, therefore the second term in the last inequality is bounded and for δ → 0 the integral over Qδ tends to 0, as the measure of the set Qδ \{u1 = u2 } tends to 0. Since ∂f /∂s > c > 0, it follows c|u1 − u2 | ≤ |f (t, u1 ) − f (t, u2 )| ≤ δ. Now noticing that {f (·, u1 ) − f (·, u2 ) ≥ 0} = {u1 − u2 ≥ 0} and letting δ → 0 gives ∂t (u1 − u2 )+ dxdt ≤ Lipu (g) (u1 − u2 )+ dxdt. Q

Q

18

Therefore, for any 0 < t < T we have t (u1 (t) − u2 (t))+ dx ≤ (u1 (0) − u2 (0))+ dx + Lipu (g) (u1 (τ ) − u2 (τ ))+ dxdτ. Ω

0





+

As u1 (0) = u2 (0) in Ω, we apply Gronwall’s Lemma to infer (u1 − u2 ) = 0 a.e. in Q. Inverting the roles  of u1 and u2 concludes the proof.

Acknowledgements The authors would like to thank their PhD advisor, David Kinderlehrer, for his valuable suggestions and continual support. This research was partially supported by NSF grant DMS 0305794.

References [1] M. Agueh, Existence of solutions to degenerate parabolic equations via the Monge-Kantorovich theory, Ph.D. Thesis, Georgia Institute of Technology, 2002. [2] L. Borland, Microscopic dynamics of the nonlinear Fokker-Planck equation: A phenomenological model, Phys. Rev. E 57, 6 (1998), 6634-6642. [3] H. Brezis, Analyse Fonctionnelle; th´ eorie et applications, MASSON, Paris (1983). [4] L. A. Caffarelli, Allocation maps with general cost functions, Partial Diff. Eq. and Appl, P. Marcellini, G. G. Talenti, E. Vesintini, editors, Lecture Notes in Pure and Applied Mathematics, 177, Marcel Dekker, Inc., New York (1996), 29-35. [5] M. Chipot, D. Kinderlehrer, M. Kowalczyk, A variational principle for molecular motors, Meccanica, 38, (2003), no. 5, 505-518. [6] G. Drazer, H.S. Wio, C. Tsallis, Exact time-dependent solutions for anomalous diffusion with absorption, Granular Matter 3, Springer-Verlag, (2001), 105-107. [7] L. C. Evans, Partial Differential Equations and Monge-Kantorovich Mass Transfer, Current Developments in Mathematics, ed. S. T. Yau, (1997). [8] W. Gangbo, R.J. McCann, Optimal maps in Monge’s mass transport problems, C. R. Acad. Sci. Paris, 321 (1995), 1653-1658. [9] W. Gangbo, R. J. McCann, The geometry of optimal transportation, Acta Math. 177, 2, (1996), 113-161. [10] C. W. Gardiner, Handbook of stochastic methods, 2nd ed., Springer-Verlag, Berlin, Heidelberg (1985). [11] R. Jordan, D. Kinderlehrer, F. Otto, The Variational Formulation of the Fokker-Plank Equation, SIAM J. of Math. Anal. 29 (1998), 1-17. [12] D. Kinderlehrer, N. Walkington, Approximations of Parabolic Equations based upon Wasserstein’s Variational Principle, Math. Mod. Num. Anal., 33.4 (1999), 837-852. [13] F. Otto, Doubly degenerate diffusion equations as steepest descent, preprint of the University of Bonn (1996). [14] F. Otto, Dynamics of labyrinthine pattern formation in magnetic fluids: a mean field theory, Arch. Rat. Mech. Anal. 141 (1998), 63-103. [15] F. Otto, The geometry of the dissipative evolution equations: the porous medium equation, Comm. Partial Diff. Eq. 8 (1999). [16] I. T. Pedron, R. S. Mendes, L. C. Malacarne, E. K. Lenzi, Nonlinear anomalous diffusion equation and fractal dimension: Exact generalized Gaussian solution, Phys. Rev. E 65 041108 (2002). [17] I. T. Pedron, R. S. Mendes, L. C. Malacarne, E. K. Lenzi, N -dimensional nonlinear FokkerPlanck equation with time-dependent coefficients, Phys. Rev. E 65 052101 (2002).

19

[18] L. Petrelli, Notes on a Wasserstein metric convergence method for Fokker-Planck equations with point controls, CNA Technical Report 04-CNA-002 (2004). [19] R. E. Showalter, Monotone Operators in Banach Space and Nonlinear Partial Differential Equations, Mathematical Surveys and Monographs, vol. 49, AMS, (1997). [20] A. Tudorascu, One-phase Stefan problems; a mass transfer approach, Adv. Math. Sci. Appl. (to appear) (2004). [21] J. L. Vasquez, An introduction to the mathematical theory of the porous medium equation, Shape optimization and free boundaries, M.C. Delfour, G. Sabidussi (editors), Kluwer, Dordrecht (1992), 347-389.

20