arXiv:1106.4424v2 [math.NA] 1 Dec 2011 -al [email protected] France. E-mail: 3, 92101, Cedex BP Nantes Nantes Noë, Centrale 44321 la Ecole de 6183, rue CNRS 1 UMR GeM, Université, LUNAM Nouy A. Spain. [email protected] (Valencia), E-mail: Patriarca del Alfara 46115 55 Comp Bartolomé la San Herrera, de Cardenal y CEU Matemáticas Universidad Físicas, Ciencias, de Departamento Falcó A. A Research National Herrera. PRCEU-UCH30/1 (French Cardenal the ANR CEU by dad and the ANR-2010-BLAN-0904) by and supported COSI-006 is work This Falcó Antonio Spaces Banach Convex Tensor in Nonlinear Problems for Decomposition Generalized Proper Abstract later inserted be should acceptance and receipt of date the sprildffrnileutosaiigfo tcatcc stochastic from arising equations t differential dimensional partial high sc in as in defined interest problems of growing solution a numerical receiving are methods Tensor-based Introduction 1 Algori Greedy Problems, High-Dimensional Decomposition, Keywords (2000) Classification Subject fun Mathematics convex a of minimization space. the Banach tensor with mathemati Decompositi associated a Generalized give problems Proper we of updated paper, and te this progressive of In of construction problems. priori such a of Decom the solution Generalized for dim Proper introduced high called recently in methods been defined of problems family of A solution spaces. numerical the for ing wl eisre yteeditor) the by inserted be (will No. manuscript Noname esrbsdmtosaerciigagoigitrs nsc in interest growing a receiving are methods Tensor-based olna ovxPolm,Tno aahsae,Poe Ge Proper spaces, Banach Tensor Problems, Convex Nonlinear · nhn Nouy Anthony 65K10,49M29 utación, luu eg Fokker-Planck (e.g. alculus srapoiain fthe of approximations nsor no rdc pcs such spaces, product ensor thms etfi optn o the for computing ientific n o atclrclass particular a for ons ninltno product tensor ensional oiin ehd have methods positions ec,gat ANR-2010- grants gency, a nlsso family a of analysis cal toa vrareflexive a over ctional nvriéd Nantes de Université , rn fteUniversi- the of grant 0 etfi comput- ientific neralized 2 equations) or quantum mechanics (e.g. Schrödinger equation), stochastic parametric partial differential equations in uncertainty quantification with functional approaches, and many mechanical or physical models involving extra parameters (for parametric analyses),.... For such problems, classical approximation methods based on the a pri- ori selection of approximation bases suffer from the so called “curse of dimensionality” associated with the exponential (or factorial) increase in the dimension of approxi- mation spaces. Tensor-based methods consist in approximating the solution u ∈ V of a problem, where V is a tensor space generated by d vector spaces Vj (assume e.g. nj 1 Vj = R ) , using separated representations of the form
m u u (1) (d) (j) ≈ m = wi ⊗ . . . ⊗ wi , wi ∈ Vj (1) Xi=1 (j) where ⊗ represents the Kronecker product. The functions wi are not a priori selected but are chosen in an optimal way regarding some properties of u. A first family of numerical methods based on classical constructions of tensor approximations [21,17,33] have been recently investigated for the solution of high- dimensional partial differential equations [18,3,20,26]. They are based on the system- atic use of tensor approximations inside classical iterative solvers. Another family of methods, called Proper Generalized Decomposition (PGD) methods [25,9,31,32,15], have been introduced for the direct construction of representations of type (1). PGD methods introduce alternative definitions of tensor approximations, not based on nat- ural best approximation problems, for the approximation to be computable without a priori information on the solution u. The particular structure of approximation sets allows the interpretation of PGDs as generalizations of Proper Orthogonal Decompo- sition (or Singular Value Decomposition, or Karhunen-Loève Decomposition) for the a priori construction of a separated representation um of the solution. They can also be interpreted as a priori model reduction techniques in the sense that they provide a way for the a priori construction of optimal reduced bases for the representation of the solution. Several definitions of PGDs have been proposed. Basic PGDs are based on a progressive construction of the sequence um, where at each step, an additional d (k) u elementary tensor ⊗k=1wm is added to the previously computed decomposition m−1 [24,2,28]. Progressive definitions of PGDs can thus be considered as Greedy algorithms [35] for constructing separated representations [6,1]. A possible improvement of these progressive decompositions consists in introducing some updating steps in order to capture an approximation of the optimal decomposition, which would be obtained by defining the whole set of functions simultaneously (and not progressively). For many applications, these updating strategies allow recovering good convergence properties of separated representations [29,32,31].
In [6], convergence results are given for the progressive Proper Generalized Decom- position in the case of the high-dimensional Laplacian problem. In [15], convergence is proved in the more general setting of linear elliptic variational problems in tensor Hilbert spaces. The progressive PGD is interpreted as a generalized singular value decomposition with respect to the metric induced by the operator, which is not neces-
1 More precisely, V is the closure with respect to a norm k · k of the algebraic tensor space d d (j) (j) V = a Vj = span v : v ∈ Vj and 1 ≤ j ≤ d Nj=1 nNj=1 o 3 sarily a crossnorm on the tensor product space.
In this paper, we propose a theoretical analysis of progressive and updated Proper Generalized Decompositions for a class of problems associated with the minimization of an elliptic and differentiable functional J,
J(u) = min J(v), v∈V where V is a reflexive tensor Banach space. In this context, progressive PGDs consist in defining a sequence of approximations um ∈ V defined by
um = um−1 + zm, zm ∈S1 where S1 is a tensor subset with suitable properties (e.g. rank-one tensors subset, Tucker tensors subset, ...), and where zm is an optimal correction in S1 of um−1, defined by
J(um−1 + zm)= min J(um−1 + z) z∈S1 Updated progressive PGDs consist in correcting successive approximations by using the information generated in the previous steps. At step m, after having computed an optimal correction zm ∈ S1 of um−1, a linear (or affine) subspace Um ⊂ V such that um−1 + zm ∈ Um is generated from the previously computed information, and the next approximation um is defined by
J(um)= min J(v) ≤ J(um−1 + zm) v∈Um
The outline of the paper is as follows. In section 2, we briefly recall some classical properties of tensor Banach spaces. In particular, we introduce some assumptions on the weak topology of the tensor Banach space in order for the (updated) progressive PGDs to be well defined (properties of subsets S1). In section 3, we introduce a class of convex minimization problems on Banach spaces in an abstract setting. In section 4, we introduce and analyze the progressive PGD (with or without updates) and we provide some general convergence results. While working on this paper, the authors became aware of the work [7], which provides a convergence proof for the purely progressive PGD when working on tensor Hilbert spaces. The present paper can be seen as an extension of the results of [7] to the more general framework of tensor Banach spaces and to a larger family of PGDs, including updating strategies and a general selection of tensor subsets S1. In section 5, we present some classical examples of applications of the present results: best approximation in Lp tensor spaces (generalizing the multidi- mensional singular value decomposition to Lp spaces), solution of p-Laplacian problem, and solution of elliptic variational problems (involving inequalities or equalities).
2 Tensor Banach spaces
d We first consider the definition of the algebraic tensor space a j=1 Vj generated from Banach spaces Vj (1 ≤ j ≤ d) equipped with norms k·kj . As underlying field we choose N 4
R C d , but the results hold also for . The suffix ‘a’ in a j=1 Vj refers to the ‘algebraic’ nature. By definition, all elements of N d V := a Vj Oj=1 v d (j) (j) are finite linear combinations of elementary tensors = j=1 v v ∈ Vj . A typical representation format is the Tucker or tensor subspace format N d (j) u = ai b , (2) ij i I X∈ Oj=1 where I = I1 × . . . × Id is a multi-index set with Ij = {1,...,rj}, rj ≤ dim(Vj ), (j) b ∈ V i ∈ I are linearly independent (usually orthonormal) vectors, and ai ∈ R. ij j j j Here, i are the components of i = (i ,...,i ). The data size is determined by the j 1 d numbers rj collected in the tuple r := (r1,...,rd). The set of all tensors representable by (2) with fixed r is
there are subspaces Uj ⊂ Vj such that Tr(V) := v ∈ V : d (3) dim(U )= r and v ∈ U := a U . j j j=1 j To simplify the notations, the set of rank-one tensors (elemNentary tensors) will be denoted by V V d (k) (k) R1( ) := T(1,...,1)( )= ⊗k=1w : w ∈ Vk . n o By definition, we then have V = span R1(V). We also introduce the set of rank-m tensors defined by m Rm(V) := zi : zi ∈ R1(V) . ( ) Xi=1 V We say that k·k is a Banach tensor space if there exists an algebraic tensor space V V V V and a norm k·k on such that k·k is the completion of with respect to the norm k·k, i.e. d k·k d V k·k := k·k Vj = a Vj . j=1 Oj=1 O V V If k·k is a Hilbert space, we say that k·k is a Hilbert tensor space.
2.1 Topological properties of Tensor Banach spaces
V V V V r Observe that span R1( ) is dense in k·k. Since R1( ) ⊂ Tr( ) for all ≥ (1, 1,..., 1), V V then span Tr( ) is also dense in k·k. d Any norm k·k on a j=1 Vj satisfying N d (j) d (j) (j) v = kv kj for all v ∈ Vj (1 ≤ j ≤ d) (4) j=1 j=1
O Y is called a crossnorm . 5
d (j) d (j) Remark 1 Eq. (4) implies the inequality k j=1 v k . j=1 kv kj which is equiv- alent to the continuity of the tensor product mapping N Q d d
: × Vj , k·kj −→ a Vj , k·k , (5) j=1 O Oj=1 (1) (d) d (j) given by ⊗ (v ,...,v ) = ⊗j=1v , where (X, k · k) denotes a vector space X equipped with norm k · k. As usual, the dual norm to k·k is denoted by k·k∗. If k·k is a crossnorm and also ∗ d ∗ k·k is a crossnorm on a j=1 Vj , i.e. ∗ d (j) N d (j) ∗ (j) ∗ ϕ = kϕ kj for all ϕ ∈ Vj (1 ≤ j ≤ d) , (6) j=1 j=1
O Y k·k is called a reasonable crossnorm. Now, we introduce the following norm.
Definition 1 Let Vj be Banach spaces with norms k·kj for 1 ≤ j ≤ d. Then for v V d ∈ = a j=1 Vj , we define the norm k·k∨ by
N ϕ(1) ⊗ ϕ(2) ⊗ . . . ⊗ ϕ(d) (v) (j) ∗ kvk := sup : 0 6= ϕ ∈ Vj , 1 ≤ j ≤ d . (7) ∨ d (j) ∗ kϕ k j=1 j Q The following proposition has been proved in [14]. V Proposition 1 Let k·k be a Banach tensor space with a norm satisfying k · k & k·k∨ on V. Then the set Tr(V) is weakly closed.
2.2 Examples
2.2.1 The Bochner spaces
Our first example, the Bochner spaces, are a generalization of the concept of Lp-spaces to functions whose values lie in a Banach space which is not necessarily the space R or C. s Let X be a Banach space endowed with a norm k · kX . Let I ⊂ R and µ a finite p measure on I (e.g. a probability measure). Let us consider the Bochner space Lµ(I; X), with 1 ≤ p< ∞, defined by
p p Lµ(I; X)= v : I → X : kv(x)kX dµ(x) < ∞ , ZI and endowed with the norm 1/p p kvk∆p = kv(x)kX dµ(x) ZI We now introduce the tensor product space V = X ⊗ Lp (I). For 1 ≤ p< ∞, k·k∆p k·k∆p µ the space Lp (I; X) can be identified with V (see Section 7, Chapter 1 in [10]). µ k·k∆p Moreover, the following proposition can be proved (see Proposition 7.1 in [10]): 6
p Proposition 2 For 1 ≤ p< ∞, the norm k·k∆p satisfies k·k∆p & k·k∨ on X⊗aLµ(I). By Propositions 2 and 1, we then conclude: p p Corollary 1 For 1 ≤ p< ∞, the set Tr X ⊗a Lµ(I) is weakly closed in Lµ(I; X). In p p p p particular, for K ⊂ Rk, we have that Tr L (K) ⊗ L (I) and R L (K) ⊗ L (I) ν a µ 1 ν a µ are weakly closed sets in Lp (K × I) . ν⊗µ 2.2.2 The Sobolev spaces
d d Let Ω = Ω1 × . . . × Ωd ⊂ R , with Ωk ⊂ R. Let α ∈ N denote a multi-index and d α u α1 αd u u |α| = k=1 αk. D ( ) = ∂x1 ...∂xd ( ) denotes a partial derivative of (x1,...,xd) of order |α|. For a fixed 1 ≤ p< ∞, we introduce the Sobolev space P Hm,p(Ω)= {u ∈ Lp(Ω) : Dα(u) ∈ Lp(Ω), 0 ≤ |α|≤ m} equipped with the norm v α v k km,p = kD ( )kLp(Ω) 0≤|Xα|≤m m,p We let Vk = H (Ωk), endowed with norms k · km,p;k defined by m j p kwkm,p;k = k∂xk (w)kL (Ωk). jX=0 Then we have the following equality d Hm,p(Ω)= Hm,p(Ω ) . k·km,p j j=1 O A first result is the following.
Proposition 3 For 1
d m,p d (k) (k) m,p R1 a H (Ω ) = ⊗ w : w ∈ H (Ω ) , j k=1 k j=1 O n o m,p is weakly closed in (H (Ω), k · km,p). To prove the above proposition we need the following two lemmas.
d p Lemma 1 Assume 1
Lemma 2 Assume 1
d d αk p k⊗k=1 wkkm,p = k∂xk (wk)kL (Ωk) 0≤|Xα|≤m kY=1 d αk p ≤ k∂xk (wk)kL (Ωk) d α∈{0X,...,m} kY=1 d m d j = k∂ (w )k p = kw k , xk k L (Ωk ) k m,p;k kY=1 jX=0 kY=1 d m,p we have ⊗k=1wk ∈ H (Ω). d m,p Conversely, if ⊗k=1wk ∈ H (Ω), then d α d k⊗k=1 wkkm,p = kD (⊗k=1wk)kLp(Ω) < ∞ 0≤|Xα|≤m α d which implies that kD (⊗k=1wk)kLp(Ω) < ∞ for all α such that 0 ≤ |α|≤ m. Taking α = (0,..., 0), we obtain
d d p p k⊗k=1 wkkL (Ω) = kwkkL (Ωk) < ∞ kY=1 p and therefore kwkkL (Ωk) < ∞ for all k. Now, for k ∈ {1 ...d}, taking α =(..., 0, j, 0,...) such that αk = j, with 1 ≤ j ≤ m, and αl = 0 for l 6= k, we obtain α d j p p p kD (⊗l=1wl)kL (Ω) = k∂xk wkkL (Ωk) kwlkL (Ωl) lY6=k j m,p and then k∂xk wkkp,Ωk < ∞ for all j ∈ {1,...,m}. Therefore wk ∈ H (Ωk) for all k ∈ {1 ...d}. ⊓⊔
Proof of Proposition 3 For m = 0 the proposition follows from Lemma 1. Now, assume m ≥ 1, and let us consider a sequence
d m,p {zn}n∈N ⊂ R1 a H (Ω ) j Oj=1 that weakly converges to an element z ∈ Hm,p(Ω). Since
d d m,p p R1 a H (Ω ) ⊂ R1 a L (Ω ) , j j j=1 j=1 O O z d p we have ∈ R1 a j=1 L (Ωj ) because, from Lemma 1, the latter set is weakly p p z d closed in L (Ω). Therefore,N there exist wk ∈ L (Ωk) such that = ⊗k=1wk. Since 8
m,p m,p z ∈ H (Ω), from Lemma 2, wk ∈ H (Ωk) for 1 ≤ k ≤ d, and therefore z = d d m,p ⊗k=1wk ∈ R1 a j=1 H (Ωj ) . ⊓⊔ From [14] it followsN the following statement.
d r m,2 m,2 Proposition 4 The set T a j=1 H (Ωj ) is weakly closed in H (Ω). N
3 Optimization of functionals over Banach spaces
Let V be a reflexive Banach space, endowed with a norm k · k. We denote by V ∗ the dual space of V and we denote by h·, ·i : V ∗ × V → R the duality pairing. We consider the optimization problem
J(u) = min J(v) (π) v∈V where J : V → R is a given functional.
3.1 Some useful results on minimization of functionals over Banach spaces
In the sequel, we will introduce approximations of (π) by considering an optimization on subsets M ⊂ V , i.e.
inf J(v) (8) v∈M
We here recall classical theorems for the existence of a minimizer (see e.g. [13]). We recall that a sequence vm ∈ V is weakly convergent if limm→∞hϕ, vmi exists for ∗ all ϕ ∈ V . We say that (vm)m∈N converges weakly to v ∈ V if limm→∞hϕ, vmi = hϕ, vi ∗ for all ϕ ∈ V . In this case, we write vm ⇀ v.
Definition 2 A subset M ⊂ V is called weakly closed if vm ∈ M and vm ⇀ v implies v ∈ M.
Note that ‘weakly closed’ is stronger than ‘closed’, i.e., M weakly closed ⇒ M closed.
Definition 3 We say that a map J : V −→ R is weakly sequentially lower semicon- tinuous (respectively, weakly sequentially continuous) in M ⊂ V if for all v ∈ M and for all vm ∈ M such that vm ⇀ v, it holds J(v) ≤ lim infm→∞ J(vm) (respectively, J(v) = limm→∞ J(vm)).
If J′ : V −→ V ∗ exists as Gateaux derivative, we say that J′ is strongly continuous ′ ′ ∗ when for any sequence vn ⇀ v in V it holds that J (vn) → J (v) in V . Recall that the convergence in norm implies the weak convergence. Thus, J weakly sequentially (lower semi)continuous in M ⇒ J (lower semi)continuous in M. It can be shown (see Proposition 41.8 and Corollary 41.9 in [37]) the following result.
Proposition 5 Let V be a reflexive Banach space and let J : V → R be a functional, then the following statements hold. 9
(a) If J is a convex and lower semicontinuous functional, then J is weakly sequentially lower semicontinuous. (b) If J′ : V −→ V ∗ exists on V as Gateaux derivative and is strongly continuous (or compact), then J is weakly sequentially continuous.
Finally, we have the following two useful theorems.
Theorem 1 Assume V is a reflexive Banach space, and assume M ⊂ V is bounded and weakly closed. If J : M → R ∪ {∞} is weakly sequentially lower semicontinuous, then problem (8) has a solution.
Proof Let α = infv∈A J(v) and {vn} ⊂ A be a minimizing sequence. Since A is bounded, {vn}n∈N is a bounded sequence in a reflexive Banach space and therefore,
there exists a subsequence {vnk }k∈N that converges weakly to an element u ∈ V . Since A is weakly closed, u ∈ A and since J is weakly sequentially lower semicontin-
uous, J(u) ≤ lim infk→∞ J(vnk ) = α. Therefore, J(u) = α and u is solution of the minimization problem. ⊓⊔
We now remove the assumption that M is bounded by adding a coercivity condition on J.
Theorem 2 Assume V is a reflexive Banach space, and M ⊂ V is weakly closed. If J : M → R∪{∞} is weakly sequentially lower semicontinuous and coercive on M, then problem (8) has a solution.
Proof Pick an element v0 ∈ M such that J(v0) 6= ∞ and define M0 = {v ∈ M : J(v) ≤ J(v0)}. Since J is coercive, M0 is bounded. Since M is weakly closed and J is weakly sequentially lower semicontinuous, M0 is weakly closed. The initial problem is then
equivalent to infv∈M0 J(v), which admits a solution from Theorem 1. ⊓⊔
3.2 Convex optimization in Banach spaces
From now one, we will assume that the functional J satisfies the following assumptions.
(A1) J is Fréchet differentiable, with Fréchet differential J′ : V → V ∗. (A2) J is elliptic, i.e. there exist α> 0 and s> 1 such that for all v,w ∈ V ;
hJ′(v) − J′(w), v − wi ≥ αkv − wks (9)
In the following, s will be called the ellipticity exponent of J.
Lemma 3 Under assumptions (A1)-(A2), we have
(a) For all v,w ∈ V, α J(v) − J(w) ≥hJ′(w), v − wi + kv − wks. (10) s
(b) J is strictly convex. (c) J is bounded from below and coercive, i.e. limkvk→∞ J(v)=+∞. 10
Proof (a) For all v,w ∈ V ,
1 d 1 J(v) − J(w)= J(w + t(v − w))dt = hJ′(w + t(v − w)), v − widt dt Z0 Z0 1 = hJ′(w), v − wi + hJ′(w + t(v − w)) − J′(w), v − widt Z0 1 α ≥hJ′(w), v − wi + kt(v − w)ksdt 0 t Zα = hJ′(w), v − wi + kv − wks s (b) From (a), we have for v 6= w,
J(v) − J(w) > hJ′(w), v − wi
(c) Still from (a), we have for all v ∈ V , α α J(v) ≥ J(0) + hJ′(0), vi + kvks ≥ J(0) − kJ′(0)kkvk + kvks s s which gives the coercivity and the fact that J is bounded from below. ⊓⊔
The above properties yield the following classical result.
Theorem 3 Under assumptions (A1)-(A2), the problem (π) admits a unique solution u ∈ V which is equivalently characterized by
hJ′(u), vi = 0 ∀v ∈ V (11)
Proof We here only give a sketch of proof of this very classical result. J is continuous and a fortiori lower semicontinuous. Since J is convex and lower semicontinuous, it is also weakly sequentially lower semicontinuous (Proposition 5(a)). The existence of a solution then follows from Theorem 2. The uniqueness is given by the strict convexity of J, and the equivalence between (π) and (11) classically follows from the differentiability of J. ⊓⊔
Lemma 4 Assume that J satisfies (A1)-(A2). If {vm} ⊂ V is a sequence such that J(vm) −→ J(u), where u is the solution of (π), then vm → u, i.e. m→∞
ku − vmk −→ 0 m→∞
Proof By the ellipticity property (10) of J, we have
′ α s α s J(vm) − J(u) ≥hJ (u), vm − ui + ku − vmk = ku − vmk . (12) s s Therefore, α s ku − vmk ≤ J(vm) − J(u) −→ 0, s m→∞ which ends the proof. ⊓⊔ 11
4 Progressive Proper Generalized Decompositions in Tensor Banach Spaces
4.1 Definition of progressive Proper Generalized Decompositions
We now consider the minimization problem (π) of functional J on a reflexive tensor V V R Banach space V = k·k. Assume that we have a functional J : k·k −→ satisfying V (A1)-(A2) and a weakly closed subset S1 in k·k such that
(B1) S1 ⊂ V, with 0 ∈S1, (B2) for each v ∈S1 we have λv ∈S1 for all λ ∈ R, and V (B3) span S1 is dense in k·k. By using the notation introduced in Section 2.2 we give the following examples. V p p Example 1 Consider k·k = Lµ(I; X) and S1 = Tr X ⊗a Lµ(I) .