Tikhonov Regularization for Weighted Total Least Squares Problems✩
Total Page:16
File Type:pdf, Size:1020Kb
View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Elsevier - Publisher Connector Applied Mathematics Letters 20 (2007) 82–87 www.elsevier.com/locate/aml Tikhonov regularization for weighted total least squares problems✩ Yimin Wei a,b,∗, Naimin Zhangc, Michael K. Ngd, Wei Xue a School of Mathematical Sciences, Fudan University, Shanghai 200433, PR China b Key Laboratory of Mathematics for Nonlinear Sciences (Fudan University), Ministry of Education, PR China c Mathematics and Information Science College, Wenzhou University, Wenzhou 325035, PR China d Department of Mathematics, Hong Kong Baptist University, Kowloon Tong, Hong Kong e Department of Computing and Software, McMaster University, Hamilton, Ont., Canada L8S 4L7 Received 15 August 2003; received in revised form 15 January 2006; accepted 1 March 2006 Abstract In this work, we study and analyze the regularized weighted total least squares (RWTLS) formulation. Our regularization of the weighted total least squares problem is based on the Tikhonov regularization. Numerical examples are presented to demonstrate the effectiveness of the RWTLS method. c 2006 Elsevier Ltd. All rights reserved. Keywords: Tikhonov regularization; Total least squares; Weighted regularized total least squares; Matrix differentiation; Lagrange multiplier 1. Introduction In this work, we study the regularized weighted total least squares (RWTLS) formulation. Our regularization of the weighted total least squares problem is based on the Tikhonov regularization [1]. For the total least squares (TLS) problem [2], the truncation approach has already been studied by Fierro et al. [3]. In [4], Golub et al. has considered the Tikhonov regularization approach for TLS problems. They derived a new regularization method in which stabilization enters the formulation in a natural way, and that is able to produce regularized solutions with superior properties for certain problems in which the perturbations are large. In the present work, we focus on RWTLS problems. We show that the RWTLS solution is closely related to the Tikhonov solution to the weighted least squares solution. Our work is organized as follows. In Section 2, we introduce the RWTLS formulation and study its regularizing properties. Computational methods are described in Section 3. In Section 4, numerical examples are presented to demonstrate the RWTLS method. ✩ Research supported in part by the National Natural Science Foundation of China under grant 10471027 and Shanghai Education Committee, and Hong Kong RGC Grants Nos 7046/03P, 7035/04P and 7035/05P, and HKBU FRGs. ∗ Corresponding author at: School of Mathematical Sciences, Fudan University, Shanghai 200433, PR China. E-mail addresses: [email protected], ymwei [email protected] (Y. Wei), [email protected] (N. Zhang), [email protected] (M.K. Ng). 0893-9659/$ - see front matter c 2006 Elsevier Ltd. All rights reserved. doi:10.1016/j.aml.2006.03.004 Y. Wei et al. / Applied Mathematics Letters 20 (2007) 82–87 83 2. The regularized weighted total least squares A general version of Tikhonov’s formulation for the linear weighted total least squares (WTLS) problem takes the form [5] ˜ ˜ ˜ ˜ min U[(A, b) − (A, b)]V F subject to b = Ax, DxS ≤ δ, (1) x where D is the regularization matrix, V = diag(W,γ)with γ being a non-zero constant, U and W are nonsingular 2 = T δ matrices, S is a symmetric positive definite matrix with y S y Sy,and is a positive constant. By using the Lagrange multiplier formulation, this problem can be rewritten as follows: L( ˜, ,μ)= [( , ) − ( ˜, ˜)] 2 + μ( 2 − δ2), ˜ = ˜ , A x U A b A b V F Dx S subject to b Ax (2) where μ is the Lagrange multiplier, and μ is equal to zero if the inequality constraint becomes equality. The solution x¯δ to this problem is different from the solution xWTLS to ˜ ˜ ˜ ˜ min U[(A, b) − (A, b)]V F subject to b = Ax, (3) x for δ less than DxWTLS2. Before we show the properties of the solution to (2), we have the following results about the matrix differentiation for the matrices A, A˜, W and U. Lemma 1. ∂ tr(W T ATU TU AW˜ ) ∂ tr(W T A˜TU TUAW) (i) = U TUAWWT (ii) = U TUAWWT ∂ A˜ ∂ A˜ ∂ tr(W T A˜TU TU AW˜ ) ∂(bTU TU Ax˜ ) (iii) = 2U TU AWW˜ T (iv) = U TUbxT ∂ A˜ ∂ A˜ ∂(xT A˜TU TUb) ∂(xT A˜TU TU Ax˜ ) (v) = U TUbxT (vi) = 2U TU Axx˜ T. ∂ A˜ ∂ A˜ Proof. Since (i) is equivalent to (ii), (iv) is equivalent to (v), and (vi) is a special case of (iii), we only give the proofs of (i) and (iii). Let Z be a p × q matrix of differentiable functions of the m × n matrix X.If ∂ Z = (mn) + ( (mn))T , = ,..., , = ,..., GEij H C Eij F i 1 m j 1 n ∂xij then ∂zij ( ) ( ) = GT E pq H T + F(E pq )TC, i = 1,...,p, j = 1,...,q, ∂ X ij ij and the converse is also true (see p. 57, Theorem 7.1 in [6]), where G = (gij) is a p ×m matrix, H = (hij) is an n ×q = ( ) × = ( ) × (kl) ( , ) matrix, C cij is a p n matrix, F fij is an m q matrix Eij is a k-by-l zero matrix except the i j -entry being equal to one. ˜ ˜ T T T ∂ tr(Y AW) ∂ ˜ ∂(Y AW)ii For (i), we consider Y = W A U U and we have = (Y AW)ii = .Since ∂ A˜ ∂ A˜ i i ∂ A˜ ˜ ∂( ˜ ) ˜ ∂(Y AW) = Y AW ij = T T ∂ tr(Y AW) = T T = T T ˜ YEijW and ˜ Y EijW , we obtain ˜ i Y Eii W Y W . The result follows. ∂ Aij ∂ A ∂ A ˜ ˜ ∂[(U AW)T(U AW)] T T T ˜ ˜ T For (iii), we find that ˜ = W E U U AW + (U AW) UEijW, and therefore we have ∂ Aij ij ˜ T ˜ ∂[(U AW) (U AW)]ii T ˜ T T T ˜ T = U U AW E W + U U AW Eii W . It follows that ∂ A˜ ii 84 Y. Wei et al. / Applied Mathematics Letters 20 (2007) 82–87 ∂ tr[(U AW˜ )T(U AW˜ )] ∂[(U AW˜ )T(U AW˜ )] = ii = U TU AW˜ ET W T + U TU AW˜ E W T ∂ ˜ ∂ ˜ ii ii A i A i = 2U TU AWW˜ T. Theorem 1. The RWTLS solution to (1) with the inequality constraint replaced by equality is a solution to the problem − − (ATU TUA+ αW T W 1 + β DT SD)x = ATU TUb, (4) where the parameters α and β are given by −γ 2b − Ax2 μ α = U TU ,β= ( + γ 2 2 ) 1 x −T −1 (5) + γ 2x2 γ 2 W W 1 W −T W −1 and μ is the Lagrange multiplier in (2). The two parameters are related by 1 βδ2 = (Ub)TU(b − Ax) + α, (6) γ 2 and the weighted TLS residual satisfies [( , ) − ( ˜, ˜)] 2 =−α. U A b A b V F (7) Proof. We characterize the solution to (1) by setting the partial derivatives of L(A˜, x,μ)to zero. Using Lemma 1,the differentiation of L(A˜, x,μ)with respect to A˜ yields U AWW˜ T − UAWWT − γ rx˜ T = 0, (8) where r˜ = γ U(b − Ax˜ ) = γ U(b − b˜). Moreover, the differentiation of L(A˜, x,μ) with respect to the entries in x yields −γ A˜TU Tr˜ + μDT SDx = 0or(γ 2 A˜TU TU A˜ + μDT SD)x = γ 2 A˜TU TUb. (9) By using (8) and (9),wehave ATU TUA = (U A˜ − γ rx˜ TW −T W −1)T(U A˜ − γ rx˜ TW −T W −1) = ˜T T ˜ + γ 2˜2 −T −1 T −T −1 A U U A r 2W W xx W W − − − − − μDT SDxxTW T W 1 − μW T W 1xxT DT SD ˜T T T T −T −1 T and A U Ub = A U Ub + γ W W xr˜ Ub. By using the assumption that DxS = δ and gathering the above terms, we obtain (5) with μ 2 2 2 2 T 2 2 α = μδ − γ ˜r x − − − γ r˜ Ub and β = (1 + γ x − − ). 2 W T W 1 γ 2 W T W 1 In order to obtain the expression for α, we first rewrite r˜ as ˜ = γ ( − ˜ ) = γ ( − − γ −1 ˜ T −T −1 ) = γ ( − ) − γ 2 ˜ 2 r U b Ax U b Ax U rx W W x U b Ax r x W −T W −1 from which we obtain the relation γ U(b − Ax) r˜ = . (10) 1 + γ 2x2 W −T W −1 From (9),wehave γ xT A˜TU Tr˜ (γ Ub−˜r)Tr˜ μ = = . (11) xT DT SDx δ2 By inserting (10) and (11) into the expression for α, we obtain (5).Eq.(6) is proved by multiplying β by δ2 and inserting (10) and (11). Y. Wei et al. / Applied Mathematics Letters 20 (2007) 82–87 85 Finally, we note from (8) that UAW − U AW˜ =−γ rx˜ TW −T and therefore we have (UAW,γUb) − (U AW˜ ,γU Ax˜ ) = (−γ rx˜ TW −T , r˜). It follows that [( , ) − ( ˜, ˜)] 2 =γ ˜ T −T 2 +˜2 U A b A b V F rx W F r 2 γ 2b − Ax2 = ( + γ 2 2 )˜2 = U TU =−α. 1 x −T −1 r W W 2 1 + γ 2x2 W −T W −1 The next theorem tells us the relationship between the RWTLS solution and the WTLS solution in (3) without the regularization. Theorem 2. For a given value of δ, the RWTLS solution x RWTLS(δ) is related to the solution xWTLS to the weighted total least squares problem without the regularization as follows: δ solution αβ ∂α δ<DxWTLSS xRWT LS(δ) = xWTLS α<0and ∂δ > 0 β>0 2 δ ≥DxWTLSS xRWT LS(δ) = xWTLS α =−σmin((UAW,γUb)) β = 0 Here σmin((UAW,γUb)) is the smallest singular value of the matrix (UAW,γUb).