Ridge/Lasso Regression, Model Selection

Ridge/Lasso Regression, Model Selection

Ridge/Lasso Regression Model Selection Ridge/Lasso Regression, Model selection Xuezhi Wang Computer Science Department Carnegie Mellon University 10701-recitation, Apr 22 Lasso Ridge/Lasso Regression Model Selection Outline 1 Ridge/Lasso Regression Linear Regression Regularization Probabilistic Intepretation 2 Model Selection Variable Selection Model selection Lasso Linear Regression Ridge/Lasso Regression Regularization Model Selection Probabilistic Intepretation Outline 1 Ridge/Lasso Regression Linear Regression Regularization Probabilistic Intepretation 2 Model Selection Variable Selection Model selection Lasso Linear Regression Ridge/Lasso Regression Regularization Model Selection Probabilistic Intepretation Linear Regression Data X: N × P matrix, Target y: N × 1 vector N samples, each sample has P features Want to find θ so that y and Xθ are as close as possible Pick θ that minimizes the cost function 1 X 1 L = (y − X θ)2 = jjy − Xθjj2 2 i i 2 i use gradient descent t+1 t @L t X θj = θj − step ∗ = θj − step ∗ (yi − Xi θ)(−Xij ) @θj i Lasso Linear Regression Ridge/Lasso Regression Regularization Model Selection Probabilistic Intepretation Linear Regression Matrix form: 1 X 1 L = (y − X θ)2 = jjy − Xθjj2 2 i i 2 i 1 = (y − Xθ)>(y − Xθ) 2 1 = (y >y − y >Xθ − θ>X >y + θ>X >Xθ) 2 Take derivative w.r.t. θ @L 1 = (−2X >y + 2X >Xθ) = 0 @θ 2 Hence we get θ = (X >X)−1X >y Lasso Linear Regression Ridge/Lasso Regression Regularization Model Selection Probabilistic Intepretation Linear Regression Comparison of iterative methods and matrix methods: matrix methods achieve solution in a single step, but can be infeasible for real-time data, or large amount of data. iterative methods can be used in large practical problems, but need to decide learning rate Any problems? Data X is an N × P matrix Usually N > P, i.e., number of data points larger than feature dimensions. And usually X is of full column rank. Under this case X >X have rank P, i.e., invertible What if X has less than full column rank? Lasso Linear Regression Ridge/Lasso Regression Regularization Model Selection Probabilistic Intepretation Outline 1 Ridge/Lasso Regression Linear Regression Regularization Probabilistic Intepretation 2 Model Selection Variable Selection Model selection Lasso Linear Regression Ridge/Lasso Regression Regularization Model Selection Probabilistic Intepretation Regularization: `2 norm Ridge Regression: 1 X 2 2 min (yi − Xi θ) + λjjθjj2 θ 2 i Solution is given by: θ = (X >X + λI)−1X >y Results in a solution with small θ Solves the problem that X >X is not invertible Lasso Linear Regression Ridge/Lasso Regression Regularization Model Selection Probabilistic Intepretation Regularization: `1 norm Lasso Regression: 1 X 2 min (yi − Xi θ) + λjjθjj1 θ 2 i Solution is given by taking subgradient: X (yi − Xi θ)(−Xij ) + λtj i where tj is the subgradient of `1 norm, tj = sign(θj ) if θj 6= 0; tj 2 [−1; 1] otherwise Sparse solution, i.e., θ will be a vector with more zero coordinates. Good for high-dimensional problems Lasso Linear Regression Ridge/Lasso Regression Regularization Model Selection Probabilistic Intepretation Solving Lasso regression Efron et al. proposed LARS (least angle regression) which computes the LASSO path efficiently Forward stagewise algorithm Assume X is standardized and y is centered choose small Start with initial residual r = y, and θ1 = ::: = θP = 0 Find the predictor Zj (jth column of X) most correlated with r > Update θj θj + δj , where δj = · sign(Zj r) Set r r − δj Zj , repeat steps 2 and 3 Lasso Linear Regression Ridge/Lasso Regression Regularization Model Selection Probabilistic Intepretation Comparison of Ridge and Lasso regression: Two-dimensional case: Lasso Linear Regression Ridge/Lasso Regression Regularization Model Selection Probabilistic Intepretation Comparison of Ridge and Lasso regression: Higher dimensional case: Lasso Linear Regression Ridge/Lasso Regression Regularization Model Selection Probabilistic Intepretation Choosing λ Standard practice now is to use cross-validation Lasso Linear Regression Ridge/Lasso Regression Regularization Model Selection Probabilistic Intepretation Outline 1 Ridge/Lasso Regression Linear Regression Regularization Probabilistic Intepretation 2 Model Selection Variable Selection Model selection Lasso Linear Regression Ridge/Lasso Regression Regularization Model Selection Probabilistic Intepretation Probabilistic Intepretation of Linear regression Assume yi = Xi θ + i , where is the random noise. Assume ∼ N (0; σ2) 2 1 (yi − Xi θ) p(yi jXi ; θ) = p exp{− g 2πσ 2σ2 Since data points are i.i.d, we have the data likelihood N PN 2 Y (yi − Xi θ) L(θ) = p(y jX ; θ) / exp{− i=1 g i i 2σ2 i=1 The log likelihood is: PN (y − X θ)2 `(θ) = − i=1 i i + const 2σ2 Maximizing the log-likelihood is equivalent to minimize PN 2 i=1(yi − Xi θ) , i.e., the loss function in LR! Lasso Linear Regression Ridge/Lasso Regression Regularization Model Selection Probabilistic Intepretation Probabilistic Intepretation of Ridge regression Assume a Gaussian prior on θ ∼ N (0; τ 2I), i.e., p(θ) / exp{−θ>θ=2τ 2g Now get the MAP estimate of θ PN (y − X θ)2 p(θjX; y) / p(yjX; θ)p(θ) = exp{− i=1 i i g exp{−θ>θ=2τ 2g 2σ2 The log likelihood is: PN (y − X θ)2 `(θjX; y) = − i=1 i i − θ>θ=2τ 2 + const 2σ2 1 P 2 2 which matches minθ 2 i (yi − Xi θ) + λjjθjj2, where λ is a constant associated with σ2; τ 2. Lasso Linear Regression Ridge/Lasso Regression Regularization Model Selection Probabilistic Intepretation Probabilistic Intepretation of Lasso regression iid Assume a Laplace prior on θi ∼ Laplace(0; t), i.e., p(θi ) / exp{−|θi j=tg Now get the MAP estimate of θ PN 2 (yi − Xi θ) X p(θjX; y) / p(yjX; θ)p(θ) = exp{− i=1 g exp{− jθ j=tg 2σ2 i i The log likelihood is: PN 2 (yi − Xi θ) X `(θjX; y) = − i=1 − jθ j=t + const 2σ2 i i 1 P 2 which matches minθ 2 i (yi − Xi θ) + λjjθjj1, where λ is a constant associated with σ2; t. Lasso Ridge/Lasso Regression Variable Selection Model Selection Model selection Outline 1 Ridge/Lasso Regression Linear Regression Regularization Probabilistic Intepretation 2 Model Selection Variable Selection Model selection Lasso Ridge/Lasso Regression Variable Selection Model Selection Model selection Variable Selection Consider "best" subsets, order O(2P ) (combinatorial explosion) Stepwise selection A new variable may be added into the model even with a small improvement in LMS When applying stepwise to a perturbation of the data, probably have different set of variables enter into the model at each stage LASSO produces sparse solutions, which takes care of model selection we can even see when variables jump into the model by looking at the LASSO path Lasso Ridge/Lasso Regression Variable Selection Model Selection Model selection Outline 1 Ridge/Lasso Regression Linear Regression Regularization Probabilistic Intepretation 2 Model Selection Variable Selection Model selection Lasso Ridge/Lasso Regression Variable Selection Model Selection Model selection Example Suppose you have data Y1; :::; Yn and you want to model the distribution of Y . Some popular models are: the Exponential distribution: f (y; θ) = θe−θy the Gaussian distribution: f (y; u; σ2) ∼ N (u; σ2) ... How do you know which model is better? Lasso Ridge/Lasso Regression Variable Selection Model Selection Model selection AIC Suppose we have models M1; :::; Mk where each model is a set of densities: Mj = fp(y; θj ): θj 2 Θj g We have data Y1; :::; Yn drawn from some density f (not necessarily drawn from these models). Define ^ AIC(j) = `j (θj ) − 2dj ^ where `j (θj ) is the log-likelihood, and θj is the parameter that maximizes the log-likelihood. dj is the dimension of Θj . Lasso Ridge/Lasso Regression Variable Selection Model Selection Model selection BIC Bayesian Information Criterion We choose j to maximize d BIC = ` (θ^) − j log n j j j 2 which is similar to AIC but the penalty is harsher, hence BIC tends to choose simpler models. Lasso Ridge/Lasso Regression Variable Selection Model Selection Model selection Simple example Let Y1; :::; Yn ∼ N (µ, 1) we want to compare two model: M0 : N (0; 1) and M1 : N (u; 1) Lasso Ridge/Lasso Regression Variable Selection Model Selection Model selection Simple example: AIC The log-likelihood is 2 Y −(Yi −u) =2 X 2 ` = log e = − (Yi − u) =2 i i X 2 AIC0 = − Yi =2 − 0 i X X n AIC = − (Y − Y¯ )2=2 − 2 = − Y 2=2 + Y¯ 2 − 2 1 i i 2 i i we choose model 1 if AIC1 > AIC0 i.e., X n X − Y 2=2 + Y¯ 2 − 2 > − Y 2=2 i 2 i i i q ¯ 4 or Y > n . Lasso Ridge/Lasso Regression Variable Selection Model Selection Model selection Simple example: BIC X 0 X BIC = − Y 2=2 − log n = − Y 2=2 0 i 2 i i i X 1 X 1 AIC = − (Y −Y¯ )2=2− log n = − Y 2=2+n=2Y¯ 2− log n 1 i 2 i 2 i i we choose model 1 if BIC1 > BIC0 i.e., X 1 X − Y 2=2 + n=2Y¯ 2 − log n > − Y 2=2 i 2 i i i q ¯ log n or Y > n . Lasso Ridge/Lasso Regression Variable Selection Model Selection Model selection Comparison Generally speaking, AIC/CV finds the most predictive model BIC find the true model with high probability, i.e., BIC assumes that one of the models is true and that you are trying to find the model most likely to be true in the Bayesian sense. Lasso.

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    27 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us