Neural Networks

Neural Networks

Neural Networks Aarti Singh Machine Learning 10-701/15-781 Nov 29, 2010 Slides Courtesy: Tom Mitchell 1 Logistic Regression Assumes the following functional form for P(Y|X): Logistic function applied to a linear function of the data (z) Logistic function logit (or Sigmoid): z Features can be discrete or continuous! 2 Logistic Regression is a Linear Classifier! Assumes the following functional form for P(Y|X): Decision boundary: 1 1 (Linear Decision Boundary) 3 Training Logistic Regression How to learn the parameters w0, w1, … wd? Training Data Maximum (Conditional) Likelihood Estimates Discriminative philosophy – Don’t waste effort learning P(X), focus on P(Y|X) – that’s all that matters for classification! 4 Optimizing concave/convex function • Conditional likelihood for Logistic Regression is concave • Maximum of a concave function = minimum of a convex function Gradient Ascent (concave)/ Gradient Descent (convex) Gradient: Update rule: Learning rate, >0 5 Logistic Regression as a Graph Sigmoid Unit d d d Neural Networks to learn f: X Y • f can be a non-linear function • X (vector of) continuous and/or discrete variables • Y (vector of) continuous and/or discrete variables • Neural networks - Represent f by network of logistic/sigmoid units, we will focus on feedforward networks: Output layer, Y Sigmoid Unit Hidden layer, H Input layer, X d d d /activation function (also linear, threshold) Differentiable Forward Propagation for prediction Sigmoid unit: 1-Hidden layer, 1 output NN: oh Prediction – Given neural network (hidden units and weights), use it to predict the label of a test point Forward Propagation – Start from input layer For each subsequent layer, compute output of sigmoid unit M(C)LE Training for Neural Networks • Consider regression problem f:XY , for scalar Y y = f(x) + assume noise N(0,), iid deterministic • Let’s maximize the conditional data likelihood Learned neural network MAP Training for Neural Networks • Consider regression problem f:XY , for scalar Y y = f(x) + noise N(0,σ) deterministic Gaussian P(W) = N(0,σI) 2 ln P(W) c i wi E – Mean Square Error d For Neural Networks, E[w] no longer convex in w Using all training data D y l l l l l l l l l y Error Gradient for a Sigmoid Unit Sigmoid Unit d d d y l l l l l l l y l l l l l l l y l l y l l l l l y l l l y l l l l l l l y l l l l l Error Gradient for 1-Hidden layer, 1- output neural network see Notes.pdf (MLE) Using Forward propagation y = target output (label) l l l y l l k k of output unit k ok(h) = unit output l l l l (obtained by forward propagation) wij = wt from i to j l Note: if i is input variable, oi = xi l lol Objective/Error no longer convex in weights How to avoid overfitting? Regularization – train neural network by maximize M(C)AP Early stopping Regulate # hidden units – prevents overly complex models ≡ dimensionality reduction w0 left strt right up Artificial Neural Networks: Summary • Actively used to model distributed computation in brain • Highly non-linear regression/classification • Vector-valued inputs and outputs • Potentially millions of parameters to estimate - overfitting • Hidden layers learn intermediate representations – how many to use? • Prediction – Forward propagation • Gradient descent (Back-propagation), local minima problems • Mostly obsolete – kernel tricks are more popular, but coming back in new form as deep belief networks (probabilistic interpretation).

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    27 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us