Semi-Supervised Learning Supervised Learning P G

Semi-Supervised Learning Supervised Learning P G

Semi‐Supervised Learning Aarti Singh Machine Learning 10-701/15-781 April 19, 2010 Slides Courtesy: Jerry Zhu 1 HW, Exam & Project • HW 5: out today, due April 26 (Monday) @ beginning of class • PjtProject PtPoster SiSession: May 4 (Tues day ), 3‐6 pm, NSH Atri um • Project Report: May 5 (Wednesday) @ midnight by email • Exam: May 7 (Friday), 5:30‐8:30 pm, DH 2302 Supervised Learning Feature Space Label Space Goal: Optimal predictor (Bayes Rule) depends on unknown PXY, so instead learn a good prediction rule from training data Learning algorithm Labeled Labeled and Unlabeled data “Crystal” “Needle” “Empty” “0” “1” “2” … Human expert/ “Sports” Special equipment/ “News” Experiment “Science” … Cheap and abundant ! Expensive and scarce ! Example: Hard to obtain labels Example: Hard to obtain labels Free‐of‐cost labels? Luis von Ahn: Games with a purpose (ReCaptcha) Word rejected by OCR (Optical Character Recogintion) You provide a free lbl!label! Semi‐Supervised learning Learning algorithm Supervised learning (SL) “Crystal” Semi-Supervised learning (SSL) Goal: Learn a better prediction rule than based on labeled data alone. Semi‐Supervised learning in Humans Can unlabeled data help? Positive labeled data Negative labeled data Unlabeled data SidDiiBdSupervised Decision Boundary Semi-Supervised Decision Boundary Assume each class is a coherent group (e.g. Gaussian) Then unlabeled data can help identify the boundary more accurately. Can unlabeled data help? “0” “1” “2” … 7 11 2 9 2 44 88 33 5 “Similar” data points have “similar” labels Self‐training Self‐training Example Propagating 1-NN Some SSL Algorithms . GtithdGenerative methods – assume a modlfdel for p (x,y)d) and max iiimize joint likelihood Mixture models . Multi-view methods – multiple independent learners that agree on prediction for unlabeled data Co-training . Graph-based methods – assume the target function p(y|x) is smooth wrt a graph or manifold Grap h/Manif old Regu lar iza tion Some SSL Algorithms . GtithdGenerative methods – assume a modlfdel for p (x,y)d) and max iiimize joint likelihood Mixture models . Multi-view methods – multiple independent learners that agree on prediction for unlabeled data Co-training . Graph-based methods – assume the target function p(y|x) is smooth wrt a graph or manifold Grap h/Manif old Regu lar iza tion Mixture Models Mixture Models <> 1/2 Mixture Models Mixture Models Mixture Models Mixture Models Mixture Models Gaussian Mixture Models EM for Gaussian Mixture Models Assumption for GMMs Assumption for GMMs Assumption for GMMs Other issues: Identifiability, EM local optima Related: Cluster and Label Some SSL Algorithms . GtithdGenerative methods – assume a modlfdel for p (x,y)d) and max iiimize joint likelihood Mixture models . Multi-view methods – multiple independent learners that agree on prediction for unlabeled data Co-training . Graph-based methods – assume the target function p(y|x) is smooth wrt a graph or manifold Grap h/Manif old Regu lar iza tion Two views of an Instance Two views of an Instance Two views of an Instance Co‐training Algorithm Blum & Mitchell’98 Co‐training Multi‐view learning Multi‐view learning Each of the k learners is good The k learners agree on unlabeled data Some SSL Algorithms . GtithdGenerative methods – assume a modlfdel for p (x,y)d) and max iiimize joint likelihood Mixture models . Multi-view methods – multiple independent learners that agree on prediction for unlabeled data Co-training . Graph-based methods – assume the target function p(y|x) is smooth wrt a graph or manifold Grap h/Manif old Regu lar iza tion Graph Regularization Assumption: Similar unlabeled data have similar labels. Graph Regularization Similarity Graphs: Model local neighborhood relations between data points Graph Regularization GhPiGraph Prior: If data points i and j are similar (i.e. weight wij is large), then their labels are similar fi = fj Loss on labeled data Graph based smoothness prior (mean square,0-1) on labeled and unlabeled data Graph Regularization Loss on labeled data Graph based smoothness prior on labeled and unlabeled data From previous lecture, recall the second term is simply the min-cut objective. If binary label, can be solved by min-cut on a modified graph - add source and sink nodes with large weight to labeled examples. Blum & Chawla’01 Source +1 Sink -1 Semi‐Supervised Learning . Generative methods – Mixture models . Multi-view methods – Co-training . Graph-based methods – Manifold Regularization . Semi-Supervised SVMs – assume unlabeled data from different classes have large margin . Many other methods SSL alggpporithms can use unlabeled data to help improve prediction accuracy if data satisfies appropriate assumptions.

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    45 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us