Machine Learning: Basic Principles

Machine Learning: Basic Principles

T-61.6020 Special Course in Computer and Information Science Machine Learning: Basic Principles Sparse Kernel Machines 20070226 | Janne Argillander, [email protected] T-61.6050 Special Course in Computer and Information Science Contents ° Introduction ° Maximum margin classifier ° Overlapping class distributions ° Multiclass classifier ° Relevance Vector Machines ° Outroduction 2 20070226 | Janne Argillander, [email protected] T-61.6050 Special Course in Computer and Information Science Contents ° Introduction ° Maximum margin classifier ° Overlapping class distributions ° Multiclass classifier ° Relevance Vector Machines ° Outroduction 3 20070226 | Janne Argillander, [email protected] T-61.6050 Special Course in Computer and Information Science ° Argorithms based on the non-linear kernels are nice, but because the kernel function k(xn,xm) has to be evaluated for possible pairs of the training points the computational complexity is high. ‰ Some kind of ”reduction” has to be done ° We say solution to be sparse when only the subset of the training data points are used for evaluation. 4 20070226 | Janne Argillander, [email protected] T-61.6050 Special Course in Computer and Information Science Which boundary should we pick? 5 20070226 | Janne Argillander, [email protected] T-61.6050 Special Course in Computer and Information Science Contents ° Introduction ° Maximum margin classifier ° Overlapping class distributions ° Multiclass classifier ° Relevance Vector Machines ° Outroduction 6 20070226 | Janne Argillander, [email protected] T-61.6050 Special Course in Computer and Information Science ° Support Vector Machine (SVM) approaches the problem stated in the previous slide through the concept of the margin ° The margin is the smallest distance between the decision boundary and any of the training samples ° The points with the smallest margin define the decision boundary and are called as support vectors 7 20070226 | Janne Argillander, [email protected] T-61.6050 Special Course in Computer and Information Science 8 20070226 | Janne Argillander, [email protected] T-61.6050 Special Course in Computer and Information Science Properties of SVM ° SVM does not provide posterior probabilities. The output value can be seen as a distance from the decision boundary to the test point (measured in the feature space) – Considering the retrieval task (most probable first), how do we sort the results? – Can we combine two different methods or models? ° Can be used for classifying, regression and to data reduction – In the classifying task the samples are usually annotated as a positive or negative examples. SVM accepts also examples without annotations. ° SVM is fundamentally a two-class classifier – What if we have more classes? 9 20070226 | Janne Argillander, [email protected] T-61.6050 Special Course in Computer and Information Science Properties of SVM ° The naive solution is slow O(N3), but with the help of lagrange multipliers task is only from O(N) to O(N2) ° Handles non-línearly separable data set in the linearly separable feature space. – This space is defined by the nonlinear kernel function K. – Dimensionality of the feature space can be high ° Traditional SVM does not like outliers 10 20070226 | Janne Argillander, [email protected] T-61.6050 Special Course in Computer and Information Science Contents ° Introduction ° Maximum margin classifier ° Overlapping class distributions ° Multiclass classifier ° Relevance Vector Machines ° Outroduction 11 20070226 | Janne Argillander, [email protected] T-61.6050 Special Course in Computer and Information Science Overlapping class distributions ° Penalty that increases with the distance (slack variables) ° -SVM J Method that can handle data sets with few outliers L Parameters have to be estimated - Cross-validation over the hold-out set 12 20070226 | Janne Argillander, [email protected] T-61.6050 Special Course in Computer and Information Science Contents ° Introduction ° Maximum margin classifier ° Overlapping class distributions ° Multiclass classifier ° Relevance Vector Machines ° Outroduction 13 20070226 | Janne Argillander, [email protected] T-61.6050 Special Course in Computer and Information Science SVM is fundamendally a two-class classifier Let us consider problem with K classes ° One-versus-rest approach (1-BGM) – Training: Ck marked as positive, rest K-1 classes as negative ‰ Training sets will be inbalanced (if we had symmetry now it is lost) (10 ’faces’, 100000 non faces) ‰ What if we had similar classes? ‰ Can lead to very very slow models, O(KN2) – O(K2N2) ° One-versus-one approach – Training: 2-Class SVMs on all possible pair of classes 14 20070226 | Janne Argillander, [email protected] T-61.6050 Special Course in Computer and Information Science Multiclass = problems? One-versus-rest One-versus-one HINT: Generally multiclass classification is an open issue. Solve it and become famous! 15 20070226 | Janne Argillander, [email protected] T-61.6050 Special Course in Computer and Information Science Contents ° Introduction ° Maximum margin classifier ° Overlapping class distributions ° Multiclass classifier ° Relevance Vector Machines ° Outroduction 16 20070226 | Janne Argillander, [email protected] T-61.6050 Special Course in Computer and Information Science On the way to even more sparse solutions ° Some limitations of the SVM – Outputs of an SVM represent decisions – Slack variables are found by using a hold-out method – Positive definite kernels – SVM is just not cool anymore? ° Relevance Vector Machine (RVM) has derived from the SVM – Shares many (good) characteristics of the SVM – No restriction to positive definite kernels – No cross-validation needed – Number of relevance vectors << Number of support vectors ‰ faster, sparser – When training the RVM model we have to calculate inverse of M x M matrix (where M is amount of basis functions) ‰ O(M3) ‰ RVM training takes more time 17 20070226 | Janne Argillander, [email protected] T-61.6050 Special Course in Computer and Information Science v-SVM RVM 18 20070226 | Janne Argillander, [email protected] T-61.6050 Special Course in Computer and Information Science Function approximation (RVM) sinc(x) sinc(x) + gaussian noise Tipping, M. E. (2000). The Relevance Vector Machine 19 20070226 | Janne Argillander, [email protected] T-61.6050 Special Course in Computer and Information Science Contents ° Introduction ° Maximum margin classifier ° Overlapping class distributions ° Multiclass classifier ° Relevance Vector Machines ° Outroduction 20 20070226 | Janne Argillander, [email protected] T-61.6050 Special Course in Computer and Information Science Software packages and good to check ° Command line tool SVMLIGHT http://svmlight.joachims.org/ SVMTorch Multi III http://www.torch.ch/ – one-versus-rest approach ° Do it yourself LIBSVM http://www.csie.ntu.edu.tw/~cjlin/libsvm/ ° Interfaces to LIBSVM – Matlab, R, Weka. Labview,... ° http://www.kernel-machines.org/ ° Google 21 20070226 | Janne Argillander, [email protected].

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    21 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us