JournalofMachineLearningResearch6(2005)1019–1041 Submitted 11/04; Revised 3/05; Published 7/05 Gaussian Processes for Ordinal Regression Wei Chu
[email protected] Zoubin Ghahramani
[email protected] Gatsby Computational Neuroscience Unit University College London London, WC1N 3AR, UK Editor: Christopher K. I. Williams Abstract We present a probabilistic kernel approach to ordinal regression based on Gaussian processes. A threshold model that generalizes the probit function is used as the likelihood function for ordinal variables. Two inference techniques, based on the Laplace approximation and the expectation prop- agation algorithm respectively, are derived for hyperparameter learning and model selection. We compare these two Gaussian process approaches with a previous ordinal regression method based on support vector machines on some benchmark and real-world data sets, including applications of ordinal regression to collaborative filtering and gene expression analysis. Experimental results on these data sets verify the usefulness of our approach. Keywords: Gaussian processes, ordinal regression, approximate Bayesian inference, collaborative filtering, gene expression analysis, feature selection 1. Introduction Practical applications of supervised learning frequently involve situations exhibiting an order among the different categories, e.g. a teacher always rates his/her students by giving grades on their overall performance. In contrast to metric regression problems, the grades are usually discrete and finite. These grades are also different from the class labels in classification problems due to the existence of ranking information. For example, grade labels have the ordering F < D < C < B < A. This is a learning task of predicting variables of ordinal scale, a setting bridging between metric regression and classification referred to as ranking learning or ordinal regression.