<<

Loss Functions for Preference Levels: Regression with Discrete Ordered Labels Jason D. M. Rennie Nathan Srebro Massachusetts Institute of Technology University of Toronto Comp. Sci. and Artificial Intelligence Laboratory Department of Computer Science Cambridge, MA 02139, USA Toronto, Ontario M5S 3G4, Canada [email protected] [email protected]

Abstract ization of (as in, e.g., ), which can be seen as a degenerate case in which only two We consider different types of loss functions for levels, “positive” and “negative”, are available. As with bi- discrete , i.e. fitting labels that nary regression, we learn a real-valued predictor z(x) (e.g. may take one of several discrete, but ordered, val- in linear binary regression, we would learn a linear function ues. These types of labels arise when preferences of the features), minimizing some loss loss(z(x); y) on the are specified by selecting, for each item, one of sev- target labels. Common choices for the are the eral rating “levels”, e.g. one through five stars. We logistic loss (as in logistic regression), and the hinge loss (dis- present two general threshold-based constructions tance from the classification margin) used in Support Vector which can be used to generalize loss functions for Machines. Here, we consider various generalizations to these binary labels, such as the logistic and hinge loss, loss functions suitable for multiple-level discrete ordinal la- and another generalization of the logistic loss based bels. on a probabilistic model for discrete ordered labels. on the 1 Million MovieLens set Threshold-based approaches indicate that one of our construction is a signifi- Crammer and Singer [2002] suggest a generalization of the cant improvement over previous classification- and algorithm for discrete ordinal labels: instead of regression-based approaches. the single threshold of the perceptron, they use K − 1 thresh- olds to separate the real line to regions corresponding to K 1 Introduction possible rating levels. Shashua and Levin [2003] suggest a similar generalization to Support Vector Machine (SVM): the In many systems, users specify preferences by selecting, for single margin constraints (for each observation) of standard each item, one of several rating “levels”, e.g. one though five SVMs are replaced with a pair of margin constraints on the “stars”. When learning to predict further preferences, these thresholds bounding the “correct” region (the region corre- rating levels serve as target labels (responses). This type of sponding to the target label). discrete ordered labels differs from more standard types of When slack is allowed, Shashua and Levin’s approach can target labels encountered in other problems: be seen as regularized regression with a specific generaliza- binary labels (as in classification tasks), discrete, unordered tion to the hinge loss, which we describe in Section 3.1 as labels (as in multi-class classification tasks) and continuous the immediate-threshold generalization of the hinge loss. In real-valued labels (as in typical regression tasks). Rating lev- Section 3.2 we discuss a different generalization, the all- els are discrete with a finite number of possibilities, like class threshold generalization, where constraints (and slack) are labels in multiclass classification. However, unlike a standard considered for all K − 1 thresholds and not only those im- multiclass classification setting, the labels are ordered—a rat- mediately bounding the correct region. We argue that such a ing of “three stars” is between a rating of “two stars” and a generalization better penalizes which violate mul- rating of “four stars”. tiple thresholds and present experimental evidence suggesting Two obvious approaches for handling discrete ordinal la- it might be more appropriate. We also discuss how other loss bels are (1) treating the different rating levels as unrelated functions, such as the logistic loss or smooth variants of the classes and learning to predict them as in a multiclass classifi- hinge loss, can also be generalized in the same way. cation setting, and (2) treating them as a real-valued responses and using a standard regression setting with a loss function Probabilistic approaches such as sum-squared error. However, neither of these reflects Other than these threshold-based generalizations, we also the specific structure of discrete ordinal labels. suggest a different generalization to logistic regression, which we term “ordistic regression” (Section 4), that, like logistic 1.1 Ordinal regression regression, can be viewed as fitting a conditional probabil- In this paper we view fitting rating levels as a regression prob- ity model P (y|x). We note that Chu and Ghahramani [2004] lem with discrete ordered labels. We view this as a general- recently suggested a different generalization to the logistic conditional model for discrete ordinal labels. In this paper, we focus on L2-regularized linear , 0 Probabilistic models for discrete ordinal response have also where z(x) = w x+w0 is a linear (or more accurately, affine) been studied in the literature [McCullagh, 1980; function of x ∈ Rd, parametrized by a weight vector w ∈ Rd Fu and Simpson, 2002]. However, the models suggested are and a bias term w0 ∈ . We seek a linear predictor that much more complex, and even just evaluating the likelihood minimizes a trade-off between the overall training loss and of a predictor is not straight-forward. On the other hand, in the (Euclidean) of the weights: the ordistic model both the log-likelihood and its derivatives X λ can be easily computed, using calculations similar to those J(w) = loss(w0xt + w ; yt) + |w|2 (1) 0 2 used in standard (binary) logistic regression. t where λ is a trade-off parameter set using cross-validation. 1.2 Other approaches We briefly mention another approach suggested for handling 2.1 Binary Regression discrete ordinal ratings. Herbrich et al. [2000] suggest ex- We first review common loss functions used with binary la- tracting from the rating levels binary comparison relation- bels (i.e. in a binary classification setting), where y ∈ ±1. ships on the rated items and thus mapping the problem to a These serve as a basis for our more general loss functions partial ordering problem. Herbrich et al. then study a gen- for discrete ordinal labels. We go into some detail regarding eralized SVM for learning from binary comparison relation- aspects of these loss functions which will be relevant in our ships. A drawback of this approach is the number of order constructions in the following sections. constraints on T items with observed labels can be of order T 2, even though the original input to the problem (the ob- Zero-one error served labels) is only linear in T . Our objective in binary regression is to be able to correctly predict a binary label. The obvious way of predicting a bi- 0 1.3 Specific contribution nary label from a real-valued predictor z(x) = w x + w0 is thresholding the predictor, and predicting sign(z(x)). The The main contribution of this paper is studying, in a system- simplest conceivable loss function is a loss function counting atic way, different loss functions for discrete ordinal regres- the number of prediction errors: sion. Since our main interest is in how to handle discrete ordinal labels, we focus on regularized linear prediction in 0 if yz > 0 loss(z; y) = (2) a simple learning setting, which we clarify in Section 2. In 1 if yz ≤ 0 Section 2.1 we review various loss functions for binary la- bels and discuss their properties. In Section 3 we present However, this simple loss function is problematic for several the immediate-threshold and all-threshold constructions men- reasons: tioned above, using the loss functions from the previous sec- • It is not convex, and minimizing it is a difficult (in fact, tions as building blocks. In Section 4 we present the ordistic NP-hard) optimization problem. model which generalizes the logistic. In Section 5 we com- • It is not continuous, let alone differentiable, and so even pare the various methods through experiments using the dif- local optimization is difficult. ferent loss functions, and compare them also to standard mul- ticlass and sum-squared-error regression approaches. • It is insensitive to the magnitude of z, and so also to the We have already used the immediate-threshold and all- magnitude of w. Regularizing w is therefore meaning- threshold generalizations of the hinge-loss in our work on less, as shrinking w and w0 towards zero would yield collaborative prediction using Maximum Margin Matrix Fac- the same error, but with a regularization term approach- torizations [Srebro et al., 2005]. Here, we present these con- ing zero. structions in detail and more generally, as well as the ordistic Margin model. The third problem can be addressed by requiring not only that z predict y correctly, but that it does so with a margin: 2 Preliminaries 0 if yz ≥ 1 loss(z; y) = (3) Since our main object of interest is how to handle discrete or- 1 if yz < 1 dinal labels, we focus on a simple learning setting in which we can demonstrate and with various loss func- This modified loss function is sensitive to the magnitude of z, t t tions. We are given a training set (x , y )t=1...T of T rated and therefore also to the magnitude of w. Summing this loss items, where for each item, xt ∈ Rd is a feature vector de- function corresponds to counting the number of violations of t 0 scribing the item and y is the rating level for the item. We the constraints y(w x + w0) ≥ 1. Rewriting these constraints 0 want to predict preferences of future items. We do so by as y( w x + w0 ) ≥ 1 , we can interpret 1 as a geometrical d |w| |w| |w| |w| learning a prediction mapping z(x): R → R such that margin around the separating hyperplane, specified by its nor- for an item with feature vector x, z(x) corresponds as well 0 mal w . Minimizing the loss (3) as well as the L regularizer as possible to the appeal of the item (i.e. is high if the item |w| 2 is highly rated and low otherwise). We investigate different |w| can therefore be interpreted as maximizing the separation 1 loss functions loss(z; y) for measuring the goodness of the margin M = |w| while minimizing the number of training correspondence between z(xt) and the target rating level yt. points not classified correctly with a margin of at least M. 3 3 3 3 3 3

2 2 2 2 2 2

1 1 1 1 1 1

0 0 0 0 0 0 −1 0 1 2 −1 0 1 2 −1 0 1 2 −1 0 1 2 −1 0 1 2 −1 0 1 2

Figure 1: Different margin penalty functions f(yz) (left to right): (1) sign agreement, (2) margin agreement, (3) hinge, (4) smooth hinge, (5) modified , (6) logistic.

Hinge loss but such a linear (at least) dependence is unavoidable in a Minimizing the margin loss (3) might be a good ideal, but convex loss function. The modified least squares goes be- this loss function is still non-convex and non-continuous. The yond this necessary dependence on the magnitude of the er- common approach to large-margin classification is therefore ror, and introduces an unnecessary (from the point of view of to minimize the hinge loss: convexity) quadratic dependence, further deviating from the zero/one margin error. lossh(z; y) = h(yz) (4) Logistic regression where h(z) is the hinge function: Another common loss function, which can also be written as 0 if z ≥ 1 a function of the classification margin yz, is the logistic loss: h(z) = max(0, 1 − z) = (5) 1 − z if z < 1 lossg(z; y) = g(yz) (8) −z This is the loss function typically minimized in soft-margin g(z) = log(1 + e ) (9) Support Vector Machine (SVM) classification. In the context The logistic loss can be viewed as a negative conditional of SVMs, the hinge loss is usually written as a sum over mar- log-likelihood lossg(z; y) = − log P (z|y) for a logistic con- t 0 yz gin violations ξ included in the constraints y(w x + w0) ≥ ditional model P (y|z) ∝ e (i.e. Y is a Bernoulli random 1 − ξt. variable, with natural parameter z). The predictor z(x) = 0 An important property of the hinge-loss is that it is an upper w x + w0 minimizing the summed logistic loss is thus the bound on the zero-one misclassification error (2), and thus maximum conditional likelihood among the para- 0 large-margin bounds bounding its value metric class of conditional models P (y|x) ∝ ey(w x+w0). In- on examples not in the training set also bounds the value of troducing an L2 regularizer as in (1) corresponds to maxi- the zero-one misclassification error, which is perhaps the true mum a-posteriori (MAP) estimation with a Gaussian prior on object of interest. the weight vector w. Smoothed hinge loss As discussed above, logistic regression corresponds to Other loss functions share properties of the hinge, but are eas- maximum conditional likelihood estimation for a conditional parametric model P (y|x). It is worth noting that this para- ier to minimize since they have a smooth derivative. We in- 0 troduce “smooth” hinge loss as an approximation to the hinge metric family of conditional models P (y|x) ∝ ey(w x+w0) that is easier to minimize: is exactly the family of conditional distributions P (y|x) for joint distributions P (y, x) where X|Y follows a multivariate 0 if z ≥ 1  spherical Gaussian distribution with which does not h(z) = (1 − z)2/2 if 0 < z < 1 (6) depend on Y , and a mean which does depend on Y , i.e.:  0.5 − z if z ≤ 0 1 |x−µ |2 P (x|y) ∝ e 2σ2 y (10) Modified least squares where µ , µ ∈ d and σ ∈ , as well as the prior proba- [ ] −1 1 R R Zhang and Oles 2001 suggest a different loss function with bility P (Y = 1) are the parameters of the joint distribution a smooth derivative, in which the hinge function is replaced model. with a truncated quadratic: For our purposes below, it would also be useful to view the yz 0 if z ≥ 1 conditional model P (y|z) ∝ e similarly, as the conditional h(z) = 2 (7) distribution arising from a joint distribution P (y, z) in which (1 − z) if z < 1 1 P (y = 1) = 2 and Z|Y ∼ N(1,Y ), i.e.: The modified least squares loss based on (7) is much more 1 |z−y|2 sensitive to and large errors then the hinge loss using P (z|y) ∝ e 2 . (11) (5) or smoothed hinge loss using (6). The margin error (3), which we might want to view as a Loss as a function of classification penalties We note that non-convex “ideal”, does not pay any attention to the magni- all loss functions discussed here can be viewed as penalties tude of the error, and penalizes all errors equally. This allows loss(y; z) = f(yz) imposed on the classification margins yz, for a few outliers fairly cheaply, but leads to a non-convex and differ only in the margin penalty function f(·) used. The objective. The hinge loss, as well as the smoothed hinge, in- various margin penalty functions discussed are shown in Fig- troduce a linear dependence on the magnitude of the error, ure 1. segment, namely z < 0, corresponding to negative labels y = −1, and a semi-infinite segment, z > 0 corresponding to positive labels. In fact, the bias term in the binary setting can be viewed as specifying a threshold. The K − 1 thresholds replace this single bias term / threshold. We describe two different constructions for loss functions

y=1 y=2 y=3 y=4 y=5 y=1 y=2 y=3 y=4 y=5 based on such thresholds. The constructions differ in how predictors outside the segment segment corresponding to the Figure 2: Threshold-based loss constructions: (left) “correct” label (or too close to its edges) are penalized. Both immediate-threshold, and (right) all-threshold. Both plots constructions are based on combining penalties for threshold show a generalization of the hinge loss for an example with violation, where each threshold violation is penalized using label y = 4. some margin penalty function f(·). 3.1 Immediate-Threshold 2.2 Discrete Ordinal Regression For the immediate-threshold construction, we consider, for So far, we discussed loss functions for binary labels y = ±1. each labeled example (x, y), only the two thresholds defining However, the main topic of this paper is loss functions for the “correct” segment (θy−1, θy), and penalize violations of discrete ordinal labels. In this scenario, the labels y can take these thresholds: on K distinct values which we denote {1, 2,...,K}. In the next section, we present different constructions for loss(z; y) = f(z − θy−1) + f(θy − z), (12) generalizing margin-based loss function for binary labels, of the form loss(z; y) = f(yz), to discrete ordinal labels. We where z = z(x) is the predictor output for the example. Fig- do so by introducing K − 1 thresholds, instead of the single ure 2 gives an example visualization for y = 4. Note that if f threshold (zero) in the binary case. is an upper bound on the binary classification zero-one error, We also introduce a generalization of the logistic loss in then the immediate-threshold loss is an upper bound on the which the joint probabilistic model P (y, z) is generalized to ordinal regression zero-one error. The immediate-threshold a mixture of K unit-variance Gaussians (instead of a mixture loss is ignorant of whether multiple thresholds are crossed. of two unit-variance Gaussians as in the binary case). 3.2 All-Threshold 2.3 Beyond Feature-Based Regression In a multi-class classification setting, all mistakes are equal; Although the main setting we focus on is a feature-based lin- there is no reason to prefer one incorrect label over another. ear regression setting, the loss functions we describe are ap- This is not true in ordinal regression where there is a distinct plicable in more general settings where a real-valued predic- ordering to the labels. It is better to predict ‘4’ than ‘1’ if the tor needs to be related to a discrete ordinal label. In fact, our true label is ‘5’. This is evident in the evaluation criterion. We study originated in matrix-completion approaches to collabo- use absolute error for ordinal regression, which counts rative prediction (completing unobserved entries in a partially the sum of distances between the true and predicted labels. observed matrix of user preferences), where the “predictors” Immediate-threshold bounds zero-one error, but not (neces- are entries in a proposed matrix [Srebro et al., 2005]. sarily) mean absolute error. We focus on using explicit features, but We introduce a construction that bounds mean absolute er- we note that all methods discussed here can also be ”- ror. The all-threshold loss is a sum of all threshold violation ized”, as in Support Vector Machines. Both the immediate- penalties. If the binary loss function bounds zero-one error, threshold and all-threshold constructions with a hinge loss then the all-threshold loss bounds mean absolute error. De- −1 if l < y can also be seen as generalizations of SVMs and can be stated fine s(l; y) = . Then the all-threshold loss is as quadratic programs to which optimization techniques typi- +1 if l ≥ y cally employed in learning SVMs apply (in fact, Shashua and T −1 Levin [2003] introduce the immediate-threshold construction X   in the context of SVMs). loss(z; y) = f s(l; y)(θl − z) . (13) l=1 3 Threshold-Based Constructions where f(·) is some margin penalty function. Figure 2 gives To extend binary loss to the case of discrete ordinal regres- an example visualization for y = 4. Note that the slope of the loss increases each time a threshold is crossed. Thus, solu- sion, we introduce K − 1 thresholds θ1 < θ2 < ··· < θK−1 partitioning the real line to K segments. The exterior seg- tions are encouraged that minimize the number of thresholds that are crossed. ments are semi-infinite, and for convenience we denote θ0 = −∞ and θ = +∞. Each of the K segment corresponds to K 3.3 Learning Thresholds one of the K labels and a predictor value of θy−1 < z < θy (i.e. in the yth segment) corresponds to a rating of y. This Fitting an ordinal regression models involves fitting the pa- generalizes the binary case, where we used a single thresh- rameters of the predictor, e.g. z(x) = w0x, as well as the old, namely zero, separating the real line into a semi-infinite thresholds θ1, . . . , θK−1. Learning the thresholds, rather than 0 fixing them to be equally spaced, allows us to capture the dif- P (y|x) ∝ zy(x) = wyx + wy0. This conditional model cor- ferent ways in which users use the available ratings, and alle- responds to a joint distribution P (y, x) where X|Y follows viates the need for per-user rating normalization. In a setting a unit-variance spherical Gaussian distribution with mean o d in which multiple users are considered concurrently, e.g. in µY ∈ R . This model differs from the ordistic model in that collaborative prediction, a different set of thresholds can be the of the K Gaussians are allowed to fall in arbitrary learned for each user. positions in Rd. On the other hand, in the ordistic model, we model Z, rather then X, as a Gaussian mixture. An alternate 4 A Probabilistic Generalization view of the ordistic model would be to view X as a Gaussian mixture, but in this case, all means would be constrained to Recall that the logistic loss for binary labels can be viewed be co-linear, since the same weight vector w is used for all as the negative conditional log-likelihood lossl(y; z) = yz labels. This constraint captures the core difference between a − log P (y|z) for a conditional model P (y|z) ∝ e standard softmax model and the ordistic model: the ordistic corresponding to a mixture-of-Gaussians joint distribution model constrains the different labels to correspond to differ- P (y, z), as described in Section 2.1. Here, we generalize the ent extents along the same direction (hence collinear means), logistic by generalizing this probabilistic model to a mixture rather then arbitrary variations in different directions. of K Gaussians, resulting in a similar simple form for the conditional model P (y|z). We refer to this model, and the 4.2 Derivatives resulting loss, as the “ordistic” model. As with the logistic model, the derivatives of the ordistic loss Unlike the threshold-based constructions which are have a simple form. In order to simplify the derivatives, it will parametrized by K − 1 thresholds, the ordistic model is be useful to refer to expectation and probabilities with respect parametrized by the K means µ1 < µ2 < ··· < µK , to the joint probabilistic (Gaussian mixture) model for (Z,Y ) and possibly also the K prior probabilities p1, . . . , pK > 0, P discussed above. The gradient with respect to z, from which pi = 1. Considering a joint probability model in which the gradient with respect to the weights can be calculated, is: P (y) = py and Z|Y ∼ N(µY , 1), the conditional distribu- P (y|z) P 2  tion is: ∂losso(z; y) i µi exp µiz + (πi − µi /2) = −µy + 2 P 2 −(z−µy ) /2 ∂z i exp (µiz + (πi − µi /2)) pye P (y|z) = 2 X P −(z−µi) /2 = −µy + µiPr (Y = i|Z = z; µ, π) i pie exp µ z + (π − µ2/2) i y y y = −µ + E [µ |Z = z] (15) = P 2 (14) y µ,π Y i exp (µiz + (πi − µi /2)) Similarly, the derivative with respect to the log-priors πi where πi = log pi. If we simplify the model by fixing µ 1 and the means i can be calculated as: pi = K , these terms drop from the conditional distribution. The ordistic loss, with parameters (µi, πi) is obtained by con- ∂losso(z; y) = Pr (Y = i|Z = z; µ, π) − δy,i (16) sidering the negative log-likelihood: ∂πi

µ z+(π −µ2 /2) ∂losso(z; y) e( y y y ) = (Pr (Y = i|Z = z; µ, π) − δy,i)(z − µi) ∂µ losso(y; z) = − log P (y|z) = − log 2 i P (µiz+(πi−µi /2)) i e where δy,i is one if y = i and zero otherwise. If all means µ1, . . . , µK are allowed to vary freely, regular- izing the weight vector w is meaningless, since rescaling it 5 Experiments can be corrected by rescaling the means. In order to impose To determine the appropriateness of the different construc- a fixed scale, we fix the extreme means to µ = −1 and 1 tions discussed earlier, we conducted experiments on a well- µ = 1. Unlike threshold-based models, a bias term w is K 0 known collaborative filtering data set. We implemented the meaningful since the location of the extreme means is fixed. two threshold-based constructions discussed in Section 3. We The overall number of parameters for an ordistic model with also implemented multi-class classification and sum-squared fixed p = 1 is therefore the same as the number of parame- i K error regression constructions to compare against. ters in the threshold constructions (K − 1 parameters for the We used the “1 Million” MovieLens data set for evaluation. K − 2 means and the bias term, or K − 1 parameters for the The data set contains 1,000,209 rating entries, made by 6040 thresholds, in addition to the weight vector). Allowing priors users on 3952 movies. Similar to the work of Crammer and introduces K − 1 additional parameters. Singer [2002] and Shashua and Levin [2003], we used the rat- The ordistic degenerates to the logistic when K = 2. ings of the top 200 users as “features” to predict the ratings of 4.1 Difference from soft-max multi-class the remaining users. To deal with “missing” ratings, we sub- tracted the user’s mean rating and filled-in empty values with classification zero. We used the remaining 5840 users’ ratings as labels A common generalization of logistic regression to multi-class for ordinal regression. For each user, we used one randomly 0 classification is to learn a predictor zi(x) = wix + wi0 for selected rating for testing, another for validation and the re- each class and fit a soft-max conditional model in which maining ratings for training. We limited our experiments to Multi-class Imm-Thresh All-Thresh Test MAE Test MAE Test MAE Mod. Least Squares 0.7486 0.7491 0.6700 (1.74e-18) Smooth Hinge 0.7433 0.7628 0.6702 (6.63e-17) Logistic 0.7490 0.7248 0.6623 (7.29e-22) Multi-class Imm-Thresh All-Thresh Test ZOE Test ZOE Test ZOE Mod. Least Squares 0.5606 0.5807 0.5509 (7.68e-02) Smooth Hinge 0.5594 0.5839 0.5512 (1.37e-01) Logistic 0.5592 0.5699 0.5466 (2.97e-02)

Table 1: Mean absolute error (MAE) and zero-one error (ZOE) results on MovieLens. For each construction/loss and error type, we selected the regularization parameter with lowest validation error. Numbers in parentheses are p-values for all-threshold versus the next best construction. As a baseline comparison, simple sum-squared-error (L2) regression achieved test MAE of 1.3368 and test ZOE of 0.7635. the top 2000 movies to ensure a minimum of 10 ratings per Acknowledgements movie. This gave us test and validation sets of size 5,840 and Jason Rennie was supported in part by the DARPA CALO a training set of 769,659 ratings. project. For each method (combination of construction method and margin penalty), and of values of the regularization pa- rameter λ, we fit weight and threshold vectors for each user References by minimizing the convex objective (1) using conjugate gra- [Chu and Ghahramani, 2004] W. Chu and Z. Ghahramani. dient descent. We calculated mean absolute error (MAE) and Gaussian processes for ordinal regression. Technical re- zero-one error (ZOE) between predicted and actual ratings on port, University College London, 2004. the validation set and used the regularization parameter with [Crammer and Singer, 2002] Koby Crammer and Yoram the lowest validation set MAE/ZOE for test set evaluation. Singer. PRanking with . In Advances in Neural Table 1 shows test MAE and ZOE for various construc- Processing Systems 14, 2002. tions and margin penalty functions. Across all penalty func- tions, all-threshold yields the lowest MAE. The differences [Fu and Simpson, 2002] Limin Fu and Douglas G. Simpson. are highly significant according to a nonparametric, two- Conditional models for ordinal response data: simul- tailed binomial test—the largest p-value is 6.63e-17. Inter- taneous logistic regression and generalized score estingly, all-threshold also yields lower ZOE, although the tests. Journal of Statistical Planning and Inference, pages comparison with multi-class classification is not conclusive 201–217, Nov 2002. (p-values around 0.03–0.1). The difference in ZOE compared [Herbrich et al., 2000] Ralf Herbrich, Thore Graepel, and to immediate-threshold is highly significant, with p-values at Klaus Obermayer. Large margin rank boundaries for or- most 7.84e-06 (not shown). dinal regression. In Advances in Large Margin Classifiers, Results indicate that the choice of construction is more im- pages 115–132. MIT Press, 2000. portant than penalty function—all-threshold with the worst- [McCullagh, 1980] Peter McCullagh. Regression models for performing penalty function yields lower MAE and ZOE than . Journal of the Royal Statistical Society. Se- the best non-all-threshold combination. However, it appears ries B (Methodological), 42(2):109–142, 1980. that the logistic loss tends to work best; in particular, the differences in MAE between logistic and other penalty func- [Shashua and Levin, 2003] Amnon Shashua and Anat Levin. tions (for the all-threshold construction) are significant at the Ranking with large margin principle: Two approaches. In p = 0.01 level (largest p-value is 9.52e-03) according to the Advances in Neural Information Processing Systems 15, two-tailed binomial test. 2003. [Srebro et al., 2005] Nathan Srebro, Jason D. M. Rennie, 6 Conclusion and Tommi S. Jaakkola. Maximum-margin matrix factor- We presented a study of discrete ordinal regression in a gen- ization. In Lawrence K. Saul, Yair Weiss, and Leon´ Bot- eral, loss function based framework. We began with bi- tou, editors, Advances in Neural Information Processing nary classification and described two threshold-based con- Systems 17. MIT Press, Cambridge, MA, 2005. structions as generalizations of binary loss functions. We [Zhang and Oles, 2001] Tong Zhang and Frank J. Oles. Text also discussed an alternate generalization of logistic regres- categorization based on regularized linear classification sion. We conducted experiments on MovieLens using the two methods. Information Retrieval, 4:5–31, 2001. threshold-based constructions and found that the all-threshold construction outperformed multi-class classification and sim- ple regression methods, as well as the immediate-threshold construction.