Recommendations As Treatments: Debiasing Learning and Evaluation

Recommendations As Treatments: Debiasing Learning and Evaluation

Recommendations as Treatments: Debiasing Learning and Evaluation Tobias Schnabel, Adith Swaminathan, Ashudeep Singh, Navin Chandak, Thorsten Joachims Cornell University, Ithaca, NY, USA TBS49, FA234, AS3354, NC475, TJ36 @CORNELL.EDU f g Abstract mate the effect of new interventions (e.g. a new treatment policy or a new set of recommendations) despite incom- Most data for evaluating and training recom- plete and biased data due to self-selection or experimenter- mender systems is subject to selection biases, bias. By connecting recommendation to causal inference either through self-selection by the users or from experimental and observational data, we derive a prin- through the actions of the recommendation sys- cipled framework for unbiased evaluation and learning of tem itself. In this paper, we provide a principled recommender systems under selection biases. approach to handle selection biases by adapting models and estimation techniques from causal in- The main contribution of this paper is four-fold. First, ference. The approach leads to unbiased perfor- we show how estimating the quality of a recommendation mance estimators despite biased data, and to a system can be approached with propensity-weighting tech- matrix factorization method that provides sub- niques commonly used in causal inference (Imbens & Ru- stantially improved prediction performance on bin, 2015), complete-cases analysis (Little & Rubin, 2002), real-world data. We theoretically and empirically and other problems (Cortes et al., 2008; Bickel et al., 2009; characterize the robustness of the approach, and Sugiyama & Kawanabe, 2012). In particular, we derive find that it is highly practical and scalable. unbiased estimators for a wide range of performance mea- sures (e.g. MSE, MAE, DCG). Second, with these estima- tors in hand, we propose an Empirical Risk Minimization 1. Introduction (ERM) framework for learning recommendation systems under selection bias, for which we derive generalization er- Virtually all data for training recommender systems is sub- ror bounds. Third, we use the ERM framework to derive a ject to selection biases. For example, in a movie rec- matrix factorization method that can account for selection ommendation system users typically watch and rate those bias while remaining conceptually simple and highly scal- movies that they like, and rarely rate movies that they able. Fourth, we explore methods to estimate propensities do not like (Pradel et al., 2012). Similarly, when an ad- in observational settings where selection bias is due to self- placement system recommends ads, it shows ads that it be- selection by the users, and we characterize the robustness lieves to be of interest to the user, but will less frequently of the framework against mis-specified propensities. display other ads. Having observations be conditioned on the effect we would like to optimize (e.g. the star rating, the Our conceptual and theoretical contributions are validated probability of a click, etc.) leads to data that is Missing Not in an extensive empirical evaluation. For the task of evalu- At Random (MNAR) (Little & Rubin, 2002). This creates ating recommender systems, we show that our performance a widely-recognized challenge for evaluating recommender estimators can be orders-of-magnitude more accurate than systems (Marlin & Zemel, 2009; Myttenaere et al., 2014). standard estimators commonly used in the past (Bell et al., 2007). For the task of learning recommender systems, we We develop an approach to evaluate and train recommender show that our new matrix factorization method substan- systems that remedies selection biases in a principled, prac- tially outperforms methods that ignore selection bias, as tical, and highly effective way. Viewing recommendation well as existing state-of-the-art methods that perform joint- from a causal inference perspective, we argue that exposing likelihood inference under MNAR data (Hernandez-Lobato´ a user to an item in a recommendation system is an inter- et al., 2014). This is especially promising given the concep- vention analogous to exposing a patient to a treatment in a tual simplicity and scalability of our approach compared to medical study. In both cases, the goal is to accurately esti- joint-likelihood inference. We provide an implemention of our method, as well as a new benchmark dataset, online1. Proceedings of the 33 rd International Conference on Machine Learning, New York, NY, USA, 2016. JMLR: W&CP volume 1https://www.cs.cornell.edu/∼schnabts/mnar/ 48. Copyright 2016 by the author(s). Recommendations as Treatments: Debiasing Learning and Evaluation 2. Related Work Horror Romance Drama Horror Romance Drama Horror Romance Drama 1 1 1 1 1 1 1 1 1 1 1 1 1 5 1 3 p p/10 p/2 1 1 1 1 1 1 Lovers 1 Horror 1 1 Past work that explicitly dealt with the MNAR nature of 1 1 1 1 1 1 1 1 1 1 recommendation data approached the problem as missing- 1 1 1 1 1 1 1 1 1 1 5 3 p/10 p p/2 1 1 1 1 1 Lovers 1 1 data imputation based on the joint likelihood of the miss- 1 1 1 Romance 1 ing data model and the rating model (Marlin et al., 2007; 1 1 1 ´ 1 5 1 5 2 5 5 3 3 1 1 1 Marlin & Zemel, 2009; Hernandez-Lobato et al., 2014). � � � 1 Lovers 1 1 Horror 1 1 1 This has led to sophisticated and highly complex meth- 1 1 1 1 1 1 1 5 5 5 5 3 1 1 1 ods. We take a fundamentally different approach that 1 1 1 Lovers 1 1 treats both models separately, making our approach mod- Romance 1 ular and scalable. Furthermore, our approach is robust to Figure 1. Movie-Lovers toy example. Top row: true rating matrix Y , propensity matrix P , observation indicator matrix O. Bottom mis-specification of the rating model, and we characterize row: two rating prediction matrices Y^ and Y^ , and intervention how the overall learning process degrades gracefully un- 1 2 indicator matrix Y^3. der a mis-specified missing-data model. We empirically compare against the state-of-the-art joint likelihood model (Hernandez-Lobato´ et al., 2014) in this paper. set of users are “horror lovers” who rate all horror movies 5 and all romance movies 1. Similarly, there is a subset of Related but different from the problem we consider is “romance lovers” who rate just the opposite way. How- recommendation from positive feedback alone (Hu et al., ever, both groups rate dramas as 3. The binary matrix 2008; Liang et al., 2016). Related to this setting are also O 0; 1 U×I in Figure1 shows for which movies the alternative approaches to learning with MNAR data (Steck, 2 f g users provided their rating to the system, [Ou;i = 1] 2010; 2011; Lim et al., 2015), which aim to avoid the prob- , [Yu;i observed]. Our toy example shows a strong correla- lem by considering performance measures less affected by tion between liking and rating a movie, and the matrix P selection bias under mild assumptions. Of these works, the describes the marginal probabilities Pu;i = P (Ou;i = 1) approach of Steck(2011) is most closely related to ours, with which each rating is revealed. For this data, consider since it defines a recall estimator that uses item popular- the following two evaluation tasks. ity as a proxy for propensity. Similar to our work, Steck (2010; 2011) and Hu et al.(2008) also derive weighted ma- 3.1. Task 1: Estimating Rating Prediction Accuracy trix factorization methods, but with weighting schemes that are either heuristic or need to be tuned via cross validation. For the first task, we want to evaluate how well a predicted In contrast, our weighted matrix factorization method en- rating matrix Y^ reflects the true ratings in Y . Standard eval- joys rigorous learning guarantees in an ERM framework. uation measures like Mean Absolute Error (MAE) or Mean Squared Error (MSE) can be written as: Propensity-based approaches have been widely used in causal inference from observational studies (Imbens & Ru- U I ^ 1 X X ^ bin, 2015), as well as in complete-case analysis for missing R(Y ) = δu;i(Y; Y ) ; (1) U I data (Little & Rubin, 2002; Seaman & White, 2013) and in · u=1 i=1 survey sampling (Thompson, 2012). However, their use in for an appropriately chosen δ (Y; Y^ ). matrix completion is new to our knowledge. Weighting ap- u;i ^ ^ proaches are also widely used in domain adaptation and co- MAE: δu;i(Y; Y ) = Yu;i Yu;i ; (2) j − j 2 variate shift, where data from one source is used to train for MSE: δu;i(Y; Y^ ) = (Yu;i Y^u;i) ; (3) a different problem (e.g., Huang et al., 2006; Bickel et al., − Accuracy: δu;i(Y; Y^ ) = 1 Y^u;i =Yu;i : (4) 2009; Sugiyama & Kawanabe, 2012). We will draw upon f g this work, especially the learning theory of weighting ap- Since Y is only partially known, the conventional practice proaches in (Cortes et al., 2008; 2010). is to estimate R(Y^ ) using the average over only the ob- served entries, 3. Unbiased Performance Estimation for 1 X R^naive(Y^ ) = δu;i(Y; Y^ ) : (5) Recommendation (u; i):Ou;i =1 jf gj (u;i):Ou;i=1 Consider a toy example adapted from Steck(2010) to il- We call this the naive estimator, and its naivety leads to lustrate the disastrous effect that selection bias can have a gross misjudgment for the Y^ and Y^ given in Figure1. on conventional evaluation using a test set of held-out rat- 1 2 Even though Y^ is clearly better than Y^ by any reasonable ings.

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    10 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us