Using Implicit Feedback for Task Recommendation In
Total Page:16
File Type:pdf, Size:1020Kb
Signals in the Silence: Models of Implicit Feedback in a Recommendation System for Crowdsourcing Christopher H. Lin∗ Ece Kamar and Eric Horvitz Department of Computer Science and Engineering Microsoft Research University of Washington Redmond, WA 98052 Seattle, WA 98195 feckamar,[email protected] [email protected] Abstract al. 1992). A collaborative filtering methodology can handle the inherent diversity of crowdsourcing tasks; the creativity We exploit the absence of signals as informative observations in the context of providing task recommendations in crowd- encouraged by crowdsourcing makes it difficult to create ex- sourcing. Workers on crowdsourcing platforms do not pro- plicit profiles using a fixed set of features for tasks and users. vide explicit ratings about tasks. We present methods that en- Observing the behaviors of users in the crowdsourcing able a system to leverage implicit signals about task prefer- marketplace provides evidence about users’ preferences in ences. These signals include types of tasks that have been the form of implicit feedback. Previous work in collabora- available and have been displayed, and the number of tasks tive filtering uses positive implicit feedback in the context of workers select and complete. In contrast to previous work, a TV recommendation system, by representing the degree to we present a general model that can represent both positive which a viewer likes a certain TV show with the number of and negative implicit signals. We introduce algorithms that times that viewer watches the show (Hu, Koren, and Volin- can learn these models without exceeding the computational complexity of existing approaches. Finally, using data from a sky 2008). While we demonstrate that this approach can be high-throughput crowdsourcing platform, we show that rea- used to generate crowdsourcing recommendations, we show soning about both positive and negative implicit feedback can that it has a significant shortcoming: the prior approach does improve the quality of task recommendations. not address negative implicit feedback. For example, it does not model when viewers may dislike certain shows, leading to a system that only learns the shows that users like and not Introduction the shows they dislike. Although the growth of crowdsourcing has facilitated an un- With crowdsourcing, worker behavior in the marketplace precedented explosion of open-call work, such growth has provides both positive and negative implicit feedback that also brought to the fore new inefficiencies and opportuni- can be used to learn more comprehensive models of work- ties for optimization. For example, the large number of tasks ers’ preferences. Workers who complete large quantities of presented in an unorganized way in marketplaces like Ama- a task provide a signal that they have a preference for that zon’s Mechanical Turk, where over 1,300 types of tasks can task, whereas workers who do not work on tasks that have 1 be available at a given time , can make it challenging for high availability on the system provide a signal that they crowdworkers to identify tasks that they enjoy or that they may not desire those tasks. We present a general-purpose have special competency with solving. Such problems with model that can represent both positive and negative implicit task identification can reduce overall throughput, accuracy, feedback. Our model is expensive to learn with existing ap- and engagement in a crowdsourcing system. proaches. To address this computational challenge, we pro- We present methods that help crowdworkers discover pose a modified coordinate-descent algorithm for tractable tasks that they have not yet worked on, but would be in- learning. We show that the complexity of this approach does terested in performing. A key challenge in crowdsourcing is not exceed the complexity of learning simpler models that that users do not typically provide explicit feedback about consider only positive implicit feedback. In addition, we their preferences by asserting which tasks they like and dis- propose an alternative sampling-based approach that can like through ratings. To help workers find desirable tasks, generate recommendations efficiently by combining multi- we explore methods for leveraging the absence of explicit ple models learned independently via sampling subsets of signals in task recommendation for crowdsourcing. We take the data. inaction in the face of displayed information on available We make the following contributions: tasks as evidence of preferences. Rather than take a content- based approach, we build a collaborative filter (Goldberg et • We develop two new methods for incorporating negative implicit feedback into a predictive modeling system in a ∗Research was performed while the author was an intern at Mi- computationally efficient way. crosoft Research. Copyright c 2014, Association for the Advancement of Artificial • We present what we believe to be the first application of Intelligence (www.aaai.org). All rights reserved. recommendation systems to crowdsourcing. We demon- 1retrieved February 2014 strate how implicit signals can be extracted from the logs of a crowdsourcing platform. Harnessing Implicit Positive Feedback for • We evaluate the proposed approaches on a dataset col- Task Recommendation in Crowdsourcing lected from a live crowdsourcing platform and show that We use the terms workers and users interchangeably. Users the addition of negative implicit feedback is beneficial. may have different backgrounds and capabilities and may have preferences about which types of tasks to work on. A Background on Collaborative Filtering user works on an abstract task by completing at least one Two common types of methods used in collaborative filter- of a collection of instances of that task posted by a single ing systems are neighborhood-based approaches and matrix requester. factorization approaches. Neighborhood-based approaches We define the problem of task recommendation in crowd- (Herlocker et al. 1999) estimate ratings based on either the sourcing as follows: Given a set of users and a set of tasks, ratings of like-minded users or the ratings for similar items we would like to compile, for each user, a personalized rec- (Linden, Smith, and York 2003; Sarwar et al. 2001). Re- ommendation list with tasks that the user has not worked on cent work has shown that matrix factorization approaches yet but that we predict the user would like to work on in have superior performance compared to neighborhood- the future. Since users cannot express explicitly which tasks based models (Koren, Bell, and Volinsky 2009), so we focus they like and which tasks they dislike, we use implicit sig- on matrix factorization models but we present a comparison nals about their behaviors as evidence about their interests. of these models in the experimental section. Specifically, we take the number of times a worker com- We first review matrix factorization approaches for ex- pletes a given task as a positive signal, such that more com- plicit feedback domains. Then, we move on to approaches pletions signal a stronger preference for the task. for harnessing implicit feedback. Let m be the number of IFMF can be applied to the problem of recommending users and let n be the number of items. Let R be an m × n crowdsourcing tasks as follows: Let cij, the implicit posi- matrix where most entries are missing and each non-missing tive signal, be the number of times user i worked on task j. Let the weighting function f be f(c ) = 1+c . Intuitively, rij 2 R is an explicit rating given by user i to item j. The ij ij goal is to learn an m × k matrix U and a k × n matrix V that for every task j that user i has worked on, pij is set to 1, de- minimize the loss function noting that user i likes task j; and for any task j that the user has not worked on, pij is 0, denoting that user i dislikes task X 2 2 2 L(U; V ) = (rij − uivj) + λ(jjUjj2 + jjV jj2) j. By weighting the positive entries of P with the number of i;j tasks a user has completed, we are able to express our con- fidence in each pij = 1, and when we learn the factors U where ui denotes row i of U, vj denotes column j of V , and V , we try harder to fit the entries we have higher confi- and λ denotes a regularization parameter. Then, R^ = UV dence in. For instance, if a user has worked on a given task contains the desired predictions. k can be picked by cross 10,000 times, then we are confident that the user likes that validation, and U and V can be estimated by a variety task (p = 1), and we would very much like to predict p^ = 1. of methods, including stochastic gradient descent (Funk 2006) and alternating least squares (Bell and Koren 2007; Incorporating Implicit Negative Feedback Zhou et al. 2008). Learning U and V exactly is a non-convex IFMF treats all tasks that a user has not worked on the same minimization problem (Srebro and Jaakkola 2003). by assigning the minimum weight to the corresponding en- When explicit feedback is not available, implicit feedback tries. Since this model cannot represent negative implicit can be used to make predictions about users. We general- feedback, the loss function has little emphasis on learn- ize the implicit feedback matrix factorization model (IFMF) ing about the tasks that users dislike. We introduce a novel proposed by Hu, Koren, and Volinsky to use any function model, IFMF2, to address this shortcoming of the IFMF ap- of some positive implicit feedback signal as confidences in proach. IFMF2 uses IFMF as its building block and extends matrix factorization. Given a set of positive implicit feed- it in the following way: Let p = 1 if c > 0 and p = 0 back signals c ≥ 0, we fit and predict an m × n binary ij ij ij ij otherwise.