<<

Collusion-resistant, Incentive-compatible Feedback Payments

Radu Jurca Boi Faltings Ecole Polytechnique F´ed´erale de Lausanne Ecole Polytechnique F´ed´erale de Lausanne (EPFL) (EPFL) Artificial Intelligence Laboratory Artificial Intelligence Laboratory CH-1015 Lausanne, Switzerland CH-1015 Lausanne, Switzerland radu.jurca@epfl.ch boi.faltings@epfl.ch

ABSTRACT etc.) that can only be observed after the purchase. This pre- Online reputation mechanisms need honest feedback to func- viously unavailable information allows the buyers to make tion effectively. Self-interested agents report the truth better, more efficient decisions. only when explicit rewards offset the potential gains ob- A key ingredient for all reputation mechanisms is hon- tained from lying. Feedback payment schemes (monetary est feedback. Human users exhibit high levels of altruistic rewards for submitted feedback) can make truth-telling ra- (i.e., honest) reporting, despite empirical evidence that ly- tional based on the correlation between the reports of dif- ing can bring external benefits [8, 21]. Nevertheless, the fu- ferent buyers. ture online economy may be dominated by rational, utility In this paper we investigate incentive-compatible payment maximizing software agents, that will exploit such misre- mechanisms that are also resistant to : groups of porting opportunities. Hence the need for designing reputa- agents cannot collude on a lying without suffer- tion mechanisms that are incentive-compatible: i.e., rational ing monetary losses. We analyze several scenarios, where, agents find it in their best interest to report the truth. for example, some or all of the agents collude. For each Fundamental results in the literature scenario we investigate both existential and implementation [7, 5] show that side payments can be designed to create the problems. Throughout the paper we use automated mecha- incentive for agents to report their private opinions truth- nism design to compute the best possible mechanism for a fully. The best such payment schemes have been constructed given setting. based on proper scoring rules [15, 12, 2], and exploit the cor- relation between the observations of different buyers about the same good. Categories and Subject Descriptors Miller, Resnick and Zeckhauser [18] adapt these results to I.2.11 [Artificial Intelligence]: Distributed Artificial In- online feedback forums. A central processing facility (the telligence reputation mechanism) scores every submitted feedback by comparing it with another report (called the reference re- General Terms port) about the same good. They prove the existence of general incentive-compatible payment mechanisms that cre- Algorithms, Design, Economics ate an equilibrium where the return when reporting honestly is better by at least an arbitrary margin δ. Keywords Jurca and Faltings [14] use an identical setting to apply reputation mechanisms, mechanism design, incentive com- automated mechanism design [3, 20]. Incentive-compatible patibility, collusion resistance payments are computed by solving an optimization prob- lem with the objective of minimizing the required budget. The simplicity of specifying payments through closed-form 1. INTRODUCTION scoring rules is sacrificed for significant gains in efficiency. Users increasingly resort to online feedback forums (rep- Intuitively, payment mechanisms encourage truth-telling utation mechanisms) for obtaining information about the because reporters expect to get paid according to how well products or services they intend to purchase. The testi- their feedback improves the current predictor of the refer- monies of previous buyers disclose hidden, experience-related ence report. Every feedback report modifies the reputation [19], product attributes (e.g., quality, reliability, ease of use, information, which acts as a predictor for future observa- tions. The payment received by the reporter reflects the quality of the updated predictor, tested against the refer- ence report. Assuming that the reference report is honest, Permission to make digital or hard copies of all or part of this work for every agent has the incentive to update the current reputa- personal or classroom use is granted without fee provided that copies are tion such that it mirrors her subjective beliefs. Agents thus not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. To copy otherwise, to report honestly, and truth-telling is a . republish, to post on servers or to redistribute to lists, requires prior specific For example, consider the owner (she) of a new house who permission and/or a fee. needs some plumbing work done. There are good and bad E C’07, J une 11–15, 2007, San D iego, California, USA. plumbers; the one chosen by our owner has a fairly good Copyright 2007 ACM 978-1-59593-653-0/07/0006 ...$5.00.

200

reputation that predicts high quality work with probabil- Third, we move to more realistic scenarios where only ity 75%. Once the work gets done, the owner will have a a fraction of the agents may collude. Colluders may not new (private) belief regarding the reputation of the plumber. transfer money among themselves, but may fully coordinate If she is happy with the service, she believes the plumber their reporting strategies. We show that honest reporting should have an even better reputation (that predicts, for ex- can be made a dominant strategy when (a) the coalition ample, good service with probability 87%1). On the other size is small enough, and (b) the non-colluders are reporting hand, if the owner is dissatisfied with the service, she be- honestly. lieves that the plumber should have a lower reputation (the Finally, we consider what happens when a single strate- probability of good service should only be 39%). An on- gic entity controls a number of fake online identities. As line reputation mechanism asks the owner to share feedback colluders can now transfer money among themselves, the about the plumber, and proposes the following payment payment mechanism must ensure that the cumulative rev- rule: “the submitted report is paid only if it matches the enue of the coalition is maximized by the honest reports. report submitted by another client about the same plumber. The result is positive, and given that non-colluders report A negative report will be paid $2.62, while a positive report honestly, appropriate payments elicit truthful information will be paid $1.54”. One can verify that honest reporting from the coalition as a whole. maximizes the expected payment of the owner, regardless of An important ingredient of our solution is to use sev- the actual experience2. eral reference reports when computing the feedback pay- While honest reporting is a Nash Equilibrium (NEQ), so is ments. This not only decreases the total cost of incentive- always reporting negative (or positive) feedback. Moreover, compatibility (a contribution of our previous work [14]), but the expected payoff from these constant reporting strategies also allows the design of mechanisms where honest reporting ($2.62 or $1.54 respectively) are both higher than the ex- is the unique equilibrium. pected payoff from the honest equilibrium. Unfortunately, After mentioning related work, this paper proceeds as fol- the existence of multiple equilibria is not an isolated problem lows. Section 2 formally introduces our model, Section 3 specific to our example. In previous work [13] we show that introduces incentive-compatible payment mechanisms and all binary incentive-compatible payment mechanisms suffer presents some of their properties. Sections 4 through 7 each from the same drawback: there are lying equilibria that gen- treat one collusion scenario. Finally we discuss future work erate higher expected payoffs than the truthful equilibrium. and conclude. This brings forth the problem of collusion. Rational agents have no reason to report truthfully, if they can do 1.1 Related Work better by coordinating on a lying equilibrium with higher Our work relates to the literature on (computational) payoff. Hence the motivation of this paper: we investigate mechanism design, , and incentive incentive-compatible payment mechanisms that are also re- contracts for principal-(multi)agent settings. The literature sistant to collusion. The extent and the power of the coali- on mechanism design (see [11] for a survey) and implemen- tion influences the complexity and difficulty of the design tation theory (see [10] for a survey) addresses the design problem. of mechanisms and institutions that satisfy certain prop- We will consider four collusion scenarios: First, we con- erties, given that the agents using the mechanism behave sider complete coalitions (all agents may be part of the coali- strategically. The main difference between mechanism de- tion) where agents may not redistribute revenues, and may sign and implementation theory is that of multiple equilib- only collude on symmetric strategies (all colluders report ac- ria. In mechanism design literature, the goal of the designer cording to the same strategy). We obtain a positive result is to find the mechanism that has the desired outcome as and show that by using several reference reports, it is possi- an equilibrium. In the implementation literature, on the ble to construct a payment mechanism with a unique honest other hand, the mechanism is required to have only the de- reporting symmetric equilibrium. sired equilibria. From this perspective, our results are closer Second, we consider a close to worst case scenario, where to the implementation theory, since in our quest for collu- all agents may collude, and coordinate on different report- sion resistance we look for payment mechanisms that induce ing strategies (every agent may report according to a dif- honest reporting as the only (or in some sense the best) equi- ferent strategy). Unsurprisingly, the result we obtain here librium. is negative: regardless of the number of reference reports, Computational mechanism design [6] extends the classi- no incentive-compatible payment mechanism has a unique cal literature by looking for equilibria that are also com- honest reporting equilibrium. putationally attractive. In the feedback reporting setting,

1 however, the main computational problem resides in com- The rationale behind these numbers will become apparent puting the mechanism itself (i.e., the payments), and not in in Section 2 executing it. In this paper, we rely on automated mechanism 2If the owner experiences good service from the plumber, she expects that some other client also gets good service with design [3] to compute the best mechanism for each context. probability 87%. Assuming that the other client reports The objective of the designer is to minimize the budget re- truthfully, the owner’s expected payment is: .87 · 1.54 + .13 · quired to pay for feedback, while enforcing the incentive- 0=1.34 if she reports good service, or .87·0+.13·2.62 = 0.34 compatibility constraints. The design problem is a linear if she reports bad service; Likewise, if the owner experiences optimization problem and can be solved efficiently [14]. A bad service, she expects that the reference report will be related method is incremental mechanism design [4] where negative with probability 1 − .39 = .61. In this case, her expected payment is: .39 · 1.54 + .61 · 0=0.6 if she reports the mechanism is solved iteratively, by incrementally adding good service, or .39 · 0+.61 · 2.62 = 1.6 if she reports bad new constraints. service. In both cases, honest reporting is better than lying A number of papers discuss incentive contracts that a by $1. principal should offer to several agents whose effort levels are

201

private. The reward received by each agent depends on the reputation mechanism) is a function τ : {0, 1}×{0, 1}N−1 → + output observed by the principal, and on the declarations of R ,whereτ(αi,α−i) ≥ 0 is the amount paid to buyer i when other agents. [9], [17], and [16] show that efficient contracts she reports αi ∈{0, 1} and the other N − 1 buyers report N−1 exist that are also incentive-compatible and collusion-proof. α−i ∈{0, 1} . The reports α−i are also called the refer- While the feedback reporting problem is similar, it differs in ence reports of agent i, since they constitute the reference one major aspect: the mechanism designer (i.e., the princi- for computing the payment for agent i. Payments are non- pal) does not observe a direct signal which is correlated to negative because most online forums do not have the means the reporters’ (i.e., agents’) private information. to impose punishments on the reporters. As the order of reports is not important, we can sim- plify the payment mechanism by assuming that τ(αi,α−i)= 2. THE MODEL ∗ ∗ τ(αi,α−i) for all α−i and α−i that contain the same num- We consider an online market where a number of rational ber of positive reports. A more compact description of the buyers (or agents) experience the same product (or service). payment mechanism is thus given by the amounts τ(α, n) The quality of the product remains fixed, and defines the where n ∈{0, 1,...,N − 1} is the number of positive re- product’s (unknown) type. Θ is the finite set of possible θ ports submitted by the reference reporters. types, and denotes a member of this set. We assume that The payoff expected by agent i depends on the distri- all buyers share a common belief regarding the prior proba- bution of the reference reports. If the other agents report bility Pr[θ], that the product is of type θ. θ∈Θ Pr[θ]=1. honestly, the distribution of the reference reports can be After the purchase, every buyer perceives a binary signal computed from the prior beliefs, and the true observation, about the quality (i.e., true type) of the product. Quality oi ∈{0, 1} of agent i. The probability that exactly n posi- N − signals are denoted as 1 (high quality) and 0 (low quality), tive reports were submitted by the other 1 agents is:  meaning that the buyer was satisfied, respectively dissatis- Pr[n|oi]= Pr[n|θ]Pr[θ|oi]; (1) fied with the product. Every product type is characterized θ∈Θ by a different probability distribution over the signals per- ceived by the buyers. Let Pr[1|θ] be the probability that the where Pr[n|θ] is given by the binomial distribution, and Pr θ|o agent buying a product of type θ is satisfied (i.e., observes [ i] can be computed from Bayes’ Law:   the quality signal 1). Pr[1|θ1] = Pr[1|θ2] for all θ1 = θ2 ∈ Θ, N −   Pr n|θ 1 Pr |θ n − Pr |θ N−1−n and Pr[1|·] is assumed common knowledge. [ ]= n [1 ] 1 [1 ] ; A central reputation mechanism asks every buyer to sub-  Pr[oi|θ]Pr[θ] mit feedback. Buyers are assumed rational, and not con- P [θ|oi]= ; Pr[oi]= Pr[oi|θ]Pr[θ]; Pr[oi] strained to report the truth. The set of pure reporting θ∈Θ strategies of a buyer is A = {(a0,a1)|a0,a1 ∈{0, 1}},where a =(a0,a1) denotes the strategy according to which the Astrategyprofilea is a vector (ai)i=1,...,N , prescribing buyer announces a0 ∈{0, 1} when she observes low quality, the reporting strategy ai ∈ A for each agent i. We will and a1 ∈{0, 1} when she observes high quality. We will sometimes use the notation a =(ai,a−i), where a−i is the often call the reports 0 and 1 as the negative, respectively strategy profile for all agents except i; i.e., a−i =(aj ), for the positive report. j =1,...,i− 1,i+1,...,N. Given the profile of reporting To ease the notation, we name the four members of the strategies (ai,a−i), let π[n, a−i] describe the belief of agent set A as the h(onest), l(ie), all1andall0strategies: i regarding the distribution of the reference reports, when: • h =(0, 1) is the honest reporting strategy; • n out of the other N −1 agents observe the high quality signal, 1 • l =(1, 0) is the strategy of always lying: the buyer reports 1 instead of 0 and 0 instead of 1; • the other N − 1 agents are reporting according to the strategy profile a−i; • all1=(1, 1) is the strategy of always reporting 1; Given n and a−i, agent i believes with probability • all0=(0, 0) is the strategy of always reporting 0; π[n, a−i](x)thatx reference reports are positive. If ai(oi) ∈ {0, 1} is the value of the report prescribed by strategy ai The reputation mechanism pays buyers for the submitted given the true observation oi, the expected payoff to agent reports. The amount received by buyer i can depend on any i is: information available to the reputation mechanism: namely, N−1 N−1 the reports submitted by other buyers, and the common V (ai,a−i|oi)= Pr[n|oi] π[n, a−i](x)τ(ai(oi),x); (2) knowledge regarding the environment (probability distribu- n=0 x=0 tion over types, and conditional probability distributions of quality signals). Let N be the total number of reports avail- Before moving to the next section, we can now justify the able to the reputation mechanism. N is finite, either because numerical values chosen for the example presented in the the total number of buyers is finite, or because the repu- introduction. The two possible types of the plumber are tation mechanism cannot wait indefinitely to receive more good (θG)andbad (θB), the good type being more likely reports. than the bad type: Pr[θG]=0.8andPr[θB ]=0.2. The Note that the reputation mechanism (a) does not know good plumber provides good service with high probability the true type of the product, and (b) cannot purchase the Pr[1|θG]=0.9; the bad plumber provides good service with product in order to get some first-hand experience regarding the lower probability, Pr[1|θB ]=0.15. The prior reputa- its quality. tion of the plumber predicts a good service with probability: Discarding from the notation the dependence on the com- Pr[θG]Pr[1|θG]+Pr[θB ]Pr[1|θB ]=0.75. However, the pos- mon knowledge, a payment mechanism (employed by the terior reputation depends on the agent’s actual experience.

202

If the agent observes 1, the posterior belief regarding the to an agent that truthfully reports 1, and the expected pay- type of the plumber will be: Pr[θG|1] = 1 − Pr[θB|1] = 0.96 ment to an agent that truthfully reports 0: (computed by Bayes’ Law), and the probability that the N−1 N−1 plumber provides good service to another client is: Pr[1|1] = W = Pr[1] Pr[n|1]τ(1,n)+Pr[0] Pr[n|0]τ(0,n); (5) Pr[1|θG]Pr[θG|1] + Pr[1|θB ]Pr[θB |1] = 0.87. Likewise, if n=0 n=0 the agent observes 0, the posterior belief regarding the where Pr[1] (respectively Pr[0]) are the prior probabilities type of the plumber is: Pr[θG|0] = 1 − Pr[θB |0] = 0.32, that the agent will perceive high (respectively low) quality, and the probability that the plumber provides good ser-  and are defined as: Pr[oi]= θ∈Θ Pr[oi|θ]Pr[θ]. vice to another client is: Pr[1|0] = Pr[1|θG]Pr[θG|0] + The payment scheme that minimizes the budget required Pr[1|θB ]Pr[θB |0] = 0.39. to pay for one honest report therefore solves the linear op- timization problem: 3. INCENTIVE-COMPATIBLE PAYMENT LP 1. MECHANISMS N−1 N−1 A payment mechanism is incentive-compatible when hon- min W = Pr[1] Pr[n|1]τ(1,n)+Pr[0] Pr[n|0]τ(0,n); est reporting is a Nash Equilibrium (NEQ): i.e., no agent can n=0 n=0 gain by lying when other agents report honestly. Formally, N−1  let (hi,h−i) be the strategy profile where all agents report s.t. Pr n| τ ,n − τ ,n ≥ δ i [ 1] (1 ) (0 ) ; honestly. It is optimal for agent to report the truth if and n=0 only if, for any observation oi, the honest report maximizes the agent’s expected payoff: N−1  Pr[n|0] τ(0,n) − τ(1,n) ≥ δ; n=0 V (hi,h−i|oi) >V(ai,h−i|oi) τ(0,n),τ(1,n) ≥ 0; ∀n = {0, 1,...,N − 1}; for any reporting strategy ai ∈ A\{h}, and any observation oi ∈{0, 1}. Although numerical algorithms can efficiently solve LP 1, Since reference reports are truthful, the expected payoff the analytical solution helps us gain additional insights to agent i is: about the structure of incentive-compatible payment mech- anisms. It turns out that LP 1 has a simple solution (details N−1 in Appendix A) where: V (hi,h−i|oi)= Pr[n|oi]τ(oi,n); n=0 τ(0,n)=0, ∀n = n1; τ(1,n)=0, ∀n = n2 Pr[n2|0] + Pr[n2|1] τ(0,n1)=δ ; and the incentive-compatibility constraints become: Pr[n2|1]Pr[n1|0] − Pr[n2|0]Pr[n1|1] Pr[n1|0] + Pr[n1|1] N−1 N−1 τ(1,n2)=δ ; Pr[n2|1]Pr[n1|0] − Pr[n2|0]Pr[n1|1] Pr[n|oi]τ(oi,n) > Pr[n|oi]τ(1 − oi,n); (3) n=0 n=0 Pr[n|1] Pr[n|0] n1 =argmin ; n2 =argmin n Pr[n|0] n Pr[n|1] for oi =0, 1. Practical mechanisms require certain margins for truth- telling [14]. Honest reporting must be better than lying by Intuitively, the optimal payment mechanism does not pay at least some margin δ, chosen by the mechanism designer to the negative or positive report of an agent unless the ref- offset the external benefits an agent might obtain by lying. erence reports contain exactly n1, respectively n2 positive δ Rewriting (3) to account for the margin ,anincentive- reports. The values n1 and n2 are chosen such that the pos- compatible payment mechanism satisfies the constraints: terior belief of the reporter regarding the reference reports N−1  changes the most: Pr n| τ ,n − τ ,n ≥ δ [ 1] (1 ) (0 ) ; • Pr[n1|0] increases the most with respect to Pr[n1|1]: n=0 e.g., n1 =argminn Pr[n|1]/P r[n|0]; N−1  (4) Pr[n|0] τ(0,n) − τ(1,n) ≥ δ; • Pr[n2|1] increases the most with respect to Pr[n2|0]: n=0 e.g., n2 =argminn Pr[n|0]/P r[n|1];

The values of τ(0,n1)andτ(1,n2) are then computed to guarantee the margin δ for honest reporting. formalizing the intuition that it is more profitable to report 3 positively (respectively negatively) when observing high (re- A similar property holds for all payment mechanisms spectively low) quality. that satisfy the incentive compatibility constraints: there must be at least two values of the reference reports, n1 = n2, [15], and [18] show that it is possible to construct pay- such that: ment mechanisms that satisfy the constraints in (4), based τ ,n >τ ,n ,Prn | >Prn | , on scoring rules. Jurca and Faltings [14] build on this ex- (0 1) (1 2) [ 1 0] [ 1 1] istence result and describe an algorithm that computes the τ(1,n2) >τ(0,n2),Pr[n2|1] >Pr[n2|0]; optimal (i.e., budget minimizing) payment mechanism. We 3One might wish, for example, to design a mechanism that will use this latter approach in this paper, for the obvious minimizes the expected budget paid to all N buyers. In practical advantages of designing an incentive compatible this case, the objective function of the problem LP 1 is: W¯ N Pr n n · τ ,n− N − n · τ ,n reputation mechanism as cheaply as possible. = n=0 [ ] (1 1) + ( ) (0 ) ,where The expected payment to an honest reporter (in the truth- Pr[n] is the prior probability that n out of N buyers observe ful NEQ) is the weighted sum between the expected payment high quality;

203

When τ(0,n1), τ(1,n2), τ(1,n1)andτ(0,n2) are scaled ap- • all0 (always reporting 0) is not NEQ when a rational propriately, a rational agent prefers the ‘bet’ on n1 when she agent would rather report 1 instead of 0 given that all all observes low quality, and the ‘bet’ on n2 when she observes other agents follow 0; high quality. τ(1, 0) >τ(0, 0); (7) It is exactly this property that makes it impossible to design an incentive-compatible mechanism that has hon- est reporting as the unique (or the most preferred) NEQ • l ie with only one reference report. n1 and n2 are constrained ( ) is not NEQ when at least one agent (either ob- to take the values 0, respectively 1, since by Bayes’ Law, serving 1 or 0) would rather report the truth. Given Pr[0|0] >Pr[0|1] and Pr[1|1] >Pr[1|0]. This results in that other agents always lie, N − 1 − n reference re- positive payments τ(0, 0) >τ(0, 1) and τ(1, 1) >τ(1, 0) (as ports will be positive whenever n high quality signals pointed out in the example from the introduction), thus the were actually observed: constant reporting strategies (always reporting 1 or always reporting 0) are also Nash Equilibria. Honest reporting is rewarded by a linear combination of τ(0, 0) and τ(1, 1), so at N−1   least one of the constant reporting strategies is more attrac- either Pr[n|0] τ(0,N − 1 − n) − τ(1,N − 1 − n) > 0; n=0 tive than truth-telling. [13] formally develops this result. (8) Using several reference reports decreases the budget re- N−1   quired to pay the reporters [14], and sometimes allows to or Pr[n|1] τ(1,N − 1 − n) − τ(0,N − 1 − n) > 0; design incentive-compatible payments where honesty is the n=0 only (or the most attractive) NEQ. In the following sections we explore some of the settings where such results apply. The objective function (5), and the constraints (4), (6), (7) and (8) define the optimal incentive-compatible payment mechanism that is also collusion-resistant in the sense ex- 4. NON-TRANSFERABLE UTILITIES, NO plained in the beginning of this section (i.e., honest reporting COORDINATION is the unique pure-strategy symmetric NEQ). To compute the payments, the mechanism designer must solve two linear The simplest collusion scenario (from the perspective of optimization problems, one corresponding to each branch of the mechanism designer) is to assume that agents (a) can the constraint (8). only coordinate once (before any of them purchases the A collusion-resistant mechanism is easier to find when the product) on the same (pure) reporting strategy, and (b) they number of reports available to the reputation mechanism cannot transfer payments from one another. Intuitively, this is higher. The minimum number of reference reports that setting characterizes anonymous feedback forums where the guarantee the existence of a collusion-proof payment mech- colluders do not have side-channels for exchanging informa- anism depends on the distributions Pr[n|·], and on the con- tion. The absence of communication channels is not an un- text. derlying assumption about the physical world, but rather a For the example provided in the introduction, it takes contextual implication: most online buyers do not know who N = 4 reports to design a collusion-resistant payment mech- is going to buy the same product in the immediate future, anism. The expected distribution over reference reports and therefore cannot synchronize their reports. is: Pr[0 ...3|0] = [0.4179, 0.2297, 0.1168, 0.2356] when the Nevertheless, the agents collude in the sense that they plumber provides bad service, and Pr[0 ...3|1] = [0.0255, all have one access to a trusted oracle that gives them a 0.0389, 0.2356, 0.7] when the plumber provides good service. reporting strategy. For example, the role of the oracle might Both probability distributions are computed according to be played by a trustworthy site that analyzes the reporting Eq. (1). The incentive-compatible, collusion-resistant pay- strategies and recommends the best one. ments are the following: τ(0, 0) = τ(0, 2) = 0, τ(0, 1) = The lack of coordination between colluders considerably 12.37, τ(0, 3) = ε, τ(1, 0) = ε, τ(1, 1) = τ(1, 3) = 0, and simplifies the problem of the mechanism designer. The only τ(1, 2) = 6.29. ε can take any value greater than 0, and the supplementary constraint on the incentive-compatible pay- guaranteed margin for truth-telling is δ =1. ment mechanism is to ensure that none of the pure symmet- ric strategy profiles is a NEQ. The set of pure strategies is finite (and contains 3 ly- 5. NON-TRANSFERABLE UTILITIES, ing strategies) therefore we can exhaustively enumerate the FULL COORDINATION constraints that prevent the corresponding symmetric lying The next collusion scenario we are considering is when strategy profiles to be NEQ. Since agents cannot transfer the N agents can use side-channels to coordinate their re- payments from one another, the constraints on the payments porting strategies, but they cannot transfer payments from should simply provide incentives for deviating from the col- one another. Unlike the previous setting, here each of the lusion strategy: N agents can have a different reporting strategy. The col- lusion strategy profile a =(ai), i =1,...,N is no longer • all1 (always reporting 1) is not NEQ when a rational symmetric, and prescribes that agent i reports according to agent would rather report 0 instead of 1 given that all the strategy ai ∈ A. other agents follow all1: We distinguish between two cases, where the communica- tion (and therefore the coordination on the reporting strat- egy profile) happens before or after the agents perceive the τ(0,N − 1) >τ(1,N − 1); (6) quality signals from the product they purchase. In both cases, however, we obtain negative results.

204

Proposition 1. When agents communicate and coordi- Similarly, a(n2 + 2) is not a NEQ iff either τ(0,n2 +1)> nate their reports after perceiving the quality signals, strict τ(1,n2 +1) (impossible), or τ(1,n2 +2) >τ(0,n2 +2). Con- incentive-compatible payment mechanisms do not exist. tinuing this argument we find that τ(1,N− 1) >τ(0,N− 1) which makes a(N) (i.e., all agents report 1) a Nash equilib- Proof. Consider two settings, that are identical except rium. Hence the result of the proposition.  for the observation of agent i. In setting A, agent i ob- serves oi = 0, in setting B, agent i observes oi =1;inboth A B n Proposition 2 holds regardless of the number of reports, and , the other agents observe high quality signals. N An incentive-compatible mechanism requires i to report 0 , available to the reputation mechanism. A full coalition in setting A, and 1 in setting B. Assume all other agents can always find a lying reporting strategy profile that is a report truthfully; during the communication phase (happen- Nash equilibrium. By definition such a coalition is stable, ing after signals have been perceived) agent i learns in both i.e., no colluder has the incentives to deviate from the col- settings that the reference reports contain n positive reports. lusion strategy. An incentive-compatible payment mechanism requires that: The obvious question is whether such coalitions are also profitable. Unless the collusion strategy brings every agent • τ(0,n) >τ(1,n) - honest reporting is strictly better at least the payoff expected from the honest equilibrium, for i in setting A ; there may be reasons to believe that the coalition will never form. Profitable coalitions require lying Nash equilibria • τ(1,n) >τ(0,n) - honest reporting is strictly better that pareto-dominate the honest one. A payment mecha- for i in setting B; nism where such equilibria do not exist, is, in some sense, collusion-resistant.  Clearlythisisimpossible. Take for example the incentive-compatible payment The previous proposition formalizes the intuition that scheme that solves LP 1, with the additional constraints truth-telling may only be an ex-ante Nash equilibrium. The that n1 =0and n2 = N − 1. A stable coalition can form reference reports must be unknown to the agent in order to on the strategy profile a(n2 +1)(or a(n1)), where n2 +1 allow the design of incentive-compatible payments. When (respectively n1) agents report 1 and the others report 0, the communication takes place before the agents observe the regardless of their observation. This equilibrium, however, signals, incentive-compatible payments do exist, but they al- does not pareto-dominate the truthful one: the agents that ways accept lying equilibria as well: report 0 do not get any reward, whereas they do get re- warded in the honest equilibrium. Proposition 2. When agents communicate and coordi- The payment mechanism can be further improved by set- nate their reports before perceiving the quality signals, no ting τ(0,n1 − 1) = τ(1,n2 +1)=ε,whereε is some small payment mechanism has a unique honest reporting Nash value. This modification eliminates the equilibria a(n2 +1) equilibrium. and a(n1) and instead introduces the equilibria a(n2 +2) and a(n1 − 1). Both these equilibria are extremely unattractive Proof. The proof shows that a full coalition can always (some agents get paid ε, while others don’t get paid at all) find a profile of constant reporting strategies, a =(ai), i = and are dominated by the honest one. a 1,...,N, ai ∈{all0,all1} that is a NEQ. For any given strategy profile, , either of the following We define the family of reporting strategy profiles a(n)= linear constraints makes the payment mechanism resistant a (ai)wheren out of N agents always report 1, and the other against a coalition on : N − n agents always report 0: i.e., ∗ ∗ V (ai,a−i|oi) Prn | τ ,n >τ ,n Pr n | > that [ 1 0] [ 1 1], (0 1) (1 1), [ 2 1] eliminate lying pareto-optimal equilibria. This algorithm Pr[n2|0] and τ(1,n2) >τ(0,n2). a n resembles the incremental mechanism design described by With non-transferable utilities, the strategy profile ( 2 + Conitzer and Sandholm [4] for social choice problems. As 1) is not a NEQ if and only if one of the n2 + 1 agents that should report 1 would rather report 0: part of future research we plan to look for heuristics that help a designer select a small set of strategies that generate τ(0,n2) >τ(1,n2); enough constraints to make honest reporting pareto optimal.

or one of the N − n2 − 1 agents that should report 0 would rather report 1: 6. NON-TRANSFERABLE UTILITIES, PARTIAL COORDINATION τ(1,n2 +1)>τ(0,n2 +1); The setting described in Section 5 is very close to the worst-case collusion scenario that may be observed in on- The first inequality cannot be true by the choice of n2; line reputation mechanisms. The coalition comprises all re- therefore, it must be that τ(1,n2 +1)>τ(0,n2 +1). porters, and the coordination mechanisms are perfect.

205

In most practical applications, not all agents can collude. Formally, let us take the subset c = {0,...,N − k} (this Some agents are altruistic in nature and report honestly for subset exists because N − k

N−k   6.1 The marginal cost of collusion resistance Pr[n|0] τ(0,n+ c) − τ(1,n+ c) ≥ δ; Incentive-compatible payments that are resistant to coali- n=0 (10) tions of size k, must satisfy the constraints in (10). As k in- N−k   creases, the design problem becomes more constrained, and Pr n| τ ,n c − τ ,n c ≥ δ [ 1] (1 + ) (0 + ) ; therefore, the budget required by the reputation mechanism n=0 is likely to grow. In this section we study the dependence of for all integers c ∈{0,...,k− 1}. the expected budget on the size of the maximum tolerated The objective function 5, and the set of constraints coalition. (10) form a linear optimization problem that defines the For a given context, the optimization problem that defines incentive-compatible payments that are resistant to a coali- the payments τk(·, ·) that are resistant to coalitions of size tion of size k. k is: The remaining question is how large may the collud- ing fraction be, such that collusion-resistant, incentive- LP 2. compatible mechanisms exist. N−1 N−1 min P r[1] Pr[n|1]τk(1,n)+Pr[0] Pr[n|0]τk(0,n); Proposition 3. When more than k agents collude, with n=0 n=0 2k>N, no incentive-compatible payment mechanism can N−k   make truth-telling the dominant strategy for the colluders. s.t. Pr[n|0] τk(0,n+ c) − τk(1,n+ c) ≥ δ; n=0 Proof. The intuition behind the proof is the following: N−k   When 2k>N,thek − 1 colluders submit at least as many Pr[n|1] τk(1,n+ c) − τk(0,n+ c) ≥ δ; reports as the remaining N − k honest reporters. Therefore, n=0 any sequence of honest reports, can be ‘corrected’ by a care- ∀c ∈{0,...k− 1}, fully chosen sequence of colluding reports, such that lying is τk(0,n),τk(1,n) ≥ 0; ∀n = {0, 1,...,N − 1}; profitable.

206

a number of fake online identities (or sybils [1]). From Table 1: Distribution of the maximum coalition the agent’s perspective, the individual revenues obtained by bound. kˆ = N/2 is the theoretical bound. Distribution of max coalition size (in %) over each sybil is irrelevant; the objective of the agent is to max- [k,ˆ kˆ − 1,...,1] imize the cumulated revenue obtained by all sybils. N =6, kˆ =3 [99.9, 0.08, 0.02] N =11, kˆ =5 [99.42, 0.44, 0.1, 0.04, 0] The fact that utilities are transferable, makes the prob- N =16, kˆ =8 [98.14, 0.68, 0.5, 0.36, 0.22, 0.06, 0.04, 0] lem of the mechanism designer significantly harder. In all N =21, kˆ =10 [97.32, 0.82, 0.52, 0.44, 0.4, 0.3, 0.1, 0.06, 0.04, 0] N =26, kˆ =13 [94.96, 1.48, 0.98, 0.62, 0.46, 0.44, 0.3, 0.28, 0.32, previous scenarios, the constraints that made an incentive- 0.06, 0.06, 0.04, 0] compatible mechanism collusion-resistant ensured that lying coalitions are unstable: at least one of the colluders is bet- ter off by deviating from the colluding strategy. However, in this context the agents that suffer from following the collud- 20 ing strategy may be rewarded by the others. The necessary N=11 (and sufficient) condition for collusion resistance requires N=16 that the cumulated revenue of the coalition is maximized N=21 15 N=26 when reporting the truth. Another difference from the settings in Sections 5 and 6 is that colluders coordinate their reporting strategy after ob- serving the quality signals. This assumption is supported by 10 the interpretation that one strategic entity controls several

relative cost fake online identities. Concretely, we are looking for a payment mechanism with k 5 the following property: whenever colluding agents observe c high quality signals, their cumulated revenue is maximized when reporting c positive reports. An underlying assump- tion is that non-colluders (the other N − k agents) are re- r 0 porting honestly. The revenue of the coalition that reports 0 0.1 0.2 0.3 0.4 0.5 (out of k) can be computed as follows. The r colluders that colluding fraction report positively are rewarded τ(1,r−1+n), while the k −r colluders that report negatively are rewarded τ(0,r+ n); n is the number of positive reports submitted by the (hon- Figure 1: The relative cost of the mechanism as we est) non-colluders. The expected revenue of the coalition is increase the resistance to colluding fraction. therefore: N−k  V (r|c)= Pr[n|c] r · τ(1,r− 1+n)+(k − r) · τ(0,r+ n) ; We numerically solved the optimal payment mechanism n=0 for 5000 problems, generated randomly in the following way: where Pr[n|c] is the probability that n out of N − k agents • the set of possible types is randomly chosen between observe high quality signals, given that c out of k positive 2 and 20; signals have already been observed. Honest reporting is the best strategy for the coalition, • θ p θ for each type, , the probability, ( ), that the buyers when for all c ∈{0,...k},argmaxr V (r|c)=c: observe high quality is randomly chosen between 0 and 1; N−k  We considered mechanisms for 6, 11, 16, 21 and 26 reports. Pr[n|c] c · τ(1,c− 1+n)+(k − c) · τ(0,c+ n) n=0 (11) As further evidence that the bound set by Proposition 3 is tight, Table 1 shows the distribution of the maximum col- −r · τ(1,r− 1+n) − (k − r) · τ(0,r+ n) ≥ δ; lusion threshold among the problems we have solved. For N example, when = 26 reports, approximately 95% of the The cheapest incentive-compatible, collusion-resistant problems accept the theoretical bound described by Propo- payment mechanism minimizes the objective function (5) sition 3. under the linear constraints (11): Figure 1 plots the relative cost of the collusion-resistant mechanism (i.e., divided by the cost of the mechanism that is LP 3. not collusion-resistant) as we increase the colluding fraction. It can be seen that the cost starts increasing exponentially N−1 N−1 min W Pr Pr n| τ ,n Pr Pr n| τ ,n when the payment mechanism must deter coalitions of more = [1] [ 1] (1 )+ [0] [ 0] (0 ); n=0 n=0 than one third of the agent population. This suggests that s.t. is true, ∀c, r ∈{ ,...k},c r the practical bound on the coalition size should be around (11) 0 = one third of the number N of reporters. τ(0,n),τ(1,n) ≥ 0; ∀n = {0, 1,...,N − 1};

7. TRANSFERABLE UTILITIES, PARTIAL We used numerical simulations to evaluate (a) the maxi- COORDINATION mum size of the tolerated coalitions, and (b) the marginal As a last scenario we assume that colluding agents can cost of increasing collusion resistance. As in Section 6, we redistribute the revenues among themselves. This will typ- generated 5000 random problems and computed the optimal ically be the case when the same strategic agent controls payments for N = 6,11,16,21 and 26 reports. For each case,

207

agents, each controlling several fake identities, try to manip- Table 2: Distribution of the maximum tolerable ulate the reporting mechanism. It is encouraging to see that coalition size. Distribution of max coalition size separately, in each scenario we can achieve relatively high N =6 [5 : 100%] N =11 [10 : 99.14%, 9:0.36%, 8:0.28%, 7:0.14%] collusion resistance at acceptable costs. As future work, we N =16 [15 : 97.56%, 14 : 0.44%, 13 : 0.32%, 12 : 0.52%] plan to use a combination of the two techniques to make the N =21 [20 : 96.04%, 19 : 0.52%, 18 : 0.36%, 17 : 0.42%] N =26 [25 : 94.24%, 24 : 0.66%, 23 : 0.58%, 22 : 0.46%] mechanisms even better.

8. CONCLUSION AND FUTURE WORK As feedback forums and reputation mechanisms become 20 increasingly important sources of information, explicit mea- N=11 sures must guarantee that honest reporting is in the best N=16 interest of the participants. Previous work shows that it is N=21 15 N=26 possible to construct payment mechanisms that reward hon- est reports higher (in expectation) than false ones. Truth- telling thus becomes a Nash equilibrium. Unfortunately, such mechanisms also have other equilib- 10 ria where reporters lie. This creates collusion opportunities,

relative cost since several agents can coordinate their lies in order to im- prove their revenues. In this paper we addressed the design 5 of incentive-compatible payments that are also resistant to collusion. For each of the four considered collusion scenario, we defined a linear optimization problem that allows the automated design of the cheapest payment mechanism. 0 0 0.2 0.4 0.6 0.8 1 In Section 4 we showed that incentive-compatible pay- colluding fraction ments can be efficiently constructed such that honest re- porting is the unique pure strategy symmetric equilibrium. Figure 2: The relative cost of the mechanism as we The results can be easily extended so that honest reporting increase the colluding fraction (setting with trans- becomes the pareto-optimal equilibrium. However, we only ferable utilities). treated pure strategies. Preventing mixed-strategy symmet- ric equilibria from becoming NEQ requires non-linear con- straints, that make the design problem computationally dif- ficult. Since reputation mechanisms need to compute such we gradually increased the coalition size (i.e., k)from1to payments for every context, we believe that the design prob- N − 1. lem should be kept simple, on the expense of precision. An Table 2 shows the distribution of the maximum coalition interesting question, therefore, is if we can find simple (e.g., size that can be deterred by the payment mechanism. This linear constraints), heuristic extensions to the linear pro- threshold is most of the time equal to N − 1, meaning that gram of Section 4 that (a) make unlikely the existence of one anonymous honest report is enough to design incentive- mixed strategy symmetric equilibria, or (b) make unlikely compatible, collusion-resistant payments. The result might the existence of mixed strategy symmetric equilibria that be surprising when related to the bound of Proposition 3: an pareto-dominate the honest equilibrium. apparently more difficult collusion scenario allows the design For the scenario in Section 5, we proved that any of payments that are resistant to bigger coalition fractions. incentive-compatible payment mechanism also accepts ly- The explanation resides in the choice of the solution con- ing equilibria, and suggested an iterative approach for find- cept for the scenario in Section 6. There, honest reporting ing the payments where truth-telling is not dominated by is the dominant strategy, so that each colluder reports the any of the lying equilibria. As future work, we plan to de- truth regardless of the reports of the other colluders. In scribe practical algorithms that guide mechanism designers the present scenario, on the other hand, individual collud- in choosing the best direction for incrementally improving ers are allowed to lie, if the coalition as a whole reports the the mechanism. truth. The constraints of LP 3 are therefore feasible for In Section 6 we took advantage of the assumption that higher coalition fractions. some agents will inherently report the truth, and con- The marginal cost of collusion resistance is however higher structed an incentive-compatible mechanism with much than in Section 6. Figure 2 plots the relative cost of the stronger collusion guarantees. When less than half of the collusion-resistant mechanism (i.e., divided by the cost of population colludes, it is theoretically possible to construct the mechanism that is not collusion-resistant) as we increase payments that make honest reporting the dominant strat- the tolerated coalition fraction. The cost grows linearly for egy for the colluders. Numerical simulations show, however, coalitions that span up to one half of the population; for that the practical bound is closer to one third of the popu- larger coalitions, the cost grows exponentially. Nonetheless, lation: preventing coalition fractions greater than one third by comparing Figures 2 and 1, we see that for the same requires exponentially higher budget. An interesting open coalition size, the collusion-resistant payments are cheaper question is if we can increase the theoretical (and practi- if we assume a setting with non-transferable utilities. cal) bound by requiring honest reporting to be the unique One last thing we would like to point out is that realistic (or pareto-optimal) Nash equilibrium. The constraints that collusion scenarios are most likely a combination between define the corresponding mechanisms lead to non-linear op- the settings presented in Sections 6 and 7: several strategic timization problems that do not scale well. We plan to use

208

approximations and heuristic methods to further investigate [15] M. Kandori and H. Matsushima. Private observation, this question. communication and collusion. Econometrica, Finally, Section 7 described incentive-compatible pay- 66(3):627–652, 1998. ments that are resistant to sybil attacks: i.e., the same [16] S. Li and K. Balachandran. Collusion proof transfer strategic agents creates several fake identities in order to payment schemes with multiple agents. Review of manipulate the payment mechanism. The designer can en- Quantitative Finance and Accounting, 15:217–233, sure that the set of reports submitted by the coalition re- 2000. flects the aggregated experience of the coalitions. Individual [17] C. Ma. Unique implementation of incentive contracts colluders do not necessarily report the truth, but overall, the with many agents. Review of Economic Studies, pages reputation mechanism obtains correct information. 555–572, 1988. [18] N. Miller, P. Resnick, and R. Zeckhauser. Eliciting 9. REFERENCES Informative Feedback: The Peer-Prediction Method. [1] A. Cheng and E. Friedman. Sybilproof reputation Management Science, 51:1359 –1373, 2005. mechanisms. In Proceeding of the Workshop on [19] A. Parasuraman, V. Zeithaml, and L. Berry. A Economics of Peer-to-Peer Systems (P2PECON), Conceptual Model of Service Quality and Its pages 128–132, 2005. Implications for Future Research. Journal of [2] R. T. Clemen. Incentive contracts and strictly proper Marketing, 49:41–50, 1985. scoring rules. Test, 11:167–189, 2002. [20] T. Sandholm. Automated mechanism design: A New [3] V. Conitzer and T. Sandholm. Complexity of Application Area for Search Algorithms. In mechanism design. In Proceedings of the Uncertainty Proceedings of the International Conference on in Artificial Intelligence Conference (UAI), 2002. Principles and Practice of Constraint Programming, [4] V. Conitzer and T. Sandholm. Incremental 2003. Mechanism Design. In Proceedings of the IJCAI, 2007. [21] E. White. Chatting a Singer Up the Pop Charts. The [5] J. Cr´emer and R. P. McLean. Optimal Selling Wall Street Journal, October 15, 1999. Strategies under Uncertainty for a Discriminating Monopolist When Demands Are Interdependent. APPENDIX Econometrica, 53(2):345–61, 1985. [6] R. K. Dash, N. R. Jennings, and D. C. Parkes. A. ANALYTICAL SOLUTION Computational-mechanism design: A call to arms. FOR INCENTIVE-COMPATIBLE IEEE Intelligent Systems, pages 40–47, November PAYMENT MECHANISMS 2003. Special Issue on Agents and Markets. For solving LP 1, let us write the corresponding dual prob- [7] C. d’Aspremont and L.-A. Grard-Varet. Incentives lem: and Incomplete Information. Journal of Public max δy0 + δy1; Economics, 11:25–45, 1979. s.t. P r[n|0]y0 − Pr[n|1]y1 ≤ Pr[0]Pr[n|0] [8] A. Harmon. Amazon Glitch Unmasks War of Pr[n|1]y1 − Pr[n|0]y0 ≤ Pr[1]Pr[n|1] Reviewers. The New York Times, February 14, 2004. ∀n ∈{0,...,N − 1}; [9] B. Holmstr¨om. Moral Hazard in Teams. Bell Journall where y0 (respectively y1) is the dual variable corresponding of Economics, 13:324–340, 1982. to the constraint where the agent observes 0 (respectively 1). [10] M. Jackson. A crash course in implementation theory. By dividing the first set of constraints with Pr[n|0] and the Social Choice and Welfare, 18(4):655–708, 2001. second set of constraints with Pr[n|1], we have: [11] M. Jackson. Encyclopedia of Life Support Systems, y0 − y1Pr[n|1]/P r[n|0] ≤ Pr[0], ∀n ∈{0,...,N − 1}; chapter Mechanism Theory. EOLSS Publishers, y1 − y0Pr[n|0]/P r[n|1] ≤ Pr[1], ∀n ∈{0,...,N − 1}; Oxford UK, 2003. [12] S. Johnson, J. Pratt, and R. Zeckhauser. Efficiency Clearly, among the 2(N−1) constraints of the dual problem, Despite Mutually Payoff-Relevant Private Information: n Pr[n|1] only two are active, corresponding to: 1 =argminn Pr[n|0] , The Finite Case. Econometrica, 58:873–900, 1990. Pr[n|0] and n2 =argmin . It follows that only two of the [13] R. Jurca and B. Faltings. Enforcing Truthful n Pr[n|1] τ ,n τ ,n Strategies in Incentive Compatible Reputation variables of LP 1 (i.e., (0 1)and (1 2)) have positive Mechanisms. In Internet and Network Economics, values. These values can be computed by solving the system of linear equations: volume 3828 of LNCS, pages 268 – 277. 2005. [14] R. Jurca and B. Faltings. Minimum payments that Pr[n1|0]τ(0,n1) − Pr[n2|0]τ(1,n2)=δ; reward honest reputation feedback. In Proceedings of − Pr[n1|1]τ(0,n1)+Pr[n2|1]τ(1,n2)=δ; the ACM Conference on Electronic Commerce,Ann Arbor, Michigan, USA, June 11-15 2006.

209