Dynamics of Profit-Sharing Games

Total Page:16

File Type:pdf, Size:1020Kb

Dynamics of Profit-Sharing Games Proceedings of the Twenty-Second International Joint Conference on Artificial Intelligence Dynamics of Profit-Sharing Games John Augustine, Ning Chen, Edith Elkind, Angelo Fanelli, Nick Gravin, Dmitry Shiryaev School of Physical and Mathematical Sciences Nanyang Technological University, Singapore Abstract information about other players. Such agents may simply re- spond to their current environment without worrying about An important task in the analysis of multiagent sys- the subsequent reaction of other agents; such behavior is said tems is to understand how groups of selfish players to be myopic. Now, coalition formation by computationally can form coalitions, i.e., work together in teams. In limited agents has been studied by a number of researchers in this paper, we study the dynamics of coalition for- multi-agent systems, starting with the work of [Shehory and mation under bounded rationality. We consider set- Kraus, 1999] and [Sandholm and Lesser, 1997]. However, tings where each team’s profit is given by a concave myopic behavior in coalition formation received relatively lit- function, and propose three profit-sharing schemes, tle attention in the literature (for some exceptions, see [Dieck- each of which is based on the concept of marginal mann and Schwalbe, 2002; Chalkiadakis and Boutilier, 2004; utility. The agents are assumed to be myopic, i.e., Airiau and Sen, 2009]). In contrast, myopic dynamics of they keep changing teams as long as they can in- non-cooperative games is the subject of a growing body of crease their payoff by doing so. We study the prop- research (see, e.g. [Fabrikant et al., 2004; Awerbuch et al., erties (such as closeness to Nash equilibrium or to- 2008; Fanelli et al., 2008]). tal profit) of the states that result after a polyno- mial number of such moves, and prove bounds on In this paper, we merge these streams of research and ap- the price of anarchy and the price of stability of the ply techniques developed in the context of analyzing the dy- corresponding games. namics of non-cooperative games to coalition formation set- tings. In doing so, we depart from the standard model of games with transferable utility, which allows the players in 1 Introduction a team to share the payoff arbitrarily: indeed, such flexibility Cooperation and collaborative task execution are fundamen- will necessitate a complicated negotiation process whenever a tally important both for human societies and for multiagent player wants to switch teams. Instead, we consider three pay- systems. Indeed, many tasks are too complicated or resource- off models that are based on the concept of marginal utility, consuming to be executed by a single agent, and a collective i.e., the contribution that the player makes to his current team. effort is needed. Such settings are usually modeled using the Each of the payoff schemes, when combined with a coopera- framework of cooperative games, which specify the amount tive game, induces a non-cooperative game, whose dynamics of payoff that each subset of agents can achieve: when the can then be studied using the rich set of tools developed for game is played the agents split into teams (coalitions), and such games in recent years. the payoff of each team is divided among its members. We will now describe our payment schemes in more detail. The standard framework of cooperative game theory is We assume that we are given a concave cooperative game, static, i.e., it does not explain how the players arrive at a i.e., the values of the teams are given by a submodular func- particular set of teams and a payoff distribution. However, tion; the submodularity property means that a player is more understanding the dynamics of coalition formation is impor- useful when he joins a smaller team, and plays an important tant for many applications of cooperative games, and game role in our analysis. In our first scheme, the payment to each theorists have studied bargaining and coalition formation in agent is given by his marginal utility for his current team; by cooperative environments (see, e.g. [Chatterjee et al., 1993; submodularity, the total payment to the team members never Moldovanu and Winter, 1995; Yan, 2003]). Much of this exceeds the team’s value. This payment scheme rewards each research assumes that the agents are fully rational, i.e., can agent according to the value he creates; we will therefore call predict the consequences of their actions and maximize their these games Fair Value games. Our second scheme takes into (expected) utility based on these predictions. However, full account the history of the interaction: we keep track of the or- rationality is a strong assumption that is unlikely to hold in der in which the players have joined their teams, and pay each many real-life scenarios: first, the agents may not have the agent his marginal contribution to the coalition formed by the computational resources to infer their optimal strategies, and players who joined his current team before him. This ensures second, they may not be sophisticated enough to do so, or lack that the entire payoff of each team is fully distributed among 37 its members. Moreover, due to the submodularity property S by changing the strategy of player i from si to si, i.e., a player’s payoff never goes down as long as he stays with (S−i,si)=(s1,s2,...,si−1,si,si+1,...,sn). the same team. This payoff scheme is somewhat reminiscent Nash Equilibria and Dynamics Given a strategy profile of the reward schemes employed in industries with strong la- S =(s1,s2,...,sn), a strategy si ∈ Σi is an improvement bor unions; we will therefore refer to these games as Labor move for player i if ui(S−i,si) >ui(S); further, si is called Union games. Our third scheme can be viewed as a hybrid α i ui(S−i,s ) > (1 + α)ui(S) of the first two: it distributes the team’s payoff according to an -improvement move for if i , α> sb ∈ the players’ Shapley values, i.e., it pays each player his ex- where 0. A strategy i Σi is a best response for player i S pected marginal contribution to a coalition formed by its pre- in state if it yields the maximum possible payoff given u S ,sb ≥ decessors when players are reordered randomly; the resulting the strategy choices of the other players, i.e., i( −i i ) u S ,s s ∈ α games are called Shapley games. i( −i i) for any i Σi.An -best response move is both an α-improvement and a best response move. Our Contributions We study the equilibria and dynamics A (pure) Nash equilibrium is a strategy profile in which of the three games described above. We establish that all our every player plays her best response. Formally, S = games admit a Nash equilibrium. Further, we argue that in (s1,s2,...,sn) is a Nash equilibrium if for all i ∈ N and all these games the total profit of the system in a Nash equi- for any strategy s ∈ Σi we have ui(S) ≥ ui(S−i,s ).We librium, i.e., the sum of teams’ values, is within a factor of 2 i i denote the set of all (pure) Nash equilibria of a game G by from optimal. In addition, for the first two classes of games, NE(G). A profile S =(s ,...,sn) is called an α-Nash equi- a natural dynamic process quickly converges to a state that is 1 librium if no player can improve his payoff by more than a almost as good as a Nash equilibrium, and the best Nash equi- factor of (1+α) by deviating, i.e., (1+α)ui(S) ≥ ui(S−i,si) librium is in fact an optimal state. We also show that Labor for any i ∈ N and any ui ∈ Σi. The set of all α-Nash equilib- Union games have other desirable properties. α ria of G is denoted by NE (G).Inastrong Nash equilibrium, Related Work The games studied in this paper belong to no group of players can improve their payoffs by deviating, the class of potential games, introduced by Monderer and i.e., S =(s1,...,sn) is a strong Nash equilibrium if for all Shapley [1996]. In potential games, any sequence of im- I ⊆ N S s ,...,s and any strategy vector =(1 n) such that provements by players converges to a pure Nash equilibrium. si = si for i ∈ N \ I,ifui(S ) >ui(S) for some i ∈ I, then However, the number of steps can be exponential in the de- uj(S ) <uj(S) for some j ∈ I. scription of the game. The complexity of computing (ap- Let Δi(S) be the improvement in the player’s payoff if proximate) Nash equilibrium in various subclasses of poten- b he performs his best response, i.e., Δi(S)=ui(S−i,si ) − tial games such as congestion games, cut games, or party b ui(S), where s is the best response of player i in state affiliation games has received a lot of attention in recent i S. For any Z ⊆ N let ΔZ (S)= Δi(S), and let years [Fabrikant et al., 2004; Skopalik and Vocking,¨ 2008; i∈Z Δ(S)=ΔN (S).ANash dynamic (respectively, α-Nash Bhalgat et al., 2010]. Related issues are how long it takes dynamic) is any sequence of best response (respectively, α- for some form of best response dynamics to reach an equilib- best response) moves. A basic Nash dynamic (respectively, rium [Chien and Sinclair, 2007; Ackermann et al., 2008],or basic α-Nash dynamic) is any Nash dynamic (respectively, how good are the states reached after a polynomial number of α-Nash dynamic) such that at each state S the player i that steps [Awerbuch et al., 2008; Fanelli et al., 2008].
Recommended publications
  • Game Theory Lecture Notes
    Game Theory: Penn State Math 486 Lecture Notes Version 2.1.1 Christopher Griffin « 2010-2021 Licensed under a Creative Commons Attribution-Noncommercial-Share Alike 3.0 United States License With Major Contributions By: James Fan George Kesidis and Other Contributions By: Arlan Stutler Sarthak Shah Contents List of Figuresv Preface xi 1. Using These Notes xi 2. An Overview of Game Theory xi Chapter 1. Probability Theory and Games Against the House1 1. Probability1 2. Random Variables and Expected Values6 3. Conditional Probability8 4. The Monty Hall Problem 11 Chapter 2. Game Trees and Extensive Form 15 1. Graphs and Trees 15 2. Game Trees with Complete Information and No Chance 18 3. Game Trees with Incomplete Information 22 4. Games of Chance 24 5. Pay-off Functions and Equilibria 26 Chapter 3. Normal and Strategic Form Games and Matrices 37 1. Normal and Strategic Form 37 2. Strategic Form Games 38 3. Review of Basic Matrix Properties 40 4. Special Matrices and Vectors 42 5. Strategy Vectors and Matrix Games 43 Chapter 4. Saddle Points, Mixed Strategies and the Minimax Theorem 45 1. Saddle Points 45 2. Zero-Sum Games without Saddle Points 48 3. Mixed Strategies 50 4. Mixed Strategies in Matrix Games 53 5. Dominated Strategies and Nash Equilibria 54 6. The Minimax Theorem 59 7. Finding Nash Equilibria in Simple Games 64 8. A Note on Nash Equilibria in General 66 Chapter 5. An Introduction to Optimization and the Karush-Kuhn-Tucker Conditions 69 1. A General Maximization Formulation 70 2. Some Geometry for Optimization 72 3.
    [Show full text]
  • CSC304 Lecture 4 Game Theory
    CSC304 Lecture 4 Game Theory (Cost sharing & congestion games, Potential function, Braess’ paradox) CSC304 - Nisarg Shah 1 Recap • Nash equilibria (NE) ➢ No agent wants to change their strategy ➢ Guaranteed to exist if mixed strategies are allowed ➢ Could be multiple • Pure NE through best-response diagrams • Mixed NE through the indifference principle CSC304 - Nisarg Shah 2 Worst and Best Nash Equilibria • What can we say after we identify all Nash equilibria? ➢ Compute how “good” they are in the best/worst case • How do we measure “social good”? ➢ Game with only rewards? Higher total reward of players = more social good ➢ Game with only penalties? Lower total penalty to players = more social good ➢ Game with rewards and penalties? No clear consensus… CSC304 - Nisarg Shah 3 Price of Anarchy and Stability • Price of Anarchy (PoA) • Price of Stability (PoS) “Worst NE vs optimum” “Best NE vs optimum” Max total reward Max total reward Min total reward in any NE Max total reward in any NE or or Max total cost in any NE Min total cost in any NE Min total cost Min total cost PoA ≥ PoS ≥ 1 CSC304 - Nisarg Shah 4 Revisiting Stag-Hunt Hunter 2 Stag Hare Hunter 1 Stag (4 , 4) (0 , 2) Hare (2 , 0) (1 , 1) • Max total reward = 4 + 4 = 8 • Three equilibria ➢ (Stag, Stag) : Total reward = 8 ➢ (Hare, Hare) : Total reward = 2 1 2 1 2 ➢ ( Τ3 Stag – Τ3 Hare, Τ3 Stag – Τ3 Hare) 1 1 1 1 o Total reward = ∗ ∗ 8 + 1 − ∗ ∗ 2 ∈ (2,8) 3 3 3 3 • Price of stability? Price of anarchy? CSC304 - Nisarg Shah 5 Revisiting Prisoner’s Dilemma John Stay Silent Betray Sam
    [Show full text]
  • Potential Games. Congestion Games. Price of Anarchy and Price of Stability
    8803 Connections between Learning, Game Theory, and Optimization Maria-Florina Balcan Lecture 13: October 5, 2010 Reading: Algorithmic Game Theory book, Chapters 17, 18 and 19. Price of Anarchy and Price of Staility We assume a (finite) game with n players, where player i's set of possible strategies is Si. We let s = (s1; : : : ; sn) denote the (joint) vector of strategies selected by players in the space S = S1 × · · · × Sn of joint actions. The game assigns utilities ui : S ! R or costs ui : S ! R to any player i at any joint action s 2 S: any player maximizes his utility ui(s) or minimizes his cost ci(s). As we recall from the introductory lectures, any finite game has a mixed Nash equilibrium (NE), but a finite game may or may not have pure Nash equilibria. Today we focus on games with pure NE. Some NE are \better" than others, which we formalize via a social objective function f : S ! R. Two classic social objectives are: P sum social welfare f(s) = i ui(s) measures social welfare { we make sure that the av- erage satisfaction of the population is high maxmin social utility f(s) = mini ui(s) measures the satisfaction of the most unsatisfied player A social objective function quantifies the efficiency of each strategy profile. We can now measure how efficient a Nash equilibrium is in a specific game. Since a game may have many NE we have at least two natural measures, corresponding to the best and the worst NE. We first define the best possible solution in a game Definition 1.
    [Show full text]
  • Lecture Notes
    GRADUATE GAME THEORY LECTURE NOTES BY OMER TAMUZ California Institute of Technology 2018 Acknowledgments These lecture notes are partially adapted from Osborne and Rubinstein [29], Maschler, Solan and Zamir [23], lecture notes by Federico Echenique, and slides by Daron Acemoglu and Asu Ozdaglar. I am indebted to Seo Young (Silvia) Kim and Zhuofang Li for their help in finding and correcting many errors. Any comments or suggestions are welcome. 2 Contents 1 Extensive form games with perfect information 7 1.1 Tic-Tac-Toe ........................................ 7 1.2 The Sweet Fifteen Game ................................ 7 1.3 Chess ............................................ 7 1.4 Definition of extensive form games with perfect information ........... 10 1.5 The ultimatum game .................................. 10 1.6 Equilibria ......................................... 11 1.7 The centipede game ................................... 11 1.8 Subgames and subgame perfect equilibria ...................... 13 1.9 The dollar auction .................................... 14 1.10 Backward induction, Kuhn’s Theorem and a proof of Zermelo’s Theorem ... 15 2 Strategic form games 17 2.1 Definition ......................................... 17 2.2 Nash equilibria ...................................... 17 2.3 Classical examples .................................... 17 2.4 Dominated strategies .................................. 22 2.5 Repeated elimination of dominated strategies ................... 22 2.6 Dominant strategies ..................................
    [Show full text]
  • Game Theoretic Interaction and Decision: a Quantum Analysis
    games Article Game Theoretic Interaction and Decision: A Quantum Analysis Ulrich Faigle 1 and Michel Grabisch 2,* 1 Mathematisches Institut, Universität zu Köln, Weyertal 80, 50931 Köln, Germany; [email protected] 2 Paris School of Economics, University of Paris I, 106-112, Bd. de l’Hôpital, 75013 Paris, France * Correspondence: [email protected]; Tel.: +33-144-07-8744 Received: 12 July 2017; Accepted: 21 October 2017; Published: 6 November 2017 Abstract: An interaction system has a finite set of agents that interact pairwise, depending on the current state of the system. Symmetric decomposition of the matrix of interaction coefficients yields the representation of states by self-adjoint matrices and hence a spectral representation. As a result, cooperation systems, decision systems and quantum systems all become visible as manifestations of special interaction systems. The treatment of the theory is purely mathematical and does not require any special knowledge of physics. It is shown how standard notions in cooperative game theory arise naturally in this context. In particular, states of general interaction systems are seen to arise as linear superpositions of pure quantum states and Fourier transformation to become meaningful. Moreover, quantum games fall into this framework. Finally, a theory of Markov evolution of interaction states is presented that generalizes classical homogeneous Markov chains to the present context. Keywords: cooperative game; decision system; evolution; Fourier transform; interaction system; measurement; quantum game 1. Introduction In an interaction system, economic (or general physical) agents interact pairwise, but do not necessarily cooperate towards a common goal. However, this model arises as a natural generalization of the model of cooperative TU games, for which already Owen [1] introduced the co-value as an assessment of the pairwise interaction of two cooperating players1.
    [Show full text]
  • Potential Games and Competition in the Supply of Natural Resources By
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by ASU Digital Repository Potential Games and Competition in the Supply of Natural Resources by Robert H. Mamada A Dissertation Presented in Partial Fulfillment of the Requirements for the Degree Doctor of Philosophy Approved March 2017 by the Graduate Supervisory Committee: Carlos Castillo-Chavez, Co-Chair Charles Perrings, Co-Chair Adam Lampert ARIZONA STATE UNIVERSITY May 2017 ABSTRACT This dissertation discusses the Cournot competition and competitions in the exploita- tion of common pool resources and its extension to the tragedy of the commons. I address these models by using potential games and inquire how these models reflect the real competitions for provisions of environmental resources. The Cournot models are dependent upon how many firms there are so that the resultant Cournot-Nash equilibrium is dependent upon the number of firms in oligopoly. But many studies do not take into account how the resultant Cournot-Nash equilibrium is sensitive to the change of the number of firms. Potential games can find out the outcome when the number of firms changes in addition to providing the \traditional" Cournot-Nash equilibrium when the number of firms is fixed. Hence, I use potential games to fill the gaps that exist in the studies of competitions in oligopoly and common pool resources and extend our knowledge in these topics. In specific, one of the rational conclusions from the Cournot model is that a firm’s best policy is to split into separate firms. In real life, we usually witness the other way around; i.e., several firms attempt to merge and enjoy the monopoly profit by restricting the amount of output and raising the price.
    [Show full text]
  • Online Learning and Game Theory
    Online Learning and Game Theory Your guide: Avrim Blum Carnegie Mellon University [Algorithmic Economics Summer School 2012] Itinerary • Stop 1: Minimizing regret and combining advice. – Randomized Wtd Majority / Multiplicative Weights alg – Connections to minimax-optimality in zero-sum games • Stop 2: Bandits and Pricing – Online learning from limited feedback (bandit algs) – Use for online pricing/auction problems • Stop 3: Internal regret & Correlated equil – Internal regret minimization – Connections to correlated equilib. in general-sum games • Stop 4: Guiding dynamics to higher-quality equilibria – Helpful “nudging” of simple dynamics in potential games Stop 1: Minimizing regret and combining expert advice Consider the following setting… Each morning, you need to pick one of N possible routes to drive to work. Robots But traffic is different each day. R Us 32 min Not clear a priori which will be best. When you get there you find out how long your route took. (And maybe others too or maybe not.) Is there a strategy for picking routes so that in the long run, whatever the sequence of traffic patterns has been, you’ve done nearly as well as the best fixed route in hindsight? (In expectation, over internal randomness in the algorithm) Yes. “No-regret” algorithms for repeated decisions A bit more generally: Algorithm has N options. World chooses cost vector. Can view as matrix like this (maybe infinite # cols) World – life - fate Algorithm At each time step, algorithm picks row, life picks column. Alg pays cost for action chosen. Alg gets column as feedback (or just its own cost in the “bandit” model). Need to assume some bound on max cost.
    [Show full text]
  • Potential Games with Continuous Player Sets
    Potential Games with Continuous Player Sets William H. Sandholm* Department of Economics University of Wisconsin 1180 Observatory Drive Madison, WI 53706 [email protected] www.ssc.wisc.edu/~whs First Version: April 30, 1997 This Version: May 20, 2000 * This paper is based on Chapter 4 of my doctoral dissertation (Sandholm [29]). I thank an anonymous referee and Associate Editor, Josef Hofbauer, and seminar audiences at Caltech, Chicago, Columbia, Harvard (Kennedy), Michigan, Northwestern, Rochester, Washington University, and Wisconsin for their comments. I especially thank Eddie Dekel and Jeroen Swinkels for their advice and encouragement. Financial support from a State Farm Companies Foundation Dissertation Fellowship is gratefully acknowledged. Abstract We study potential games with continuous player sets, a class of games characterized by an externality symmetry condition. Examples of these games include random matching games with common payoffs and congestion games. We offer a simple description of equilibria which are locally stable under a broad class of evolutionary dynamics, and prove that behavior converges to Nash equilibrium from all initial conditions. W e consider a subclass of potential games in which evolution leads to efficient play. Finally, we show that the games studied here are the limits of convergent sequences of the finite player potential games studied by Monderer and Shapley [22]. JEL Classification Numbers: C72, C73, D62, R41. 1. Introduction Nash equilibrium is the cornerstone of non-cooperative game theory, providing a necessary condition for stable behavior among rational agents. Still, to justify the prediction of Nash equilibrium play, one must explain how players arrive at a Nash equilibrium; if equilibrium is not reached, the fact that it is self-sustaining becomes moot.
    [Show full text]
  • Noisy Directional Learning and the Logit Equilibrium*
    CSE: AR SJOE 011 Scand. J. of Economics 106(*), *–*, 2004 DOI: 10.1111/j.1467-9442.2004.000376.x Noisy Directional Learning and the Logit Equilibrium* Simon P. Anderson University of Virginia, Charlottesville, VA 22903-3328, USA [email protected] Jacob K. Goeree California Institute of Technology, Pasadena, CA 91125, USA [email protected] Charles A. Holt University of Virginia, Charlottesville, VA 22903-3328, USA [email protected] PROOF Abstract We specify a dynamic model in which agents adjust their decisions toward higher payoffs, subject to normal error. This process generates a probability distribution of players’ decisions that evolves over time according to the Fokker–Planck equation. The dynamic process is stable for all potential games, a class of payoff structures that includes several widely studied games. In equilibrium, the distributions that determine expected payoffs correspond to the distributions that arise from the logit function applied to those expected payoffs. This ‘‘logit equilibrium’’ forms a stochastic generalization of the Nash equilibrium and provides a possible explanation of anomalous laboratory data. ECTED Keywords: Bounded rationality; noisy directional learning; Fokker–Planck equation; potential games; logit equilibrium JEL classification: C62; C73 I. Introduction Small errors and shocks may have offsetting effects in some economic con- texts, in which case there is not much to be gained from an explicit analysis of stochasticNCORR elements. In other contexts, a small amount of randomness can * We gratefully acknowledge financial support from the National Science Foundation (SBR- 9818683U and SBR-0094800), the Alfred P. Sloan Foundation and the Dutch National Science Foundation (NWO-VICI 453.03.606).
    [Show full text]
  • Strong Nash Equilibria in Games with the Lexicographical Improvement
    Strong Nash Equilibria in Games with the Lexicographical Improvement Property Tobias Harks, Max Klimm, and Rolf H. M¨ohring Preprint 15/07/2009 Institut f¨ur Mathematik, Technische Universit¨at Berlin harks,klimm,moehring @math.tu-berlin.de { } Abstract. We introduce a class of finite strategic games with the property that every deviation of a coalition of players that is profitable to each of its members strictly decreases the lexicographical order of a certain function defined on the set of strategy profiles. We call this property the Lexicographical Improvement Property (LIP) and show that it implies the existence of a generalized strong ordinal potential function. We use this characterization to derive existence, efficiency and fairness properties of strong Nash equilibria. We then study a class of games that generalizes congestion games with bottleneck objectives that we call bot- tleneck congestion games. We show that these games possess the LIP and thus the above mentioned properties. For bottleneck congestion games in networks, we identify cases in which the potential function associated with the LIP leads to polynomial time algorithms computing a strong Nash equilibrium. Finally, we investigate the LIP for infinite games. We show that the LIP does not imply the existence of a generalized strong ordinal potential, thus, the existence of SNE does not follow. Assuming that the function associated with the LIP is continuous, however, we prove existence of SNE. As a consequence, we prove that bottleneck congestion games with infinite strategy spaces and continuous cost functions possess a strong Nash equilibrium. 1 Introduction The theory of non-cooperative games is used to study situations that involve rational and selfish agents who are motivated by optimizing their own utilities rather than reaching some social optimum.
    [Show full text]
  • Cooperative and Non-Cooperative Game Theory Olivier Chatain
    HEC Paris From the SelectedWorks of Olivier Chatain 2014 Cooperative and Non-Cooperative Game Theory Olivier Chatain Available at: https://works.bepress.com/olivier_chatain/9/ The Palgrave Encyclopedia of Strategic Management Edited by Mie Augier and David J. Teece (http://www.palgrave.com/strategicmanagement/information/) Entry: Cooperative and non-cooperative game theory Author: Olivier Chatain Classifications: economics and strategy competitive advantage formal models methods/methodology Definition: Cooperative game theory focuses on how much players can appropriate given the value each coalition of player can create, while non-cooperative game theory focuses on which moves players should rationally make. Abstract: This article outlines the differences between cooperative and non-cooperative game theory. It introduces some of the main concepts of cooperative game theory as they apply to strategic management research. Keywords: cooperative game theory coalitional game theory value-based strategy core (solution concept) Shapley value; biform games Cooperative and non-cooperative game theory Game theory comprises two branches: cooperative game theory (CGT) and non-cooperative game theory (NCGT). CGT models how agents compete and cooperate as coalitions in unstructured interactions to create and capture value. NCGT models the actions of agents, maximizing their utility in a defined procedure, relying on a detailed description of the moves 1 and information available to each agent. CGT abstracts from these details and focuses on how the value creation abilities of each coalition of agents can bear on the agents’ ability to capture value. CGT can be thus called coalitional, while NCGT is procedural. Note that ‘cooperative’ and ‘non-cooperative’ are technical terms and are not an assessment of the degree of cooperation among agents in the model: a cooperative game can as much model extreme competition as a non-cooperative game can model cooperation.
    [Show full text]
  • What Is Game Theory Trying to Accomplish?
    1 What Is Game Theory Trying to Accomplish? 1 Introduction The language of game theory—coalitions, payo¤s, markets, votes— suggests that it is not a branch of abstract mathematics; that it is moti- vated by and related to the world around us; and that it should be able to tell us something about that world. Most of us have long realized that game theory and the ‘‘real world’’ (it might better be called the complex world) have a relationship that is not entirely comfortable; that it is not clear just what it is that we are trying to do when we build a game- theoretic model and then apply solution concepts to it. This is the subject I would like to explore in this paper. I might add that much the same questions apply to economic theory, at least the kind that those of us working in mathematical economics see most often; and that much of my paper will apply, mutatis mutandis, to economic theory as well. There is a branch of philosophy that deals with theory in the social sciences, so some of the things I have to say are unquestionably old hat. But I am not trying to be particularly original: I am only trying to open this topic, which I think concerns all of us, for discussion, and to suggest a particu- lar point of view. No doubt others have thought about these questions more thoroughly and deeply than I have, and are better versed in the history and philosophy of science in general. I will be grateful to anybody who sets me straight when I err, and who gives me references for the things I get right.
    [Show full text]