Minimax Group Fairness: Algorithms and Experiments

Total Page:16

File Type:pdf, Size:1020Kb

Minimax Group Fairness: Algorithms and Experiments Minimax Group Fairness: Algorithms and Experiments Emily Diana Wesley Gill Michael Kearns University of Pennsylvania University of Pennsylvania University of Pennsylvania Amazon AWS AI Amazon AWS AI Amazon AWS AI [email protected] [email protected] [email protected] Krishnaram Kenthapadi Aaron Roth Amazon AWS AI University of Pennsylvania [email protected] Amazon AWS AI [email protected] ABSTRACT of the most intuitive and well-studied group fairness notions, and We consider a recently introduced framework in which fairness in enforcing it one often implicitly hopes that higher error rates is measured by worst-case outcomes across groups, rather than by on protected or “disadvantaged” groups will be reduced towards the more standard differences between group outcomes. In this the lower error rate of the majority or “advantaged” group. But framework we provide provably convergent oracle-efficient learn- in practice, equalizing error rates and similar notions may require ing algorithms (or equivalently, reductions to non-fair learning) artificially inflating error on easier-to-predict groups — without for minimax group fairness. Here the goal is that of minimizing necessarily decreasing the error for the harder to predict groups — the maximum error across all groups, rather than equalizing group and this may be undesirable for a variety of reasons. errors. Our algorithms apply to both regression and classification For example, consider the many social applications of machine settings and support both overall error and false positive or false learning in which most or even all of the targeted population is negative rates as the fairness measure of interest. They also support disadvantaged, such as predicting domestic situations in which relaxations of the fairness constraints, thus permitting study of the children may be at risk of physical or emotional harm [8]. While tradeoff between overall accuracy and minimax fairness. We com- we might be interested in ensuring that our predictions are roughly pare the experimental behavior and performance of our algorithms equally accurate across racial groups, income levels, or geographic across a variety of fairness-sensitive data sets and show empirical location, if this can only be achieved by raising lower group error cases in which minimax fairness is strictly and strongly preferable rates without lowering the error for any other population, then to equal outcome notions. arguably we will have only worsened overall social welfare, since this is not a setting where we can argue that we are “taking from CCS CONCEPTS the rich and giving to the poor.” Similar arguments can be made in other high-stakes applications, such as predictive modeling for • Theory of computation ! Design and analysis of algorithms. medical care. In these settings it might be preferable to consider KEYWORDS the alternative fairness criterion of achieving minimax group error, in which we seek not to equalize error rates, but to minimize the Fair Machine Learning; Minimax Fairness; Game Theory largest group error rate — that is, to make sure that the worst-off ACM Reference Format: group is as well-off as possible. Emily Diana, Wesley Gill, Michael Kearns, Krishnaram Kenthapadi, and Aaron Minimax group fairness, which was recently proposed by [21] in Roth. 2021. Minimax Group Fairness: Algorithms and Experiments. In Pro- the context of classification, has the property that any model that ceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society (AIES achieves it Pareto dominates (at least weakly, and possibly strictly) ’21), May 19–21, 2021, Virtual Event, USA. ACM, New York, NY, USA, 11 pages. an equalized-error model with respect to group error rates — that https://doi.org/10.1145/3461702.3462523 0 is, if 68 is the error rate on group 8 in a minimax solution, and 6 is the common error rate for all groups in a solution that equalizes 0 1 INTRODUCTION group error rates, then 6 ≥ 68 for all 8. If one or more of these Machine learning researchers and practitioners have often focused inequalities is strict, it constitutes a proof that equalized errors can on achieving group fairness with respect to protected attributes only be achieved by deliberately inflating the error of one or more such as race, gender, or ethnicity. Equality of error rates is one groups in the minimax solution. Put another way, one technique for finding an optimal solution subject to an equality of group error Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed rates constraint is to first find a minimax solution, and thento for profit or commercial advantage and that copies bear this notice and the full citation artifically inflate the error rate on any group that does not saturate on the first page. Copyrights for components of this work owned by others than ACM the minimax constraint — an “optimal algorithm” that makes plain must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a the deficiencies of equal error solutions. fee. Request permissions from [email protected]. In contrast to approaches that learn separate models for each AIES ’21, May 19–21, 2021, Virtual Event, USA protected group — which also satisfy minimax error, simply by © 2021 Association for Computing Machinery. ACM ISBN 978-1-4503-8473-5/21/05...$15.00 https://doi.org/10.1145/3461702.3462523 optimizing each group independently — (e.g. [11, 30]), the minimax set of non-convex losses. In applying similar techniques to a fair approach we use here has two key advantages: classification domain, our conceptual contributions include the • The minimax approach does not require that groups of in- ability to handle overlapping groups. terest be disjoint, which is a requirement for the approach Our work builds upon the notion of minimax fairness proposed of learning a different model for each group. This allows for by Martinez et al. [21] in the context of classification. Their algo- protecting groups defined by intersectional characteristics as rithm – Approximate Projection onto Start Sets (APStar) – is also in [17, 18], protecting (for example) not just groups defined an iterative process which alternates between finding a model that by race or gender alone, but also by combinations of race minimizes risk and updating group weightings using a variant of and gender. gradient descent, but they provide no convergence analysis. More • The minimax approach does not require that protected at- importantly, APStar relies on knowledge of the base distributions tributes be given as inputs to the trained model. This can be of the samples (while our algorithms do not) and it does not provide extremely important in domains (like credit and insurance) the capability to relax the minimax constraint and explore an error in which using protected attributes as features for prediction vs. fairness tradeoff. [21] analyze only the classification setting, but is illegal. we provide theory and perform experiments in both classification and regression settings. Because our meta-algorithm is easily exten- Our primary contributions are as follows: sible, we are able to generalize to non-disjoint groups and various (1) First, we propose two algorithms. The first finds a minimax error types (with an emphasis on false positive and false negative group fair model from a given statistical class, and the second rates). navigates tradeoffs between a relaxed notion of minimax Achieving minimax fairness over a large number of groups has fairness and overall accuracy. been proposed by [20] as a technique for achieving fairness when (2) Second, we prove that both algorithms converge and are or- protected group labels are not available. Our work relates to [20] acle efficient — meaning that they can be viewed as efficient in much the same way as it relates to [21], in that [20] is purely reductions to the problem of unconstrained (nonfair) learn- empirical, whereas we provide a formal analysis of our algorithm, ing over the same class. We also study their generalization as well as a version of the algorithm that allows us to relax the properties. minimax constraint and explore performance tradeoffs. (3) Third, we show how our framework can be easily extended Technically, our work uses a similar approach to [1, 2], which also to incorporate different types of error rates – such as false reduce a “fair” classification problem to a sequence of unconstrained positive and negative rates – and explain how to handle cost-sensitive classification problems via the introduction of a zero- overlapping groups. sum game. Crucial to this line of work is the connection between (4) Finally, we provide a thorough experimental analysis of our no regret learning and equilibrium convergence, originally due to two algorithms under different prediction regimes. In this [14]. section, we focus on the following: • We start with a demonstration of the learning process of 2 FRAMEWORK AND PRELIMINARIES learning a fair predictor from the class of linear regression We consider pairs of feature vectors and labels ¹G ,~ º= , where G models. This setting matches our theory exactly, because 8 8 8=1 8 is a feature vector, divided into groups f퐺 , . ,퐺 g. We choose weighted least squares regression is a convex problem, 1 a class 퐻 of models (either classification or regression models), and so we really do have efficient subroutines for the mapping features to predicted labels, with loss function ! taking unconstrained (nonfair) learning problem. values in »0, 1¼, average population error: • We conduct an exploration of the fairness vs. accuracy = tradeoff for regression and highlight an example in which 1 Õ n¹ℎº = !¹ℎ¹G º,~ º our minimax algorithms provide a substantial Pareto im- = 8 8 provement over the equality of error rates notion.
Recommended publications
  • Uniqueness and Symmetry in Bargaining Theories of Justice
    Philos Stud DOI 10.1007/s11098-013-0121-y Uniqueness and symmetry in bargaining theories of justice John Thrasher Ó Springer Science+Business Media Dordrecht 2013 Abstract For contractarians, justice is the result of a rational bargain. The goal is to show that the rules of justice are consistent with rationality. The two most important bargaining theories of justice are David Gauthier’s and those that use the Nash’s bargaining solution. I argue that both of these approaches are fatally undermined by their reliance on a symmetry condition. Symmetry is a substantive constraint, not an implication of rationality. I argue that using symmetry to generate uniqueness undermines the goal of bargaining theories of justice. Keywords David Gauthier Á John Nash Á John Harsanyi Á Thomas Schelling Á Bargaining Á Symmetry Throughout the last century and into this one, many philosophers modeled justice as a bargaining problem between rational agents. Even those who did not explicitly use a bargaining problem as their model, most notably Rawls, incorporated many of the concepts and techniques from bargaining theories into their understanding of what a theory of justice should look like. This allowed them to use the powerful tools of game theory to justify their various theories of distributive justice. The debates between partisans of different theories of distributive justice has tended to be over the respective benefits of each particular bargaining solution and whether or not the solution to the bargaining problem matches our pre-theoretical intuitions about justice. There is, however, a more serious problem that has effectively been ignored since economists originally J.
    [Show full text]
  • Fair Division Michael Meurer Boston Univeristy School of Law
    Boston University School of Law Scholarly Commons at Boston University School of Law Faculty Scholarship 1999 Fair Division Michael Meurer Boston Univeristy School of Law Follow this and additional works at: https://scholarship.law.bu.edu/faculty_scholarship Part of the Gaming Law Commons Recommended Citation Michael Meurer, Fair Division, 47 Buffalo Law Review 937 (1999). Available at: https://scholarship.law.bu.edu/faculty_scholarship/530 This Article is brought to you for free and open access by Scholarly Commons at Boston University School of Law. It has been accepted for inclusion in Faculty Scholarship by an authorized administrator of Scholarly Commons at Boston University School of Law. For more information, please contact [email protected]. BOSTON UNIVERSITY SCHOOL OF LAW WORKING PAPER SERIES, LAW & ECONOMICS WORKING PAPER NO. 99-10 FAIR DIVISION MICHAEL MEURER THIS PAPER APPEARS IN 47 BUFFALO LAW REVIEW 937 AND IS REPRINTED BY PERMISSION OF THE AUTHOR AND BUFFALO LAW REVIEW This paper can be downloaded without charge at: The Boston University School of Law Working Paper Series Index: http://www.bu.edu/law/faculty/papers The Social Science Research Network Electronic Paper Collection: http://papers.ssrn.com/paper.taf?abstract_id=214949 BOOK REVIEW Fair Division † MICHAEL J. MEURER appearing in 47 BUFFALO LAW REVIEW 937-74 (1999) INTRODUCTION Fair division is a fundamental issue of legal policy. The law of remedies specifies damages and rules of contribution that apportion liability among multiple defendants.1 Probate law specifies
    [Show full text]
  • Endgame Solving
    Endgame Solving: The Surprising Breakthrough that Enabled Superhuman Two-Player No- Limit Texas Hold 'em Play Sam Ganzfried Assistant Professor, Computer Science, Florida International University, Miami FL PhD, Computer Science Department, Carnegie Mellon University, 2015 [email protected] 1 2 3 4 5 Milestones • Opponent Modelling in Poker, Billings et.al., ‘98 • Abstraction Methods for Game-Theoretic Poker, Shi/Littman, ‘00 • Approximating Game-Theoretic Optimal Strategies for Full-scale Poker, Billings et.al., ‘03 • Optimal Rhode Island Poker, Gilpin/Sandholm, ‘05 • Annual Computer Poker Competition ‘06-Present • EGT/automated abstraction algorithms, Gilpin/Sandholm ‘06-‘08 • Regret Minimization in Games with Incomplete Information, Zinkevich et.al., ‘07 • Man vs. Machine limit Texas hold ‘em competitions ‘08-’09 • Computer Poker & Imperfect Information Symposium/Workshop ‘12-Present • Heads-up Limit Hold'em Poker is Solved, Bowling et.al., ‘15 • Brains vs. AI no-limit Texas hold ‘em competition ’15 • First Computer Poker Tutorial ‘16 • DeepStack: Expert-Level Artificial Intelligence in No-Limit Poker ’17 • Second Brains vs. AI no-limit Texas hold ‘em competition ‘17 6 Scope and applicability of game theory • Strategic multiagent interactions occur in all fields – Economics and business: bidding in auctions, offers in negotiations – Political science/law: fair division of resources, e.g., divorce settlements – Biology/medicine: robust diabetes management (robustness against “adversarial” selection of parameters in MDP) – Computer
    [Show full text]
  • Publications of Agnes Cseh
    Publications of Agnes Cseh This document lists all peer-reviewed publications of Agnes Cseh, Chair for Algorithm Engineering, Hasso Plattner Institute, Potsdam, Germany. This listing was automatically generated on October 6, 2021. An up-to-date version is available online at hpi.de/friedrich/docs/publist/cseh.pdf. Journal articles [1] Cseh, Á., Juhos, A., Pairwise Preferences in the Stable Marriage Problem. In: ACM Transactions on Economics and Computation (TEAC) 9, pp. 1–28, 2021. [2] Cseh, Á., Kavitha, T., Popular matchings in complete graphs. In: Algorithmica, pp. 1–31, 2021. [3] Andersson, T., Cseh, Á., Ehlers, L., Erlanson, A., Organizing time exchanges: Lessons from matching markets. In: American Economic Journal: Microeconomics 13, pp. 338–73, 2021. [4] Cseh, Á., Fleiner, T., The Complexity of Cake Cutting with Unequal Shares. In: ACM Transactions on Algorithms 16, pp. 1–21, 2020. An unceasing problem of our prevailing society is the fair division of goods. The problem of proportional cake cutting focuses on dividing a heterogeneous and divisible resource, the cake, among n players who value pieces according to their own measure function. The goal is to assign each player a not necessarily connected part of the cake that the player evaluates at least as much as her proportional share. In this article, we investigate the problem of proportional division with unequal shares, where each player is entitled to receive a predetermined portion of the cake. Our main contribution is threefold. First we present a protocol for integer demands, which delivers a proportional solution in fewer queries than all known proto- cols.
    [Show full text]
  • On Achieving Fairness and Stability in Many-To-One Matchings
    On Achieving Fairness and Stability in Many-to-One Matchings Shivika Narang1, Arpita Biswas2, and Y Narahari1 1Indian Institute of Science (shivika, narahari @iisc.ac.in) 2Harvard University ([email protected]) Abstract The past few years have seen a surge of work on fairness in social choice literature. A major portion of this work has focused on allocation settings where items must be allocated in a fair manner to agents having their own individual preferences. In comparison, fairness in two- sided settings where agents have to be matched to other agents, with preferences on both sides, has received much less attention. Moreover, two-sided matching literature has mostly focused on ordinal preferences. This paper initiates the study of finding a stable many-to-one matching, under cardinal valuations, while satisfying fairness among the agents on either side. Specifically, motivated by several real-world settings, we focus on leximin optimal fairness and seek leximin optimality over many-to-one stable matchings. We first consider the special case of ranked valuations where all agents on each side have the same preference orders or rankings over the agents on the other side (but not necessarily the same valuations). For this special case, we provide a complete characterisation of the space of stable matchings. This leads to FaSt, a novel and efficient algorithm to compute a leximin optimal stable matching under ranked isometric valuations (where, for each pair of agents, the valuation of one agent for the other is the same). The running time of FaSt is linear in the number of edges.
    [Show full text]
  • Efficiency and Sequenceability in Fair Division of Indivisible Goods With
    Efficiency and Sequenceability in Fair Division of Indivisible Goods with Additive Preferences Sylvain Bouveret, Michel Lemaître To cite this version: Sylvain Bouveret, Michel Lemaître. Efficiency and Sequenceability in Fair Division of Indivisible Goods with Additive Preferences. COMSOC-2016, Jun 2016, Toulouse, France. hal-01293121 HAL Id: hal-01293121 https://hal.archives-ouvertes.fr/hal-01293121 Submitted on 5 Apr 2016 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Efficiency and Sequenceability in Fair Division of Indivisible Goods with Additive Preferences Sylvain Bouveret and Michel Lemaˆıtre Abstract In fair division of indivisible goods, using sequences of sincere choices (or picking sequences) is a natural way to allocate the objects. The idea is the following: at each stage, a designated agent picks one object among those that remain. This paper, restricted to the case where the agents have numerical additive preferences over objects, revisits to some extent the seminal paper by Brams and King [9] which was specific to ordinal and linear order preferences over items. We point out similarities and differences with this latter context. In particular, we show that any Pareto- optimal allocation (under additive preferences) is sequenceable, but that the converse is not true anymore.
    [Show full text]
  • Fair Division with Bounded Sharing
    Fair Division with Bounded Sharing Erel Segal-Halevi∗ Abstract A set of objects is to be divided fairly among agents with different tastes, modeled by additive utility-functions. If the objects cannot be shared, so that each of them must be entirely allocated to a single agent, then fair division may not exist. How many objects must be shared between two or more agents in order to attain a fair division? The paper studies various notions of fairness, such as proportionality, envy-freeness and equitability. It also studies consensus division, in which each agent assigns the same value to all bundles | a notion that is useful in truthful fair division mechanisms. It proves upper bounds on the number of required sharings. However, it shows that finding the minimum number of sharings is NP-hard even for generic instances. Many problems remain open. 1 Introduction What is a fair way to allocate objects without monetary transfers? When the objects are indivisible, it may be impossible to allocate them fairly | consider a single object and two people. A common approach to this problem is to look for an approximately- fair allocation. There are several definitions of approximate fairness, the most common of which are envy-freeness except one object (EF1) and maximin share (MMS). An alternative solution is to \make objects divisible" by allowing randomization and ensure that the division is fair ex-ante. While approximate or ex-ante fairness are reasonable when allocating low-value objects, such as seats in a course or in a school, they are not suitable for high-value objects, e.g., houses or precious jewels.
    [Show full text]
  • Tutorial on Fair Division Ulle Endriss Institute for Logic, Language and Computation University of Amsterdam
    Fair Division COST-ADT School 2010 Tutorial on Fair Division Ulle Endriss Institute for Logic, Language and Computation University of Amsterdam " # COST-ADT Doctoral School on Computational Social Choice Estoril, Portugal, 9–14 April 2010 (http://algodec.org) Ulle Endriss 1 Fair Division COST-ADT School 2010 Table of Contents Introduction . 3 Fairness and Efficiency Criteria . 8 Divisible Goods: Cake-Cutting Procedures . 32 Indivisible Goods: Combinatorial Optimisation . 48 Conclusion . 68 Ulle Endriss 2 Fair Division COST-ADT School 2010 Introduction Ulle Endriss 3 Fair Division COST-ADT School 2010 Fair Division Fair division is the problem of dividing one or several goods amongst two or more agents in a way that satisfies a suitable fairness criterion. • Traditionally studied in economics (and to some extent also in mathematics, philosophy, and political science); now also in computer science (particularly multiagent systems and AI ). • Abstract problem, but immediately relevant to many applications. Ulle Endriss 4 Fair Division COST-ADT School 2010 Fair Division and Social Choice Fair division can be considered a problem of social choice: • A group of agents each have individual preferences over a collective agreement (the allocation of goods to be found). • But: in fair division preferences are often assumed to be cardinal (utility functions) rather than ordinal (as in voting) • And: fair division problems come with some internal structure often absent from other social choice problems (e.g., I will be indifferent between allocations giving me the same set of goods) Ulle Endriss 5 Fair Division COST-ADT School 2010 The Problem Consider a set of agents and a set of goods.
    [Show full text]
  • Fair Division and Cake Cutting
    Fair Division and Cake Cutting Subhash Suri February 27, 2020 1 Fair Division • Questions about fairness have tantalized thinkers for millennia, going back to the times of ancient Greeks and the Bible. In one of the famous biblical stories, King Solomon attempts to fairly resolve a maternity dispute between two women, both claiming to the true mother of a baby, by asking the guards to cut the baby in two and give each woman a half. The baby is finally given to the woman who would rather give the whole baby to the other woman than have it be cut in half. • One of the classic early versions of fair division occurs in cutting a cake to be split between two children. A commonly accepted protocol (solution) for this problem is called the \you cut, I choose:" one child cuts the cake into two \seemingly" equal halves, and the second child gets to pick the piece. 1. Intuitively, this seems like a fair division, but is there a formal way to prove it? 2. If so, what assumptions about the cake (goods) are needed in this algorithm? 3. Can this protocol be extended to n-way division, for any n ≥ 3? 4. Are there more than one way to define fairness? 5. These are some of the questions we plan to study in this lecture. • Cake-cutting may seem like a problem in recreational mathematics, but ideas behind these algorithms have been applied to a number of important real-world problems, including international negotiations and divorce settlements, and there are even com- mercial companies, such as Fair Outcomes, based on cake cutting algorithms.
    [Show full text]
  • Math 180: Course Summary Spring 2011, Prof. Jauregui
    Math 180: Course Summary Spring 2011, Prof. Jauregui • Probability { terminology: experiments, outcomes, events, sample space, probabilities { operations: OR, AND, NOT { mutually exclusive events { independent events, multiplication rule ∗ Case: People v. Collins { rules of probability ∗ addition rule, inclusion{exclusion ∗ subset rule, complement rule { coin flip examples (e.g., at least two heads out in 10 flips) { birthday problem { fifty-fifty fallacy { possibility of multiple lottery winners { \law of averages" fallacy { Conditional probability (definition and formula) ∗ multiplication rule for conditional probabilities { Bayes' formula ∗ disease testing, true/false negatives/positives ∗ general misunderstanding of conditional probability ∗ applications to DNA matching, determination of guilt { Prosecutor's fallacy, Sally Clark case { Defense attorney's fallacy, DNA matching, O.J. Simpson trial { Monty Hall Problem { expected value, variance, standard deviation { Bernoulli trials, special formulas for mean and variance • Statistics { bell-shaped curve/normal distribution, interpretation of µ and σ { 68 − 95 − 99:7 rule { z-scores, using the table { gathering data, computing mean and median, quartiles, IQR { Simpson's paradox (batting averages, alleged discrimination case) 1 { simple random samples ∗ sources of bias (opportunity sample, response bias, sampling bias, etc.) { yes/no surveys ∗ p andp ^, mean and standard deviation ofp ^ ∗ margin-of-error, confidence intervals • Decision theory { Drawing and solving decision trees (branches,
    [Show full text]
  • Fair Division of a Graph Sylvain Bouveret Katar´Ina Cechlarov´ A´ LIG - Grenoble INP, France P.J
    Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence (IJCAI-17) Fair Division of a Graph Sylvain Bouveret Katar´ına Cechlarov´ a´ LIG - Grenoble INP, France P.J. Safˇ arik´ University, Slovakia [email protected] [email protected] Edith Elkind Ayumi Igarashi Dominik Peters University of Oxford, UK University of Oxford, UK University of Oxford, UK [email protected] [email protected] [email protected] Abstract to members of an academic department, where there are de- pendencies among tasks (e.g., dealing with incoming foreign We consider fair allocation of indivisible items un- students has some overlap with preparing study programmes der an additional constraint: there is an undirected in foreign languages, but not with fire safety). graph describing the relationship between the items, and each agent’s share must form a connected sub- Our contribution We propose a framework for fair division graph of this graph. This framework captures, e.g., under connectivity constraints, and investigate the complexity fair allocation of land plots, where the graph de- of finding good allocations in this framework according to scribes the accessibility relation among the plots. three well-studied solution concepts: proportionality, envy- We focus on agents that have additive utilities for freeness (in conjunction with completeness), and maximin the items, and consider several common fair di- share guarantee. We focus on additive utility functions. vision solution concepts, such as proportionality, For proportionality and envy-freeness, we obtain hardness envy-freeness and maximin share guarantee. While results even for very simple graphs: finding proportional al- finding good allocations according to these solution locations turns out to be NP-hard even for paths, and finding concepts is computationally hard in general, we de- complete envy-free allocations is NP-hard both for paths and sign efficient algorithms for special cases where the for stars.
    [Show full text]
  • Procedurally Fair and Stable Matching∗
    Procedurally Fair and Stable Matching¤ Bettina Klausy Flip Klijnz November 2004 Abstract: We motivate procedural fairness for matching mechanisms and study two procedu- rally fair and stable mechanisms: employment by lotto (Aldershof et al., 1999) and the random order mechanism (Roth and Vande Vate, 1990, Ma, 1996). For both mechanisms we give vari- ous examples of probability distributions on the set of stable matchings and discuss properties that di®erentiate employment by lotto and the random order mechanism. Finally, we consider an adjustment of the random order mechanism, the equitable random order mechanism, that combines aspects of procedural and \endstate" fairness. Keywords: procedural fairness, random mechanism, stability, two-sided matching. JEL classi¯cation: C78, D63 1 Introduction The marriage model describes a two-sided, one-to-one matching market without money where the two sides of the market for instance are workers and ¯rms (job matching) or medical students and hospitals (matching of students to internships). We use the common terminology in the literature and refer to one side of the market as \men" and to the other as \women." An outcome for a marriage market is called a matching, which can simply be described by a collection of single agents and \married" pairs (consisting of one man and one woman). Loosely speaking, a matching is stable if all agents have acceptable spouses and there is no couple whose members both like each other better than their current spouses. Gale and Shapley (1962) formalized this notion of stability for marriage markets and provided an algorithm to calculate stable matchings. These classical results (Gale and Shapley, 1962) inspired many researchers to study stability not only for the marriage model, but for more general models as well.
    [Show full text]