Recoverable Mutual Exclusion for Non-Volatile Main Memory Systems

Total Page:16

File Type:pdf, Size:1020Kb

Recoverable Mutual Exclusion for Non-Volatile Main Memory Systems View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by University of Waterloo's Institutional Repository RGLock: Recoverable Mutual Exclusion for Non-Volatile Main Memory Systems by Aditya Ramaraju A thesis presented to the University of Waterloo in fulfillment of the thesis requirement for the degree of Master of Applied Science in Electrical and Computer Engineering (Computer Software) Waterloo, Ontario, Canada, 2015 ©Aditya Ramaraju 2015 AUTHOR'S DECLARATION I hereby declare that I am the sole author of this thesis. This is a true copy of the thesis, including any required final revisions, as accepted by my examiners. I understand that my thesis may be made electronically available to the public. Aditya Ramaraju. ii Abstract Mutex locks have traditionally been the most popular concurrent programming mechanisms for inter- process synchronization in the rapidly advancing field of concurrent computing systems that support high- performance applications. However, the concept of recoverability of these algorithms in the event of a crash failure has not been studied thoroughly. Popular techniques like transaction roll-back are widely known for providing fault-tolerance in modern Database Management Systems. Whereas in the context of mutual exclusion in shared memory systems, none of the prominent lock algorithms (e.g., Lamport’s Bakery algorithm, MCS lock, etc.) are designed to tolerate crash failures, especially in operations carried out in the critical sections. Each of these algorithms may fail to maintain mutual exclusion, or sacrifice some of the liveness guarantees in presence of crash failures. Storing application data and recovery information in the primary storage with conventional volatile memory limits the development of efficient crash-recovery mechanisms since a failure on any component in the system causes a loss of program data. With the advent of Non-Volatile Main Memory technologies, opportunities have opened up to redefine the problem of Mutual Exclusion in the context of a crash-recovery model where processes may recover from crash failures and resume execution. When the main memory is non-volatile, an application’s entire state can be recovered from a crash using the in-memory state near-instantaneously, making a process’s failure appear as a suspend/resume event. This thesis proceeds to envision a solution for the problem of mutual exclusion in such systems. The goal is to provide a first-of-its-kind mutex lock that guarantees mutual exclusion and starvation freedom in emerging shared-memory architectures that incorporate non-volatile main memory (NVMM). iii Acknowledgements This thesis wouldn’t have been possible without the expertise, guidance, and the incredible patience of my advisor Prof. Wojciech Golab. The precision and non-dualism in everything that he says and writes is something I have continuously strived to inculcate throughout the duration of my MASc program. I shall always remain as the greatest admirer of his vast knowledge and skills in many areas, among which his technical writing is a class apart. Few people can inspire interest and passion towards the subject of Distributed Computing better than he does. I am deeply indebted to him for mentoring, educating, employing, and in a way, tirelessly parenting me during the last two years that made a significant difference in my life and shaped me as a better individual. I sincerely doubt that I will ever be able to express my appreciation fully, but I owe him my eternal gratitude. Every academic and professional achievement I ever make, starting with RGLock, is only a dedication and tribute to him. I owe my gratitude to Prof. Rajay Vedaraj at VIT University for all the knowledge sharing, constant encouragement, and inspiration he provided ever since my undergrad days. My sincere thanks to my friend Kartik Josyula for motivating me to pursue higher studies in Canada, and special thanks to Ankur Agarwal and Krunal Barot, my supervisors at IBM India, for duly encouraging me in this regard. I am also very thankful to my close friends, particularly Youcef Tebbal and Nicholas Petroff, for making my time in Waterloo so wonderful with as many thought-provoking discussions as the numerous mindless debates we ran into. I must also acknowledge the important role of my brother, Ramakrishna Rao Ramaraju, in being the primary source of inspiration for everything I have ever learnt about computers, not to mention the readily available financial assistance he provided whenever required. Finally, I am ever grateful to my parents for all their support and the sacrifices they have made for providing me with high quality education at only the best schools in every stage, of which, home remains my most favorite. iv Dedication To Annaya, and To my Guru. v Table of Contents AUTHOR'S DECLARATION ...................................................................................................................... ii Abstract ........................................................................................................................................................ iii Acknowledgements ...................................................................................................................................... iv Dedication ..................................................................................................................................................... v Table of Contents ......................................................................................................................................... vi List of Figures ............................................................................................................................................ viii List of Algorithms ........................................................................................................................................ ix Chapter 1 Introduction .................................................................................................................................. 1 1.1 Preliminaries ....................................................................................................................................... 2 1.2 Background & Motivation .................................................................................................................. 3 1.2.1 Spin-lock algorithms .................................................................................................................... 3 1.2.2 Crash-recoverable Mutex ............................................................................................................. 6 1.3 Summary of Results .......................................................................................................................... 10 Chapter 2 Related Work .............................................................................................................................. 11 Chapter 3 Model Specification ................................................................................................................... 17 3.1 Hardware Considerations .................................................................................................................. 17 3.2 Formalism ......................................................................................................................................... 18 vi 3.3 Crash-recoverable Execution ............................................................................................................ 19 Chapter 4 Recoverable Mutual Exclusion .................................................................................................. 22 Chapter 5 Algorithm ................................................................................................................................... 24 5.1 RGLock ............................................................................................................................................. 24 5.2 Swap and Store ................................................................................................................................. 30 5.3 Crash-recovery Procedures ............................................................................................................... 32 Chapter 6 Correctness ................................................................................................................................. 35 6.1 Notational Conventions .................................................................................................................... 35 6.2 Preliminaries ..................................................................................................................................... 36 6.3 Recoverable Mutual Exclusion ......................................................................................................... 41 Chapter 7 Conclusion and Future Work ..................................................................................................... 54 7.1 Summary ........................................................................................................................................... 54 7.2 Future Research ................................................................................................................................ 56 vii List of Figures Figure 3.1 Failure-free and crash-recoverable passages. ............................................................................ 21 Figure 5.1 Phase transitions of a process
Recommended publications
  • Scalable NUMA-Aware Blocking Synchronization Primitives
    Scalable NUMA-aware Blocking Synchronization Primitives Sanidhya Kashyap, Changwoo Min, Taesoo Kim The rise of big NUMA machines 'i The rise of big NUMA machines 'i The rise of big NUMA machines 'i Importance of NUMA awareness NUMA node 1 NUMA node 2 NUMA oblivious W1 W2 W3 W4 W6 W5 W6 W5 W3 W1 W4 L W2 File Importance of NUMA awareness NUMA node 1 NUMA node 2 NUMA oblivious W1 W2 W3 W4 W6 W5 W6 W5 W3 W1 W4 NUMA aware/hierarchical L W2 W1 W6 W2 W3 W4 W5 File Importance of NUMA awareness NUMA node 1 NUMA node 2 NUMA oblivious W1 W2 W3 W4 W6 W5 W6 W5 W3 W1 W4 NUMA aware/hierarchical L W2 W1 W6 W2 W3 W4 W5 File Idea: Make synchronization primitives NUMA aware! Lock's research efforts and their use Lock's research efforts Linux kernel lock Dekker's algorithm (1962) adoption / modification Semaphore (1965) Lamport's bakery algorithm (1974) Backoff lock (1989) Ticket lock (1991) MCS lock (1991) HBO lock (2003) Hierarchical lock – HCLH (2006) Flat combining NUMA lock (2011) Remote Core locking (2012) Cohort lock (2012) RW cohort lock (2013) Malthusian lock (2014) HMCS lock (2015) AHMCS lock(2016) Lock's research efforts and their use Lock's research efforts Linux kernel lock Dekker's algorithm (1962) adoption / modification Semaphore (1965) Lamport's bakery algorithm (1974) Backoff lock (1989) Ticket lock (1991) MCS lock (1991) HBO lock (2003) Hierarchical lock – HCLH (2006) Flat combining NUMA lock (2011) Remote Core locking (2012) NUMA- Cohort lock (2012) aware RW cohort lock (2013) locks Malthusian lock (2014) HMCS lock (2015) AHMCS lock(2016)
    [Show full text]
  • Non-Scalable Locks Are Dangerous
    Non-scalable locks are dangerous Silas Boyd-Wickizer, M. Frans Kaashoek, Robert Morris, and Nickolai Zeldovich MIT CSAIL Abstract more cores. The paper presents a predictive model of non-scalable lock performance that explains this phe- Several operating systems rely on non-scalable spin locks nomenon. for serialization. For example, the Linux kernel uses A third element of the argument is that critical sections ticket spin locks, even though scalable locks have better which appear to the eye to be very short, perhaps only theoretical properties. Using Linux on a 48-core ma- a few instructions, can nevertheless trigger performance chine, this paper shows that non-scalable locks can cause collapses. The paper’s model explains this phenomenon dramatic collapse in the performance of real workloads, as well. even for very short critical sections. The nature and sud- den onset of collapse are explained with a new Markov- Naturally we argue that one should use scalable locks [1, based performance model. Replacing the offending non- 7, 9], particularly in operating system kernels where scalable spin locks with scalable spin locks avoids the the workloads and levels of contention are hard to con- collapse and requires modest changes to source code. trol. As a demonstration, we replaced Linux’s spin locks with scalable MCS locks [9] and re-ran the software that caused performance collapse. For 3 of the 4 benchmarks 1 Introduction the changes in the kernel were simple. For the 4th case the changes were more involved because the directory It is well known that non-scalable locks, such as simple cache uses a complicated locking plan and the directory spin locks, have poor performance when highly con- cache in general is complicated.
    [Show full text]
  • Concurrency: Better Critical Section Solutions
    University of New Mexico Concurrency: Better Critical Section Solutions Prof. Patrick G. Bridges University of New Mexico Fetch-And-Add Atomically increment a value while returning the old value at a particular address. 1 int FetchAndAdd(int *ptr) { 2 int old = *ptr; 3 *ptr = old + 1; 4 return old; 5 } Fetch-And-Add Hardware atomic instruction (C-style) University of New Mexico Ticket Lock Ticket lock can be built with fetch-and add. ▪ Ensure progress for all threads. → fairness 1 typedef struct __lock_t { 2 int ticket; 3 int turn; 4 } lock_t; 5 6 void lock_init(lock_t *lock) { 7 lock->ticket = 0; 8 lock->turn = 0; 9 } 10 11 void lock(lock_t *lock) { 12 int myturn = FetchAndAdd(&lock->ticket); 13 while (lock->turn != myturn) 14 ; // spin 15 } 16 void unlock(lock_t *lock) { 17 FetchAndAdd(&lock->turn); 18 } University of New Mexico So Much Spinning Hardware-based spin locks are simple and they work. In some cases, these solutions can be quite inefficient. ▪ Any time a thread gets caught spinning, it wastes an entire time slice doing nothing but checking a value. How To Avoid Spinning? We’ll need OS Support too! University of New Mexico A Simple Approach: Just Yield When you are going to spin, give up the CPU to another thread. ▪ OS system call moves the caller from the running state to the ready state. ▪ The cost of a context switch can be substantial and the starvation problem still exists. 1 void init() { 2 flag = 0; 3 } 4 5 void lock() { 6 while (TestAndSet(&flag, 1) == 1) 7 yield(); // give up the CPU 8 } 9 10 void unlock() { 11 flag = 0; 12 } Lock with Test-and-set and Yield University of New Mexico Using Queues: Sleeping Instead of Spinning Queue to keep track of which threads are waiting to enter the lock.
    [Show full text]
  • Multicore Semantics & Programming
    Multicore Semantics & Programming Exercise sheet (Tim Harris’ section) Each question carries equal marks, for a total of 50% of the course total Core exercises (Part II and ACS/Part III) 1. Locks and performance A machine has 4 processors, each with its own cache. A program runs with 1 thread per processor, and 1 mutual exclusion lock used by all of the threads. Each thread must acquire the lock, execute a short critical section, and then release the lock. For each of the following locks (i) describe the cache line transfers that occur in a simple MESI cache protocol in the best-case execution in which threads happen to only attempt to acquire the lock when it is available, and (ii) describe any additional cache line transfers that may occur if threads attempt to acquire the lock at the same time. Test-and-set lock. Test-and-test-and-set lock. MCS lock. You may assume that, by chance, the thread on processor 1 successfully acquires the lock first, then processors 2, 3, and 4. Also, you may assume that each QNode structure for the MCS lock is a single cache line in size. 2. Lock implementation (i) In the slides, the pseudo-code for the MCS acquireMCS operation shows that the new QNode is only linked into the queue after performing a CAS operation on lock->tail. This makes the releaseMCS operation more complicated because it may need to wait in the while loop watching qn->next. Therefore it is tempting to look for ways to update the links between the QNodes before updating the tail of the list in lock->tail.
    [Show full text]
  • 28. Locks Operating System: Three Easy Pieces
    28. Locks Operating System: Three Easy Pieces Youjip Won 1 Locks: The Basic Idea Ensure that any critical section executes as if it were a single atomic instruction. An example: the canonical update of a shared variable balance = balance + 1; Add some code around the critical section 1 lock_t mutex; // some globally-allocated lock ‘mutex’ 2 … 3 lock(&mutex); 4 balance = balance + 1; 5 unlock(&mutex); Youjip Won 2 Locks: The Basic Idea Lock variable holds the state of the lock. available (or unlocked or free) No thread holds the lock. acquired (or locked or held) Exactly one thread holds the lock and presumably is in a critical section. Youjip Won 3 The semantics of the lock() lock() Try to acquire the lock. If no other thread holds the lock, the thread will acquire the lock. Enter the critical section. This thread is said to be the owner of the lock. Other threads are prevented from entering the critical section while the first thread that holds the lock is in there. Youjip Won 4 Pthread Locks - mutex The name that the POSIX library uses for a lock. Used to provide mutual exclusion between threads. 1 pthread_mutex_t lock = PTHREAD_MUTEX_INITIALIZER; 2 3 Pthread_mutex_lock(&lock); // wrapper for pthread_mutex_lock() 4 balance = balance + 1; 5 Pthread_mutex_unlock(&lock); We may be using different locks to protect different variables Increase concurrency (a more fine-grained approach). Youjip Won 5 Building A Lock Efficient locks provided mutual exclusion at low cost. Building a lock need some help from the hardware and the OS. Youjip Won 6 Evaluating locks – Basic criteria Mutual exclusion Does the lock work, preventing multiple threads from entering a critical section? Fairness Does each thread contending for the lock get a fair shot at acquiring it once it is free? (Starvation) Performance The time overheads added by using the lock Youjip Won 7 Controlling Interrupts Disable Interrupts for critical sections One of the earliest solutions used to provide mutual exclusion Invented for single-processor systems.
    [Show full text]
  • A Flexible and Scalable Affinity Lock for the Kernel
    A Flexible and Scalable Affinity Lock for the Kernel Benlong Zhang, Junbin Kang, Tianyu Wo, Yuda Wang and Renyu Yang State Key Laboratory of Software Development Environment Beihang University Beijing, P.R. of China fzhangbl, kangjb, woty, wangyuda, [email protected] Abstract—A number of NUMA-aware synchronization algo- large access latency discrepancies among the cores within a rithms have been proposed lately to stress the scalability in- single processor, because the latency to a nearby private cache efficiencies of existing locks. However their presupposed local can be as much as 40 times that of the local cache [3]. lock granularity, a physical processor, is often not the optimum configuration for various workloads. This paper further explores This paper extends the discussion of NUMA-aware locks the design space by taking into consideration the physical affinity and presents FSL: a flexible and scalable affinity lock for the between the cores within a single processor, and presents FSL kernel. The new lock design relies on the idea that the first- to support variable and finely tuned group size for different layer granularity of an affinity lock should be dynamically con- lock contexts and instances. The new design provides a uniform figured. It provides a uniform model to build affinity locks and model for the discussion of affinity locks and can completely explores the design space consistently and comprehensively. In subsume the previous NUMA-aware designs because they have this more general discussion, the group size in an affinity lock only discussed one special case of the model. The interfaces of the new scheme are kernel-compatible and thus largely facilitate should be as small as possible to maximize the exploitation kernel incorporation.
    [Show full text]
  • Non-Scalable Locks Are Dangerous
    Non-scalable locks are dangerous Silas Boyd-Wickizer, M. Frans Kaashoek, Robert Morris, and Nickolai Zeldovich MIT CSAIL Abstract more cores. The paper presents a predictive model of non-scalable lock performance that explains this phe- Several operating systems rely on non-scalable spin locks nomenon. for serialization. For example, the Linux kernel uses A third element of the argument is that critical sections ticket spin locks, even though scalable locks have better which appear to the eye to be very short, perhaps only theoretical properties. Using Linux on a 48-core ma- a few instructions, can nevertheless trigger performance chine, this paper shows that non-scalable locks can cause collapses. The paper’s model explains this phenomenon dramatic collapse in the performance of real workloads, as well. even for very short critical sections. The nature and sud- den onset of collapse are explained with a new Markov- Naturally we argue that one should use scalable locks [1, based performance model. Replacing the offending non- 7, 9], particularly in operating system kernels where scalable spin locks with scalable spin locks avoids the the workloads and levels of contention are hard to con- collapse and requires modest changes to source code. trol. As a demonstration, we replaced Linux’s spin locks with scalable MCS locks [9] and re-ran the software that caused performance collapse. For 3 of the 4 benchmarks 1 Introduction the changes in the kernel were simple. For the 4th case the changes were more involved because the directory It is well known that non-scalable locks, such as simple cache uses a complicated locking plan and the directory spin locks, have poor performance when highly con- cache in general is complicated.
    [Show full text]
  • Lock-Based Data Structures on Gpus with Independent Thread Scheduling
    Lock-based Data Structures on GPUs with Independent Thread Scheduling vorgelegt von Phillip Grote Matrikelnummer: 371570 dem Fachgebiet Datenbanksysteme und Informationsmanagement der Fakultät IV - Elektrotechnik und Informatik der Technischen Universität Berlin Bachelor Thesis February 3, 2020 Gutachter: Prof. Dr. Volker Markl Betreuer: Clemens Lutz Zeitraum: September 2019 - Februar 2020 Erklärung Hiermit erkläre ich, dass ich die vorliegende Arbeit selbstständig und eigenhändig sowie ohne unerlaubte fremde Hilfe und ausschließlich unter Verwendung der aufgeführten Quellen und Hilfsmittel angefertigt habe. I hereby declare that the thesis submitted is my own, unaided work, completed without any unpermitted external help. Only the sources and resources listed were used. Die selbständige und eigenhändige Anfertigung versichert an Eides statt: The independent and unaided completion of the thesis is affirmed by affidavit: Berlin, den 02.02.2020 Phillip Grote Abstract GPUs support thousands of concurrent threads and are able to accelerate throughput oriented workloads by an order of magnitude. In order to implement highly concurrent data structures, efficient synchronization is necessary. However, correct and efficient synchronization requires knowledge of the underlying hardware. Especially since the handling of control flow on GPUs can lead to livelock conditions, which prevent the system from making forward progress. In this thesis we investigate efficient synchronization techniques on GPUs with Independent Thread Scheduling. By exploring how the hardware handles control flow, we are able to show that Independent Thread Scheduling prevents SIMT deadlocks from happening. However, we identify situations in which livelocks occur on GPUs with Independent Thread Scheduling. To address this problem, we present a technique that prevents such livelocks from occurring.
    [Show full text]
  • A Program Logic for Concurrent Objects Under Fair Scheduling
    A Program Logic for Concurrent Objects under Fair Scheduling Hongjin Liang Xinyu Feng School of Computer Science and Technology & Suzhou Institute for Advanced Study University of Science and Technology of China [email protected] [email protected] Abstract In addition to linearizability, a safety property, object imple- mentations are expected to also satisfy progress properties. The Existing work on verifying concurrent objects is mostly concerned non-blocking progress properties, such as wait-freedom and lock- with safety only, e.g., partial correctness or linearizability. Although freedom which have been studied a lot (e.g., [5, 10, 16, 24]), guar- there has been recent work verifying lock-freedom of non-blocking antee the termination of the method calls independently of how the objects, much less efforts are focused on deadlock-freedom and threads are scheduled. Unfortunately these properties are too strong starvation-freedom, progress properties of blocking objects. These to be satisfied by algorithms with blocking synchronization. For properties are more challenging to verify than lock-freedom because clients using lock-based objects, a delay of a thread holding a lock they allow the progress of one thread to depend on the progress of will block other threads requesting the lock. Thus their progress another, assuming fair scheduling. relies on the assumption that every thread holding the lock will We propose LiLi, a new rely-guarantee style program logic for eventually be scheduled to release it. This assumption requires fair together verifying linearizability and progress for concurrent objects scheduling, i.e., every thread gets eventually executed. As summa- under fair scheduling.
    [Show full text]
  • 15-418/618, Spring 2018 Exam 2 Practice SOLUTIONS
    15-418/618, Spring 2018 Exam 2 Practice SOLUTIONS April, 2018 Warm Up: Miscellaneous Short Problems Problem 1. (21 points): A. (3 pts) Consider the following code where two threads are running on different cores of a two-core processor that features relaxed memory consistency (assume all variables start out with value 0). The system allows write-after-write and read-after-write memory operations to be reordered. (You should assume that the compiler itself performs no instruction reordering). Thread 1 Thread 2 ============================================= A = 1 while (flag == 0) {} flag = 1 print A print A Is it possible for Thread 1 to print the value 0 on this system? Is it possible for Thread 2 print the value 0? In both cases, please justify your answer. Solution: It is not possible for thread 1 to print 0, since the statement print A has a program order dependency on the proceeding statement A=1. Even if a system features relaxed memory consistency, it will not ever violate program order. It is possible for thread 2 to print 0. Since the system relaxes write after write ordering, thread 2 may observe the write to flag prior to observing the write to A. (Relaxed consistency troubled students on the midterm so as advertised we brought it back up on Exam 2 again.) B. (3 pts) Your job is to implement an iPhone app that continuously processes images from the camera. To provide a good experience, the app must run at a 30 fps. While developing the app, you always test it on a benchmark that contains 100 frames.
    [Show full text]
  • High Performance Synchronization Algorithms for Multiprogrammed Multiprocessors
    High Performance Synchronization Algorithms for Multiprogrammed Multiprocessors Robert W. Wisniewski, Leonidas I. Kontothanassis, and Michael L. Scott Department of Computer Science University of Rochester Rochester, NY 14627-0226 fbob,kthanasi,[email protected] Abstract Most busy-wait synchronization algorithms assume a dedicated machine with one process per processor. The Scalable busy-wait synchronization algorithms are es- introduction of multiprogramming can cause severe per- sential for achieving good parallel program performance formance degradation for these algorithms, especially on large scale multiprocessors. Such algorithms include for the variants designed to scale well on dedicated ma- mutual exclusion locks, reader-writer locks, and bar- chines. Performance degrades in the presence of multi- rier synchronization. Unfortunately, scalable synchro- programming under the following circumstances: nization algorithms are particularly sensitive to the ef- fects of multiprogramming: their performance degrades • A process is preempted while holding a lock. sharply when processors are shared among different ap- This situation arises in both mutual exclusion and plications, or even among processes of the same applica- reader-writer locks when a process is preempted in tion. In this paper we describe the design and evaluation its critical section. It has been addressed by several of scalable scheduler-conscious mutual exclusion locks, researchers [2, 4, 7, 15]. reader-writer locks, and barriers, and show that by shar- ing information across the kernel/application interface • A process is preempted while waiting for a lock and we can improve the performance of scheduler-oblivious then is handed the lock while still preempted. This implementations by more than an order of magnitude. situation can arise in locks that enforce a predeter- mined ordering, either for the sake of fairness or to minimize contention on large-scale machines [21].
    [Show full text]
  • Progress of Concurrent Objects with Partial Methods
    Progress of Concurrent Objects with Partial Methods ∗ HONGJIN LIANG and XINYU FENG , Nanjing University, China and University of Science and Technology of China, China Various progress properties have been proposed for concurrent objects, such as wait-freedom, lock-freedom, starvation-freedom and deadlock-freedom. However, none of them applies to concurrent objects with partial methods, i.e., methods that are supposed not to return under certain circumstances. A typical example is the lock_acquire method, which must not return when the lock has already been acquired. In this paper we propose two new progress properties, partial starvation-freedom (PSF) and partial deadlock- freedom (PDF), for concurrent objects with partial methods. We also design four patterns to write abstract specifications for PSF or PDF objects under strongly or weakly fair scheduling, so that these objects contextually refine the abstract specifications. Our Abstraction Theorem shows the equivalence between PSF (orPDF)and the progress-aware contextual refinement. Finally, we generalize the program logic LiLi to have a new logicto verify the PSF (or PDF) property and linearizability of concurrent objects. CCS Concepts: • Theory of computation → Hoare logic; Program specifications; Program verifica- tion; Abstraction; Additional Key Words and Phrases: Concurrent Objects, Progress, Refinement, Partial Methods ACM Reference Format: Hongjin Liang and Xinyu Feng. 2018. Progress of Concurrent Objects with Partial Methods. Proc. ACM Program. 20 Lang. 2, POPL, Article 20 (January 2018), 31 pages. https://doi.org/10.1145/3158108 1 INTRODUCTION A concurrent object consists of shared data and a set of methods which provide an interface for client threads to access the shared data.
    [Show full text]