Proportional Share Scheduling for Uniprocessor and Multiprocessor Systems

Total Page:16

File Type:pdf, Size:1020Kb

Proportional Share Scheduling for Uniprocessor and Multiprocessor Systems Group Ratio Round-Robin: O(1) Proportional Share Scheduling for Uniprocessor and Multiprocessor Systems Bogdan Caprita, Wong Chun Chan, Jason Nieh, Clifford Stein,∗ and Haoqiang Zheng Department of Computer Science Columbia University Email: {bc2008, wc164, nieh, cliff, hzheng}@cs.columbia.edu Abstract We present Group Ratio Round-Robin (GR3), the first pro- work has been done to provide proportional share schedul- portional share scheduler that combines accurate propor- ing on multiprocessor systems, which are increasingly com- tional fairness scheduling behavior with O(1) scheduling mon especially in small-scale configurations with two or overhead on both uniprocessor and multiprocessor systems. four processors. Over the years, a number of scheduling GR3 uses a simple grouping strategy to organize clients mechanisms have been proposed, and much progress has into groups of similar processor allocations which can be been made. However, previous mechanisms have either su- more easily scheduled. Using this strategy, GR3 combines perconstant overhead or less-than-ideal fairness properties. the benefits of low overhead round-robin execution with a We introduce Group Ratio Round-Robin (GR3), the 3 novel ratio-based scheduling algorithm. GR introduces a first proportional share scheduler that provides constant novel frontlog mechanism and weight readjustment algo- fairness bounds on proportional sharing accuracy with O(1) 3 rithm to operate effectively on multiprocessors. GR pro- scheduling overhead for both uniprocessor and small-scale vides fairness within a constant factor of the ideal general- multiprocessor systems. In designing GR3, we observed ized processor sharing model for client weights with a fixed that accurate, low-overhead proportional sharing is easy to upper bound and preserves its fairness properties on multi- achieve when scheduling a set of clients with equal pro- 3 processor systems. We have implemented GR in Linux cessor allocations, but is harder to do when clients require and measured its performance. Our experimental results very different allocations. Based on this observation, GR3 3 show that GR provides much lower scheduling overhead uses a simple client grouping strategy to organize clients and much better scheduling accuracy than other schedulers into groups of similar processor allocations which can be commonly used in research and practice. more easily scheduled. Using this grouping strategy, GR3 combines the benefits of low overhead round-robin execu- 1 Introduction tion with a novel ratio-based scheduling algorithm. GR3 uses the same basic uniprocessor scheduling al- Proportional share resource management provides a flexible gorithm for multiprocessor scheduling by introducing the and useful abstraction for multiplexing processor resources notion of a frontlog. On a multiprocessor system, a client among a set of clients. Proportional share scheduling has a may not be able to be scheduled to run on a processor be- clear colloquial meaning: given a set of clients with asso- cause it is currently running on another processor. To pre- ciated weights, a proportional share scheduler should allo- serve its fairness properties, GR3 keeps track of a frontlog cate resources to each client in proportion to its respective per client to indicate when the client was already running weight. However, developing processor scheduling mech- but could have been scheduled to run on another processor. anisms that combine good proportional fairness scheduling It then assigns the client a time quantum that is added to its behavior with low scheduling overhead has been difficult to allocation on the processor it is running on. The frontlog achieve in practice. For many proportional share scheduling ensures that a client receives its proportional share alloca- mechanisms, the time to select a client for execution grows tion while also taking advantage of any cache affinity by with the number of clients. For server systems which may continuing to run the client on the same processor. service large numbers of clients, the scheduling overhead of 3 algorithms whose complexity grows linearly with the num- GR provides a simple weight readjustment algo- ber of clients can waste more than 20 percent of system re- rithm that takes advantage of its grouping strategy. On sources [3] for large numbers of clients. Furthermore, little a multiprocessor system, proportional sharing is not fea- sible for some client weight assignments, such as having ∗also in Department of IEOR one client with weight 1 and another with weight 2 on a USENIX Association 2005 USENIX Annual Technical Conference 337 two-processor system. By organizing clients with similar on some metric of fairness, and has a time complexity that weights into groups, GR3 adjusts for infeasible weight as- grows with the number of clients. The other way is to ad- signments without the need to order clients, resulting in just the size of a client’s time quantum so that it runs longer lower scheduling complexity than previous approaches [7]. for a given allocation, as is done in weighted round-robin We have analyzed GR3 and show that with only O(1) (WRR). This is fast, providing constant time complexity overhead, GR3 provides fairness within O(g2) of the ideal scheduling overhead. However, allowing a client to mo- Generalized Processor Sharing (GPS) model [16], where g, nopolize the resource for a long period of time results in ex- the number of groups, grows at worst logarithmically with tended periods of unfairness to other clients which receive the largest client weight. Since g is in practice a small con- no service during those times. The unfairness is worse with stant, GR3 effectively provides constant fairness bounds skewed weight distributions. with only O(1) overhead. Moreover, we show that GR3 GR3 is a proportional share scheduler that matches uniquely preserves its worst-case time complexity and fair- with O(1) time complexity of round-robin scheduling but ness properties for multiprocessor systems. provides much better proportional fairness guarantees in We have implemented a prototype GR3 processor practice. At a high-level, the GR3 scheduling algorithm scheduler in Linux, and compared it against uniproces- can be briefly described in three parts: sor and multiprocessor schedulers commonly used in prac- tice and research, including the standard Linux sched- 1. Client grouping strategy: Clients are separated into uler [2], Weighted Fair Queueing (WFQ) [11], Virtual-Time groups of clients with similar weight values. The k Round-Robin (VTRR) [17], and Smoothed Round-Robin group of order is assigned all clients with weights 2k 2k+1 − 1 k ≥ 0 (SRR) [9]. We have conducted both simulation studies and between to ,where . kernel measurements on micro-benchmarks and real ap- 2. Intergroup scheduling: Groups are ordered in a list 3 plications. Our results show that GR can provide more from largest to smallest group weight, where the group than an order of magnitude better proportional sharing ac- weight of a group is the sum of the weights of all curacy than these other schedulers, in some cases with more clients in the group. Groups are selected in a round- than an order of magnitude less overhead. These results robin manner based on the ratio of their group weights. 3 demonstrate that GR can in practice deliver better propor- If a group has already been selected more than its pro- tional share control with lower scheduling overhead than portional share of the time, move on to the next group 3 these other approaches. Furthermore, GR is simple to im- in the list. Otherwise, skip the remaining groups in plement and easy to incorporate into existing scheduling the group list and start selecting groups from the be- frameworks in commodity operating systems. ginning of the group list again. Since the groups with This paper presents the design, analysis, and evalua- larger weights are placed first in the list, this allows 3 tion of GR . Section 2 describes the uniprocessor schedul- them to get more service than the lower-weight groups ing algorithm. Section 3 describes extensions for multipro- at the end of the list. cessor scheduling, which we refer to as GR3MP. Section 4 analyzes the fairness and complexity of GR3. Section 5 3. Intragroup scheduling: From the selected group, a presents experimental results. Section 6 discusses related client is selected to run in a round-robin manner that work. Finally, we present some concluding remarks and accounts for its weight and previous execution history. directions for future work. Using this client grouping strategy, GR3 separates scheduling in a way that reduces the need to schedule enti- 2 GR3 Uniprocessor Scheduling ties with skewed weight distributions. The client grouping Uniprocessor scheduling, the process of scheduling a time- strategy limits the number of groups that need to be sched- multiplexed resource among a set of clients, has two basic uled since the number of groups grows at worst logarithmi- steps: 1) order the clients in a queue, 2) run the first client in cally with the largest client weight. Even a very large 32-bit the queue for its time quantum, which is the maximum time client weight would limit the number of groups to no more interval the client is allowed to run before another schedul- than 32. The client grouping strategy also ensures that all ing decision is made. We refer to the units of time quanta clients within a group have weight within a factor of two. As a result, the intragroup scheduler never needs to sched- as time units (tu) rather than an absolute time measure such 3 as seconds. A scheduler can therefore achieve proportional ule clients with skewed weight distributions. GR groups sharing in one of two ways. One way, often called fair are simple lists that do not need to be balanced; they do not queueing [11, 18, 28, 13, 24, 10] is to adjust the frequency require any use of more complex balanced tree structures.
Recommended publications
  • Queue Scheduling Disciplines
    White Paper Supporting Differentiated Service Classes: Queue Scheduling Disciplines Chuck Semeria Marketing Engineer Juniper Networks, Inc. 1194 North Mathilda Avenue Sunnyvale, CA 94089 USA 408 745 2000 or 888 JUNIPER www.juniper.net Part Number:: 200020-001 12/01 Contents Executive Summary . 4 Perspective . 4 First-in, First-out (FIFO) Queuing . 5 FIFO Benefits and Limitations . 6 FIFO Implementations and Applications . 6 Priority Queuing (PQ) . 6 PQ Benefits and Limitations . 7 PQ Implementations and Applications . 8 Fair Queuing (FQ) . 9 FQ Benefits and Limitations . 9 FQ Implementations and Applications . 10 Weighted Fair Queuing (WFQ) . 11 WFQ Algorithm . 11 WFQ Benefits and Limitations . 13 Enhancements to WFQ . 14 WFQ Implementations and Applications . 14 Weighted Round Robin (WRR) or Class-based Queuing (CBQ) . 15 WRR Queuing Algorithm . 15 WRR Queuing Benefits and Limitations . 16 WRR Implementations and Applications . 18 Deficit Weighted Round Robin (DWRR) . 18 DWRR Algorithm . 18 DWRR Pseudo Code . 19 DWRR Example . 20 DWRR Benefits and Limitations . 24 DWRR Implementations and Applications . 25 Conclusion . 25 References . 26 Textbooks . 26 Technical Papers . 26 Seminars . 26 Web Sites . 27 Copyright © 2001, Juniper Networks, Inc. List of Figures Figure 1: First-in, First-out (FIFO) Queuing . 5 Figure 2: Priority Queuing . 7 Figure 3: Fair Queuing (FQ) . 9 Figure 4: Class-based Fair Queuing . 11 Figure 5: A Weighted Bit-by-bit Round-robin Scheduler with a Packet Reassembler . 12 Figure 6: Weighted Fair Queuing (WFQ)—Service According to Packet Finish Time . 13 Figure 7: Weighted Round Robin (WRR) Queuing . 15 Figure 8: WRR Queuing Is Fair with Fixed-length Packets . 17 Figure 9: WRR Queuing is Unfair with Variable-length Packets .
    [Show full text]
  • Scheduling Algorithms
    Scheduling in OQ architectures Scheduling and QoS scheduling Andrea Bianco Telecommunication Network Group [email protected] http://www.telematica.polito.it/ Andrea Bianco – TNG group - Politecnico di Torino Computer Networks Design and Management - 1 Scheduling algorithms • Scheduling: choose a packet to transmit over a link among all packets stored in a given buffer (multiplexing point) • Mainly look at QoS scheduling algorithms – Choose the packet according to QoS needs N OUTPUT inputs BUFFER Andrea Bianco – TNG group - Politecnico di Torino Computer Networks Design and Management - 2 Pag. 1 Scheduling in OQ architectures Output buffered architecture • Advantage of OQ (Output Queued) architectures – All data immediately transferred to output buffers according to data destination – It is possible to run QoS scheduling algorithms independently for each output link • In other architectures, like IQ or CIOQ switches, problems become more complex – Scheduling to satisfy QoS requirements and scheduling to maximize the transfer data from inputs to outputs have conflicting requirements Andrea Bianco – TNG group - Politecnico di Torino Computer Networks Design and Management - 3 QoS scheduling algorithms • Operate over multiplexing points • Micro or nano second scale • Easy enough to be implemented in hardware at high speed • Regulate interactions among flows • Single traffic relation (1VP/1VC) • Group of traffic relations (more VC/1VP o more VC with similar QoS needs) • QoS classes • Strictly related and dependent from buffer management
    [Show full text]
  • Tradeoffs Between Low Complexity, Low Latency, and Fairness with Deficit Round-Robin Schedulers 683
    IEEE/ACM TRANSACTIONS ON NETWORKING, VOL. 12, NO. 4, AUGUST 2004 681 Tradeoffs Between Low Complexity, Low Latency, and Fairness With Deficit Round-Robin Schedulers Luciano Lenzini, Enzo Mingozzi, and Giovanni Stea Abstract—Deficit Round-Robin (DRR) is a scheduling algorithm The existing work-conserving scheduling algorithms are devised for providing fair queueing in the presence of variable commonly classified as sorted-priority or frame-based. Sorted- length packets. The main attractive feature of DRR is its sim- priority algorithms associate a timestamp with each queued plicity of implementation: in fact, it can exhibit @IA complexity, provided that specific allocation constraints are met. However, packet and transmit packets by increasing timestamp order. according to the original DRR implementation, meeting such con- On the other hand, frame-based algorithms divide time into straints often implies tolerating high latency and poor fairness. In frames, and select which packet to transmit on a per-frame this paper, we first derive new and exact bounds for DRR latency basis. Within the frame-based class, round-robin algorithms and fairness. On the basis of these results, we then propose a service the various flows cyclically, and within a round each novel implementation technique, called Active List Queue Method (Aliquem), which allows a remarkable gain in latency and fairness flow is serviced for up to a given quantum, so that the frame to be achieved, while still preserving @IA complexity. We show length can vary up to a given maximum. Sorted-priority algo- that DRR achieves better performance metrics than those of other rithms generally exhibit better latency and fairness properties round-robin algorithms such as Pre-Order Deficit Round-Robin compared to round-robin ones, but have a higher complexity, and Smoothed Round-Robin.
    [Show full text]
  • Fairness in a Data Center
    University of New Hampshire University of New Hampshire Scholars' Repository Doctoral Dissertations Student Scholarship Winter 2012 Fairness in a data center MIkkel Hagen University of New Hampshire, Durham Follow this and additional works at: https://scholars.unh.edu/dissertation Recommended Citation Hagen, MIkkel, "Fairness in a data center" (2012). Doctoral Dissertations. 694. https://scholars.unh.edu/dissertation/694 This Dissertation is brought to you for free and open access by the Student Scholarship at University of New Hampshire Scholars' Repository. It has been accepted for inclusion in Doctoral Dissertations by an authorized administrator of University of New Hampshire Scholars' Repository. For more information, please contact [email protected]. Fairness in a data center BY Mikkel Hagen M.S. Computer Science, University of New Hampshire (2008) B.S. Computer Science/Biology/Psychology, St. Bonaventure University (2004) DISSERTATION Submitted to the University of New Hampshire in partial fulfillment of the requirements for the degree of Doctor of Philosophy in Computer Science December 2012 UMI Number: 3537816 All rights reserved INFORMATION TO ALL USERS The quality of this reproduction is dependent upon the quality of the copy submitted. In the unlikely event that the author did not send a complete manuscript and there are missing pages, these will be noted. Also, if material had to be removed, a note will indicate the deletion. Dissertation Publishing UMI 3537816 Published by ProQuest LLC 2013. Copyright in the Dissertation held by the Author. Microform Edition © ProQuest LLC. All rights reserved. This work is protected against unauthorized copying under Title 17, United States Code. uest ProQuest LLC 789 East Eisenhower Parkway P.O.
    [Show full text]
  • Hardware Implementation of a Scheduler for High Performance Switches with Quality of Service Support
    See discussions, stats, and author profiles for this publication at: https://www.researchgate.net/publication/253896547 Hardware implementation of a scheduler for high performance switches with Quality of Service support Article in Proceedings of SPIE - The International Society for Optical Engineering · May 2009 DOI: 10.1117/12.821522 CITATIONS READS 0 52 4 authors, including: Félix B. Tobajas Valentín de Armas Sosa Universidad de Las Palmas de Gran Canaria Universidad de Las Palmas de Gran Canaria 81 PUBLICATIONS 212 CITATIONS 63 PUBLICATIONS 151 CITATIONS SEE PROFILE SEE PROFILE Roberto Sarmiento Universidad de Las Palmas de Gran Canaria 202 PUBLICATIONS 1,075 CITATIONS SEE PROFILE Some of the authors of this publication are also working on these related projects: NAutILES(Novel Autonomous Intelligent Localisable Endoscopy System) Link: https://nautilesorg.wordpress.com View project All content following this page was uploaded by Félix B. Tobajas on 21 January 2015. The user has requested enhancement of the downloaded file. Hardware implementation of a scheduler for high performance switches with Quality of Service support R. Arteaga, F. Tobajas, V. De Armas and R. Sarmiento Institute for Applied Microelectronics (IUMA) Departamento de Ingeniería Electrónica y Automática (DIEA) University of Las Palmas de Gran Canaria, Campus de Tafira, 35017, Las Palmas de Gran Canaria. ABSTRACT In this paper, the hardware implementation of a scheduler with QoS support is presented. The starting point is a Differentiated Service (DiffServ) network model. Each switch of this network classifies the packets in flows which are assigned to traffic classes depending of its requirements with an independent queue being available for each traffic class.
    [Show full text]
  • Performance Evaluation of Different Scheduling Algorithms in Wimax
    PERFORMANCE EVALUATION OF DIFFERENT SCHEDULING ALGORITHMS IN WIMAX Ala’a Z. Al-Howaide 1, Ahmad S. Doulat 2, Yaser M. Khamayseh 3 Department of Computer Science, Jordan University of Science and Technology, Irbid, Jordan 1 2 3 [email protected] , [email protected] , [email protected] ABSTRACT Worldwide Interoperability for Microwave Access (WiMAX) networks were expected to be the main Broadband Wireless Access (BWA) technology that provided several services such as data, voice, and video services including different classes of Quality of Services (QoS), which in turn were defined by IEEE 802.16 standard. Scheduling in WiMAX became one of the most challenging issues, since it was responsible for distributing available resources of the network among all users; this leaded to the demand of constructing and designing high efficient scheduling algorithms in order to improve the network utilization, to increase the network throughput, and to minimize the end-to-end delay. In this study, we presenedt a simulation study to measure the performance of several scheduling algorithms in WiMAX, which were Strict Priority algorithm, Round-Robin (RR), Weighted Round Robin (WRR), Weighted Fair Queuing (WFQ), Self-Clocked Fair (SCF), and Diff-Serv Algorithm. KEYWORDS WiMAX,Quality of Services (QoS), Strict Priority (SP), Round Robin (RR), Weighted Round Robin (WRR), Weighted Fair Queuing (WFQ), Self-Clocked Fair (SCF) Queuing, Diff-Serv (DS). 1. INTRODUCTION WiMAX is a telecommunication protocol that provides fixed and mobile internet access, in which this protocol combines a number of wireless technologies that have emerged from IEEE to face the rapid demand of higher data rate and longer transmission range in wireless access and to enable a high speed connection to the Internet in terms of multimedia service, trade, commerce, education, research and other applications.
    [Show full text]
  • Simulating Strict Priority Queueing, Weighted Round Robin, And
    International Journal on Advances in Networks and Services, vol 10 no 1 & 2, year 2017, http://www.iariajournals.org/networks_and_services/ 1 Simulating Strict Priority Queueing, Weighted Round Robin, and Weighted Fair Queueing with NS-3 Robert Chang and Vahab Pournaghshband Advanced Network and Security Research Laboratory Computer Science Department California State University, Northridge Northridge, California, USA [email protected] [email protected] Abstract—Strict priority queueing, weighted fair queueing, and weighted round robin are amongst the most popular differen- tiated service queueing disciplines widely used in practice to ensure quality of service for specific types of traffic. In this paper, we present the design and implementation of these three methods in Network Simulator 3 (ns-3). ns-3 is a discrete event network simulator designed to simulate the behavior of computer networks and internet systems. Utilizing our implementations will provide users with the opportunity to research new solutions to existing problems that were previously not available to solve with the existing tools. We believe the ease of configuration and use of our modules will make them attractive tools for further research. By comparing the behavior of our modules with expected outcomes derived from the theoretical behavior of each queueing algorithm, we were able to verify the correctness of Figure 1. ns-3’s simulation network architecture [3] our implementation in an extensive set of experiments. These format. ns-3 has an object oriented design, which facilitates implementations can be used by the research community to investigate new properties and applications of differentiated rapid coding and extension. It also includes automatic memory service queueing.
    [Show full text]
  • Packet Scheduling in Wireless Systems by Token Bank Fair Queueing Algorithm
    Packet scheduling in wireless systems by Token Bank Fair Queueing Algorithm by WILLIAM KING WONG B.A.Sc, The University of Western Ontario, 1991 M.A.Sc, The University of Western Ontario, 1993 M.Sc., The University of Western Ontario, 1994 A THESIS SUBMITTED IN PARTIAL FULFILMENT OF THE REQUIREMENTS FOR THE DEGREE OF DOCTOR OF PHILOSOPHY in THE FACULTY OF GRADUATE STUDffiS ( ELECTRICAL AND COMPUTER ENGINEERING ) THE UNIVERSITY OF BRITISH COLUMBIA July 2005 © William King Wong, 2005 ABSTRACT The amalgamation of wireless systems is a vision for the next generation wireless systems where multimedia applications will dominate. One of the challenges for the operators of next generation networks is the delivery of quality-of-service (QoS) to the end users. Wireless scheduling algorithms play a critical and significant role in the user perception of QoS. However, the scarcity of wireless resources and its unstable channel condition present a challenge to providing QoS. To that end, this thesis proposes a novel scheduling algorithm Token Bank Fair Queuing (TBFQ) that is simple to implement and has the soft tolerance characteristic which is suitable under wireless constraints. TBFQ combines both token- based policing mechanism and fair throughput prioritization rule in scheduling packets over packet- switched wireless systems. This results in a simple design and yet achieves better performance and fairness than existing wireless scheduling strategies. The characterization of the algorithm, which leads to its tight bound performance, is conducted through numerical analysis that recursively determines the steady-state probability distribution of the bandwidth allocation process and ultimately the probability distribution of input queue.
    [Show full text]
  • Multi-Queue Fair Queuing Mohammad Hedayati, University of Rochester; Kai Shen, Google; Michael L
    Multi-Queue Fair Queuing Mohammad Hedayati, University of Rochester; Kai Shen, Google; Michael L. Scott, University of Rochester; Mike Marty, Google https://www.usenix.org/conference/atc19/presentation/hedayati-queue This paper is included in the Proceedings of the 2019 USENIX Annual Technical Conference. July 10–12, 2019 • Renton, WA, USA ISBN 978-1-939133-03-8 Open access to the Proceedings of the 2019 USENIX Annual Technical Conference is sponsored by USENIX. Multi-Queue Fair Queueing Mohammad Hedayati Kai Shen Michael L. Scott Mike Marty University of Rochester Google University of Rochester Google Abstract ric within a few microseconds. GPUs and machine learning Modern high-speed devices (e.g., network adapters, storage, accelerators may offload computations that run just a few mi- accelerators) use new host interfaces, which expose multiple croseconds at a time [30]. At the same time, the proliferation software queues directly to the device. These multi-queue in- of multicore processors has necessitated architectures tuned terfaces allow mutually distrusting applications to access the for independent I/O across multiple hardware threads [4, 36]. device without any cross-core interaction, enabling through- These technological changes have shifted performance bot- put in the order of millions of IOP/s on multicore systems. tlenecks from hardware resources to the software stacks that Unfortunately, while independent device access is scalable, manage them. In response, it is now common to adopt a multi- it also introduces a new problem: unfairness. Mechanisms queue architecture in which each hardware thread owns a that were used to provide fairness for older devices are no dedicated I/O queue, directly exposed to the device, giving longer tenable in the wake of multi-queue design, and straight- it an independent path over which to send and receive re- forward attempts to re-introduce it would require cross-core quests.
    [Show full text]
  • Research Article Packet Scheduling in HighSpeed Networks Using Improved Weighted Round Robin
    Research Journal of Applied Sciences, Engineering and Technology 7(10): 2046-2057, 2014 DOI:10.19026/ajfst.7.498 ISSN: 2040-7459; e-ISSN: 2040-7467 © 2014 Maxwell Scientific Publication Corp. Submitted: June 25, 2013 Accepted: July 08, 2013 Published: March 15, 2014 Research Article Packet Scheduling in High-speed Networks Using Improved Weighted Round Robin Guikai Liu and Qing Li School of Computer Science and Engineering, Hunan University of Science and Technology, Xiangtan, 411201, China Abstract: A variety of applications with different QoS requirements are supported simultaneously in the high-speed packet-switched networks, packet scheduling algorithms play a critical role in guaranteeing the performance of routing and switching devices. This study presents a simple, fair, efficient and easily implementary scheduling algorithm, called Successive Minimal-weight Round Robin (SMRR). In each round, SMRR provides the same service opportunity, which is equivalent to the minimal weight of the current round, for all active data flows. On the basis of the concept of Latency-Rate (LR) servers, we obtain the upper bound on the latency of SMRR and WRR (Weighted Round Robin) respectively and the results indicate that SMRR makes a significant improvement on the latency bound in comparison to WRR. We also discuss the fairness and implementation complexity of SMRR and the theoretical analysis shows that SMRR preserves the good implementation complexity of O (1) with respect to the number of flows and has better fairness than WRR. Keywords: Complexity, high-speed networks, latency bound, minimal weight, packet scheduling, relative fairness, Weighted Round Robin (WRR) INTRODUCTION SRR (Smoothed Round Robin) (Guo, 2001), the scheduler simply serves all non-empty queues in a There are many kinds of services with different round-robin manner.
    [Show full text]
  • Efficient Fair Queuing Using Deficit Round-Robin
    IEEEIACM TKANSACTIONS ON NETWORKING, VOL. 4, NO. 3, JUNE 1996 315 Efficient Fair Queuing Using Deficit Round-Robin M. Shreedhar and George Varghese, Member, IEEE Abstract- Fair queuing is a technique that allows each flow in [3], and the uncapitalized “fair queuing” to refer to the passing through a network device to have a fair share of net- generic idea. work resources. Previous schemes for fair queuing that achieved In this paper, we shall define an isolation mechanism that nearly perfect fairness were expensive to implement; specifically, the work required to process a packet in these schemes was achieves nearly perfect fairness (in terms of throughput), and 0 (log(71 ) ), where 71. is the number of active flows. This is expen- takes O(1) processing work per packet. Our scheme is simple sive at high speeds. On the other hand, cheaper approximations of (and, therefore, inexpensive) to implement at high speeds at a fair queuing reported in the literature exhibit unfair behavior. In router or gateway. Furthermore, we provide analytical results this paper, we describe a new approximation of fair queuing, that that do not depend on assumptions about traffic distributions. we call deficit round-robin. Our scheme achieves nearly perfect fairness in terms of throughput, requires only O(1) work to Flows: Our intent is to provide firewalls between different process a packet, and is simple enough to implement in hard- packet streams. We formalize the intuitive notion of a packet ware. Deficit round-robin is also applicable to other scheduling stream using the more precise notion of ajow [18].
    [Show full text]
  • Improved Waiting Time of Tasks Scheduled Under Preemptive Round Robin
    Improving Waiting Time of Tasks Scheduled Under Preemptive Round Robin Using Changeable Time Quantum Saad Zagloul Rida* Safwat Helmy Hamad ** Samih Mohemmed Mostafa *** *Faculty of Science, Mathematics Department, South Valley University, Egypt, Quena. ** Faculty of Computer and Information Sciences, Ain Shams University, Egypt, Cairo. *** Faculty of Science, Computer Science Department, South Valley University, Egypt, Quena. ABSTRACT Minimizing waiting time for tasks waiting in the queue for execution is one of the important scheduling cri- teria which took a wide area in scheduling preemptive tasks. In this paper we present Changeable Time Quan- tum (CTQ) approach combined with the round-robin algorithm, we try to adjust the time quantum according to the burst times of the tasks in the ready queue. There are two important benefits of using (CTQ) approach: minimizing the average waiting time of the tasks, consequently minimizing the average turnaround time, and keeping the number of context switches as low as possible, consequently minimizing the scheduling overhead. In this paper, we consider the scheduling problem for preemptive tasks, where the time costs of these tasks are known a priori. Our experimental results demonstrate that CTQ can provide much lower scheduling overhead and better scheduling criteria. General Terms: Algorithm, round-robin, CPU length. Keywords: Processor sharing, proportional sharing, residual time, survived tasks, cyclic queue. 1. INTRODUCTION Modern operating systems (OS) nowadays have become more complex than ever before. They have evolved from a single task, single user architecture to a multitasking environment in which tasks run in a con- current manner. Allocating CPU to a task requires careful attention to assure fairness and avoid task starvation for CPU.
    [Show full text]