PCC: Re-Architecting Congestion Control for Consistent High Performance

Total Page:16

File Type:pdf, Size:1020Kb

PCC: Re-Architecting Congestion Control for Consistent High Performance PCC: Re-architecting Congestion Control for Consistent High Performance Mo Dong and Qingxi Li, University of Illinois at Urbana-Champaign; Doron Zarchy, Hebrew University of Jerusalem; P. Brighten Godfrey, University of Illinois at Urbana-Champaign; Michael Schapira, Hebrew University of Jerusalem https://www.usenix.org/conference/nsdi15/technical-sessions/presentation/dong This paper is included in the Proceedings of the 12th USENIX Symposium on Networked Systems Design and Implementation (NSDI ’15). May 4–6, 2015 • Oakland, CA, USA ISBN 978-1-931971-218 Open Access to the Proceedings of the 12th USENIX Symposium on Networked Systems Design and Implementation (NSDI ’15) is sponsored by USENIX PCC: Re-architecting Congestion Control for Consistent High Performance Mo Dong*, Qingxi Li*, Doron Zarchy**, P. Brighten Godfrey*, and Michael Schapira** *University of Illinois at Urbana-Champaign **Hebrew University of Jerusalem Abstract a very difficult task within TCP’s rate control architec- ture, which we refer to as : certain TCP and its variants have suffered from surprisingly poor hardwired mapping predefined packet-level events are hardwired to certain performance for decades. We argue the TCP family has predefined control responses. TCP reacts to events that little hope of achieving consistent high performance due can be as simple as “one packet loss” (TCP New Reno) to a fundamental architectural deficiency: hardwiring or can involve multiple signals like “one packet loss and packet-level events to control responses. We propose RTT increased by x%” (TCP Illinois). Similarly, the con- Performance-oriented Congestion Control (PCC), a new trol response might be “halve the rate” (New Reno) or a congestion control architecture in which each sender more complex action like “reduce the window size w to continuously observes the connection between its ac- f ( RTT)w” (Illinois). The defining feature is that the tions and empirically experienced performance, enabling ∆ control action is a direct function of packet-level events. it to consistently adopt actions that result in high perfor- mance. We prove that PCC converges to a stable and A hardwired mapping has to make assumptions about fair equilibrium. Across many real-world and challeng- the network. Take a textbook event-control pair: a packet ing environments, PCC shows consistent and often 10 loss halves the congestion window. TCP assumes that × performance improvement, with better fairness and sta- the loss indicates congestion in the network. When bility than TCP. PCC requires no router hardware support the assumption is violated, halving the window size can or new packet format. severely degrade performance (e.g. if loss is random, rate should stay the same or increase). It is fundamentally hard to formulate an “always optimal” hardwired map- 1 Introduction ping in a complex real-world network because the actual optimal response to an event like a loss (i.e. decrease In the roughly 25 years since its deployment, TCP’s rate or increase? by how much?) is sensitive to network congestion control architecture has been notorious for conditions. And modern networks have an immense di- degraded performance. TCP performs poorly on lossy versity of conditions: random loss and zero loss, shallow links, penalizes high-RTT flows, underutilizes high queues and bufferbloat, RTTs of competing flows vary- bandwidth-delay product (BDP) connections, cannot ing by more than 1000 , dynamics due to mobile wire- handle rapidly changing networks, can collapse under × less or path changes, links from Kbps to Gbps, AQMs, data center incast [24] and incurs very high latency with software routers, rate shaping at gateways, virtualization bufferbloat [28] in the network. layers and middleboxes like firewalls, packet inspectors As severe performance problems have accumulated and load balancers. These factors add complexity far be- over time, protocol “patches” have addressed prob- yond what can be summarized by the relatively simplistic lems in specific network conditions such as high BDP assumptions embedded in a hardwired mapping. Most links [31,52], satellite links [23,42], data center [18,55], unfortunately, when its assumptions are violated, TCP wireless and lossy links [38,39], and more. However, the still rigidly carries out the harmful control action. fact that there are so many TCP variants suggests that each is only a point solution: they yield better perfor- We propose a new congestion control architecture: mance under specific network conditions, but break in Performance-oriented Congestion Control (PCC). PCC’s others. Worse, we found through real-world experiments goal is to understand what rate control actions improve that in many cases these TCP variants’ performance is performance based on live experimental evidence, avoid- still far away from optimal even in the network condi- ing TCP’s assumptions about the network. PCC sends at tions for which they are specially engineered. Indeed, a rate r for a short period of time, and observes the re- TCP’s low performance has impacted industry to the ex- sults (e.g. SACKs indicating delivery, loss, and latency tent that there is a lucrative market for special-purpose of each packet). It aggregates these packet-level events high performance data transfer services [1, 2, 11, 13]. into a utility function that describes an objective like Thus, the core problem remains largely unsolved: “high throughput and low loss rate”. The result is a sin- achieving consistently high performance over complex gle numerical performance utility u. At this point, PCC real-world network conditions. We argue this is indeed has run a single “micro-experiment” that showed send- 1 USENIX Association 12th USENIX Symposium on Networked Systems Design and Implementation (NSDI ’15) 395 TCP Packet Level ing at rate r produced utility u. To make a rate control Events Hardwired decision, PCC runs multiple such micro-experiments: it Action 1 Network Action 2 tries sending at two different rates, and moves in the di- PCC Packet-level rection that empirically results in greater performance Events Performance objective Move to Sending Aggregate U1 Y rate R1 utility. This is effectively A/B testing for rate control rate R1 Network and is the core of PCC’s decisions. PCC runs these Performance U1>U2? objective Sending Aggregate micro-experiments continuously (on every byte of data, Network U2 N Move to rate R2 rate R2 not on occasional probes), driven by an online learn- ing algorithm that tracks the empirically-optimal send- Figure 1: The decision-making structure of TCP and PCC. ing rate. Thus, rather than making assumptions about f might traverse a shallow buffer on a high-BDP • the potentially-complex network, PCC adopts the actions link, with the loss due to bad luck in statistical mul- that empirically achieve consistent high performance. tiplexing rather than high link utilization. Then, PCC’s rate control is selfish in nature, but surpris- backing off a little is sufficient. ingly, using a widely applicable utility function, com- There may be a higher-rate competing flow. Then, • peting PCC senders provably converge to a fair equi- f should maintain its rate and let the other back off. librium (with a single bottleneck link). Indeed, experi- There may be random non-congestion loss some- • ments show PCC achieves similar convergence time to where along the path. Then, f should maintain or TCP with significantly smaller rate variance. Moreover, increase its rate. the ability to express different objectives via choice of Classically, TCP assumes a packet loss indicates non- the utility function (e.g. throughput or latency) provides negligible congestion, and that halving its rate will im- a flexibility beyond TCP’s architecture. prove network conditions. However, this assumption With handling real-world complexity as a key goal, is false and will degrade performance in three of the we experimentally evaluated a PCC implementation in four scenarios above. Fundamentally, picking an opti- large-scale and real-world networks. Without tweak- mal predefined and hardwired control response is hard ing its control algorithm, PCC achieves consistent high because for the same packet-level events, a control re- performance and significantly beats specially engineered sponse optimal under one network scenario can deci- TCPs in various network environments: (a.) in the wild mate performance in even a slightly different scenario. on the global commercial Internet (often more than 10 × The approach taken by a large number of TCP vari- the throughput of TCP CUBIC); (b.) inter-data center ants is to use more sophisticated packet-level events and networks (55.23 vs. TCP Illinois); (c.) emulated satel- × control actions. But this does not solve the fundamen- lite Internet links (1717 vs TCP Hybla); (d.) unreliable × tal problem, because the approach still hardwires pre- lossy links (1010 37 vs Illinois); (e.) unequal RTT of − × determined events to predetermined control responses, competing senders (an architectural cure to RTT un- thus inevitably embedding unreliable assumptions about fairness); (f.) shallow buffered bottleneck links (up to the network. When the unreliable assumptions are vi- 45 higher performance, or 13 less buffer to reach × × olated by the complexity of the network, performance 90% throughput); (g.) rapidly changing networks (1414 × degrades severely. For example, TCP Illinois [38] uses vs CUBIC, 5.6 vs Illinois). PCC performs similar to × both loss and delay to form an event-control mapping, ICTCP [55] in (h.) the incast scenario in data centers. but its throughput collapses with even a small amount of Though it is a substantial shift in architecture, PCC can random loss, or when network conditions are dynamic be deployed by only replacing the sender-side rate con- ( 4). More examples are in 5. § § trol of TCP. It can also deliver real data today with a user- Most unfortunately, if some control actions are indeed space implementation at speedier.net/pcc. harming performance, TCP can still blindly “jump off 2 PCC Architecture the cliff”, because it does not notice the control action’s actual effect on performance.
Recommended publications
  • A Letter to the FCC [PDF]
    Before the FEDERAL COMMUNICATIONS COMMISSION Washington, DC 20554 In the Matter of ) ) Amendment of Part 0, 1, 2, 15 and 18 of the ) ET Docket No. 15­170 Commission’s Rules regarding Authorization ) Of Radio frequency Equipment ) ) Request for the Allowance of Optional ) RM­11673 Electronic Labeling for Wireless Devices ) Summary The rules laid out in ET Docket No. 15­170 should not go into effect as written. They would cause more harm than good and risk a significant overreach of the Commission’s authority. Specifically, the rules would limit the ability to upgrade or replace firmware in commercial, off­the­shelf home or small­business routers. This would damage the compliance, security, reliability and functionality of home and business networks. It would also restrict innovation and research into new networking technologies. We present an alternate proposal that better meets the goals of the FCC, not only ensuring the desired operation of the RF portion of a Wi­Fi router within the mandated parameters, but also assisting in the FCC’s broader goals of increasing consumer choice, fostering competition, protecting infrastructure, and increasing resiliency to communication disruptions. If the Commission does not intend to prohibit the upgrade or replacement of firmware in Wi­Fi ​ ​ devices, the undersigned would welcome a clear statement of that intent. Introduction We recommend the FCC pursue an alternative path to ensuring Radio Frequency (RF) compliance from Wi­Fi equipment. We understand there are significant concerns regarding existing users of the Wi­Fi ​ spectrum, and a desire to avoid uncontrolled change. However, we most strenuously advise against prohibiting changes to firmware of devices containing radio components, and furthermore advise against allowing non­updatable devices into the field.
    [Show full text]
  • Evaluation of Priority Scheduling and Flow Starvation for Thin Streams with FQ-Codel
    2015 European Conference on Networks and Communications (EuCNC) Evaluation of Priority Scheduling and Flow Starvation for Thin Streams with FQ-CoDel Eduard Grigorescu, Chamil Kulatunga, Gorry Nicolas Kuhn Fairhurst Télécom Bretagne, IRISA School of Engineering, University of Aberdeen, UK [email protected] {eduard, chamil, gorry}@erg.abdn.ac.uk Abstract— Bufferbloat is the result of oversized buffers and algorithms, managing packet scheduling and isolation/capacity induced high end-to-end latency experienced by applications allocation among flows, can be introduced. As one example of across the Internet. This additional delay can adversely impact a scheme that mixes both classes, FlowQueue-CoDel (FQ- thin streams that frequently exchange small amounts of data, but CoDel) [7] is a scheduling scheme that features prioritization have stringent latency requirements. Active Queue Management and flow isolation. FQ-CoDel creates one sub-queue per flow (AQM) techniques, such as Controlled Delay (CoDel), can control and applies CoDel on each of them. The awareness of the the queuing delay in a network device to ensure low latency by dropping packets to indicate incipient congestion. FlowQueue- latency resulting from over-provisioned buffers has been CoDel (FQ-CoDel) is a scheduling scheme that creates one sub- accompanied by an increase in real-time applications such as queue per flow and applies CoDel on each of them. FQ-CoDel Voice over Internet Protocol, gaming or financial trading features: (1) priority scheduling for low-rate traffic; (2) flow applications. As one example, the latency experienced by isolation; (3) queue management with CoDel. First, this paper gamers can directly impact the perceived value of the network fills a gap in the understanding of FQ-CoDel by analyzing what service [10].
    [Show full text]
  • AQM Algorithms and Their Interaction with TCP Congestion Control Mechanisms
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by Universidad Carlos III de Madrid e-Archivo Grado Universitario en Ingenier´ıaTelem´atica 2016/2017 Trabajo Fin de Grado Control de Congesti´onTCP y mecanismos AQM Sergio Maeso Jim´enez Tutor/es Celeste Campo V´azquez Carlos Garc´ıaRubio Legan´es,2 de Octubre de 2017 Esta obra se encuentra sujeta a la licencia Creative Commons Reconocimiento - No Comercial - Sin Obra Derivada Control de Congesti´onTCP y mecanismos AQM By Sergio Maeso Jim´enez Directed By Celeste Campo V´azquez Carlos Garc´ıaRubio A Dissertation Submitted to the Department of Telematic Engineering in Partial Fulfilment of the Requirements for the BACHELOR'S DEGREE IN TELEMATICS ENGINEERING Approved by the Supervising Committee: Chairman Marta Portela Garc´ıa Chair Carlos Alario Hoyos Secretary I~naki Ucar´ Marqu´es Deputy Javier Manuel Mu~noz Garc´ıa Grade: Legan´es,2 de Octubre de 2017 iii iv Acknowledgements I would like to thanks my tutors Celeste Campo and Carlos Garcia for all the support they gave me while I was doing this thesis with them. To my parents, who believe in me against all odds. v vi Abstract In recent years, the relevance of delay over throughput has been particularly emphasized. Nowadays our networks are getting more and more sensible to latency due to the proliferation of applications and services like VoIP, IPTV or online gaming where a low delay is essential for a proper performance and a good user experience. Most of this unnecessary delay is created by the misbehaviour of many buffers that populate Internet.
    [Show full text]
  • Auto-Tuning Active Queue Management
    2017 9th International Conference on Communication Systems and Networks (COMSNETS) Auto-Tuning Active Queue Management Joe H. Novak Sneha Kumar Kasera University of Utah University of Utah Abstract not require pre-specification of even those parameters including 3 Active queue management (AQM) algorithms preemptively EWMA weights or protocol timer intervals that are routinely drop packets to prevent unnecessary delays through a network specified in protocol design and implementation. while keeping utilization high. Many AQM ideas have been When we view the network link in terms of utilization proposed, but none have been widely adopted because these rely and the queue corresponding to the link in terms of delay on pre-specification or pre-tuning of parameters and thresholds (Fig. 1 shows a typical delay-utilization curve), we see that that do not necessarily adapt to dynamic network conditions. as utilization increases, delay also increases. At a certain point, We develop an AQM algorithm that relies only on network however, there is a very large increase in delay for only a small runtime measurements and a natural threshold, the knee on improvement in utilization. This disproportionate increase in the delay-utilization curve. We call our AQM algorithm Delay delay is of little to no value to the applications at the endpoints. Utilization Knee (DUK) based on its key characteristic of We want to avoid this unstable region of high increase in delay keeping the system operating at the knee of the delay-utilization with little increase in utilization. As a result, a natural threshold curve. We implement and evaluate DUK in the Linux kernel in becomes apparent.
    [Show full text]
  • Controlling Queuing Delays for Real-Time Communication: the Interplay of E2E and AQM Algorithms
    Controlling Queuing Delays for Real-Time Communication: The Interplay of E2E and AQM Algorithms Gaetano Carlucci Luca De Cicco Saverio Mascolo Politecnico di Bari, Italy Politecnico di Bari, Italy Politecnico di Bari, Italy [email protected] [email protected] [email protected] ABSTRACT control algorithm; the other placed in the network routers, Real-time media communication requires not only i.e. the Active Queue Management (AQM) algorithm. congestion control, but also minimization of queuing delays When designing end-to-end congestion control algorithms to provide interactivity. In this work we consider the case of for video conference, UDP is a natural choice, since real-time communication between web browsers (WebRTC) such applications cannot tolerate large latencies due to and we focus on the interplay of an end-to-end delay-based TCP packet retransmissions. Additionally, delay-based congestion control algorithm, i.e. the Google congestion congestion control algorithms are usually preferred to control (GCC), with two delay-based AQM algorithms, loss-based ones since they can detect congestion before namely CoDel and PIE, and two flow queuing schedulers, packets are lost due to buffer overflow. i.e. SFQ and Fq Codel. Experimental investigations show On the other hand, AQM schemes control the router that, when only GCC flows are considered, the end-to-end buffers by dropping the packets or marking them if ECN algorithm is able to contain queuing delays without AQMs. is used [21]. Despite the fact that many AQM algorithms Moreover the interplay of GCC flows with PIE or CoDel have been proposed over the past two decades [7], the leads to higher packet losses with respect to the case of a adoption of these algorithms has been hampered by two DropTail queue.
    [Show full text]
  • Controlling Queuing Delays for Real-Time Communication: the Interplay of E2E and AQM Algorithms
    Controlling Queuing Delays for Real-Time Communication: The Interplay of E2E and AQM Algorithms Gaetano Carlucci Luca De Cicco Saverio Mascolo Politecnico di Bari, Italy Politecnico di Bari, Italy Politecnico di Bari, Italy [email protected] [email protected] [email protected] ABSTRACT requirements can be met by employing two complementary control Real-time media communication requires not only congestion algorithms: one placed at the end points, namely the end-to-end control, but also minimization of queuing delays to provide congestion control algorithm; the other placed in the network interactivity. In this work we consider the case of real-time routers, i.e. the Active Queue Management (AQM) algorithm. communication between web browsers (WebRTC) and we focus When designing end-to-end congestion control algorithms on the interplay of an end-to-end delay-based congestion control for video conference, UDP is the natural choice, since such algorithm, i.e. the Google congestion control (GCC), with two applications cannot tolerate large latencies due to TCP packet delay-based AQM algorithms, namely CoDel and PIE, and two retransmissions. Additionally, delay-based congestion control flow queuing schedulers, i.e. SFQ and Fq_Codel. Experimental algorithms are usually preferred to loss-based ones since they can investigations show that, when only GCC flows are considered, detect congestion before packets are lost due to buffer overflow. the end-to-end algorithm is able to contain queuing delays without On the other hand, AQM schemes control the router buffers AQMs. Moreover the interplay of GCC flows with PIE or CoDel by dropping the packets or marking them if ECN is used [21].
    [Show full text]
  • Packet Mark Methodology for Reduce the Congestion
    Forn International Journal of Scientific Research in Computer Science and Engineering Review Paper VolVolVol-Vol ---1,1, IssueIssue----3333 ISSN: 2322320000––––76397639 Packet Mark Methodology for Reduce the Congestion V.Ganesh Babu and K.Saraswathi Department of CS, Government College for Women, Maddur, Mandya Dt Karnataka, INDIA Department of CS, Nehru Memorial College, Puthanampatti, Trichy Dt. Tamil Nadu, INDIA Available online at www.isroset.org Received: 17 March 2013 Revised: 22 March 2013 Accepted: 16 May 2013 Published: 30 June 2013 Abstract-Congestion is one of the major problems of a communication network in routing. The function of a routing is to guide packets through the communication network to their correct destinations. Optimal Routing has been widely studied for interconnection networks. This paper provides two methodologies ECN and QBER for reduce the congestion. Keywords- Routing, congestion, ECN, QBER 1. INTRODUCTION packet. This act is referred to as “marking” and its purpose is to inform the receiving endpoint of impending The first methodology Explicit Congestion Notification congestion. At the receiving endpoint, this congestion (ECN) is an extension to the Internet protocol and to the indication is handled by the upper layer protocol (transport Transmission Control Protocol and is defined in layer protocol) and needs to be echoed back to the RFC3168(2001). ECN allows end-to-end notification of transmitting node in order to signal it to reduce its network congestion without dropping packets. ECN is an transmission rate. optional feature that is only used when both endpoints support it and are willing to use it. It is only effective when 2.2 Operation of ECN with TCP supported by the underlying network.
    [Show full text]
  • Piece of CAKE: a Comprehensive Queue Management Solution for Home Gateways
    Piece of CAKE: A Comprehensive Queue Management Solution for Home Gateways Toke Høiland-Jørgensen Dave Täht Jonathan Morton Dept. of Computer Science Teklibre Karlstad University, Sweden Los Gatos, California Somero, Finland [email protected] [email protected] [email protected] Abstract—The last several years has seen a renewed interest compelling benefit of CAKE is that it takes state of the art in smart queue management to curb excessive network queueing solutions and integrates them to provide: delay, as people have realised the prevalence of bufferbloat in • a high-precision rate-based bandwidth shaper that in- real networks. However, for an effective deployment at today’s last mile cludes overhead and link layer compensation features for connections, an improved queueing algorithm is not enough in various link types. itself, as often the bottleneck queue is situated in legacy systems • a state of the art fairness queueing scheme that simulta- that cannot be upgraded. In addition, features such as per-user neously provides both host and flow isolation. fairness and the ability to de-prioritise background traffic are • a Differentiated Services (DiffServ) prioritisation scheme often desirable in a home gateway. with rate limiting of high-priority flows and work- In this paper we present Common Applications Kept Enhanced (CAKE), a comprehensive network queue management system conserving bandwidth borrowing behaviour. designed specifically for home Internet gateways. CAKE packs • TCP ACK filtering that increases achievable throughput several compelling features into an integrated solution, thus on highly asymmetrical links. easing deployment. These features include: bandwidth shaping CAKE is implemented as a queueing discipline (qdisc) for with overhead compensation for various link layers; reasonable DiffServ handling; improved flow hashing with both per-flow and the Linux kernel.
    [Show full text]
  • Evaluation of the Impact of Packet Drops Due to AQM Over Capacity Limited Paths
    Evaluation of the Impact of Packet Drops due to AQM over Capacity Limited Paths Eduard Grigorescu, Chamil Kulatunga, Gorry Fairhurst Electronics Research Group, School of Engineering, University of Aberdeen, Aberdeen, AB24 3UE, United Kingdom {eduard, chamil, gorry}@erg.abdn.ac.uk Abstract - For many years Internet routers have been designed and interactive applications, and impairing the quality of real-time benchmarked in ways that encourage the use of large buffers. When (e.g. voice) and near-real-time traffic (e.g. live streaming). these buffers accumulate a large standing queue, this can lead to high network path latency. Two AQM algorithms: PIE and CoDel, The solution to high latency requires two elements: 1) a have been recently proposed to reduce buffer latency by avoiding strategy to signal the transport protocol to "slow down", the drawbacks of previous AQM algorithms like RED. This paper necessary to keep queues small; and 2) packet scheduling explores the performance of these new algorithms in simulated rural broadband networks where capacity is limited. We compared between flows (sharing capacity between different types of the new algorithms using Adaptive RED as a reference. We observe network flows). These solutions become more necessary at the that to achieve a small queuing delay PIE and CoDel both increase edge of the network, where speed of links is less and therefore packet loss. We therefore explored this impact on the quality of head of line blocking becomes more significant. experience for loss-sensitive unreliable multimedia applications, such as real-time and near-real-time video. The results from One way to prevent or minimise latency is to intelligently simulations show that PIE performs better than CoDel in terms of manage the queue of packets so that a router does not build a packet loss rates affecting video quality.
    [Show full text]
  • TCP Congestion Control Macroscopic Behavior for Combinations of Source and Router Algorithms
    TCP Congestion Control Macroscopic Behavior for Combinations of Source and Router Algorithms Mulalo Dzivhani1, 2, Dumisa Ngwenya3, 4, Moshe Masonta1, Khmaies Ouahada2 Council for Scientific and Industrial Research (CSIR), Meraka Institute1, Pretoria Faculty of Engineering and Built Environment, University of Johannesburg2, Johannesburg SENTECH3, Johannesburg University of Pretoria4, Pretoria South Africa Emails: [email protected] 1, 2, [email protected] 3, 4, [email protected], [email protected] Abstract—The network side of Transmission Control Protocol Most widely used TCP variant over the decades has been TCP (TCP) congestion control is normally considered a black-box in New Reno, evolving from TCP Tahoe and TCP Reno. TCP performance analysis. However, the overall performance of New Reno has proved to be efficient in fixed low speed TCP/IP networks is affected by selection of congestion control networks and networks with low BDP. However, as shown by mechanisms implemented at the source nodes as well as those [1], TCP as implemented is showing performance in high-speed implemented at the routers. The paper presents an evaluation of macroscopic behaviour of TCP for various combinations of source wide-area networks, in wireless links and for real-time flows. algorithms and router algorithms using a Dumbbell topology. In For this reason a number of source based algorithms have been particular we are interested in the throughput and fairness index. proposed in the past two decades. TCP Cubic is one of those TCP New Reno and TCP Cubic were selected for source nodes. that has gained popularity for high speed networks. Packet First-in-First-out (PFIFO) and Controlled Delay (CoDel) mechanisms were selected for routers.
    [Show full text]
  • An Extension of the Codel Algorithm
    Proceedings of the International MultiConference of Engineers and Computer Scientists 2015 Vol II, IMECS 2015, March 18 - 20, 2015, Hong Kong Hard Limit CoDel: An Extension of the CoDel Algorithm Fengyu Gao, Hongyan Qian, and Xiaoling Yang Abstract—The CoDel algorithm has been a significant ad- II. BACKGROUND INFORMATION vance in the field of AQM. However, it does not provide bounded delay. Rather than allowing bursts of traffic, we argue that a Nowadays the Internet is suffering from high latency and hard limit is necessary especially at the edge of the Internet jitter caused by unprotected large buffers. The “persistently where a single flow can congest a link. Thus in this paper we full buffer problem”, recently exposed as “bufferbloat” [1], proposed an extension of the CoDel algorithm called “hard [2], has been observed for decades, but is still with us. limit CoDel”, which is fairly simple but effective. Instead of number of packets, this extension uses the packet sojourn time When recognized the problem, Van Jacobson and Sally as the metric of limit. Simulation experiments showed that the Floyd developed a simple and effective algorithm called RED maximum delay and jitter have been well controlled with an (Random Early Detection) in 1993 [3]. In 1998, the Internet acceptable loss on throughput. Its performance is especially Research Task Force urged the deployment of active queue excellent with changing link rates. management in the Internet [4], specially recommending Index Terms—bufferbloat, CoDel, AQM. RED. However, this algorithm has never been widely deployed I. INTRODUCTION because of implementation difficulties. The performance of RED depends heavily on the appropriate setting of at least The CoDel algorithm proposed by Kathleen Nichols and four parameters: Van Jacobson in 2012 has been a great innovation in AQM.
    [Show full text]
  • A Hybrid Loss-Delay Gradient Congestion Control Algorithm for the Internet
    A Hybrid Loss-Delay Gradient Congestion Control Algorithm for the Internet Rasool Al-Saadi School of Software and Electrical Engineering Swinburne University of Technology This thesis is submitted for the degree of Doctor of Philosophy Faculty of Science, Engineering and October 2019 Technology I dedicate this thesis to my loving parents, my beloved wife Dalia and my sons Ridha and Yousif. Declaration I declare that this thesis submitted for the degree of Doctor of Philosophy contains no ma- terial that has been accepted for the award to the candidate of any other degree or diploma, except where due reference made a in the text of the examinable outcome. To the best of the candidate’s knowledge contains no material previously published or written by another person except where due reference is made in the text of the examinable outcome. Where the work is based on joint research or publications, discloses the relative contributions of the respective creators or authors. Rasool Al-Saadi October 2019 Acknowledgements I would like to express my thanks and appreciation to all the people and institutions who helped, guided, advised and supported me during my PhD journey. I would like to thank my supervisors Prof. Grenville Armitage, Dr. Jason But and Assoc. Prof. Philip Branch for their insight, patience, friendship, encouragement, continuous support and valuable advice during my PhD candidature. I had the honour of being a student under their wise supervision. Without their support, this work would never have seen the light of day. I am greatly indebted to my family, especially my wife Dalia Al-Zubaidy, my parents, my brother and my sister for their patience, persevering support and advice.
    [Show full text]