The Medusa Proxy a Tool for Exploring User-Perceived Web Performance

Total Page:16

File Type:pdf, Size:1020Kb

The Medusa Proxy a Tool for Exploring User-Perceived Web Performance The Medusa Proxy A Tool For Exploring User-Perceived Web Performance Mimika Koletsou and Geoffrey M. Voelker Department of Computer Science and Engineering University of California, San Diego 9500 Gilman Dr., MS 0114 La Jolla, CA 92093-0114 USA g fkoletsou,voelker @cs.ucsd.edu Abstract example, it is not uncommon for users to believe that Web caches impose noticeable overhead to browsing, or for content distribu- In this paper we describe the Medusa proxy, a tool for explor- tion network (CDN) companies such as Akamai [1] to make ag- ing user-perceived Web performance. The Medusa proxy is a gressive claims about improvements in download performance. non-caching forwarding proxy used in conjunction with a user’s Currently, though, there is no convenient way for users to vali- browser. The two key features that the Medusa proxy provides date those beliefs or claims. are (1) the ability to simultaneously mirror HTTP requests from Characterizing user-perceived Web performance is challenging the browser to different Web delivery systems and directly com- because of the difficulty of generalizing conclusions about per- pare the results, and (2) the ability to transform requests, e.g., to formance results across a wide range of user environments and transform Akamai URLs back into URLs to the customer’s origin network locations. Rather than try to generalize our conclusions server. across a diverse group of users, we have developed the Medusa We show how the Medusa proxy can be used to explore user- proxy as a convenient tool with which individual users can ex- perceived Web performance using two experiments that deter- plore their Web performance, and the impact of various Web de- mine (1) to what extent using large-scale caching systems like livery systems on that performance, from their desktop. the NLANR cache hierarchy improve user-perceived latency, and The Medusa proxy is a non-caching forwarding proxy used in (2) to what extent content distribution networks like Akamai im- conjunction with a user’s browser. The Medusa proxy provides prove latency compared with using their customer’s origin server. two key features for experimenting with user-perceived Web per- We found that the NLANR cache hierarchy served 63% of HTTP formance. The first feature is the ability to simultaneously mirror requests in our workload at least as fast as the origin servers, re- HTTP requests from the browser to different Web delivery sys- sulting in a decrease of average latency of 15%. By transforming tems and directly compare the results. For example, the Medusa and mirroring Akamai URLs to both Akamai edge servers and proxy can mirror requests to a proxy cache and to the origin server their customer’s origin servers, we find that Akamai edge servers to evaluate the effectiveness of the proxy cache on reducing la- are able to serve HTTP requests an average of 5.7 times as fast tency. The second feature is the ability to transform HTTP re- as their customers. However, because requests to Akamai edge quests. For example, users can install a filter that transforms an servers are only 6% of our workload, we find that using Akamai Akamai URL into the URL to the original object on the Aka- reduces overall mean latency by only 2% for our workload. mai customer’s origin server. By transforming and mirroring re- quests, the Medusa proxy can then simultaneously download the Keywords: User-Perceived Web Performance, Web Proxy, Web object from an Akamai edge server and the customer origin server Caching, Content Distribution Networks and evaluate the effectiveness of having Akamai deliver that cus- tomer’s content to the user. In this paper we describe the features of the Medusa Web proxy 1 Introduction and its uses for exploring Web performance, such as examining HTTP request headers, mirroring requests and monitoring perfor- In this paper, we describe the Medusa proxy, a tool for explor- mance, recording and replaying traces, transforming requests for ing user-perceived Web performance. There has been extensive privacy or experimentation, validating responses from experimen- work evaluating the macroscopic impact of Web infrastructure tal delivery systems, and accelerating performance; in this paper, for improving overall Web performance, such as caching systems we focus on user-perceived latency as the primary performance (e.g., [2, 6, 7, 9, 16, 17, 19, 21, 25, 30, 32, 34]), prefetching systems metric, although the Medusa proxy can be used to explore other (e.g., [4, 11, 15, 18, 27, 29]), and to a limited extent content dis- performance metrics such as bandwidth, request rate, etc. We tribution networks (e.g., [20, 22]). Surprisingly, however, there then use the Medusa proxy to answer the following performance has been relatively little work characterizing the impact on Web and optimization questions: performance as perceived by the user of various Web delivery systems. Although we have a relatively good understanding of To what extent do large-scale cache systems like the how improvements in Web infrastructure impact overall latency, NLANR cache hierarchy improve user-perceived latency? server load, and network utilization for large groups of users, we What is the potential of using the Medusa proxy as a Web do not have a good understanding of how those improvements accelerator, returning the fastest-of-n mirrored parallel re- affect individual users at the desktop. In particular, there is no quests? convenient mechanism by which users can determine the impact new developments in Web infrastructure, such as caching, con- To what extent do content distribution networks like Akamai tent distribution, or prefetching, have on their Web browsing. For improve user-perceived latency over the performance of the 1 customer’s origin server? single user to all accessed services through a potential set of Web delivery systems. The rest of this paper is organized as follows. Section 2 dis- There has been comparatively little work experimentally evalu- cusses related work. Section 3 describes the features and im- ating the performance of content distribution networks. In one re- plementation of the Medusa proxy. Section 4 describes our ex- cent study, Johnson et al. compare the performance of two CDNs, perimental methodology and results for evaluating the NLANR, Akamai and Digital Island, in terms of their ability to select the Akamai, and Squid Web delivery systems on user-perceived Web servers with minimum latency to the client [22]. However, they performance, and Section 5 concludes. do not evaluate the impact of the CDNs on user-perceived latency, only on the relative performance of server selection within the CDN. There are tools that report user Web download performance, 2 Related Work such as Net.Medic [28], but they do not enable users to answer exploratory questions such as whether a proxy cache is degrad- The work most closely related to the Medusa proxy is the Simulta- ing their performance, or how much added value CDNs are actu- neous Proxy Evaluation (SPE) system proposed by Davison [13]. ally providing to them. Commercial systems such as the Lateral The SPE system has a similar goal in that it proposes to use mir- Line Network from Appliant.com [3] and various services from roring to compare the performance of Web delivery systems such Keynote [24] report on the health and performance of customer as proxy caches from the perspective of clients. However, the Web sites from the perspective of many distributed users. How- goals of comparing a Web delivery system with the origin servers, ever, as with the personal Web performance tools, these services or transforming requests, is not proposed (although it would seem do not compare alternative delivery systems and their impact on straightforward to incorporate). SPE was not implemented when performance. originally proposed, but Davison and Krishnan are currently im- plementing it as the Rutgers Online Proxy Evaluator (ROPE) [14]. There have been a number of studies investigating Web user behavior and client performance using passive monitoring. Cunha 3 Medusa Proxy et al. collected an early client trace at Boston University by instru- menting the Mosaic browser used by people in their department, The Medusa proxy is a non-caching forwarding proxy that sup- ports the Proxy Request form of HTTP requests. Since it was de- and derived statistical characteristics of the client workload such signed to explore user-perceived latency, it is typically configured as object size and popularity [12]. Catledge and Pitkow also col- as a personal proxy for the Web browser and it executes alongside lected a trace of Mosaic events from users at Georgia Tech, and the browser on the same machine. It can also serve as a proxy for used it to study methods of user interaction with the browser as well as user navigation patterns [8]. Kelly has recently collected multiple clients, but we have yet to explore its use in that way. For a very large anonymized client trace from WebTV, and has used portability, we implemented the Medusa proxy in Java to support widespread use. Finally, our support for HTTP 1.1 persistent and it to explore browser cache performance, reference locality, and pipelined connections is ongoing, and the Medusa proxy currently document aliasing [23]. All of these studies have used passive only supports HTTP 1.0. measurement to explore user behavior and client performance, whereas we use the Medusa proxy to compare the performance The Medusa proxy has a number of features for exploring user- of various Web delivery systems using active measurement. perceived Web performance: Krishnamurthy and Wills [26] studied end-to-end Web per- Mirroring. The Medusa proxy can be configured to mirror HTTP formance from a collection of client sites distributed around the requests from multiple incoming sources to multiple destinations.
Recommended publications
  • Managing the Performance Impact of Web Security
    Electronic Commerce Research, 5: 99–116 (2005) 2005 Springer Science + Business Media, Inc. Manufactured in the Netherlands. Managing the Performance Impact of Web Security ADAM STUBBLEFIELD and AVIEL D. RUBIN Johns Hopkins University, USA DAN S. WALLACH Rice University, USA Abstract Security and performance are usually at odds with each other. Current implementations of security on the web have been adopted at the extreme end of the spectrum, where strong cryptographic protocols are employed at the expense of performance. The SSL protocol is not only computationally intensive, but it makes web caching impossible, thus missing out on potential performance gains. In this paper we discuss the requirements for web security and present a solution that takes into account performance impact and backwards compatibil- ity. Keywords: security, web performance scalability, Internet protocols 1. Introduction Security plays an important role in web transactions. Users need assurance about the au- thenticity of servers and the confidentiality of their private data. They also need to know that any authentication material (such as a password or credit card) they send to a server is only readable by that server. Servers want to believe that data they are sending to paying customers is not accessible to all. Security is not free, and in fact, on the web, it is particularly expensive. This paper focuses on managing the performance impact of web security. The current web security standard, SSL, has a number of problems from a performance perspective. Most obviously, the cryptographic operations employed by SSL are computationally expensive, especially for the web server. While this computational limitation is slowly being overcome by faster processors and custom accelerators, SSL also prevents the content it delivers from being cached.
    [Show full text]
  • Direct Internet 3 User Manual for the Windows Operating Systems
    DIRECT INTERNET 3 User Manual for the Mac OS® Operating System Iridium Communications Inc. Rev. 2; October 29, 2010 Table of Contents 1 OVERVIEW ............................................................................................................................................1 2 HOW IT WORKS ....................................................................................................................................1 3 THE DIAL-UP CONNECTION ...............................................................................................................2 3.1 Connect ..........................................................................................................................................2 3.2 Disconnect .....................................................................................................................................5 4 DIRECT INTERNET 3 WEB ACCELERATOR ......................................................................................6 4.1 Launch ...........................................................................................................................................6 4.2 The User Interface Menu ...............................................................................................................7 4.3 Start and Stop ................................................................................................................................8 4.4 Statistics .........................................................................................................................................9
    [Show full text]
  • Combating Spyware in the Enterprise.Pdf
    www.dbebooks.com - Free Books & magazines Visit us at www.syngress.com Syngress is committed to publishing high-quality books for IT Professionals and delivering those books in media and formats that fit the demands of our cus- tomers. We are also committed to extending the utility of the book you purchase via additional materials available from our Web site. SOLUTIONS WEB SITE To register your book, visit www.syngress.com/solutions. Once registered, you can access our [email protected] Web pages. There you will find an assortment of value-added features such as free e-booklets related to the topic of this book, URLs of related Web site, FAQs from the book, corrections, and any updates from the author(s). ULTIMATE CDs Our Ultimate CD product line offers our readers budget-conscious compilations of some of our best-selling backlist titles in Adobe PDF form. These CDs are the perfect way to extend your reference library on key topics pertaining to your area of exper- tise, including Cisco Engineering, Microsoft Windows System Administration, CyberCrime Investigation, Open Source Security, and Firewall Configuration, to name a few. DOWNLOADABLE EBOOKS For readers who can’t wait for hard copy, we offer most of our titles in download- able Adobe PDF form. These eBooks are often available weeks before hard copies, and are priced affordably. SYNGRESS OUTLET Our outlet store at syngress.com features overstocked, out-of-print, or slightly hurt books at significant savings. SITE LICENSING Syngress has a well-established program for site licensing our ebooks onto servers in corporations, educational institutions, and large organizations.
    [Show full text]
  • Performance Analysis of a Client-Side Caching/Prefetching System for Web Traffic
    1 Performance Analysis of a Client-Side Caching/Prefetching System for Web Traffic Abdullah Balamash and Marwan Krunz Department of Electrical & Computer Engineering University of Arizona Tucson, AZ 85721 fbalamash,[email protected] Philippe Nain INRIA 06902 Sophia Antipolis, France [email protected] Abstract Network congestion remains one of the main barriers to the continuing success of the Internet. For Web users, congestion manifests itself in unacceptably long response times. One possible remedy to the latency problem is to use caching at the client, at the proxy server, or within the Internet. However, Web documents are becoming increasingly dynamic (i.e., have short lifetimes), which limits the potential benefit of caching. The performance of a Web caching system can be dramatically increased by integrating document prefetching (a.k.a. “proactive caching”) into its design. Although prefetching reduces the response time of a requested document, it also increases the network load, as some documents will be unnecessarily prefetched (due to the imprecision in the prediction algorithm). In this study, we analyze the confluence of the two effects through a tractable mathematical model that enables us to establish the conditions under which prefetching reduces the average response time of a requested document. The model accommodates both passive client and proxy caching along with prefetching. Our analysis is used to dynamically compute the “optimal” number of documents to prefetch in the subsequent client’s idle (think) period. In general, this optimal number is determined through a simple numerical procedure. Closed-form expressions for this optimal number are obtained for special yet important cases.
    [Show full text]
  • Web Acceleration Mechanics
    Web Acceleration Mechanics Alexander Krizhanovsky Tempesta Technologies, Inc. [email protected] Who am I? CEO at Tempesta Technologies, INC Custom software development since 2008: ● Network security: WAF, VPN, DPI etc. e.g. Positive Technologies AF, “Visionar” Gartner magic quadrant’15 ● Databases: one of the top MariaDB contributors ● Perfomance tuning Tempesta FW – Linux Application Delivery Controller Tempesta FW: Application Delivery Controller (ADC) Web accelerator Load balancing Backend connections management TLS termination Protocol downgrade Cache responses Agenda & Disclaimer Mechanics: ● HTTP connections & messages management ● HTTP decoders & parsers ● Web caches ● Network I/O ● Multitasking ● TLS The web accelerators are mentioned only as implementation examples Some software is just more familiar to me HTTP connections & messages management Server connections New connections or persistent connections (usual) HTTP keep-alive connections Keep-Alive: timeout=5, max=10000 Reestablish closed KA connection New connections if all are busy N backend connections = N backend requests in-flight DDoS and flash crowds: as many server connections as client connections Run out of port numbers HTTP/1 message pipelining Mostly unused by proxies Squid, Tempesta FW, Polipo Messages multiplexing Forwarding and reforwarding issues Security issues ● Breaking authentication ● HTTP Response splitting HTTP Response Splitting attack (Web cache poisoning) [CRLF CRLF] Client: /redir_lang.jsp?lang=foobar%0d%0aContent-Length:%200%0d%0a%0d%0a HTTP/1.1%20200%20OK%0d%0aContent-Type:%20text/html%0d%0a
    [Show full text]
  • Web Cache for Grid HTTP Access
    Web cache for grid HTTP access Duarte Meneses Fabrizio Furano [email protected] [email protected] April 1, 2014 1 1 Introduction Sites have usually several jobs and people accessing files in the grid. Even though files are often on site, these files may come from several storage elements, which can be located at any other site. Due to recent developments, many storage elements have currently an HTTP interface, implemented for example, with XRootd or with a WebDAV server. Thus, the files can be accessed with this widely known protocol using any HTTP client, instead of forcing the usage of grid-specific protocols. The use of HTTP could also enable easy caching of data using existing web caching products. In this document we explore further this possibility. The HTTP resources (files) can be re-fetched several times within the same site, ei- ther by the same user or not. This can be especially frequent with batch jobs, that pull periodically the same files from the network. To optimize the access to the data and reduce the load in the network, we are in- vestigating potential advantages and implications of deploying HTTP caches at the sites. Accessing a cached file within the site could bring a significant performance boost, pactic- ularly for sites that constantly consume external resources, i.e. resources located in other sites. Using web caches is a solution adopted often by websites for optimizing the access to it by caching its contents (possibility at several locations, close to the users). These web caches are deployed in reverse mode, and the DNS is usually pointed to it.
    [Show full text]
  • Tempesta FW Linux Application Delivery Controller
    Tempesta FW Linux Application Delivery Controller Alexander Krizhanovsky Tempesta Technologies, Inc. [email protected] Who am I? CEO & CTO at Tempesta Technologies (Seattle, WA) Developing Tempesta FW – open source Linux Application Delivery Controller (ADC) Custom software development in: ● high performance network traffic processing e.g. WAF mentioned in Gartner magic quadrant ● Databases e.g. MariaDB SQL System Versioning is coming soon (https://github.com/tempesta-tech/mariadb_10.2) Challenges Usual Web accelerators aren’t suitable for HTTP filtering Kernel HTTP accelerators are better, but they’re dead => need a hybrid of HTTP accelerator and a firewall ● Very fast HTTP parser to process HTTP floods ● Very fast Web cache to mitigate DDoS which we can’t filter out ● Network I/O optimized for massive ingress traffic ● Advanced filtering abilities at all network layers L7 DDoS mitigation Web accelerator? Application Delivery Controller (ADC) Use cases CMSes, CDNs, virtual hostings, heavy loaded Web sites, OEMs in Web security etc. Usual ADC cases: ● When you need performance ● Web content acceleration ● Web application protection ● HTTP load balancing Application layer DDoS Service from Cache Rate limit Nginx 22us 23us (Additional logic in limiting module) Fail2Ban: write to the log, parse the log, write to the log, parse the log… Application layer DDoS Service from Cache Rate limit Nginx 22us 23us (Additional logic in limiting module) Fail2Ban: write to the log, parse the log, write to the log, parse the log… - really in 21th century?! tight integration of Web accelerator and a firewall is needed Web-accelerator capabilities Nginx, Varnish, Apache Traffic Server, Squid, Apache HTTPD etc.
    [Show full text]
  • Optimasi Apache Web Server Menggunakan Varnish Web Cache Dan Reverse Proxy Nginx
    OPTIMASI APACHE WEB SERVER MENGGUNAKAN VARNISH WEB CACHE DAN REVERSE PROXY NGINX Skripsi disusun sebagai salah satu syarat untuk memperoleh gelar Sarjana Komputer Program Studi Teknik Informatika Oleh Anindya Putri Arunawati 4611416030 JURUSAN ILMU KOMPUTER FAKULTAS MATEMATIKA DAN ILMU PENGETAHUAN ALAM UNIVERSITAS NEGERI SEMARANG 2020 ii iii iv MOTTO DAN PERSEMBAHAN MOTTO • Jika yang dapat kau lakukan hanyalah merangkak, maka mulailah merangkak (Rumi) • Semakin dia yakin kepada Allah, semakin sedikit dia mengeluh dalam hidupnya (ust. Hanan Attaki) • Jangan pernah lelah untuk berdoa dan percayakan semuanya pada Allah, Allah yang mengatur segalanya (Anindya) PERSEMBAHAN Skripsi ini saya persembahkan kepada: • Kedua Orang Tua saya Bapak Achmad Daserun dan Ibu Watini yang telah mencurahkan keringatnya untuk membiayai pendidikan saya, yang selalu memberikan kasih sayang, doa, dan dukungannnya. • Adik saya, Hilal Hubdin yang telah memberikan dukungan serta doa yang terus dipanjatkan. • Teman-teman saya di jurusan Ilmu Komputer, Fakultas MIPA, serta teman-teman di Universitas Negeri Semarang. • Semua pihak yang tidak dapat disebutkan satu persatu yang telah membantu hingga terselesaikannya penulisan skripsi ini. • Almamater, Universitas Negeri Semarang. v PRAKATA Puji syukur penulis panjatkan kepada Allah Subhanahu wa ta'ala atas berkat rahmat dan hidayah-Nya penulis dapat menyelesaikan skripsi yang berjudul “Optimasi Apache Web Server Menggunakan Varnish Web Cache Dan Reverse Proxy Nginx”. Penulis menyadari bahwa penulisan skripsi ini tidak akan selesai tanpa adanya dukungan serta bantuan dari berbagai pihak. Oleh karena itu, penulis ingin menyampaikan ucapan terima kasih kepada: 1. Prof. Dr. Fathur Rokhman, M.Hum., Rektor Universitas Negeri Semarang. 2. Dr. Sugianto M.Si., Dekan Fakultas Matematika dan Ilmu Pengetahuan Alam Universitas Negeri Semarang.
    [Show full text]
  • Performance Evaluation of Raspberry Pi 3B As a Web Server
    Performance evaluation of Raspberry pi 3B as a web server Meron Istifanos and Israel Tekahun 18-06-2020 Faculty of Computing Blekinge Institute of Technology SE-371 79 Karlskrona Sweden This thesis is submitted to the Faculty of Computing at Blekinge Institute of Tech- nology in partial fulfillment of the requirements for the bachelor’s degree in software engineering. The thesis is equivalent to 10 weeks of full-time studies. Contact Information: Author(s): Meron Istifanos E-mail: [email protected] Israel Tekahun E-mail: [email protected] University advisor: Dr. Michel Nass Department of Software Engineering Faculty of Computing Internet : www.bth.se Blekinge Institute of Technology Phone : +46 455 38 50 00 SE-371 79 Karlskrona, Sweden Fax : +46 455 38 50 57 Abstract Context. During the usage of a product, evaluating its performance quality is a crucial procedure. Web servers are one of the most used technological products in today’s modern world[1]. Thus, in this thesis we will evaluate and compare the performances of two web servers. The servers that are tested in the experiment are a raspberry pi 3B and a personal HP laptop. Objectives. The main objective of the study is to evaluate the performance of a raspberry pi 3B as a web server. In order to give a clearer image of how the rasp- berry pi performs, the laptop will also be evaluated and its performance will be used as a contrast during the study. Realization. To fulfill our objective, an experiment was conducted with the help of a performance testing tool called apache bench.
    [Show full text]
  • SO De Xarxes D'àrea Local
    Gestió de xarxes d’àrea local TEMA 3: S.O. De Xarxes d'Àrea Local Introducció SO de Xarxes 3 Introducció......................................................................................................................4 3.0 Instal·lació d'un SO Lan..........................................................................................4 3.0.1 Instal·lació de Linux .......................................................................................5 3.0.1.1 Instal·lació amb els Cds o el DVD..........................................................5 3.0.1.2 sense Cds.................................................................................................5 3.1 Instal·lació d’una LAN............................................................................................6 3.1.1 LAN en DOS...................................................................................................6 3.1.2 LAN en Windows...........................................................................................7 3.1.21 Configuració TCP/IP des de consola......................................................13 3.1.3 LAN en Linux...............................................................................................14 3.1.3.1 LAN sense fils en Linux........................................................................18 3.2 Configuració automàtica de xarxes.......................................................................21 3.2.1 Bootstrap protocol.........................................................................................21
    [Show full text]
  • A Ground-Up Approach to High-Throughput Cloud Computing
    Università degli Studi di Torino Scuola di Dottorato in Scienze ed Alte Tecnologie XXVI ciclo Tesi di Dottorato A ground-up approach to High-Throughput Cloud Computing in High-Energy Physics Dario Berzano Relatore Correlatori Prof. Massimo Masera Dott. Gerardo Ganis Università degli Studi di Torino CERN Dott. Stefano Bagnasco INFN Torino CERN-THESIS-2014-376 16/04/2014 Controrelatore Prof. Marco Paganoni Università degli Studi di Milano-Bicocca Anno Accademico 2012–2013 «Simplicity is the ultimate sophistication.» Apple II commercial (1977) Table of Contents 1 Virtualization and Cloud Computing in High Energy Physics 13 1.1 Computing in the LHC era . 13 1.1.1 Hierarchy of computing centres . 16 1.1.2 Event processing paradigm . 17 1.1.3 Analysis framework and data format: ROOT . 18 1.1.4 The ALICE Experiment . 19 1.2 Virtualization and Cloud Computing . 21 1.2.1 State of the art . 23 1.3 From the Grid to clouds? . 25 1.3.1 Grid over clouds and more . 26 1.4 Overview of this work . 27 2 Building a Private Cloud out of a Grid Tier-2 29 2.1 Overview of the INFN Torino Computing Centre . 29 2.1.1 Grid VOs and ALICE . 30 2.2 Rationale of the Cloud migration . 32 2.2.1 No performance loss through virtualization . 32 2.2.2 Abstraction of resources in the Grid and the cloud . 32 2.2.3 Harmonization of diverse use cases in a private cloud . 33 2.2.4 IaaS: separating administrative domains . 34 2.2.4.1 Virtual farm administrator .
    [Show full text]
  • Automatic and Dynamic Configuration of Data Compression for Web Servers
    Automatic and Dynamic Configuration of Data Compression for Web Servers Eyal Zohar, Yahoo! Labs; Yuval Cassuto, Technion—Israel Institute of Technology https://www.usenix.org/conference/lisa14/conference-program/presentation/zohar This paper is included in the Proceedings of the 28th Large Installation System Administration Conference (LISA14). November 9–14, 2014 • Seattle, WA ISBN 978-1-931971-17-1 Open access to the Proceedings of the 28th Large Installation System Administration Conference (LISA14) is sponsored by USENIX Automatic and Dynamic Configuration of Data Compression for Web Servers Eyal Zohar∗ Yuval Cassuto Yahoo! Labs Technion - Israel Institute of Technology Haifa, Israel Department of Electrical Engineering [email protected] Haifa, Israel [email protected] Abstract depends on how compressible the data is). To maintain good quality-of-experience, system administrators must HTTP compression is an essential tool for web speed up monitor their services and adapt their configuration to and network cost reduction. Not surprisingly, it is used changing conditions on a regular basis. As web-related by over 95% of top websites, saving about 75% of web- technologies get complex, ensuring a healthy and robust page traffic. web service requires significant expertise and constant The currently used compression format and tools were attention. designed over 15 years ago, with static content in mind. The increasing complexity raises the popularity of au- Although the web has significantly evolved since and be- tomated solutions for web configuration management [7, came highly dynamic, the compression solutions have 25]. Ideally, an automatic configuration management not evolved accordingly. In the current most popular should let system administrators specify high-level de- web-servers, compression effort is set as a global and sired behaviors, which will then be fulfilled by the sys- static compression-level parameter.
    [Show full text]