Distribution, Data, Deployment: Software Architecture Convergence

Total Page:16

File Type:pdf, Size:1020Kb

Distribution, Data, Deployment: Software Architecture Convergence FEATURE: BIG DATA SOFTWARE SYSTEMS Addressing the challenges of soft- Distribution, ware for big data systems requires careful design tradeoffs spanning the distributed software, data, and deployment architectures. It also Data, requires extending traditional soft- ware architecture design knowledge to account for the tight coupling Deployment that exists in scalable systems. Scale drives a consolidation of concerns, Software Architecture so that distribution, data, and de- ployment architectural qualities can no longer be effectively considered Convergence separately. To illustrate this, we’ll use an example from our current re- in Big Data Systems search in healthcare informatics. The Challenges Ian Gorton and John Klein, Software Engineering Institute of Big Data Data-intensive systems have long been built on SQL database tech- // Big data systems present many challenges to nology, which relies primarily on vertical scaling—faster processors software architects. In particular, distributed and bigger disks—as workload or software architectures become tightly coupled storage requirements increase. SQL databases’ inherent vertical-scaling to data and deployment architectures. This limitations4 have led to new prod- causes a consolidation of concerns; designs ucts that relax many core tenets of must be harmonized across these three relational databases. Strictly de ned normalized data models, strong data architectures to satisfy quality requirements. // consistency guarantees, and the SQL standard have been replaced by schemaless and intentionally de- normalized data models, weak con- sistency, and proprietary APIs that expose the underlying data man- agement mechanisms to the pro- grammer. These NoSQL products4 typically are designed to scale hori- zontally across clusters of low-cost, THE EXPONENTIAL GROWTH of data repositories ever constructed. moderate- performance servers. They data over the last decade has fueled a Their pioneering efforts,2,3 along achieve high performance, elastic new specialization for software tech- with those of numerous other big storage capacity, and availability by nology: data-intensive, or big data, data innovators, have provided a va- partitioning and replicating datasets software systems.1 Internet-born or- riety of open source and commercial across a cluster. Prominent examples ganizations such as Google and Am- data management technologies that of NoSQL databases include Cas- azon are on this revolution’s cutting let any organization construct and sandra, Riak, and MongoDB (see the edge, collecting, managing, storing, operate massively scalable, highly sidebar “NoSQL Databases”). and analyzing some of the largest available data repositories. Distributed databases have funda- 78 IEEE SOFTWARE | PUBLISHED BY THE IEEE COMPUTER SOCIETY 0740-7459/15/$31.00 © 2015 IEEE NOSQL DATABASES The rise of big data applications has caused a signi cant ux typically, some form of directed graph. They can provide ex- in database technologies. While mature relational database ceptional performance for problems involving graph traversals technologies continue to evolve, a spectrum of databases and subgraph matching. Because ef cient graph partition- called NoSQL has emerged over the past decade. The rela- ing is an NP-hard problem, these databases tend to be less tional model imposes a strict schema, which inhibits data concerned with horizontal scaling and commonly offer ACID evolution and causes dif culties scaling across clusters. In (atomicity, consistency, isolation, durability) transactions to response, NoSQL databases have adopted simpler data mod- provide strong consistency. Examples include Neo4j (www els. Common features include schemaless records, allowing .neo4j.org) and GraphBase (http://graphbase.net). data models to evolve dynamically, and horizontal scaling, by NoSQL technologies have many implications for applica- sharding (partitioning and distributing) and replicating data tion design. Because there’s no equivalent of SQL, each tech- collections across large clusters. Figure A illustrates the four nology supports its own query mechanism. These mecha- most prominent data models, whose characteristics we sum- nisms typically make the application programmer responsible marize here. More comprehensive information is at http:// for explicitly formulating query executions, rather than relying nosql-database.org. on query planners that execute queries based on declara- Document databases (see Figure A1) store collections of tive speci cations. The programmer is also responsible for objects, typically encoded using JSON (JavaScript Object combining results from different data collections. This lack of Notation) or XML. Documents have keys, and you can build the ability to perform JOINs forces extensive denormalization secondary indexes on nonkey elds. Document formats are of data models so that JOIN-style queries can be ef ciently self-describing; a collection might include documents with executed by accessing a single data collection. When data- different formats. Leading examples are MongoDB (www bases are sharded and replicated, the programmer also must .mongodb.org) and CouchDB (http://couchdb.apache.org). manage consistency when concurrent updates occur and Key–value databases (see Figure A2) implement a distrib- must design applications to tolerate stale data due to latency uted hash map. Records can be accessed primarily through in update replication. key searches, and the value associated with each key is treated as opaque, requir- ing reader interpretation. This simple model “id”: “1”, “Name”: “John”, “Employer”: “SEI” facilitates sharding and replication to cre- “id”: “2”, “Name”: “Ian”, “Employer”: “SEI”, “Previous”: “PNNL” ate highly scalable and available systems. (1) Examples are Riak (http://riak.basho.com) “key”: “1”, value{“Name”: “John”, “Employer”: “SEI”} and DynamoDB (http://aws.amazon.com/ “key”: “2”, value{“Name”: “Ian”, “Employer”: “SEI”, “Previous”: “PNNL”} dynamodb). (2) Column-oriented databases (see Figure “row”: “1”, “Employer” “Name” A3) extend the key–value model by organiz- “SEI” “John” ing keyed records as a collection of columns, “row”: “2”, “Employer” “Name” “Previous” where a column is a key–value pair. The key “SEI” “Ian” “PNNL” (3) becomes the column name; the value can be an arbitrary data type such as a JSON Node: Employee “is employed by” Node: Employer “id”: “1”, “Name”: “John” “Name”: “SEI” document or binary image. A collection can “id”: “2”, “Name”: “Ian” contain records with different numbers of “Name”: “PNNL” columns. Examples are HBase (http://hbase “previously employed by” (4) .apache.org) and Cassandra (https:/ /cassandra.apache.org). Graph databases (see Figure A4) orga- FIGURE A. Four major NoSQL data models. (1) A document store. (2) A key– nize data in a highly connected structure— value store. (3) A column store. (4) A graph store. MAY/JUNE 2015 | IEEE SOFTWARE 79 FEATURE: BIG DATA SOFTWARE SYSTEMS mental quality constraints, defined igently evaluate candidate database lion.10 Analysis of petabytes of data by Eric Brewer’s CAP (consistency, technologies and select databases across patient populations, taken availability, partition tolerance) the- that can satisfy application require- from diverse sources such as insur- orem.5 When a network partition ments. This often leads to polyglot ance payers, public health entities, occurs (causing an arbitrary mes- persistence—using different data- and clinical studies, can reduce costs sage loss between cluster nodes), a base technologies to store different by improving patient outcomes. In system must trade consistency (all datasets in a single system, to meet addition, operational efficiencies readers see the same data) against quality attribute requirements.8 can extract new insights for disease availability (every request receives a Furthermore, as data volumes treatment and prevention. success or failure response). Daniel grow to petascale and beyond, the Across these and many other do- Abadi’s PACELC provides a practi- required hardware resources grow mains, big data systems share four cal interpretation of this theorem.6 If from hundreds to tens of thousands requirements that drive the design of a partition (P) occurs, a system must of servers. At this deployment scale, suitable software solutions. Collec- trade availability (A) against consis- many widely used software architec- tively, these requirements represent tency (C). Else (E), in the usual case ture patterns are unsuitable. Archi- a significant departure from tradi- of no partition, a system must trade tectural and algorithmic approaches tional business systems, which are latency (L) against consistency (C). that are sensitive to hardware re- relatively well constrained in terms Additional design challenges for source use can significantly reduce of data growth, analytics, and scale. scalable data-intensive systems stem overall costs. For more on this, see First, from social media sites to from the following three issues. the sidebar “Why Scale Matters.” high-resolution sensor data collec- tion in the power grid, big data sys- tems must be able to sustain write- heavy workloads.1 Because writes Big data systems must be able to sustain are costlier than reads, systems can use data sharding (partitioning and write-heavy workloads. distribution) to spread write opera- tions across disks and can use rep- lication to provide high availability. Sharding and replication introduce First, achieving high scalability and Big Data Application
Recommended publications
  • ( 12 ) United States Patent
    US010430263B2 (12 ) United States Patent ( 10 ) Patent No. : US 10 ,430 , 263 B2 Polar Seminario (45 ) Date of Patent : Oct. 1 , 2019 ( 54 ) DEVICES, SYSTEMS AND METHODS FOR ( 56 ) References Cited VALIDATING AND UPGRADING FIRMWARE IN INTELLIGENT ELECTRONIC DEVICES U . S . PATENT DOCUMENTS (71 ) Applicant: Electro Industries /Gauge Tech , 2 ,883 , 255 A 4 / 1959 Anderson Westbury , NY (US ) 2 , 987 , 704 A 6 / 1961 Gimpel et al . ( Continued ) ( 72 ) Inventor: Jorge A . Polar Seminario , Westbury, NY (US ) FOREIGN PATENT DOCUMENTS (73 ) Assignee : Electro Industries /Gauge Tech , CN 101799681 A 8 /2010 Westbury, NY (US ) WO 2009044369 A2 4 / 2009 ( * ) Notice: Subject to any disclaimer, the term of this OTHER PUBLICATIONS patent is extended or adjusted under 35 U . S . C . 154 ( b ) by 59 days. Microsoft TechNet , Using DNS Servers with DHCP, Jan . 21, 2005 , Accessed from https: // technet. microsoft . com / en -us / library / ( 21 ) Appl. No. : 15 /421 , 573 cc787034 on Jun . 9 , 2015 ; pp . 1 - 4 . (22 ) Filed : Feb . 1, 2017 (Continued ) Primary Examiner — Charles Ehne (65 ) Prior Publication Data (74 ) Attorney , Agent, or Firm — Gerald E . Hespos ; US 2017/ 0220404 A1 Aug . 3 , 2017 Michael J . Porco ; Matthew T . Hespos Related U . S . Application Data (57 ) ABSTRACT (60 ) Provisional application No . 62 /289 , 440 , filed on Feb . Apparatuses , systems, and method for validating and 1 , 2016 . upgrading firmware in an intelligent electronic device ( IED ) (51 ) Int. Ci. are provided. In one aspect of the present disclosure , an IED G06F 11 /00 ( 2006 . 01 ) is provided including at least one processor and at least one GO6F 11/ 07 (2006 .
    [Show full text]
  • PHP Beyond the Web Shell Scripts, Desktop Software, System Daemons and More
    PHP Beyond the web Shell scripts, desktop software, system daemons and more Rob Aley This book is for sale at http://leanpub.com/php This version was published on 2013-11-25 This is a Leanpub book. Leanpub empowers authors and publishers with the Lean Publishing process. Lean Publishing is the act of publishing an in-progress ebook using lightweight tools and many iterations to get reader feedback, pivot until you have the right book and build traction once you do. ©2012 - 2013 Rob Aley Tweet This Book! Please help Rob Aley by spreading the word about this book on Twitter! The suggested hashtag for this book is #phpbeyondtheweb. Find out what other people are saying about the book by clicking on this link to search for this hashtag on Twitter: https://twitter.com/search?q=#phpbeyondtheweb Contents Welcome ............................................ i About the author ...................................... i Acknowledgements ..................................... ii 1 Introduction ........................................ 1 1.1 “Use PHP? We’re not building a website, you know!”. ............... 1 1.2 Are you new to PHP? ................................. 2 1.3 Reader prerequisites. Or, what this book isn’t .................... 3 1.4 An important note for Windows and Mac users ................... 3 1.5 About the sample code ................................ 4 1.6 External resources ................................... 4 1.7 Book formats/versions available, and access to updates ............... 5 1.8 English. The Real English. .............................. 5 2 Getting away from the Web - the basics ......................... 6 2.1 PHP without a web server .............................. 6 2.2 PHP versions - what’s yours? ............................. 7 2.3 A few good reasons NOT to do it in PHP ...................... 8 2.4 Thinking about security ...............................
    [Show full text]
  • 2.3 Apache Chemistry
    UNIVERSITY OF DUBLIN TRINITY COLLEGE DISSERTATION PRESENTED TO THE UNIVERSITY OF DUBLIN,TRINITY COLLEGE IN PARTIAL FULFILMENT OF THE REQUIREMENTS FOR THE DEGREE OF MAGISTER IN ARTE INGENIARIA Content Interoperability and Open Source Content Management Systems An implementation of the CMIS standard as a PHP library Author: Supervisor: MATTHEW NICHOLSON DAVE LEWIS Submitted to the University of Dublin, Trinity College, May, 2015 Declaration I, Matthew Nicholson, declare that the following dissertation, except where otherwise stated, is entirely my own work; that it has not previously been submitted as an exercise for a degree, either in Trinity College Dublin, or in any other University; and that the library may lend or copy it or any part thereof on request. May 21, 2015 Matthew Nicholson i Summary This paper covers the design, implementation and evaluation of PHP li- brary to aid in use of the Content Management Interoperability Services (CMIS) standard. The standard attempts to provide a language indepen- dent and platform specific mechanisms to better allow content manage- ment systems (CM systems) work together. There is currently no PHP implementation of CMIS server framework available, at least not widely. The name given to the library is Elaphus CMIS. The implementation should remove a barrier for making PHP CM sys- tems CMIS compliant. Aswell testing how language independent the stan- dard is and look into the features of PHP programming language. The technologies that are the focus of this report are: CMIS A standard that attempts to structure the data within a wide range of CM systems and provide standardised API for interacting with this data.
    [Show full text]
  • Cloud Computing Zun¨Achst Heiter Bis Wolkig, Sp¨Ater Dauerhaft Bedeckt“ ”
    Ausgew¨ahlte Kapitel der Systemsoftware: Cloud Computing Zun¨achst heiter bis wolkig, sp¨ater dauerhaft bedeckt\ " Timo H¨onig Friedrich-Alexander-Universit¨at Erlangen-Nurnberg¨ Lehrstuhl Informatik 4 (Verteilte Systeme und Betriebssysteme) http://www4.informatik.uni-erlangen.de Sommersemester 2010 Cloud Computing: Agenda Zun¨achst heiter bis wolkig, sp¨ater dauerhaft bedeckt\ " Semesterabschluss Keine Wiederholung Aktive Teilnahme erwunscht¨ Themen Ph¨anomen Cloud Computing: Ursprung und Wurzeln Prinzip der Offenheit Fallstudie Twitter HS AKSS (SoSe 10) Cloud Computing: Zun¨achst heiter bis wolkig, sp¨ater dauerhaft bedeckt\1 " Eric Schmidt (Google) Search Engine Strategies Conference, San Jose, Cloud Computing: Der Anfang. We call it cloud computing (. ) 9. August 2006 HS AKSS (SoSe 10) Cloud Computing: Zun¨achst heiter bis wolkig, sp¨ater dauerhaft bedeckt\2 " Cloud Computing: Der Anfang. We call it cloud computing (. ) Eric Schmidt (Google) Search Engine Strategies Conference, San Jose, 9. August 2006 HS AKSS (SoSe 10) Cloud Computing: Zun¨achst heiter bis wolkig, sp¨ater dauerhaft bedeckt\2 " Cloud Computing: . und seine Folgen. *) Hochrechnung Gravierende Auswirkungen des Modeworts Cloud Computing\ " Forschung Wirtschaft Cloud Computing Fokus auf Technik Cloud ^=Internet, Cloud Computing ^=Internet + ? HS AKSS (SoSe 10) Cloud Computing: Zun¨achst heiter bis wolkig, sp¨ater dauerhaft bedeckt\3 " Infrastruktur, Hard- und Software-Technologie X Datenubertragungsrate,¨ Commodity Hardware, Virtualisierung Systemsoftware, Verteilte
    [Show full text]
  • 2015 Open Source Yearbook
    Opensource.com/yearbook . ........ OPENSOURCE.COM ................... Opensource.com publishes stories about creating, adopting, and sharing open source solutions. Visit Opensource.com to learn more about how the open source way is improving technologies, education, business, government, health, law, entertainment, humanitarian efforts, and more. Submit a story idea: https://opensource.com/story Email us: [email protected] Chat with us in Freenode IRC: #opensource.com Twitter @opensourceway: https://twitter.com/opensourceway Google+: https://plus.google.com/+opensourceway Facebook: https://www.facebook.com/opensourceway Instagram: https://www.instagram.com/opensourceway FROM THE EDITOR ............................. Dear Open Source Yearbook reader, The “open source” label was created back in 1998, not long after I got my start in tech publishing [1]. Fast forward to late 2014, when I was thinking about how much open source technologies, commu- nities, and business models have changed since 1998. I realized that there was no easy way—like a yearbook—to thumb through tech history to get a feel for open source. Sure, you can flip through the virtual pages of a Google search and read the “Best of” lists collected by a variety of technical publications and writers, much like you can thumb through newspapers from the 1980s to see the how big we wore our shoulder pads, neon clothing, and hair back then. But neither research method is particularly efficient, nor do they provide snapshots that show diversity within communities and moments of time. The idea behind the Open Source Yearbook is to collaborate with open source communities to collect a diverse range of stories from the year.
    [Show full text]
  • Pemanfaatan Hiphop for Php Pada Web Server
    PEMANFAATAN HIPHOP FOR PHP PADA WEB SERVER Gunawan 1, Suwandi Halim 2 STMIK Mikroskil Jl. Thamrin No. 112, 124, 140 Medan 20212 [email protected] 1, [email protected] 2 Abstrak HipHop for PHP merupakan sebuah teknologi optimasi website PHP yang dikembangkan oleh Facebook dengan mentransformasi source code PHP menjadi C++ yang optimal dan dengan mengorbankan beberapa fungsi PHP yang jarang digunakan. Artikel ini membahas penggunaan HipHop for PHP pada web server dan pengujiannya dengan cara membandingkannya dengan web server yang tidak menggunakan HipHop for PHP atau tanpa optimasi apapun. Uji coba terhadap pemakaian HipHop for PHP adalah dengan meng-compile sebuah website PHP atau kumpulan source code PHP menjadi binary files C++ hingga bisa diakses oleh sebuah web server . Juga akan dilakukan benchmark terhadap HipHop for PHP dengan menggunakan Apache Benchmark untuk mengetahui performansi HipHop for PHP dalam menangani request (permintaan) user dalam jumlah yang banyak. Dari hasil pengujian didapatkan bahwa HipHop for PHP mampu meningkatkan performansi web server untuk website PHP hingga lebih dari 50%. Kata Kunci : PHP, HipHop, web server, benchmark . 1. Pendahuluan PHP Hypertext Preprocessor merupakan salah satu bahasa scripting bersifat server-side yang cukup sering dijumpai pada pembuatan website . PHP berjalan pada web server , sehingga eksekusi script PHP bergantung pada kemampuan server yang digunakan. Semakin besar script PHP yang dieksekusi, maka penggunaan sumber daya CPU juga semakin besar, begitu pula dengan penggunaan memori. Dengan demikian, ketika semakin banyak user yang mengakses web server dan banyaknya script PHP yang harus dieksekusi, maka semakin besar pula sumber daya CPU dan memori yang digunakan, sehingga respon web server terhadap permintaan ( request ) masing-masing user akan menjadi semakin lambat.
    [Show full text]
  • Advanced Database Technologies Universität Summer 2011 Konstanz Prof
    Lecture: Advanced Database Technologies Universität Summer 2011 Konstanz Prof. Dr. Marc H. Scholl, Dr. Christian Grün. DBIS Group Advanced Database Technologies NoSQL: Not only SQL Christian Grün Database & Information Systems Group Universität Konstanz Lecture: Advanced Database Technologies Universität Summer 2011 Konstanz Prof. Dr. Marc H. Scholl, Dr. Christian Grün. DBIS Group NoSQL Introduction 30, 40 years history of well-established database technology… all in vain? Not at all! But both setups and demands have drastically changed: • main memory and CPU speed have exploded, compared to the time when System R (the mother of all RDBMS) was developed • at the same time, huge amounts of data are now handled in real-time • both data and use cases are getting more and more dynamic • social networks (relying on graph data ) have gained impressive momentum • full-texts have always been treated shabbily by relational DBMS 2 Lecture: Advanced Database Technologies Universität Summer 2011 Konstanz Prof. Dr. Marc H. Scholl, Dr. Christian Grün. DBIS Group NoSQL: Facebook Statistics royal.pingdom.com/2010/06/18/the-software-behind-facebook • 500 million users • 570 billion page views per month • 3 billion photos uploaded per month • 1.2 million photos served per second • 25 billion pieces of content (updates, comments) shared every month • 50 million server-side operations per second • 2008: 10,000 servers; 2009: 30,000 , … ☞ One RDBMS may not be enough to keep this going on! 3 Lecture: Advanced Database Technologies Universität Summer 2011 Konstanz Prof. Dr. Marc H. Scholl, Dr. Christian Grün. DBIS Group NoSQL: Facebook Architecture Memcached • distributed memory caching system • caching layer between web and database servers • based on a distributed hash table (DHT) HipHop for PHP • developed by Facebook to improve scalability • compiles PHP to C++ code , which can be better optimized • PHP runtime system was reimplemented 4 Lecture: Advanced Database Technologies Universität Summer 2011 Konstanz Prof.
    [Show full text]
  • What Is Big Data and How Facebook Uses Big Data
    International Journal of Research in Engineering and Science (IJRES) ISSN (Online): 2320-9364, ISSN (Print): 2320-9356 www.ijres.org Volume 9 Issue 5 ǁ 2021 ǁ PP. 57-63 What is Big Data and how Facebook uses Big Data Remya.S.P Assistant Professor On Contract,Department of Computer Science, University of Calicut,Kerala,India Abstract: With the advent and increased use of the internet, social media has become an integral part of people’s daily routine. In the second quarter of 2020, Facebook has emerged as the biggest social network worldwide, with over 2.9 billion monthly active users which is approximately the 1/5th of the world's total population. Launched in the year 2004, it has grown tremendously since then. With a sudden obsession in social media, the number of people on Facebook has increased enormously, producing a massive amount of data every minute. They are generating almost 500 terabytes of data every day. To store and process this huge amount of data, the concept of big data is used. Big Data refers to large volume of data which is growing exponentially with time. Big Data uses open source data warehousing platforms like hadoop, hive etc. By being massively engaged in its capacity for collecting, storing, and interpreting data, Facebook has placed Big Data at the heart of its operations. This paper presents the application of big data on face book, examining its current as well as future impact. Keyword: Big Data,Characteristics,Facebook,Types,Technology --------------------------------------------------------------------------------------------------------------------------------------- Date of Submission: 27-04-2021 Date of acceptance: 11-05-2021 --------------------------------------------------------------------------------------------------------------------------------------- I.
    [Show full text]
  • PHP Optimization Using Hip Hop Virtual Machine Chaitali Tambe, Pramod Pawar, Dashrath Mane Vivekanand Education Society Institute of Technology, Chembur
    International Journal of Advanced Research in Computer Engineering & Technology (IJARCET) Volume 4 Issue 6, June 2015 PHP Optimization Using Hip Hop Virtual Machine Chaitali Tambe, Pramod Pawar, Dashrath Mane Vivekanand education society institute of technology, Chembur Abstract— HipHop compiler is PHP based compiler created facebook.com. PHP applications that were developed by facebook in order to optimize the php code and save using HipHop terminology resulted in increased speed of resources on the facebook server. HPHPc translates the php application. Several drawbacks of standalone php code into executable c++ code. Later facebook realized that applications were eliminated by using HPHPc, one of HPHPc had some snags which included resource consumption and parallel maintenance of HPHP and which was performance. HPHPi. In order to overcome the drawbacks of HPHPc A. Limitation of HPHPc: HipHop virtual machine was developed in 2011 which is a process virtual machine based on just-in-time (JIT) HPHPc did not fully support the PHP language, compilation, serving as an execution engine for PHP and including the create_function() and eval() Hack programming languages. HHVM provides complete constructs. support for the entire PHP language. This paper shows how HHVM is superior to other compilers like PHP, PHP 7 It involved a specific time and resource consuming and how various framework uses HHVM to optimize deployment process that required a bigger than php. In addition the lock-down approach shows the 1 GB binary to be compiled and distributed to behavior of each framework with hhvm and show how many servers in short orders. performance optimization can be obtained for heavy In addition, maintaining HPHPc and HPHPi in websites.
    [Show full text]
  • Caching PHP Virtual Machine Opcodes for Reuse in a Secure, Multi-User, Enterprise Programming Environment
    PHP ACCELERATION IN THE CLOUD: Caching PHP Virtual Machine Opcodes for Reuse in a Secure, Multi-User, Enterprise Programming Environment by Leif Wendal Wickland A project submitted in partial fulfillment of the requirements for the degree of Master of Science in Computer Science MONTANA STATE UNIVERSITY Bozeman, Montana November 2010 ©COPYRIGHT by Leif Wendal Wickland 2010 All Rights Reserved ii APPROVAL of a thesis submitted by Leif Wendal Wickland This project has been read by each member of the thesis committee and has been found to be satisfactory regarding content, English usage, format, citation, bibliographic style, and consistency and is ready for submission to the Division of Graduate Education. Dr. Rocky Ross Approved for the Department of Computer Science Dr. John Paxton Approved for the Division of Graduate Education Dr. Carl A. Fox iii STATEMENT OF PERMISSION TO USE In presenting this project in partial fulfillment of the requirements for a master’s degree at Montana State University, I agree that the Library shall make it available to borrowers under rules of the Library. If I have indicated my intention to copyright this thesis by including a copyright notice page, copying is allowable only for scholarly purposes, consistent with “fair use” as prescribed in the U.S. Copyright Law. Requests for permission for extended quotation from or reproduction of this thesis in whole or in parts may be granted only by the copyright holder. Leif Wendal Wickland November 2010 iv DEDICATION and ACKNOWLEDGEMENTS To my wife and daughter who love me more than I deserve. -LW I would like to acknowledge Erich Hannan for giving me the opportunity to wait for the other shoe to drop.
    [Show full text]
  • HHVM @Richardtape
    HHVM @richardtape “It was a high-risk, high-reward bet. We now operate at such an enormous scale, we have to take enormous risks in order to survive” Jay Parikh, Facebook engineer $ HipHop for PHP (HPHPc) • A PHP transpiler (a PHP whatnow?) • Transpiles/Translates PHP into C++ • Then compiled into a binary • Then run as an executable • Regularly increased the speed of PHP Applications by 6 times Wait. Binary? Executable? What? • Caveat: IANAG (it’s a bit like I am not a lawyer) • Normal PHP • <?php your_code(); ?> -> opcodes -> interpreted (i.e. non-compiled, slow) • HPHPc • your_code(); -> Transpiled (fast!) into C++ -> binary -> executed YOU SAID COMPILE!! A new challenger approaches Hip Hop Virtual Machine Seriously. That’s actually what HHVM stands for. HHVM is a “process virtual machine” using just-in-time compilation YOU SAID COMPILE!! Don’t panic! • The ‘just-in-time’ before the word-we- no-longer say is key • It means you don’t have to do it yourself • And you just write your PHP as you normally would Process Virtual Machine? AKA an Application Virtual Machine. (Oh that’s much better.) Or a Managed Runtime Environment (Who makes these things up?) In real words it means that your code is completely abstracted away from the underlying platform which in turn means the same code will give the same results, regardless of platform. Hip Hop Byte Code I promise you I’m not just making up words. Again, that’s actually it’s name. PHP (and ‘Hack’) are converted into Hip Hop Byte Code (HHBC) which is then converted into machine code and then executed.
    [Show full text]
  • Challenges of Web Application Development: How to Optimize Client-Side Code
    Challenges of Web Application Development: How to Optimize Client-Side Code Chris Mulder S0326631 August 16, 2011 Contents 1 Introduction 3 1.1 Problem Statement . 4 1.2 Research Goals . 5 1.3 Research Methodology . 5 1.4 Structure of the Thesis . 5 2 Description of Problem Domain 6 2.1 History . 6 2.1.1 The Invention of the Web . 6 2.1.2 Browser Wars . 7 2.1.3 The HTML Standard . 8 2.1.4 JavaScript . 10 2.1.5 CSS . 13 2.2 Five Emerging Issues . 13 2.2.1 History/Back Button . 14 2.2.2 Search Engine Optimization . 14 2.2.3 Modularization, Code-generation . 15 2.2.4 Hiding Differences in Target Platforms . 16 2.2.5 Profiling/Benchmarking and Optimization . 17 2.3 Choice and Motivation . 18 3 Action Research 20 3.1 Research Setup . 20 3.2 Project Kickoff . 20 3.3 Research Phase #1: Tools Overview . 21 3.4 Research Phase #2: First Trials . 24 3.4.1 Plain JavaScript Approach . 24 3.4.2 Speed Tracer Approach . 25 3.5 Development Phase: Chosen Approach . 26 3.5.1 Navigation Timing . 26 3.5.2 Custom Measurements . 27 3.5.3 Triggering The Benchmark . 28 3.5.4 The Beacon . 29 3.5.5 Automation of Benchmarking . 30 1 3.6 Overview of Implementation . 31 3.7 Results . 31 4 Discussion 33 4.1 Challenges in Web Engineering Research . 33 4.2 Course of the Process . 33 5 Conclusion 35 6 Furture Work 37 6.1 Additions to Current Implementation .
    [Show full text]