PC2 V10 PERFORMANCE TEST FRAMEWORK a Project Presented

Total Page:16

File Type:pdf, Size:1020Kb

PC2 V10 PERFORMANCE TEST FRAMEWORK a Project Presented PC2 V10 PERFORMANCE TEST FRAMEWORK A Project Presented to the faculty of the Department of Computer Science California State University, Sacramento Submitted in partial satisfaction of the requirements for the degree of MASTER OF SCIENCE in Computer Science by Bhushan Arun Ladde FALL 2018 © 2018 Bhushan Arun Ladde ALL RIGHTS RESERVED ii PC2 V10 PERFORMANCE TEST FRAMEWORK A Project by Bhushan Arun Ladde Approved by: __________________________________, Committee Chair John Clevenger, Ph.D. __________________________________, Second Reader Doan Nguyen. Ph.D. ____________________________ Date iii Student: Bhushan Arun Ladde I certify that this student has met the requirements for format contained in the University format manual, and that this project is suitable for shelving in the Library and credit is to be awarded for the project. __________________________, Graduate Coordinator ___________________ Jinsong Ouyang, Ph.D. Date Department of Computer Science iv Abstract of PC2 V10 PERFORMANCE TEST FRAMEWORK by Bhushan Arun Ladde The PC2 Contest Control System provides for the ability of teams in a programming contest to submit runs – that is, programs which are attempts at solving a contest problem specified by the contest judges. In Version 10 (V10) of PC2, runs are submitted via web pages served by a webserver embedded in a module called the AppServer. On the server- side, AppServer automatically inserts a submitted run into a database. The Dispatcher queues run notifications and makes them available to another module called the Checker. Checker returns the result to the Dispatcher that whether run correctly solved the problem or not. The Dispatcher in turn forwards the result to the AppServer for subsequent display to the team which submitted the run. There is a Resource Manager which supports the ability of the Contest Administrator to define multiple different contests and to enter contest configuration information for each defined contest. The architecture allows only a single Dispatcher and a single Resource Manager to be running at any given time. Each V10 module runs on its own separate machine. v There are some limitations in the current system. A fundamental premise of the PC2 V10 system is that it is intended to handle large numbers of teams simultaneously participating in multiple different contests. The problem is that there is currently no mechanism for testing the performance of the system under such a large load and verify the performance capabilities of the V10 system. The proposed solution includes several important points. A. Build a large collection of programming contest problems and corresponding run submissions. B. Investigate a variety of available performance and load-testing tools. C. Implement a framework for submitting the collection of contest problems and corresponding submissions to PC2 V10 and monitoring the resulting performance using the selected tool(s). D. In the event that Jetty is found to be unable to support the desired load, investigate and propose mechanisms for substituting alternative webservers into V10 (e.g, using the nGinX webserver). _______________________, Committee Chair John Clevenger, Ph.D. _______________________ Date vi ACKNOWLEDGEMENT I would like to express my gratitude to my project supervisor Dr. John Clevenger for the useful comments, remarks, guidance and engagement through the learning process of this master’s project. I would also like to thank Dr. Doan Nguyen for taking time to review my report and giving me important feedback. Added to that, I would like to thank the Department of Computer Science at California State University, Sacramento for providing an opportunity to pursue my Masters Degree and guiding me all the way to become a successful student. vii TABLE OF CONTENTS Page Acknowledgement.………..............................................................................................vii List of Figures………………………………………………………………………….. x Chapter 1. INTRODUCTION……………………………………………………………………. 1 1.1 What is a Programming Contest?…………………………………………...........1 1.2 PC2…………………………………………………………………….................1 1.3 PC2 V10 Introduction……………………………………………………………3 1.3.1 Main Components of PC2 V10…………………………………………….4 1.3.2 Technology Stack for PC2 V10……………………………………………5 2. EXISTING PROBLEM AND PROPOSED SOLUTION…………………………....6 2.1 Performance Measurement Tool ………………………………………………...6 2.2 Architecture of JMeter…………………..……………………….........................8 3. TESTING TOOLS……………………….….............................................................10 3.1 Initial Excel Utils Project Set Up…...…………………………………………..10 3.2 JMeter Installation…………………………......................................................12 3.3 JMeter Test Plan Creation ……………………..……………...........................14 viii 3.4 Create Test User in JMeter ………………………………................................19 3.5 JVM(Java Virtual Machine) Monitor ………....................................................20 3.6 Postman ...………………...................................................................................22 4. IMPLEMENTATION STRATEGY ..………………………....................................23 4.1 Contest Data Creation Using PC2 V10 Tool.......................................................24 4.2 Steps to Implement PC2 V10 Test Plan..............................................................26 4.3 Measured Points..................................................................................................27 5. SIMULATION RESULTS…………………………..................................................31 5.1 Load Capacity Result............................................................................................31 5.2 Memory Dump Issue.............................................................................................32 5.3 Graphical Analysis of Performance……………………………….…………….32 5.4 Do Not Use XAMPP as a Database......................................................................33 6. FUTURE SCOPE…………………………................................................................36 6.1 Distributed Environment......................................................................................36 6.2 Testing on Different Platforms.............................................................................36 6.3 Security Test.........................................................................................................36 6.4 Alternative Webservers into V10…………........................................................37 References…..….……………………….........................................................................38 ix LIST OF FIGURES Figures Page 1. PC2 V10 Architecture………….….……………….....................................................4 2. JMeter Architecture Block Diagram….…....................................................................8 3. JMeter Installation Step 1………………...................................................................12 4. JMeter Installation Step 2………………...................................................................13 5. JMeter Initial Screen………………...........................................................................13 6. JMeter Test Plan Creation Step 1………………........................................................14 7. JMeter Test Plan Creation Step 2………………........................................................15 8. JMeter Test Plan Creation Step 3………………........................................................15 9. JMeter Test Plan Creation Step 4…………................................................................16 10. JMeter Test Plan Creation Step 5…………...............................................................17 11. JMeter Test Plan Creation Step 6…………...............................................................18 12. JMeter Test Plan Creation Step 7…………...............................................................18 13. JMeter Recorder Creation…..………….....................................................................19 14. JVM Test Monitor - 1….…..…………......................................................................21 15. JVM Test Monitor - 2….…..…………......................................................................22 16. JMeter PC2 V10 Login Test Plan Implementation.....................................................23 17. JMeter PC2 V10 Test Plan Implementation ……………..........................................24 18. Contest.yaml Configuration...………….....................................................................25 x 19. PC2 V10 Tool…………….....…………..................................................................26 20. Performance Graph of PC2 V10……………………………………………...……33 21. XAMPP…………...………......................................................................................34 xi 1 1. INTRODUCTION 1.1 What is a Programming Contest? A programming contest is an internet-based contest. There are two type of contests, first one is an individual programming contest where the individual programmer will get a problem statement in which he/she has to provide its individual solution, second one is group programming contest in which there are group of people who will get the problem statement. Users have to provide solution to that problem. A vast majority of problems appear in programming contest are based on mathematics and logical concepts. Once any user has submitted its solution then the automation process verifies the answer. The verification process is performed by the host machine, it is also called as judging the answer. There are a certain number of test cases to those problem
Recommended publications
  • Hardware & Software Standards
    Hardware & Software Standards Introduction This document identifies the current City of Chicago standards for its hardware and software environments, and is intended primarily for City department and vendor use. These standards do not mean that other software and hardware, which might have been previously listed as standard, may not be used or supported, but the following items should be purchased for any new initiative or growth/replacement needs. Any proposals for non-standard hardware or software purchases or questions/comments should be forwarded to the Department of Innovation and Technology (DoIT) Enterprise Architecture Board for review, and will need to be approved via the Technology Purchase Review and Approval (TPRA) process. Standards denoted with an asterisk (*) are currently under review. Platform Standards Operating System (O/S) Hardware Platform Solaris 10 (Unix) (Oracle) Sun Microsystems RedHat Linux Enterprise Server 6.x, 7.x Dell RedHat Linux Enterprise Server 7.x (PCI Services) Dell VMWare VSphere 6.5U1 Dell Windows 2012 R2 & 2016 (Standard and Enterprise) Dell Windows 7, Windows 10 Dell, Panasonic Page 1 of 6 Last Revised January 2018 Hardware & Software Standards Enterprise Services Type Windows 2008 Server All other platforms Oracle Enterprise 11gR2, 12cR1; Postgres 9.x or 10.x (EnterpriseDB or Database N/A community) Print O/S n/a File O/S n/a Email Exchange 2016 / Office365 n/a Desktops, Laptops, & Tablets Type Model Standard Users Dell OptiPlex 5050 SFF, Dell OptiPlex 7450 All-In-One Mobile User Latitude 12 Rugged Extreme Latitude 14 Rugged 5414 Latitude 12 2 in 1 with case and Doc Latitude 5480 14" Laptop 6th gen proc High-End Workstation Dell Precision T5810 Laptop Accessories Docking- For the E-5470 units, Dell Business Dock - WD15 with 130W Adapter Monitor Dell 23 Monitor – P2317H Page 2 of 6 Last Revised January 2018 Hardware & Software Standards Printing and Scanning The Department of Fleet and Facility Management (2FM) oversees print services for the City of Chicago.
    [Show full text]
  • IT CLASSIFICATION TECHNOLOGY LIST – ISSUE DATE: April 18, 2017
    IT CLASSIFICATION TECHNOLOGY LIST – ISSUE DATE: April 18, 2017 Technology Definition: A set of knowledge, skills and/or abilities, taking a significant time (e.g. 6 months) to learn, and applicable to the defined classification specification assigned. Example of Tools: These are examples only for illustration purposes and are not meant to constitute a full and/or comprehensive list. CLASSIFICATION DISCIPLINE TECHNOLOGY DEFINITION EXAMPLE OF TOOLS Omegamon, IBM Admin Tools, Log Analyzer, Relational The relational database management system provided by IBM that runs on Unix, Linux, DBA DB2 DB2 Compare, Nsynch, TSM, Universal Database Windows and z/OS platforms including DB2 Connect and related tools. Command, SQL SQL Server Mgmt. Studio, Red Gate, Vantage, SQL Server The relational database management system and related tools provided by Microsoft Corp. Tivoli, Snap Manager, Toad, Enterprise Manager, SQL ORACLE The relational database management system and related tools provided by Oracle Corp. ASE SYBASE The relational database management system and related tools provided by Sybase. Cincom SUPRA SQL – Cincom’s relational database management system provides access to data Supra 2.X in open and proprietary environments through industry-standard SQL for standalone and client/server application solutions. Open Source Open Source database management system such as Mysql. Phpadmin, mysqladmin, Vertica Hierarchical The hierarchical database management system provided by IBM that runs on z/OS mainframe IMS BMC IMS Utilities, Strobe, Omegamon Database platform including related tools. Cincom SUPRA® PDM – Cincom’s networked, hierarchical database management system provides access to your data through a Physical Data Manager (PDM) that manages the data Supra 1 structures of the physical files that store the data.
    [Show full text]
  • Functional Programming at Facebook
    Functional Programming at Facebook Chris Piro, Eugene Letuchy Commercial Users of Functional Programming (CUFP) Edinburgh, Scotland ! September "##$ Agenda ! Facebook and Chat " Chat architecture # Erlang strengths $ Setbacks % What has worked Facebook The Facebook Environment The Facebook Environment ▪ The web site ▪ More than 250 million active users ▪ More than 3.5 billion minutes are spent on Facebook each day The Facebook Environment ▪ The web site ▪ More than 250 million active users ▪ More than 3.5 billion minutes are spent on Facebook each day ▪ The engineering team ▪ Fast iteration: code gets out to production within a week ▪ Polyglot programming: interoperability with Thrift ▪ Practical: high-leverage tools win Using FP at Facebook Using FP at Facebook ▪ Erlang ▪ Chat backend (channel servers) ▪ Chat Jabber interface (ejabberd) ▪ AIM presence: a JSONP validator Using FP at Facebook ▪ Erlang ▪ Chat backend (channel servers) ▪ Chat Jabber interface (ejabberd) ▪ AIM presence: a JSONP validator ▪ Haskell ▪ lex-pass: PHP parse transforms ▪ Lambdabot ▪ textbook: command line Facebook API client ▪ Thrift binding Thrift Thrift ▪ An efficient, cross-language serialization and RPC framework Thrift ▪ An efficient, cross-language serialization and RPC framework ▪ Write interoperable servers and clients Thrift ▪ An efficient, cross-language serialization and RPC framework ▪ Write interoperable servers and clients ▪ Includes library and code generator for each language Thrift ▪ An efficient, cross-language serialization and RPC framework ▪ Write
    [Show full text]
  • Enterprise Development with Flex
    Enterprise Development with Flex Enterprise Development with Flex Yakov Fain, Victor Rasputnis, and Anatole Tartakovsky Beijing • Cambridge • Farnham • Köln • Sebastopol • Taipei • Tokyo Enterprise Development with Flex by Yakov Fain, Victor Rasputnis, and Anatole Tartakovsky Copyright © 2010 Yakov Fain, Victor Rasputnis, and Anatole Tartakovsky.. All rights reserved. Printed in the United States of America. Published by O’Reilly Media, Inc., 1005 Gravenstein Highway North, Sebastopol, CA 95472. O’Reilly books may be purchased for educational, business, or sales promotional use. Online editions are also available for most titles (http://my.safaribooksonline.com). For more information, contact our corporate/institutional sales department: (800) 998-9938 or [email protected]. Editor: Mary E. Treseler Indexer: Ellen Troutman Development Editor: Linda Laflamme Cover Designer: Karen Montgomery Production Editor: Adam Zaremba Interior Designer: David Futato Copyeditor: Nancy Kotary Illustrator: Robert Romano Proofreader: Sada Preisch Printing History: March 2010: First Edition. Nutshell Handbook, the Nutshell Handbook logo, and the O’Reilly logo are registered trademarks of O’Reilly Media, Inc. Enterprise Development with Flex, the image of red-crested wood-quails, and related trade dress are trademarks of O’Reilly Media, Inc. Many of the designations used by manufacturers and sellers to distinguish their products are claimed as trademarks. Where those designations appear in this book, and O’Reilly Media, Inc. was aware of a trademark claim, the designations have been printed in caps or initial caps. While every precaution has been taken in the preparation of this book, the publisher and authors assume no responsibility for errors or omissions, or for damages resulting from the use of the information con- tained herein.
    [Show full text]
  • Web-Engineering" Kapitel 9: Website-Analyse Und -Optimierung
    Reader zum MOOC "Web-Engineering" Kapitel 9: Website-Analyse und -Optimierung Die PDF-Datei wurde mit Hilfe des Open-Source-Werkzeugs „mwlib“ erstellt. Für weitere Informationen siehe http://code.pediapress.com/ PDF generated at: Sat, 14 Dec 2013 16:53:28 UTC Inhalt Artikel Web-Controlling 1 Web Analytics 1 Google Analytics 7 Piwik 10 AWStats 12 Webalizer 14 Logdateianalyse 16 Ganglinie 20 Konversion (Marketing) 22 Conversion-Tracking 24 Click-Through-Rate 25 Besucherzähler 26 Unique Visit 27 Referrer 28 Seitenabruf 30 Zählpixel 32 Cron 34 Do Not Track 38 Web-Monitoring 41 Munin (Software) 41 WPO: Web Performance Optimization 44 Google Page Speed 44 SPDY 45 Performnace-Test 47 Lasttest (Computer) 47 Stresstest 50 Apache JMeter 52 Dienstgütevereinbarung für Websites 53 Service-Level-Agreement 53 Webhosting 55 SMART (Projektmanagement) 58 Downtime 59 Verfügbarkeit 60 Paketumlaufzeit 62 Quellennachweise Quelle(n) und Bearbeiter des/der Artikel(s) 64 Quelle(n), Lizenz(en) und Autor(en) des Bildes 66 Artikellizenzen Lizenz 67 1 Web-Controlling Web Analytics Web Analytics (auch Web-Controlling, Web-Analyse, Datenverkehrsanalyse, Traffic-Analyse, Clickstream-Analyse, Webtracking) ist die Sammlung und Auswertung des Verhaltens von Besuchern auf Websites. Ein Analytic-Tool untersucht typischerweise, woher die Besucher kommen, welche Bereiche auf einer Internetseite aufgesucht werden und wie oft und wie lange welche Unterseiten und Kategorien angesehen werden. In Deutschland ist der Einsatz solcher Werkzeuge aus Datenschutzgründen umstritten. Sie wird vor allem zur Optimierung der Website und zur besseren Erreichung von Zielen der Website (z. B. Häufigkeit von Besuchen, Vermehrung von Seitenaufrufen, Bestellungen, Newsletter-Abonnements) eingesetzt. Grundlegend kann bei Web Analytics zwischen Auswertungsverfahren zur permanenten Messung der Site-Effektivität und Methoden zur Auffindung von Schwachpunkten in der Site und Verbesserungsmöglichkeiten unterschieden werden (siehe Methoden).
    [Show full text]
  • The Third Annual ICFP Programming Contest
    The Third Annual ICFP Programming Contest August 26 – 29, 2000 (Version 1.18) 1 The problem This year’s ICFP programming challenge is to implement a ray tracer. The input to the ray tracer is a scene description written in a simple functional language, called GML. Execution of a GML program produces zero, or more, image files, which are in PPM format. A web page of sample images, along with the GML inputs that were used to produce them, is linked off of the contest home page. The feature set of GML is organized into three tiers. Submissions must implement the first tier of features and extra credit will be given to submissions that implement the second or third tiers. Submissions will be evaluated on three scales: correctness of the produced images, run-time performance, and the tier of implemented GML features. GML has primitives for defining simple geometric objects (e.g., planes, spheres, and cubes) and lighting sources. The surface properties used to render the objects are specified as functions in GML itself. In addition to supporting scene description, GML also has a render operator that renders a scene to an image file. For each pixel in the output image, the render command must compute a color. Conceptually, this color is computed by tracing the path of the light backwards from the eye of the viewer, to where it bounced off an object, and ultimately back to the light sources. This document is organized as follows. Section 2 describes the syntax and general semantics of the modeling language. It is followed by Section 3, which describes those aspects of the language that are specific to ray tracing.
    [Show full text]
  • ICFP Programming Contest 2010 International Cars and Fuels Production
    About Contest Task Running the Contest Background Winners Future ICFP Programming Contest 2010 International Cars and Fuels Production Bertram Felgenhauer, University of Innsbruck, Austria Johannes Waldmann, HTWK Leipzig, Germany June 18–21, 2010 Felgenhauer, Waldmann ICFP Programming Contest 2010 1/31 About Contest Task Running the Contest Background Winners Future About the ICFP Programming Contest programming, problem solving, fun annual contest, since 1998 sponsored by ICFP conference/ACM 2010 contest hosted by HTWK Leipzig, Germany contest format 72 hours (June 18, 12:00 – June 21, 12:00 GMT) participation online, international teams allowed no fixed programming language lightning division (first 24 hours) Felgenhauer, Waldmann ICFP Programming Contest 2010 2/31 earn money by (efficiently) solving instances, or creating instances (with solution, which is hard to find) income tax (devaluates earnings by 1/2 per day) About Contest Task Running the Contest Background Winners Future Contest Task storyline: market for cars (= problem instance) (public) fuels (= problem solution) (private) Felgenhauer, Waldmann ICFP Programming Contest 2010 3/31 About Contest Task Running the Contest Background Winners Future Contest Task storyline: market for cars (= problem instance) (public) fuels (= problem solution) (private) earn money by (efficiently) solving instances, or creating instances (with solution, which is hard to find) income tax (devaluates earnings by 1/2 per day) Felgenhauer, Waldmann ICFP Programming Contest 2010 3/31 About Contest Task
    [Show full text]
  • ICFP 2009 Final Program
    ICFP 2009 Final Program Monday, August 31, 2009 Tuesday, September 1, 2009 Wednesday, September 2, 2009 Invited Talk (Chair: Andrew Tolmach) Invited Talk (Chair: Graham Hutton) Invited Talk (Chair: Lennart Augustsson) 9:00 Organizing Functional Code for Parallel Execution; or, 9:00 Lambda, the Ultimate TA: Using a Proof Assistant to 9:00 Commutative Monads, Diagrams and Knots foldl and foldr Considered Slightly Harmful Teach Programming Language Foundations Dan Piponi; Industrial Light & Magic Guy L. Steele, Jr.; Sun Microsystems Benjamin C. Pierce; University of Pennsylvania 10:00 Break 10:00 Break 10:00 Break Session 11 (Chair: Ralf Hinze) Session 1 (Chair: Shin-Cheng Mu) Session 6 (Chair: Xavier Leroy) 10:25 Generic Programming with Fixed Points for Mutually 10:25 Functional Pearl: La Tour D’Hano¨ı 10:25 A Universe of Binding and Computation Recursive Datatypes Ralf Hinze; University of Oxford Daniel Licata and Robert Harper; Carnegie Mellon University Alexey Rodriguez Yakushev1, Stefan Holdermans2, Andres 2 3 10:50 Purely Functional Lazy Non-deterministic Program- 10:50 Non-Parametric Parametricity L¨oh , Johan Jeuring ; 1Vector Fabrics B.V., 2Utrecht University, ming Georg Neis, Derek Dreyer, Andreas Rossberg; MPI-SWS 3Utrecht University, Open University of the Netherlands Sebastian Fischer1, Oleg Kiselyov2, Chung-chieh Shan3; 11:15 Break 10:50 Attribute Grammars Fly First-Class: How to do As- 1Christian-Albrechts University, 2FNMOC, 3Rutgers University Session 7 (Chair: Robby Findler) pect Oriented Programming in Haskell 11:15 Break
    [Show full text]
  • Cigniti Technologies Service Sheet
    Cigniti Technologies Service Sheet Cigniti Technologies is an Independent Software Testing Services Company headquartered in Irving, Texas. Cigniti offers Test Consulting, Enterprise Mobility Testing, Regression Automation, Security and Performance Testing services. Over the last 14 years, leveraging its Smart Tools, Cigniti helped Enterprises and ISVs build quality software while improving time to market and reducing cost of quality. Designed to deliver quality solutions, the service portfolio spans technologies transcending Cloud, Mobile, Social, Big Data and Enter- prise IT across verticals and ensures impeccable standards of QA. World’s 3rd Largest Independent Software Testing Services Company Dallas | Hyderabad | Atlanta | San Jose | Des Moines | London | Toronto | www.cigniti.com Service Sheet Over the last decade, there has been a lot of transformation in the philosophy of testing and several tools have been introduced in the market place. With so many choices around, it has become a real challenge to choose the right testing partner / service provider. The Testing industry has witnessed a phenomenal growth in term of offerings, tools, and infrastructure with diverse engagement models. Now, it has become a lot more important for organizations to choose the right testing partner. Cigniti delivers Independent Quality Assurance Services backed by its proprietary IP. As a strategic partner, Cigniti brings comprehensive service offerings in the Quality Assurance space which accelerate overall test efforts for its clients. Cigniti’s Service Offerings TCoE Performance Testing Security Testing Cigniti's Test Center of Excellence Cigniti’s new age performance Cigniti’s security testing services lays down roadmaps for scalable testing frameworks initiate and ensure early detection with com- frameworks that enrich business revitalize performance across prehensive vulnerability assess- outcomes with speed, skill and systems, networks and software ment and match the emerging accuracy.
    [Show full text]
  • Test Center of Excellence How Can It Be Set Up? ISSN 1866-5705
    ISSN 1866-5705 www.testingexperience.com free digital version print version 8,00 € printed in Germany 18 The Magazine for Professional Testers The MagazineforProfessional Test Center of Excellence Center Test How can itbesetup? How June 2012 Pragmatic, Soft Skills Focused, Industry Supported CAT is no ordinary certification, but a professional jour- The certification does not simply promote absorption ney into the world of Agile. As with any voyage you have of the theory through academic mediums but encour- to take the first step. You may have some experience ages you to experiment, in the safe environment of the with Agile from your current or previous employment or classroom, through the extensive discussion forums you may be venturing out into the unknown. Either way and daily practicals. Over 50% of the initial course is CAT has been specifically designed to partner and guide based around practical application of the techniques you through all aspects of your tour. and methods that you learn, focused on building the The focus of the course is to look at how you the tes- skills you already have as a tester. This then prepares ter can make a valuable contribution to these activities you, on returning to your employer, to be Agile. even if they are not currently your core abilities. This The transition into a Professional Agile Tester team course assumes that you already know how to be a tes- member culminates with on the job assessments, dem- ter, understand the fundamental testing techniques and onstrated abilities in Agile expertise through such fo- testing practices, leading you to transition into an Agile rums as presentations at conferences or Special Interest team.
    [Show full text]
  • The 20 Best 2018 Performance Tester Insights
    Whitepaper The 20 Best 2018 Performance Tester insights THE 20 BEST 2018 PERFORMANCE TESTER INSIGHTS 2nd edition 20 speakers from 10 countries 24 hours live 35 cup of coffee 800 registrants 950 stickers Whitepaper The 20 Best 2018 Performance Tester insights This past June, Neotys brought together 20 experts from around the world to share their experience and knowledge regarding techniques, challenges, and best practices associated with load testing for the modern enterprise. The gathering was delivered online as a virtual event. The second annual Performance Advisory Council (PAC) event, proved to be transformational. Attendees were treated to new and innovative practices on the art and science of performance testing. Each presentation provided a wealth of information with insights and techniques aimed at performance testing improvement. The following is a synopsis of several presentations made during the event. We’re happy to share this information in the spirit of creating a more well informed, efficient community of test practitioners. Enabling user story sharing and learning on an international basis is critical to us. Stijn Schepers Performance Testing is Not an Average Game! Stijn Schepers has over 18 years of international experience as a performance test specialist/manager using both Agile (Scrum, SAFe) and Waterfall models. He has worked in Europe, New Zealand, and Australia as a performance engineer testing and optimizing complex enterprise solutions like Oracle: Financials, Forms, Web Center, CRM Applications, JD Edwards, J2EE solutions, Vignette, SIEBEL, SAP, Cúram, IBM Websphere, and Mule. Schepers is a visionary leader with in-depth technical knowledge in concept and practice driving performance testing in the modern enterprise.
    [Show full text]
  • NCA’04) 0-7695-2242-4/04 $ 20.00 IEEE Interesting Proposals
    WALTy: A User Behavior Tailored Tool for Evaluating Web Application Performance G. Ruffo, R. Schifanella, and M. Sereno R. Politi Dipartimento di Informatica CSP Sc.a.r.l. c/o Villa Gualino Universita` degli Studi di Torino Viale Settimio Severo, 65 - 10133 Torino (ITALY) Corso Svizzera, 185 - 10149 Torino (ITALY) Abstract ticular, stressing tools make what-if analysis practical in real systems, because workload intensity can be scaled to In this paper we present WALTy (Web Application Load- analyst hypothesis, i.e., a workload emulating the activity based Testing tool), a set of tools that allows the perfor- of N users with pre-defined behaviors can be replicated, mance analysis of web applications by means of a scalable in order to monitor a set of performance parameters dur- what-if analysis on the test bed. The proposed approach is ing test. Such workload is based on sequences of object- based on a workload characterization generated from infor- requests and/or analytical characterization, but sometimes mation extracted from log files. The workload is generated they are poorly scalable by the analyst; in fact, in many by using of Customer Behavior Model Graphs (CBMG), stressing framework, we can just replicate a (set of) man- that are derived by extracting information from the web ap- ually or randomly generated sessions, losing in objectivity plication log files. In this manner the synthetic workload and representativeness. used to evaluate the web application under test is repre- The main scope of this paper is to present a methodol- sentative of the real traffic that the web application has to ogy based on the generation of a representative traffic.
    [Show full text]