On Agile Performance Requirements Specification and Testing

Total Page:16

File Type:pdf, Size:1020Kb

On Agile Performance Requirements Specification and Testing On Agile Performance Requirements Specification and Testing Chih-Wei Ho1, Michael J. Johnson2, Laurie Williams1, and E. Michael Maximilien2 1Department of Computer Science, North Carolina State University {cho, lawilli3}@ncsu.edu 2IBM Corporation {mjj, maxim}@us.ibm.com Abstract engineering activities start with performance requirements (PRs) definition [10]. However, some of Underspecified performance requirements can the performance factors are not predictable (e.g., an cause performance issues in a software system. application may be run on an inferior processor with However, a complete, upfront analysis of a software slower speed than initially anticipated) or not available system is difficult, and usually not desirable. We at the early stages of development (e.g., the number of propose an evolutionary model for performance hits of a web server). To conduct a complete upfront requirements specifications and corresponding analysis for performance issues is usually undesirable, validation testing. The principles of the model can be especially for agile practitioners. integrated into agile development methods. Using Therefore, we propose the software Performance this approach, the performance requirements and test Requirements Evolution Model (PREM) as a guideline cases can be specified incrementally, without big for PRs specification and validation. Using the upfront analysis. We also provide a post hoc model, the development can start with quick, simple examination of a development effort at IBM that had a PRs and related performance test cases. During high focus on performance requirements. The development, more precise and realistic performance examination indicates that our evolutionary model can characteristics are obtained through customer be used to specify performance requirements such that communication or performance model solving. The the level of detail is commensurate with the nature of performance characteristics can be added to the PRs, the project. Additionally, the IBM experience and the test cases can be more detailed, until the PRs indicates that test driven development-type validation are “good enough” for the project and its testing corresponding to the model can be used to functionalities. determine if performance objectives have been met. In this paper, we provide an overview of PREM and a post hoc examination of a development effort at IBM 1. Introduction that had a high focus on performance requirements. While software performance concerns include time Agile development methods are sometimes (e.g., throughput, response time) and space (e.g., criticized for not having explicit practices for eliciting memory usage, storage usage), we limit and only focus non-functional requirements (NFRs) [13]. In this on the discussion of time-related performance issues. paper, we discuss an agile approach to address the However, we believe that our model is applicable to specification and testing of an important NFR: space-related performance issues. performance. Performance is highly visible to the The rest of the paper is organized as follows: software user. A system that runs too slowly is likely Section 2 provides the background and related work to be rejected by all users. Failure to achieve some for PRs and testing tools; Section 3 gives description expected performance level might make the system of PREM; Section 4 presents an IBM experience with unusable, and the project might fail or get cancelled if PRs specification and testing; and Section 5 concludes the system performance objective is not met [18]. this paper with summary and future work. Research shows that performance issues should be dealt with early, otherwise performance problems are IBM is a trademark of International Business Machines Corporation more difficult and expensive to fix in later stages of the in the United States or other countries or both. Other company, development process [6, 7, 17]. Performance product, and service names may be trademarks or service marks of others. JUnit decorators that perform both timed and load 2. Background tests. Figure 2 shows an example of JUnitPerf performance testing code. In this section, we present some background information and related work on agile-style software //JUnit 4 timed test: only passes if it performance and performance testing tools. //finishes in 200 ms. @Test(timeout=200) public void perfTest() { //test scenario 2.1. Software performance and Agile Methods ... } At a glance, agile methods do not have explicit Figure 1: JUnit timeout parameter practices for eliciting overarching non-functional system properties. However, Auer and Beck list a //JUnitPerf load test: run 10 instances family of software efficiency patterns called Lazy //of test, with 100 ms intervals between Optimization [1], which reflects the famous quote from //them – max elapsed time is 1000 ms. Knuth that “Premature optimization is the root of all Timer timer = new ConstantTimer(100); evil in programming,”1 and the “You Aren’t Gonna Test test = new MyTest("Perf Test"); Need It (YAGNI)” philosophy. These patterns can be Test loadTest = new LoadTest(test, 10, timer); summarized as follows. Early in development, the Test timeLoadTest = system performance is estimated with a short new TimedTest(loadTest, 1000); performance assessment. Rough performance criteria Figure 2: JUnitPerf example are specified to show performance concerns in the system, and are evolved as the system matures. Tune These JUnit-based test frameworks provide an easy, performance only when the functionality works but programmatic way to write performance test cases. does not pass the performance criteria. However, in complex test cases with complicated Another camp claims that performance is mostly workloads, accurate probability distribution may be determined during the architecture stages [8]. Smith required [19]. JUnit-based test frameworks may be and Williams criticize that the “fix-it-later” attitude is insufficient in such situations. Additionally, the one of the causes of performance failures [17]. Their resulting test code would be difficult to understand. standing ground is that performance models built in To design more complicated performance test cases, architecture and early phases can predict the system one should consider more advanced performance performance. Fixing software problems later is difficult testing tools, for example, script-based (e.g., The and expensive. Therefore, performance decisions 4 Grinder [21]), user action recording (e.g., Apache should be made as early as the architecture phase. 5 JMeter ) or other commercial, high-end tools. We believe the disagreement is rooted in the Agile approaches rely heavily on testing for different philosophy of software development. For software quality assurance [9]. Although, as stated in agile practitioners, software architecture is an instance the previous section, Auer and Beck argue that of the big up-front design (BDUF) that is avoided. performance concerns should be dealt with after the Lazy Optimization aims for agility, but may not be functionality is complete [1], we posit that a sufficient for high-criticality software projects, of development team can benefit from early performance which predictability is more desirable. specification and testing. 2.2. Performance testing tools 3. Performance requirements evolution JUnit2 is the most popular unit testing framework in model the agile community. Beginning with Version 4 which was released in 2006, JUnit provides a In this section, we provide an overview of PREM. “timeout” parameter to support performance-related PREM is an evolutionary model for PRs specification. testing. A test method with timeout parameter can PREM provides guidelines on the level of detail only pass if the test is finished in the specified amount needed in a PR for development teams to specify the of time. Figure 1 shows an example of a test method necessary performance characteristics details and the with the timeout parameter. JUnitPerf3 is a library of form of validation for the PR. Before explaining the model, a short example is provided to show how a 1 “Computer Programming as an Art,” 1974 Turing Award lecture. 2 http://www.junit.org 4 http://grinder.sourceforge.net/ 3 http://www.clarkware.com/software/JUnitPerf.html 5 http://jakarta.apache.org/jmeter/ performance objective can fail because of usually specified qualitatively. As a result, they can underspecified PRs. usually be validated with qualitative evaluation. In A PR might be as simple as “The authentication the authentication process example, the development process shall be complete in 0.2 seconds.” An team might provide a prototype to the customer, and intuitive way to write a test case for this requirement is see whether the user is satisfied with the response time. to run the authentication process several times, and compute the average. However, even if the average is 3.1.2. Level 1. Successful agile testing relies heavily below 0.2 seconds, when the system is live, and on test automation [9]; human testing is undesirable concurrent authentication requests come in, the users and should be limited whenever possible. To make may experience more than 0.2 seconds of waiting time. the requirement testable with automation, the first step In this example, the performance objective is not is to provide quantitative expectations in the achieved because the workload characteristic is not specification, and promote it to a Level 1 PR. The specified in the requirement and not accounted for
Recommended publications
  • Engineering of IT Management Automation Along Task Analysis, Loops, Function Allocation, Machine Capabilities
    Engineering of IT Management Automation along Task Analysis, Loops, Function Allocation, Machine Capabilities Dissertation an der Fakultat¨ fur¨ Mathematik, Informatik und Statistik der Ludwig-Maximilians-Universitat¨ Munchen¨ vorgelegt von Ralf Konig¨ Tag der Einreichung: 12. April 2010 Engineering of IT Management Automation along Task Analysis, Loops, Function Allocation, Machine Capabilities Dissertation an der Fakultat¨ fur¨ Mathematik, Informatik und Statistik der Ludwig-Maximilians-Universitat¨ Munchen¨ vorgelegt von Ralf Konig¨ Tag der Einreichung: 12. April 2010 Tag des Rigorosums: 26. April 2010 1. Berichterstatter: Prof. Dr. Heinz-Gerd Hegering, Ludwig-Maximilians-Universitat¨ Munchen¨ 2. Berichterstatter: Prof. Dr. Bernhard Neumair, Georg-August-Universitat¨ Gottingen¨ Abstract This thesis deals with the problem, that IT management automation projects are all tackled in a different manner with a different general approach and different resulting system architecture. It is a relevant problem for at least two reasons: 1) more and more IT resources with built-in or asso- ciated IT management automation systems are built today. It is inefficient to try to solve each on their own. And 2) doing so, reuse of knowledge between IT management automation systems, as well as reuse of knowledge from other domains is severely limited. While this worked with simple stand-alone remote monitoring and remote control facilities, automation of cognitive tasks will more and more profit from existing knowledge in domains such as artificial intelligence, statistics, control theory, and automated planning. A common structure also would ease integration and coupling of such systems, delegating cognitive partial tasks, and switching between commonly defined levels of automation. So far, this problem is only partly solved.
    [Show full text]
  • Collaborative Software Performance Engineering for Enterprise Applications
    Proceedings of the 50th Hawaii International Conference on System Sciences | 2017 Collaborative Software Performance Engineering for Enterprise Applications Hendrik Müller, Sascha Bosse, Markus Wirth, Klaus Turowski Faculty of Computer Science, Otto-von-Guericke-University, Magdeburg, Germany {hendrik.mueller, sascha.bosse, markus.wirth, klaus.turowski}@ovgu.de Abstract address these challenges from a software engineering perspective under the term DevOps [2], [3]. In the domain of enterprise applications, Corresponding research activities aim at increased organizations usually implement third-party standard flexibility through shorter release cycles in order to software components in order to save costs. Hence, support frequently changing business processes. application performance monitoring activities Therefore, DevOps enables a culture, practices and constantly produce log entries that are comparable to automation that support fast, efficient and reliable a certain extent, holding the potential for valuable software delivery [4]. However, especially for collaboration across organizational borders. Taking enterprise applications, IT departments usually make advantage of this fact, we propose a collaborative use of existing standard software components instead knowledge base, aimed to support decisions of of developing solutions entirely in-house [5]. In such performance engineering activities, carried out cases, activities referred to as Dev, rather include during early design phases of planned enterprise requirements engineering, architectural design, applications. To verify our assumption of cross- customization, testing, performance-tuning and organizational comparability, machine learning deployment [6]. Here, key service levels typically algorithms were trained on monitoring logs of 18,927 include performance objectives, expressed in terms of standard application instances productively running average response times, throughput or latency, in at different organizations around the globe.
    [Show full text]
  • Proquest Dissertations
    Verification and Validation of UML and SysML Based Systems Engineering Design Models YOSR JARRAYA A THESIS IN The Department of Electrical and Computer Engineering Presented in Partial Fulfillment of the Requirements For the Degree of Doctor of Philosophy Concordia University Montréal, Québec, Canada April 2010 Yosr Jarraya, 2010 Library and Archives Bibliothèque et 1*1 Canada Archives Canada Published Heritage Direction du Branch Patrimoine de l'édition 395 Wellington Street 395, rue Wellington OttawaONK1A0N4 Ottawa ON K1A 0N4 Canada Canada Yourfìle Votre référence ISBN: 978-0-494-67341-6 Our file Notre référence ISBN: 978-0-494-67341-6 NOTICE: AVIS: The author has granted a non- L'auteur a accordé une licence non exclusive exclusive license allowing Library and permettant à la Bibliothèque et Archives Archives Canada to reproduce, Canada de reproduire, publier, archiver, publish, archive, preserve, conserve, sauvegarder, conserver, transmettre au public communicate to the public by par télécommunication ou par l'Internet, prêter, telecommunication or on the Internet, distribuer et vendre des thèses partout dans le loan, distribute and sell theses monde, à des fins commerciales ou autres, sur worldwide, for commercial or non- support microforme, papier, électronique et/ou commercial purposes, in microform, autres formats. paper, electronic and/or any other formats. The author retains copyright L'auteur conserve la propriété du droit d'auteur ownership and moral rights in this et des droits moraux qui protège cette thèse. Ni thesis. Neither the thesis nor la thèse ni des extraits substantiels de celle-ci substantial extracts from it may be ne doivent être imprimés ou autrement printed or otherwise reproduced reproduits sans son autorisation.
    [Show full text]
  • Performance Analysis of Business Process Models with Advanced Constructs David Piessens Eindhoven University of Technology, [email protected]
    Association for Information Systems AIS Electronic Library (AISeL) ACIS 2010 Proceedings Australasian (ACIS) 2010 Performance Analysis of Business Process Models with Advanced Constructs David Piessens Eindhoven University of Technology, [email protected] Queensland University of Technology Moe Thandar Wynn Queensland University of Technology, [email protected] Michael Adams Queensland University of Technology, [email protected] Boudewijn F. van Dongen Eindhoven University of Technology, [email protected] Follow this and additional works at: http://aisel.aisnet.org/acis2010 Recommended Citation Piessens, David; Queensland University of Technology; Wynn, Moe Thandar; Adams, Michael; and van Dongen, Boudewijn F., "Performance Analysis of Business Process Models with Advanced Constructs" (2010). ACIS 2010 Proceedings. 23. http://aisel.aisnet.org/acis2010/23 This material is brought to you by the Australasian (ACIS) at AIS Electronic Library (AISeL). It has been accepted for inclusion in ACIS 2010 Proceedings by an authorized administrator of AIS Electronic Library (AISeL). For more information, please contact [email protected]. 21 st Australasian Conference on Information Systems Performance Analysis of Business Process Models 1-3 Dec 2010, Brisbane Piessens et al. Performance Analysis of Business Process Models with Advanced Constructs David Piessens Eindhoven University of Technology, Eindhoven, The Netherlands Queensland University of Technology, Brisbane, Australia Email: [email protected] Moe Thandar Wynn Michael Adams Queensland University of Technology, Brisbane, Australia Email: [email protected]; [email protected] Boudewijn F. van Dongen Eindhoven University of Technology, Eindhoven, The Netherlands Email: [email protected] Abstract The importance of actively managing and analysing business processes is acknowledged more than ever in or- ganisations nowadays.
    [Show full text]
  • The Backbone of the Performance Engineering Process Software
    Performance Requirements: the Backbone of the Performance Engineering Process Alexander Podelko Oracle Performance requirements should to be tracked from system's inception through its whole lifecycle including design, development, testing, operations, and maintenance. They are the backbone of the performance engineering process. However different groups of people are involved in each stage and they use their own vision, terminology, metrics, and tools that makes the subject confusing when you go into details. The paper discusses existing issues and approaches in their relationship with the performance engineering process. What is the Problem? At first glance, the subject of performance requirements looks simple enough. Almost every book about performance has a few pages about performance requirements. Quite often a performance requirements section can be found in project documentation. However, the more you examine the area of performance requirements, the more questions and issues arise. The performance engineering process is a set of performance-related activities associated with every stage of the Software Development Life Cycle (SDLC). It may even be represented as a separate Performance Engineering Life Cycle going in parallel with the basic SDLC as on the fig.1 below [Performance12]: Software Development Performance Engineering Life Cycle Life Cycle Requirements Gathering and Analysis Performance Requirements Gathering and Analysis Architecture and Design Design for Performance and Performance Modeling Construction / Implementation Unit Performance Tests and Code Optimization Testing Performance Testing Deployment and Maintenance Performance Monitoring and Capacity Management Fig.1. Performance Engineering Life Cycle. 1 Performance requirements should be tracked from the system inception through the whole system l ifecycle including design, development, testing, operations, and maintenance.
    [Show full text]
  • Applying Performance Patterns for Requirements Analysis
    Applying Performance Patterns for Requirements Analysis Azadeh Alebrahim, paluno – The Ruhr Institute for Software Technology, Germany Maritta Heisel, paluno – The Ruhr Institute for Software Technology, Germany Performance as one of the critical quality requirements for the success of a software system must be integrated into software development from the beginning to prevent performance problems. Analyzing and modeling performance demands knowledge of performance experts and analysts. In order to integrate performance analysis into software analysis and design methods, performance-specific properties known as domain knowledge have to be identified, analyzed, and documented properly. In this paper, we propose the performance analysis method PoPeRA to guide the requirements engineer in dealing with performance problems as early as possible in requirements analysis. Our struc- tured method provides support for identifying potential performance problems using performance-specific domain knowledge attached to the requirement models. To deal with identified performance problems, we make use of performance analysis patterns to be applied to the requirement models in the requirements engineering phase. To show the application of our approach, we illustrate it with the case study CoCoME, a trading system to be deployed in supermarkets for handling sales. Categories and Subject Descriptors: I.5.2 [Pattern Recognition]: Design Methodology—pattern analysis; D.2.1 [Software Engineering]: Requirements/Specifications—Methodologies; D.2.9 [Software Engineering]: Management—Software quality assurance (SQA); D.2.11 [Software Engineering]: Software Architectures—Patterns General Terms: Design, Performance Additional Key Words and Phrases: Performance patterns, problem frames, requirements engineering, software architecture, UML ACM Reference Format: Alebrahim, A. and Heisel, M. 2015. Applying Performance Patterns for Requirements Analysis jn 2, 3, Article 1 (May 2010), 15 pages.
    [Show full text]
  • From Uml to Performance Models by Xml Transformations
    FROM UML TO PERFORMANCE MODELS BY XML TRANSFORMATIONS by Ping Gu A thesis submitted to the Faculty o f Graduate Studies and Research in partial fulfillment of the requirements for the degree o f Doctor o f Philosophy Ottawa-Carleton Institute for Electrical and Computer Engineering Faculty of Engineering Department of Systems and Computer Engineering Carleton University Ottawa, Ontario, Canada, K1S 5B6 October 2006 ©Copyright 2006, Ping Gu Reproduced with permission of the copyright owner. Further reproduction prohibited without permission. Library and Bibliotheque et Archives Canada Archives Canada Published Heritage Direction du Branch Patrimoine de I'edition 395 Wellington Street 395, rue Wellington Ottawa ON K1A 0N4 Ottawa ON K1A 0N4 Canada Canada Your file Votre reference ISBN: 978-0-494-27098-1 Our file Notre reference ISBN: 978-0-494-27098-1 NOTICE: AVIS: The author has granted a non­ L'auteur a accorde une licence non exclusive exclusive license allowing Library permettant a la Bibliotheque et Archives and Archives Canada to reproduce, Canada de reproduire, publier, archiver, publish, archive, preserve, conserve, sauvegarder, conserver, transmettre au public communicate to the public by par telecommunication ou par I'lnternet, preter, telecommunication or on the Internet, distribuer et vendre des theses partout dans loan, distribute and sell theses le monde, a des fins commerciales ou autres, worldwide, for commercial or non­ sur support microforme, papier, electronique commercial purposes, in microform, et/ou autres formats. paper, electronic and/or any other formats. The author retains copyright L'auteur conserve la propriete du droit d'auteur ownership and moral rights in et des droits moraux qui protege cette these.
    [Show full text]
  • Performance Engineering Initiatives for Early Software Test of High Availability Systems
    Performance Engineering Initiatives for Early Software Test of High Availability Systems Ed Beck Sr. Manager NDIA Test & Evaluation Conference March 1-4, 2010 Slide 0 Overview Mission Solutions Engineering, LLC • Mission Solutions Engineering (MSE) is a full service systems and software engineering provider with 40 years’ experience in delivering mission systems. • Formed by CSC to address government concerns for potential OCI, MSE comprises CSC’s former systems and software engineering support to the US Navy and Missile Defense Agency, primarily operating out of Moorestown, New Jersey and with headquarters in Arlington, Virginia. • The core of our engineering activities is the design, development and integration of mission-critical software. We have been rated at Capability Maturity Model Integration1 Level 5 for software engineering (SW-CMMI®) since 2001 and systems engineering (SE- CMMI®) since 2004. Slide 1 Overview Aegis Weapon System • The world’s most advanced shipboard Anti-Air Warfare (AAW) Weapons System • A highly integrated Combat System capable of Multi- Mission warfare • Air, Surface and Subsurface • Open Architecture migration provides foundation for the modernization effort and future war fighting capabilities U.S. Navy Photo Slide 2 Overview • Performance Engineering (PE), Integration & Test • Catalyst for PE Initiatives • Performance Engineering Framework • Summary U.S. Navy Photo Slide 3 What is Performance Engineering? • Performance Engineering is an emerging Computer Science practice comprised of the following functions: – Capacity Planning – Projection of future resource needs based on historical data and growth projections – Performance Measurements – Demonstrating that the system meets performance criteria – Performance Analysis – Characterizing software behavior and identifying anomalies • In practice, as we have defined it at MSE, Performance Engineering is the process by which software is tested and tuned with the intent of realizing the required performance.
    [Show full text]
  • Performance Engineering
    Lecture Notes in Computer Science 2047 Performance Engineering State of the Art and Current Trends Bearbeitet von Reiner Dumke, Claus Rautenstrauch, Andreas Schmietendorf, Andre Scholz 1. Auflage 2001. Taschenbuch. xiv, 349 S. Paperback ISBN 978 3 540 42145 0 Format (B x L): 15,5 x 23,3 cm Gewicht: 1140 g Weitere Fachgebiete > EDV, Informatik > Hardwaretechnische Grundlagen > Systemverwaltung & Management Zu Inhaltsverzeichnis schnell und portofrei erhältlich bei Die Online-Fachbuchhandlung beck-shop.de ist spezialisiert auf Fachbücher, insbesondere Recht, Steuern und Wirtschaft. Im Sortiment finden Sie alle Medien (Bücher, Zeitschriften, CDs, eBooks, etc.) aller Verlage. Ergänzt wird das Programm durch Services wie Neuerscheinungsdienst oder Zusammenstellungen von Büchern zu Sonderpreisen. Der Shop führt mehr als 8 Millionen Produkte. Preface The performance analysis of concrete technologies has already been discussed in a multitude of publications and conferences, but the practical application has often been neglected. An engineering procedure was comprehensively discussed for the first time during the “International Workshop on Software and Performance: WOSP 1998” in Santa Fe, NM, in 1998. Teams were formed to examine the integration of per- formance analysis, in particular into software engineering. Practical experiences from industry and new research approaches were discussed in these teams. Diverse national and international activities, e.g., the foundation of a working group within the German Association of Computer Science, followed. This book continues the discussion of performance engineering methodologies. On the one hand, it is based on selected and revised contributions to conferences that took place in 2000: • Second International Workshop on Software and Performance – WOSP 2000, 17 – 20 September 2000 in Ottawa, Canada, • First German Workshop on Performance Engineering within Software Development, May 17th in Darmstadt, Germany.
    [Show full text]
  • Evaluating Engineering System Interventions
    HANDBOOK OF ENGINEERING SYSTEMS DESIGN (PREPRINT VERSION) 1 Evaluating Engineering System Interventions Wester C.H. Schoonenberg, Amro M. Farid Abstract—Our modern life has grown to depend on many sciences. Over the past decades, engineering solutions and nearly ubiquitous large complex engineering systems. have evolved from engineering artifacts that have a Transportation, water distribution, electric power, natural single function, to systems of artifacts that optimize gas, healthcare, manufacturing and food supply are but a few. These engineering systems are characterized by an the delivery of a specific service, and then to engineer- intricate web of interactions within themselves but also ing systems that deliver services within a societal and between each other. Furthermore, they have a long-standing economic context. In order to understand engineering nature that means that any change requires an intervention systems, a holistic approach is required that assesses into a legacy system rather than a new “blank-slate” system their impact beyond technical performance. Engineering design. The interventions themselves are often costly with implications lasting many decades into the future. Conse- Systems are defined as: quently, when it comes to engineering system interventions, Definition 1. Engineering System [1] A class of systems there is a real need to “get it right”. This chapter discusses two types of engineering system interventions; namely those characterized by a high degree of technical complexity, that change system behavior and those that change system social intricacy, and elaborate processes, aimed at fulfill- structure. It then discusses the types of measurement that ing important functions in society. can be applied to evaluating such interventions.
    [Show full text]
  • BPT Part 2: a Performance Engineering Strategy Without a Strategy, Performance Engineering Is Simply an Exercise in Trial and Error
    BPT Part 2: A Performance Engineering Strategy Without a strategy, performance engineering is simply an exercise in trial and error. Following a sound strategy in the engineering effort Beyond will increase your performance engineering team’s efficiency and effectiveness. This article outlines a strategy that complements the Performance Rational Unified Process® approach and is easily customizable to Testing your project and organization, and that’s been validated by numerous clients worldwide. The templates I provide will give you a starting point for documenting your performance engineering engagement. Applying this strategy, coupled with your own experience, should significantly improve your overall effectiveness as a performance engineer. by: This is the second article in the "Beyond Performance Testing" series, R. Scott Barber which focuses on isolating performance bottlenecks and working collaboratively with the development team to resolve them. If you’re new to this series, you may want to begin by reading Part 1 the series introduction. This article is intended for all levels of users of the Rational Suite® TestStudio® system testing tool, as well as managers and other members of the development team. A Closer Look at Performance Engineering As defined in Part 1, performance engineering is the process by which software is tested and tuned with the intent of realizing the required performance. Let’s look more closely at this process. In the simplest terms, this approach can be described as shown in Figure 1. What? Detect Diagnose N ot Why? Re sol ved Resolve Figure 1: Performance engineering in its simplest terms Beyond Performance Testing - BPT Part 2: A Performance Engineering Strategy © PerfTestPlus, Inc.
    [Show full text]
  • UML Profile for Schedulability, Performance, and Time Specification
    UMLTM Profile for Schedulability, Performance, and Time Specification January 2005 Version 1.1 formal/05-01-02 An Adopted Specification of the Object Management Group, Inc. Copyright © 2001, ARTiSAN Software Tools, Inc. Copyright © 2001, I-Logix, Inc. Copyright © 2003, Object Management Group Copyright © 2001, International Business Machines Corp. Copyright © 2001, Telelogic AB Copyright © 2001, TimeSys Corporation Copyright © 2001, Tri-Pacific Software USE OF SPECIFICATION - TERMS, CONDITIONS & NOTICES The material in this document details an Object Management Group specification in accordance with the terms, conditions and notices set forth below. This document does not represent a commitment to implement any portion of this specification in any company's products. The information contained in this document is subject to change without notice. LICENSES The companies listed above have granted to the Object Management Group, Inc. (OMG) a nonexclusive, royalty-free, paid up, worldwide license to copy and distribute this document and to modify this document and distribute copies of the modified version. Each of the copyright holders listed above has agreed that no person shall be deemed to have infringed the copyright in the included material of any such copyright holder by reason of having used the specification set forth herein or having conformed any computer software to the specification. Subject to all of the terms and conditions below, the owners of the copyright in this specification hereby grant you a fully- paid up, non-exclusive,
    [Show full text]