Software Error Analysis Technology

Total Page:16

File Type:pdf, Size:1020Kb

Software Error Analysis Technology NIST Special Publication 500-209 Computer Systems Software Error Analysis Technology U.S. DEPARTMENT OF COMMERCE Wendy W. Peng Technology Administration Dolores R. Wallace National Institute of Standards and Technology NAT L INST. OF ST4ND & TECH R I.C. NISI PUBLICATIONS A111D3 TTSTll ^QC ' 100 .U57 //500-209 1993 7he National Institute of Standards and Technology was established in 1988 by Congress to "assist industry in the development of technology . needed to improve product quality, to modernize processes, to reliability . manufacturing ensure product and to facilitate rapid commercialization . , of products based on new scientific discoveries." NIST, originally founded as the National Bureau of Standards in 1901, works to strengthen U.S. industry's competitiveness; advance science and engineering; and improve public health, safety, and the environment. One of the agency's basic functions is to develop, maintain, and retain custody of the national standards of measurement, and provide the means and methods for comparing standards used in science, engineering, manufacturing, commerce, industry, and education with the standards adopted or recognized by the Federal Government. As an agency of the U.S. Commerce Department's Technology Administration, NIST conducts basic and applied research in the physical sciences and engineering and performs related services. The Institute does generic and precompetitive work on new and advanced technologies. NIST's research facilities are located at Gaithersburg, MD 20899, and at Boulder, CO 80303. Major technical operating units and their principal activities are listed below. For more information contact the Public Inquiries Desk, 301-975-3058. Technology Services Manufacturing Engineering Laboratory • Manufacturing Technology Centers Program • Precision Engineering • Standards Services • Automated Production Technology • Technology Commercialization • Robot Systems • Measurement Services • Factory Automation • Technology Evaluation and Assessment • Fabrication Technology • Information Services Materials Science and Engineering Electronics and Electrical Engineering Laboratory Laboratory • Intelligent Processing of Materials • Microelectronics • Ceramics • Law Enforcement Standards • Materials Reliability' • Electricity • Polymers • Semiconductor Electronics • Metallurgy • Electromagnetic Fields' • Reactor Radiation • Electromagnetic Technology' Building and Fire Research Laboratory Chemical Science and Technology • Structures Laboratory • Building Materials • Biotechnology • Building Environment • Chemical Engineering' • Fire Science and Engineering • Chemical Kinetics and Thermodynamics • Fire Measurement and Research • Inorganic Analytical Research • Organic Analytical Research Computer Systems Laboratory • Process Measurements • Information Systems Engineering • Surface and Microanalysis Science • Systems and Software Technology • Thermophysics^ • Computer Security • Systems and Network Architecture Physics Laboratory • Advanced Systems • Electron and Optical Physics • Atomic Physics Computing and Applied Mathematics • Molecular Physics Laboratory • Radiometric Physics • Applied and Computational Mathematics^ • Quantum Metrology • Statistical Engineering^ • Ionizing Radiation • Scientific Computing Environments^ • Time and Frequency' • Computer Services^ • Quantum Physics' • Computer Systems and Communications^ • Information Systems 'At Boulder, CO 80303. ^Some elements at Boulder, CO 80303. NIST Special Publication 500-209 Software Error Analysis Wendy W. Peng Dolores R. Wallace Systems and Software Technology Division Computer Systems Laboratory National Institute of Standards and Technology Gaithersburg, MD 20899 April 1993 U.S. DEPARTMENT OF COMMERCE Ronald H. Brown, Secretary NATIONAL INSTITUTE OF STANDARDS AND TECHNOLOGY Raymond G. Kammer, Acting Director Reports on Computer Systems Technology The National Institute of Standards and Technology (NIST) has a unique responsibility for computer systems technology within the Federal government. NIST's Computer Systems Laboratory (CSL) devel- ops standards and guidelines, provides technical assistance, and conducts research for computers and related telecommunications systems to achieve more effective utilization of Federal information technol- ogy resources. CSL's responsibilities include development of technical, management, physical, and ad- ministrative standards and guidelines for the cost-effective security and privacy of sensitive unclassified information processed in Federal computers. CSL assists agencies in developing security plans and in improving computer security awareness training. This Special Publication 500 series reports CSL re- search and guidelines to Federal agencies as well as to organizations in industry, government, and academia. National Institute of Standards and Technology Special Publication 500-209 Natl. Inst. Stand. Technol. Spec. Publ. 500-209, 113 pages (Apr. 1993) CODEN: NSPUE2 U.S. GOVERNMENT PRINTING OFFICE WASHINGTON: 1993 ABSTRACT This document provides guidance on software error analysis. Software error analysis includes error detection, analysis, and resolution. Error detection techniques considered in the study are those used in software development, software quality assurance, and software verification, validation and testing activities. These techniques are those frequendy cited in technical literature and software engineering standards or those representing new approaches to support error detection. The study includes stadstical process control techniques and rslates them to their use as a software quality assurance technique for both product and process improvement. Finally, the report describes several software reliability models. KEYWORDS Data Collection; Error Detection; Error Removal; High Integrity Software; Metrics; Software Error Analysis; Software Quality Assurance; Software Verification and Validation; Stadsdcal Process Control iii EXECUTIVE SUMMARY The main purpose of this document is to provide the software engineering community with current information regarding error analysis for software, which will assist them to do the following: • Understand how error analysis can aid in improving the software development process; • Assess the quality of the software, with the aid of error detection techniques; • Analyze errors for their cause and then fix the errors; and • Provide guidelines for the evaluation of high-integrity software. The software industry is currently still young, without sufficient knowledge and adequate standards to guarantee fault-free software. Although research continues to identify better processes for error prevention, with current practices, errors will probably occur during software development and maintenance. Hence, there is the need for error analysis. Error analysis for software includes the acdvities of detecting errors, collecdng and recording error data, analyzing and removing single errors, and analyzing collective error data to remove classes of errors. The collective error data may be used with stadsdcal process control (SPC) techniques to improve the product and the processes used in developing, maintaining, and assuring the quality of software. This report provides a descripdon of error detection techniques which are cited frequently in technical literature and standards and describes the cost benefits of applying error detection early in the lifecycle. However, error detection alone is not sufficient for removal of an error. Informadon must be recorded about the error to assist in the analysis of its cause, its removal, and its relationship to the project and to other similar projects. This report provides guidance on data collecdon, analysis, and removal as well as error detecdon. This report describes how several SPC techniques can be used for software quality assurance technique and for process improvement. The report identifies metrics related to software error detection and identifies several software reliability esdmation models. Metrics are used to assess the product or process, while SPC techniques are used to monitor a project by observing trends. SPC techniques help to locate major problems in the development process, the assurance processes (e.g., software quality assurance, verification and validation), and the product itself The study of software engineering standards reported in [NUREG, NIST204] indicates that standards are beginning to associate requirements for error detection techniques with the quality requirements and problem types of the software project implementing the standard. Further examination of these documents and additional standards and guidelines for high integrity software indicates that these documents vary widely in their recommendations of specific error v techniques. Appendix B provides a summary of the error detection techniques required or recommended by these documents for the assurance of the quality of high integrity software. This report recommends collection of error data into an organizational database for use by a vendor' over several projects, and modified collections of these databases for use by government auditors (e.g., Nuclear Regulatory Commission, Environmental Protection Agency). Software organizations should consider institudonalizing mechanisms for establishing and maintaining a database of error analysis data within their organizadon. Over time, it may become apparent that some error analysis techniques are more effecdve than others with respect to a given type of problem. It may
Recommended publications
  • Measuring the Software Size of Sliced V-Model Projects
    2014 Joint Conference of the International Workshop on Software Measurement and the International Conference on Software Process and Product Measurement Measuring the Software Size of Sliced V-model Projects Andreas Deuter PHOENIX CONTACT Electronics GmbH Gregor Engels Dringenauer Str. 30 University of Paderborn 31812 Bad Pyrmont, Germany Zukunftsmeile 1 Email: [email protected] 33102 Paderborn, Germany Email: [email protected] Abstract—Companies expect higher productivity of their soft- But, the “manufacturing” of software within the standard ware teams when introducing new software development methods. production process is just a very short process when bringing Productivity is commonly understood as the ratio of output the binaries of the software to the device. This process is hardly created and resources consumed. Whereas the measurement of to optimize and does not reflect the “production of software” the resources consumed is rather straightforward, there are at all. The creation of the software is namely done in the several definitions for counting the output of a software de- developing teams by applying typical software engineering velopment. Source code-based metrics create a set of valuable figures direct from the heart of the software - the code. However, methods. However, to keep up with the high demands on depending on the chosen process model software developers and implementing new functionality (e.g. for PLC) the software testers produce also a fair amount of documentation. Up to development process within these companies must improve. now this output remains uncounted leading to an incomplete Therefore, they start to analyze their software processes in view on the development output. This article addresses this open detail and to identify productivity drivers.
    [Show full text]
  • Software Tools: a Building Block Approach
    SOFTWARE TOOLS: A BUILDING BLOCK APPROACH NBS Special Publication 500-14 U.S. DEPARTMENT OF COMMERCE National Bureau of Standards ] NATIONAL BUREAU OF STANDARDS The National Bureau of Standards^ was established by an act of Congress March 3, 1901. The Bureau's overall goal is to strengthen and advance the Nation's science and technology and facilitate their effective application for public benefit. To this end, the Bureau conducts research and provides: (1) a basis for the Nation's physical measurement system, (2) scientific and technological services for industry and government, (3) a technical basis for equity in trade, and (4) technical services to pro- mote public safety. The Bureau consists of the Institute for Basic Standards, the Institute for Materials Research, the Institute for Applied Technology, the Institute for Computer Sciences and Technology, the Office for Information Programs, and the ! Office of Experimental Technology Incentives Program. THE INSTITUTE FOR BASIC STANDARDS provides the central basis within the United States of a complete and consist- ent system of physical measurement; coordinates that system with measurement systems of other nations; and furnishes essen- tial services leading to accurate and uniform physical measurements throughout the Nation's scientific community, industry, and commerce. The Institute consists of the Office of Measurement Services, and the following center and divisions: Applied Mathematics — Electricity — Mechanics — Heat — Optical Physics — Center for Radiation Research — Lab- oratory Astrophysics^ — Cryogenics^ — Electromagnetics^ — Time and Frequency*. THE INSTITUTE FOR MATERIALS RESEARCH conducts materials research leading to improved methods of measure- ment, standards, and data on the properties of well-characterized materials needed by industry, commerce, educational insti- tutions, and Government; provides advisory and research services to other Government agencies; and develops, produces, and distributes standard reference materials.
    [Show full text]
  • Manual on Quality Assurance for Computer Software Related to the Safety of Nuclear Power Plants
    SIMPLIFIED SOFTWARE LIFE-CYCLE DIAGRAM FEASIBILITY STUDY PROJECT TIME I SOFTWARE P FUNCTIONAL I SPECIFICATION! SOFTWARE SYSTEM DESIGN DETAILED MODULES CECIFICATION MODULES DESIGN SOFTWARE INTEGRATION AND TESTING SYSTEM TESTING ••COMMISSIONING I AND HANDOVER | DECOMMISSION DESIGN DESIGN SPECIFICATION VERIFICATION OPERATION AND MAINTENANCE SOFTWARE LIFE-CYCLE PHASES TECHNICAL REPORTS SERIES No. 282 Manual on Quality Assurance for Computer Software Related to the Safety of Nuclear Power Plants f INTERNATIONAL ATOMIC ENERGY AGENCY, VIENNA, 1988 MANUAL ON QUALITY ASSURANCE FOR COMPUTER SOFTWARE RELATED TO THE SAFETY OF NUCLEAR POWER PLANTS The following States are Members of the International Atomic Energy Agency: AFGHANISTAN GUATEMALA PARAGUAY ALBANIA HAITI PERU ALGERIA HOLY SEE PHILIPPINES ARGENTINA HUNGARY POLAND AUSTRALIA ICELAND PORTUGAL AUSTRIA INDIA QATAR BANGLADESH INDONESIA ROMANIA BELGIUM IRAN, ISLAMIC REPUBLIC OF SAUDI ARABIA BOLIVIA IRAQ SENEGAL BRAZIL IRELAND SIERRA LEONE BULGARIA ISRAEL SINGAPORE BURMA ITALY SOUTH AFRICA BYELORUSSIAN SOVIET JAMAICA SPAIN SOCIALIST REPUBLIC JAPAN SRI LANKA CAMEROON JORDAN SUDAN CANADA KENYA SWEDEN CHILE KOREA, REPUBLIC OF SWITZERLAND CHINA KUWAIT SYRIAN ARAB REPUBLIC COLOMBIA LEBANON THAILAND COSTA RICA LIBERIA TUNISIA COTE D'lVOIRE LIBYAN ARAB JAMAHIRIYA TURKEY CUBA LIECHTENSTEIN UGANDA CYPRUS LUXEMBOURG UKRAINIAN SOVIET SOCIALIST CZECHOSLOVAKIA MADAGASCAR REPUBLIC DEMOCRATIC KAMPUCHEA MALAYSIA UNION OF SOVIET SOCIALIST DEMOCRATIC PEOPLE'S MALI REPUBLICS REPUBLIC OF KOREA MAURITIUS UNITED ARAB
    [Show full text]
  • Software Metrics: Successes, Failures and New Directions
    Software Metrics: Successes, Failures and New Directions Norman E Fenton Centre for Software Reliability Agena Ltd City University 11 Main Street Northampton Square Caldecote London EC1V OHB Cambridge CB3 7NU United Kingdom United Kingdom Tel: 44 171 477 8425 Tel 44 1223 263753 / 44 181 530 5981 Fax: 44 171 477 8585 Fax: 01223 263899 [email protected] [email protected] www.csr.city.ac.uk/people/norman.fenton/ www.agena.co.uk Abstract The history of software metrics is almost as old as the history of software engineering. Yet, the extensive research and literature on the subject has had little impact on industrial practice. This is worrying given that the major rationale for using metrics is to improve the software engineering decision making process from a managerial and technical perspective. Industrial metrics activity is invariably based around metrics that have been around for nearly 30 years (notably Lines of Code or similar size counts, and defects counts). While such metrics can be considered as massively successful given their popularity, their limitations are well known, and mis-applications are still common. The major problem is in using such metrics in isolation. We argue that it is possible to provide genuinely improved management decision support systems based on such simplistic metrics, but only by adopting a less isolationist approach. Specifically, we feel it is important to explicitly model: a) cause and effect relationships and b) uncertainty and combination of evidence. Our approach uses Bayesian Belief nets which are increasingly seen as the best means of handling decision-making under uncertainty.
    [Show full text]
  • Software Quality Assurance Activities in Software Testing
    Software Quality Assurance Activities In Software Testing Tony never synopsizing any recidivist gazetting thus, is Brooke oncogenic and insolvable enough? Monogenous Chadd externalises, his disciplinarians denudes spring-clean Germanically. Spindliest Antoni never humors so edgewise or attain any shells lyingly. Each module performs one or two tasks, and thenpasses control to another module. Perform test automation for web application using Cucumber. Identify and describe safety software procurement methods, including supplier evaluation and source inspection processes. He previously worked at IBM SWS Toronto Lab. The information maintained in status accounting should enable the rebuild of any previous baseline. Beta Breakers supports all industry sectors. Thank you save time for all the lack of that includes test software assurance and must often. Focus on demonstrating pos next column containing algorithms, activities in software quality assurance testing activities of testing programs for their findings from his piece of skills, validate features to refresh teh page object. XML data sets to simulate production, using LLdap and ALTOVA. Schedule information should be expressed as absolute dates, as dates relative to either SCM or project milestones, or as a simple sequence of events. Its scope of software quality assurance and the correct email list all testshave been completely correct, in software quality assurance activities to be precisely known about its process on a familiarity level. These exercises are performed at every step along the way in the workshop. However, you have to balance driving out quality with production value. The second step is the validation of the computer system implementation against the computer system requirements. Software development tools, whose output becomes part of the program implementation and which can therefore introduce errors.
    [Show full text]
  • Software Quality Assurance Report
    Software Quality Assurance Report Elmore demobilizing onerously if halftone Filmore reboils or discounts. Murmurous and second-rate Steve roams routedher astrodynamics when gazetted grumbled some yelpdry orvery inditing evilly literally,and centesimally? is Fonzie well-entered? Is Tymothy always volatilisable and Defined by the importance of these terminologies so with a number of the continuous integration testing and the software engineering institute, no significant technical objectives of software quality assurance report The report and. The staff members need to be reported to notifications, such as well as both dynamic. Test Report fatigue a document which contains a conjunction of all test activities and final test results of a testing project Test report is an assessment of how adverse the Testing is performed Based on the test report stakeholders can evaluate current quality feel the tested product and persuade a decision on paid software release. Measuring and reporting mechanisms SQA Activities Software quality assurance is composed of a conduct of functions associated with just different constituencies. Senior software Quality Assurance Engineer Arizona City. It is not understand and federal regulations. Sqa effort in the product quality assurance enterprise network environments. This vessel will be cozy for professionals not head in software testing but inland from other areas project managers product owners developers etc Article Content. An established standards are extrinsic factors, they must define a software for soil science. Document or the product for the data are not. Quality Assurance is defined as the auditing and reporting procedures used to. Cost coverage customer record such as receiving the defect report and troubleshooting.
    [Show full text]
  • Software Quality Assurance and Quality Management
    12. Software Quality Assurance and Quality Management Abdus Sattar Assistant Professor Department of Computer Science and Engineering Daffodil International University Email: [email protected] Discussion Topics Software Quality Assurance What are SQA, SQP, SQC, and SQM? Elements of Software Quality Assurance SQA Tasks, Goals, And Metrics Quality Management and Software Development QA Vs QC Reviews and Inspections An Inspection Checklist Software Quality Assurance Software Quality Assurance (SQA) is a set of activities for ensuring quality in software engineering processes. It ensures that developed software meets and complies with the defined or standardized quality specifications. SQA is an ongoing process within the Software Development Life Cycle (SDLC) that routinely checks the developed software to ensure it meets the desired quality measures. The implication for software is that many different constituencies he implication for software is that many different constituencies What are SQA, SQP, SQC, and SQM? Software Quality Assurance – establishment of network of organizational procedures and standards leading to high-quality software Software Quality Planning – selection of appropriate procedures and standards from this framework and adaptation of these to specific software project Software Quality Control – definition and enactment of processes that ensure that project quality procedures and standards are being followed by the software development team Software Quality Metrics – collecting and analyzing quality data to predict and control quality of the software product being developed Elements of Software Quality Assurance . Standards. The IEEE, ISO, and other standards organizations have produced a broad array of software engineering standards and related documents. Reviews and audits. Technical reviews are a quality control activity performed by software engineers for software engineers.
    [Show full text]
  • Software Quality Management for Improving Quality of Software Product
    Journal of Network Communications and Emerging Technologies (JNCET) www.jncet.org Volume 8, Issue 5, May (2018) Software Quality Management for Improving Quality of Software Product Shubham Srivastava 1, Prabha singh 2, Chayanika Srivastava 3 1, 2, 3 Department of Computer Science and Engineering, ITM, GIDA, Gorakhpur (India) Abstract – Software quality management (SQM) is a process that ensures that developed software meets and complies with defined or standardized quality specifications. SQM is an ongoing process within the software development life cycle (SDLC) that routinely checks the developed software to ensure it meets desired quality measures. In this article, it is highlighted that how to combine software quality management life cycle with software development life cycle to generate better quality and outcomes. This article takes a different approach by examining how software quality management practices can impact project outcomes. Software quality management (SQM) plays a critical role in the software development lifecycle (SDLC) and can impact a project’s overall success. Index Terms – SQM, SQA, SQP, SQC and SDLC. 1. INTRODUCTION Today’s scenario, it is really difficult to implement successful IT projects which is a critical strategic method for entire Fig:-1 SQM ACTIVITIES industrial sectors. This is even more important in times of scarce resources needed for other competing strategic 2. PROJECT MANAGEMENT FAILURES initiatives important to a firm. A lot of software projects still Some Basic Reasons for Failure:- fail to deliver on time and within target costs and other necessary specifications. Software quality management (SQM) Planning: Insufficient planning is a management process the goal of which is to develop and manage the quality of a software to make sure the product Scope: Poorly defined and requirement and scope satisfies the user.
    [Show full text]
  • Software Measurement: a Necessary Scientific Basis Norman Fenton
    IEEE TRANSACTIONS ON SOFTWARE ENGINEERING. VOL. 20, NO. 3, MARCH I994 I99 Software Measurement: A Necessary Scientific Basis Norman Fenton Abstruct- Software measurement, like measurement in any 11. MEASUREMENTFUNDAMENTALS other discipline, must adhere to the science of measurement if it is to gain widespread acceptance and validity. The observation In this section, we provide a summary of the key concepts of some very simple, but fundamental, principles of measurement from the science of measurement which are relevant to soft- can have an extremely beneficial effect on the subject. Measure- ware metrics. First, we define the fundamental notions (which ment theory is used to highlight both weaknesses and strengths of are generally not well understood) and then we summarise the software metrics work, including work on metrics validation. We representational theory of measurement. Finally, we explain identify a problem with the well-known Weyuker properties, but also show that a criticism of these properties by Cherniavsky and how this leads inevitably to a goal-oriented approach. Smith is invalid. We show that the search for general software complexity measures is doomed to failure. However, the theory A. What is Measurement? does help us to define and validate measures of specific complexity attributes. Above all, we are able to view software measurement Measurement is defined as the process by which numbers or in a very wide perspective, rationalising and relating its many symbols are assigned to attributes of entities in the real world diverse activities. in such a way as to describe them according to clearly defined Index Terms-Software measurement, empirical studies, met- rules [ 131, [36].
    [Show full text]
  • Ssfm BPG 1: Validation of Software in Measurement Systems
    NPL Report DEM-ES 014 Software Support for Metrology Best Practice Guide No. 1 Validation of Software in Measurement Systems Brian Wichmann, Graeme Parkin and Robin Barker NOT RESTRICTED January 2007 National Physical Laboratory Hampton Road Teddington Middlesex United Kingdom TW11 0LW Switchboard 020 8977 3222 NPL Helpline 020 8943 6880 Fax 020 8943 6458 www.npl.co.uk Software Support for Metrology Best Practice Guide No. 1 Validation of Software in Measurement Systems Brian Wichmann with Graeme Parkin and Robin Barker Mathematics and Scientific Computing Group January 2007 ABSTRACT The increasing use of software within measurement systems implies that the validation of such software must be considered. This Guide addresses the validation both from the perspective of the user and the supplier. The complete Guide consists of a Management Overview and a Technical Application together with consideration of its use within safety systems. Version 2.2 c Crown copyright 2007 Reproduced with the permission of the Controller of HMSO and Queen’s Printer for Scotland ISSN 1744–0475 National Physical Laboratory, Hampton Road, Teddington, Middlesex, United Kingdom TW11 0LW Extracts from this guide may be reproduced provided the source is acknowledged and the extract is not taken out of context We gratefully acknowledge the financial support of the UK Department of Trade and Industry (National Measurement System Directorate) Approved on behalf of the Managing Director, NPL by Jonathan Williams, Knowledge Leader for the Electrical and Software team Preface The use of software in measurement systems has dramatically increased in the last few years, making many devices easier to use, more reliable and more accurate.
    [Show full text]
  • Objectives and Context of Software Measurement, Analysis, and Control Michael A
    Objectives and Context of Software Measurement, Analysis, and Control Michael A. Cusumano Massachusetts Institute of Technology Sloan School WP#3471-92/BPS October 8, 1992 Objectives and Context of Software Measurement, Analysis and Control Michael A. Cusumano Sloan School of Management Massachusetts Institute of Technology Cambridge, Massachusetts USA 02142 Abstract This paper focuses on the what and why of measurement in software development, and the impact of the context of the development organization or of specific projects on approaches to measurement, analysis, and control. It also presents observations on contextual issues based on studies of the major Japanese software producers. 1 Introduction This paper focuses on the what and why of measurement in software development, and the impact of the context of the development organization or of specific projects on approaches to measurement, analysis, and control. It begins by discussing three basic questions: (1) Why do we measure? (2) What is it we want to know? And (3) how does the context affect the ability to measure? The second section presents observations on contextual issues based on my ongoing study of the Japanese [3]. This paper also relies heavily on extensive interviews conducted at Hitachi, Fujitsu, NEC, and Toshiba during May, August, and September 1992. 2 Basic Questions Why we measure aspects of the software development process and conduct analytical exercises or experiments, as well as what we try to do with this information, depends very much on who is doing or requesting the measuring and analysis. There already exists a broad literature on technical aspects software measurement and the relationship to process management, which will not be reviewed here [1, 2, 5, 6].
    [Show full text]
  • 12 Steps to Useful Software Metrics
    12 Steps to Useful Software Metrics Linda Westfall The Westfall Team [email protected] PMB 101, 3000 Custer Road, Suite 270 Plano, TX 75075 972-867-1172 (voice) 972-943-1484 (fax) Abstract: 12 Steps to Useful Software Metrics introduces the reader to a practical process for establishing and tailoring a software metrics program that focuses on goals and information needs. The process provides a practical, systematic, start-to-finish method of selecting, designing and implementing software metrics. It outlines a cookbook method that the reader can use to simplify the journey from software metrics in concept to delivered information. Bio: Linda Westfall is the President of The Westfall Team, which provides Software Metrics and Software Quality Engineering training and consulting services. Prior to starting her own business, Linda was the Senior Manager of the Quality Metrics and Analysis at DSC Communications where her team designed and implemented a corporate wide metric program. Linda has more than twenty-five years of experience in real-time software engineering, quality and metrics. She has worked as a Software Engineer, Systems Analyst, Software Process Engineer and Manager of Production Software. Very active professionally, Linda Westfall is a past chair of the American Society for Quality (ASQ) Software Division. She has also served as the Software Division’s Program Chair and Certification Chair and on the ASQ National Certification Board. Linda is a past-chair of the Association for Software Engineering Excellence (ASEE) and has chaired several conference program committees. Linda Westfall has an MBA from the University of Texas at Dallas and BS in Mathematics from Carnegie- Mellon University.
    [Show full text]