Appendix A: Measuring Software Size

Total Page:16

File Type:pdf, Size:1020Kb

Appendix A: Measuring Software Size Appendix A: Measuring Software Size Software size matters and you should measure it. —Mary Gerush and Dave West Software size is the major determinant of software project effort. Effectively measuring software size is thus a key element of successful effort estimation. Failed software sizing is often the main contributor to failed effort estimates and, in consequence, failed projects. Measuring software size is a challenging task. A number of methods for mea- suring software size have been proposed over the years. Yet, none of them seems to be the “silver bullet” that solves all problems and serves all purposes. Depending on the particular situation different size metrics should typically be used. In this appendix we provide a brief overview of the common software sizing approaches: functional and structural. These approaches represent two, often- confused, perspectives on software size: customer perspective and developer per- spective. The first perspective represents value of software, whereas the second perspective represents the cost (work) of creating this value. We summarize the most relevant strengths and weaknesses of the two most popular sizing methods that represent these perspectives: function point analysis (FPA) and lines of code (LOC). In technical terms, FPA represents “what” whereas LOC represent “how” in software development. Finally, we provide guidelines for selecting appropriate sizing methods for the purpose of effort estimation. A.1 Introduction There are two aspects to sizing a system: what is the count, and what is the meaning or definition of the unit being counted? —Phillip G. Armour A. Trendowicz and R. Jeffery, Software Project Effort Estimation, 433 DOI 10.1007/978-3-319-03629-8, # Springer International Publishing Switzerland 2014 434 Appendix A: Measuring Software Size A.1.1 What Is Software Size? One of the key questions project managers need to answer is “how big is my development project going to be?” The key to answering this question is software size. The size of software is the major determinant of project volume, and thus it is commonly used as a proxy for project’s size. It is a basis for essential project management activities, such as planning, tracking, and controlling project scope and productivity. But what constitutes software size? In order to answer this question we need to answer two simple questions: what is software and what is size of software? The first question addresses multiple facets of software that need to be consid- ered. Is software the executable code delivered to the customer? Does it include user documentation? Or does software encompass all products created during the whole software development project? The second question addresses multiple abstractions of software. Is size refer- ring to abstract properties of software such as functionality or to physical properties of software project outcomes such as code and documentation? This multiple nature of software makes measuring its size an extremely challenging task. A number of software sizing metrics have been proposed to address this challenge. Yet, practitioners criticize traditional sizing techniques as being too complex and expensive. In consequence, very few organizations measure size of software or of a project in a consistent and systematic way. The software sizing solution depends on the specific situation in which we need to measure size, and in particular our needs and capabilities. Tip " Choose the right sizing method based on your objectives and capabilities. As with any other technology, software sizing requires associated introduction, applica- tion, and continuous optimization processes. A.1.2 Types of Size Metrics Briand and Wieczorek (2002) group software sizing methods into two types: problem-oriented and solution-oriented. Problem-oriented metrics represent the size of the problem that the future software product will solve, i.e., the target functionality of the system. Since these metrics typically focus on measuring elements of software functionality they are also commonly referred to as functional size metrics. Solution-oriented metrics represent the size of the software develop- ment outcomes such as specifications, the design, or implementation of software system. Since they typically focus on measuring structural properties of software products they are also commonly referred to as structural size metrics. In principle, we may say that while problem-oriented measurement methods represent a cus- tomer perspective, the solution-oriented methods represent a developer perspective. Appendix A: Measuring Software Size 435 Table A.1 Methods for estimating the functional size of software Sizing method Description Function Introduced by Albrecht (1979) the function point method measures the relative size points of software functions based on the amount of different data types processed (enter, leave, read and written to storage) by a software function. Albrecht’s approach evolved into several variants of functional size measurement methods. Several of them have been published as ISO standards: • IFPUG: ISO-20926 (ISO 2009) • NESMA: ISO-24570 (ISO 2005) • FiSMA: ISO-29881 (ISO 2008) • COSMIC: ISO-19761 (ISO 2003) • Mk II: ISO-20968 (ISO 2002) Feature Introduced by Jones (1996) “The SPR Feature Point Method” extended Albrecht’s points function point method by additionally counting algorithms and multiplying this count by relative complexity of an algorithm. This method suffers from imprecise description and lack of reference applications. It has been largely abandoned in practice Use-case Introduced by Karner (1993) Use-Case Points (UCP) are inspired and based on points function point analysis and founded on its philosophy. Inputs for UCP are use-case specifications. UCP counts instances of two elements: actors and use cases. Similar to FPA each instance is evaluated with respect to its complexity and assigned an appropriate weight. The unadjusted CUP count is equal to the weighted sum of individual counts. Similar to FPA unadjusted UCP can be modified using an adjustment factor, which is determined on the basis of several technical complexity and environmental factors Story points Introduced by Cohn (2004, 2005) for the purpose of estimating the size of agile software projects. Story Point is an abstract unit of software size that represents an intuitive mixture of development effort and complexity. Story points are assigned according to a Fibonacci number sequence, where each number is the sum of the previous two. As noted earlier, in planning poker story points may start from 1/2 and continue as 1, 2, 3, 5, 8, 13, and so on Object points A number of different object point metrics have been proposed for measuring software size. Two example measurement methods are: • Predictive Object Points (POP): Introduced by Minkiewicz (1997) predictive object points are based on counts of classes and weighted methods per class, with adjustments for the average depth of the inheritance tree and the average number of children per class. Methods are weighted by considering their type (constructor, destructor, modifier, selector, iterator) and complexity (low, average, high), giving a number of POPs in a way analogous to traditional FPs • Object Points (OP): Introduced by Banker et al. (1992) object points measure software size by counting visual widgets of the fourth generation languages in integrated CASE environments. Widgets are (1) logical system components, (2) constructs of the third generation language that is used to supplement fourth generation language code, (3) user interface screens, and (4) user report. Widgets counts are adjusted by complexity and reuse multipliers, and summed to total object points count. Other than POP object points do not consider such object-oriented concepts as inheritance or encapsulation. Instead OP considers elements of user interface (continued) 436 Appendix A: Measuring Software Size Table A.1 (continued) Sizing method Description Web objects Introduced by Reifer (2000) the Web Objects method extends the well-known Function Point Analysis in that, in addition to the five elements already considered in FPA, it counts four elements specific to Web application: multimedia files, web building blocks, scripts, and links. Web objects are measured on user requirements and Web page designs. The counting procedure is analogical to FPA and comprises counting instances of the nine elements and weighting them according to their complexity. Total size is computed as weighted sum of all individual counts Problem-Oriented Size Metrics Problem-oriented software sizing methods identify the elements of the problem, and count their instances. These instances can additionally be weighted with respect to their complexity. Total software size is then a result of the weighted sum over individual instance counts. Several methods provide an opportunity to adjust software size with respect to the particular context in which the software is developed. For this purpose they consider a set of project context factors, commonly referred to as technical or environmental factors. The method rates each factor based on the associated weights and constants defined within the method. The total adjustment coefficient is then computed as a weighted sum of ratings for all factors. Software size is then adjusted by multiplying it by the adjustment coefficient. Table A.1 provides a brief overview of common problem-oriented software size measurement methods. Solution-Oriented Size Metrics Solution-oriented software sizing
Recommended publications
  • The Principles of Sizing and Estimating Projects Using
    Principles of Sizing… www.davidconsultinggroup.com The Principles of Sizing and Estimating Projects Using International Function Point User Group (IFPUG) Function Points By David Garmus, David Consulting Group Introduction Software practitioners are frequently challenged to provide early and accurate software project estimates. It speaks poorly of the software community that the issue of accurate estimating, early in the lifecycle, has not been adequately addressed and standardized. The CHAOS Report by The Standish Group’s study on software development projects revealed: 60% of projects were behind schedule 50% were over cost, and 45% of delivered projects were unusable. At the heart of the estimating challenge are two issues: 1) the need to understand and express (as early as possible) the software problem domain and 2) the need to understand our capability to deliver the required software solution within a specified environment. Then, and only then, can we accurately predict the effort required to deliver the product. The software problem domain can be defined simply as the scope of the required software. The problem domain must be accurately assessed for its size and complexity. To complicate the situation, experience tells us that at the point in time that we need an initial estimate (early in the systems lifecycle), we cannot presume to have all the necessary information at our disposal. Therefore, we must have a rigorous process that permits a further clarification of the problem domain. Various risk factors relating to environment, technology, tools, methodologies, and people skills and motivation influence our capability to deliver. An effective estimating model, as contained in Figure 1, considers three elements: size, complexity and risk factors, to determine an estimate.
    [Show full text]
  • A Metrics-Based Software Maintenance Effort Model
    A Metrics-Based Software Maintenance Effort Model Jane Huffman Hayes Sandip C. Patel Liming Zhao Computer Science Department Computer Science Department Computer Science Department Lab for Advanced Networking University of Louisville University of Kentucky University of Kentucky [email protected] [email protected] [email protected] (corresponding author) Abstract planning a new software development project. Albrecht introduced the notion of function points (FP) to estimate We derive a model for estimating adaptive software effort [1]. Mukhopadhyay [19] proposes early software maintenance effort in person hours, the Adaptive cost estimation based on requirements alone. Software Maintenance Effort Model (AMEffMo). A number of Life Cycle Management (SLIM) [23] is based on the metrics such as lines of code changed and number of Norden/Rayleigh function and is suitable for large operators changed were found to be strongly correlated projects. Shepperd et al. [27] argued that algorithmic cost to maintenance effort. The regression models performed models such as COCOMO and those based on function well in predicting adaptive maintenance effort as well as points suggested an approach based on using analogous provide useful information for managers and maintainers. projects to estimate the effort for a new project. In addition to the traditional off-the-self models such as 1. Introduction COCOMO, machine-learning methods have surfaced recently. In [17], Mair et al. compared machine-learning Software maintenance typically accounts for at least 50 methods in building software effort prediction systems. percent of the total lifetime cost of a software system [16]. There has also been some work toward applying fuzzy Schach et al.
    [Show full text]
  • Application of Sizing Estimation Techniques for Business Critical Software Project Management
    International Journal of Soft Computing And Software Engineering (JSCSE) e-ISSN: 2251-7545 Vol.3,No.6, 2013 DOI: 10.7321/jscse.v3.n6.2 Published online: Jun 25, 2013 Application of Sizing Estimation Techniques for Business Critical Software Project Management *1 Parvez Mahmood Khan, 2 M.M. Sufyan Beg Department of Computer Engineering, J.M.I. New Delhi, India 1 [email protected] , 2 [email protected] Abstract Estimation is one of the most critical areas in software project management life cycle, which is still evolving and less matured as compared to many other industries like construction, manufacturing etc. Originally the word estimation, in the context of software projects use to refer to cost and duration estimates only with software-size almost always assumed to be a fixed input. Continued legacy of bad estimates has compelled researchers, practitioners and business organizations to draw their attention towards another dimension of the problem and seriously validate an additional component – size estimation. Recent studies have shown that size is the principal determinant of cost, and therefore an accurate size estimate is crucial to good cost estimation[10]. Improving the accuracy of size estimates is, therefore, instrumental in improving the accuracy of cost and schedule estimates. Moreover, software size and cost estimates have the highest utility at the time of project inception - when most important decisions (e.g. budget allocation, personnel allocation, etc). are taken. The dilemma, however, is that only high- level requirements for a project are available at this stage. Leveraging this high-level information to produce an accurate estimate of software size is an extremely challenging and high risk task.
    [Show full text]
  • Automating Software Quality Measurement with Standards
    Automating Software Quality Measurement with Standards Paul C. Bentz Amsterdam June 18, 2019 Why Automate? ©2019 CISQ 2 Complexity 1 Unit Level • Code style & layout • Expression complexity • Code documentation • Class or program design • Basic coding standards J • Developer level APIs JSP ASP.NET Java Java Java a 2 Technology Level Web v Services • Single language/technology layer • layer Architecture layer a Intra-technology architecture - • Intra-layer dependencies Hibernate Messaging • Inter-program invocation Struts .NET • Security vulnerabilities Spring • Development team level COBOL PL/SQL T/SQL EJB language, multi - 3 System Level SQL Server . Integration quality . Data access control Oracle Multi . Architectural compliance . SDK versioning DB2 . Risk propagation . Calibration across Sybase IMS . Application security technologies . Resiliency checks . IT organization level . Transaction integrity . Function point, . Effort estimation Technology Stack ©2019 CISQ 3 Velocity ©2019 CISQ 4 Automated Complex Toolchains • Production metrics, objects and feedback • Requirements • Design of the software and • Business metrics configuration • Update release metrics • Coding including code quality • Release plan, timing and business case and performance • Security policy and requirement • Software build and build performance • Infrastructure storage, • Release candidate database and network provisioning and configuring • Application provision and configuration • Acceptance testing • Regression testing • Security and vulnerability analysis
    [Show full text]
  • The Evolution of Software Size: a Search for Value©
    Software Engineering Technology The Evolution of Software Size: A Search for Value© Arlene F. Minkiewicz PRICE Systems, LLC Software size measurement continues to be a contentious issue in the software engineering community. This article reviews soft- ware sizing methodologies employed through the years, focusing on their uses and misuses. It covers the journey the software community has traversed in the quest for finding the right way to assign value to software solutions, highlighting the detours and missteps along the way. Readers will gain a fresh perspective on software size, what it really means, and what they can and cannot learn from history. hen I first started programming, it evolved throughout the last 25 years. It In the ’70s, the LOC measure never occurred to me to think focuses on reasons why these approaches seemed like a pretty good device. aboutW the size of the software I was were attempted, the technological or Programming languages were simple developing. This was true for several rea- human factors that were in play, and the and a fairly compelling argument could sons. First of all, when I first learned to degree of success achieved in the use of be made about the equivalence among program, software had a tactile quality each approach. Finally, it addresses some LOC. Besides, it was the only measure in through the deck of punched cards of the reasons why the software engineer- town. required to run a program. If I wanted to ing community is still searching for the In the late ’70s, RCA introduced the size the software, there was something I right way to measure software size.
    [Show full text]
  • Language-Independent Volume Measurement
    Language-independent volume measurement Edwin Ouwehand [email protected] Summer 2018, 52 pages Supervisor: Ana Oprescu Organisation supervisor: Lodewijk Bergmans Second reader: Ana Varbanescu Host organisation: Software Improvement Group, http://ww.sig.eu Universiteit van Amsterdam Faculteit der Natuurwetenschappen, Wiskunde en Informatica Master Software Engineering http://www.software-engineering-amsterdam.nl Contents Abstract 3 1 Introduction 4 1.1 Problem statement...................................... 5 1.2 Research Questions...................................... 5 1.3 Software Improvement Group................................ 6 1.4 Outline ............................................ 6 2 Background 7 2.1 Software Sizing........................................ 7 2.1.1 Function Points.................................... 7 2.1.2 Effort estimation................................... 8 2.2 Expressiveness of programming languages......................... 8 2.3 Kolmogorov Complexity................................... 8 2.3.1 Incomputability and Estimation .......................... 9 2.3.2 Applications ..................................... 9 2.4 Data compression....................................... 9 2.4.1 Compression Ratio.................................. 10 2.4.2 Lossless & Lossy................................... 10 2.4.3 Archives........................................ 10 3 Research Method 12 3.1 Methodology ......................................... 12 3.2 Data.............................................. 12 3.3 Counting
    [Show full text]
  • Quality Characteristics and Metrics for Reusable Software (Preliminary Report)
    Quality Characteristics and Metrics for Reusable Software (Preliminary Report) W. J. Salamon D. R. Wallace Prepared by the U.S. DEPARTMENT OF COMMERCE Technology Administration National Institute of Standards and Technology Gaithersburg, MD 20899 for the Department of Defense Ballistic Missile Defense Organization —Qe 100 .U56 NIST NO. 5459 1994 4 NISTIR 5459 Quality Characteristics and Metrics for Reusable Software (Preliminary Report) W. J. Salamon D. R. Wallace Prepared by the U.S. DEPARTMENT OF COMMERCE Technology Administration National Institute of Standards and Technology Gaithersburg, MD 20899 for the Department of Defense Ballistic Missile Defense Organization May 1994 U.S. DEPARTMENT OF COMMERCE Ronald H. Brown, Secretary TECHNOLOGY ADMINISTRATION Mary L Good, Under Secretary for Technology NATIONAL INSTITUTE OF STANDARDS AND TECHNOLOGY Arad Prabhakar, Director ABSTRACT This report identifies a set of quality characteristics of software and provides a summary of software metrics that are useful in measuring these quality characteristics for software products. The metrics are useful in assessing the reusability of software products. This report is preliminary. Additional research is needed to ensure the completeness of the quality characteristics and supporting metrics, and to provide guidance on using the metrics. KEYWORDS Reusable Software; Quality Characteristics; Software Reliability; Completeness; Correctness; Software Metrics ,fi> i^fi.'4»^lw(ti>bi ncwp>t «frfT' '^^m'^ki^'m m tBtis ir^m^t »mw^g ' ^ .. ^ W : IBIOL EXECUTIVE SUMMARY The Software Producibility Manufacturing Operations Development and Integration Laboratory (Software Producibility MODIL) was established at the National Institute of Standards and Technology (NIST) in 1992. The Software Producibility MODIL was one of four MODILs instituted at national laboratories by the U.S.
    [Show full text]
  • Software Development Cost Estimation Approaches – a Survey1
    Software Development Cost Estimation Approaches – A Survey1 Barry Boehm, Chris Abts University of Southern California Los Angeles, CA 90089-0781 Sunita Chulani IBM Research 650 Harry Road, San Jose, CA 95120 1 This paper is an extension of the work done by Sunita Chulani as part of her Qualifying Exam report in partial fulfillment of requirements of the Ph.D. program of the Computer Science department at USC [Chulani 1998]. Abstract This paper summarizes several classes of software cost estimation models and techniques: parametric models, expertise-based techniques, learning-oriented techniques, dynamics-based models, regression-based models, and composite-Bayesian techniques for integrating expertise- based and regression-based models. Experience to date indicates that neural-net and dynamics-based techniques are less mature than the other classes of techniques, but that all classes of techniques are challenged by the rapid pace of change in software technology. The primary conclusion is that no single technique is best for all situations, and that a careful comparison of the results of several approaches is most likely to produce realistic estimates. 1. Introduction Software engineering cost (and schedule) models and estimation techniques are used for a number of purposes. These include: § Budgeting: the primary but not the only important use. Accuracy of the overall estimate is the most desired capability. § Tradeoff and risk analysis: an important additional capability is to illuminate the cost and schedule sensitivities of software project decisions (scoping, staffing, tools, reuse, etc.). § Project planning and control: an important additional capability is to provide cost and schedule breakdowns by component, stage and activity.
    [Show full text]
  • Enhancement in Function Point Analysis
    International Journal of Software Engineering & Applications (IJSEA), Vol.3, No.6, November 2012 ENHANCEMENT IN FUNCTION POINT ANALYSIS Archana Srivastava1 , Dr. Syed Qamar Abbas2 , Dr.S.K.Singh3 1 Amity University, Lucknow, India [email protected], [email protected], 2Director, Ambalika Institute of Management & Technology , Lucknow, India 3Professor, Amity University, Lucknow, India ABSTRACT Early and accurate estimation of software size plays an important role in facilitating effort and cost estimation of software systems. One of the commonly used methodologies for software size estimation is Function Point Analysis (FPA). The purpose of Software size estimation and effort estimation techniques is to provide a useful measure of the software complexities, efforts, and costs involved in software development. Despite almost three decades of research on software estimation, the research community has yet not able to provide a reliable estimation model for End-User Development (EUD) environments. EUD essentially out-sources development effort to the end user. Hence one element of the size and effort is the additional design time expended in end-user programming. This paper discusses the concept end-user programming and enhancement of FPA by adding end-user programming as an additional General System Characteristic (GSC). KEYWORDS size estimation, end user programming, function point analysis, end user development 1. INTRODUCTION End-User Programming system aims to give some programmable system functionality to people who are not professional programmers. One of the most successful computer programs of all times is the spreadsheet applications. The primary reason for its success is that end users can program it without going into the background details of logic and programming.
    [Show full text]
  • COSMIC FFP Measurement Manual V2.2
    The COSMIC Functional Size Measurement Method Version 4.0.1 Guideline on Non-Functional & Project Requirements How to consider non-functional and project requirements in software project performance measurement, benchmarking and estimating Version 1. November 2015 Version control Date Reviewer(s) Modifications / Additions March 2014 v0.1 followed by 15 iterations November See Acknowledgements First version 1.0 published 2015 Copyright 2015. All Rights Reserved. The Common Software Measurement International Consortium (COSMIC). Permission to copy all or part of this material is granted provided that the copies are not made or distributed for commercial advantage and that the title of the publication, its version number, and its date are cited and notice is given that copying is by permission of the Common Software Measurement International Consortium (COSMIC). To copy otherwise requires specific permission. Acknowledgements Authors and reviewers who have contributed to the development of v1.0 (alphabetical order) Alain Abran, Diana Baklizky Carol Buttle École de Technologie Supérieure, TI Metricas Safety and Security Engineering Université du Québec Council Brazil Canada United Kingdom Jean-Marc Desharnais Peter Fagg Cigdem Gencel École de Technologie Supérieure, Pentad Ltd DEISER Université du Québec United Kingdom Spain Canada Barbara Kitchenham, Arlan Lesterhuis Roberto Meli Keele University Data Processing Organization United Kingdom The Netherlands Italy Dylan Ren Luca Santillo Hassan Soubra Measures Agile Metrics Ecole Supérieure
    [Show full text]
  • Software Estimation, Measurement, and Metrics Chapter 13: Software Estimation, Measurement & Metrics GSAM Version 3.0
    Part 3: Management GSAM Version 3.0 Chapter 13 Software Estimation, Measurement, and Metrics Chapter 13: Software Estimation, Measurement & Metrics GSAM Version 3.0 Contents 13.1 Chapter Overview .................................................................................. 13-4 13.2 Software Estimation ............................................................................... 13-4 13.2.1 Measuring Software Size................................................................... 13-6 13.2.2 Source Lines-of-Code Estimates ....................................................... 13-7 13.2.3 Function Point Size Estimates........................................................... 13-7 13.2.4 Feature Point Size Estimates.............................................................. 13-9 13.2.5 Cost and Schedule Estimation Methodologies/Techniques ............. 13-10 13.2.6 Ada-Specific Cost Estimation ......................................................... 13-12 13.3 Software Measurement ......................................................................... 13-13 13.3.1 Measures and Metrics, and Indicators ............................................. 13-13 13.3.2 Software Measurement Life Cycle................................................... 13-15 13.3.3 Software Measurement Life Cycle at Loral...................................... 13-16 13.4 Software Measurement Process........................................................... 13-18 13.4.1 Metrics Plan ..................................................................................
    [Show full text]
  • Software Cost Estimation and Sizing Methods: Issues and Guidelines
    THE ARTS This PDF document was made available CHILD POLICY from www.rand.org as a public service of CIVIL JUSTICE the RAND Corporation. EDUCATION ENERGY AND ENVIRONMENT Jump down to document6 HEALTH AND HEALTH CARE INTERNATIONAL AFFAIRS The RAND Corporation is a nonprofit NATIONAL SECURITY research organization providing POPULATION AND AGING PUBLIC SAFETY objective analysis and effective SCIENCE AND TECHNOLOGY solutions that address the challenges SUBSTANCE ABUSE facing the public and private sectors TERRORISM AND HOMELAND SECURITY around the world. TRANSPORTATION AND INFRASTRUCTURE WORKFORCE AND WORKPLACE Support RAND Purchase this document Browse Books & Publications Make a charitable contribution For More Information Visit RAND at www.rand.org Explore RAND Project AIR FORCE View document details Limited Electronic Distribution Rights This document and trademark(s) contained herein are protected by law as indicated in a notice appearing later in this work. This electronic representation of RAND intellectual property is provided for non- commercial use only. Permission is required from RAND to reproduce, or reuse in another form, any of our research documents. This product is part of the RAND Corporation monograph series. RAND monographs present major research findings that address the challenges facing the public and private sectors. All RAND mono- graphs undergo rigorous peer review to ensure high standards for research quality and objectivity. 4PGUXBSF $PTU&TUJNBUJPO BOE4J[JOH.FUIPET *TTVFTBOE(VJEFMJOFT 4IBSJ-BXSFODF1GMFFHFS 'FMJDJB8V 3PTBMJOE-FXJT 1SFQBSFEGPSUIF6OJUFE4UBUFT"JS'PSDF "QQSPWFEGPSQVCMJDSFMFBTFEJTUSJCVUJPOVOMJNJUFE 4HERESEARCHDESCRIBEDINTHISREPORTWASSPONSOREDBYTHE5NITED3TATES !IR&ORCEUNDER#ONTRACT& # &URTHERINFORMATIONMAY BEOBTAINEDFROMTHE3TRATEGIC0LANNING$IVISION $IRECTORATEOF0LANS (Q53!& ,IBRARYOF#ONGRESS#ATALOGING IN 0UBLICATION$ATA 0FLEEGER 3HARI,AWRENCE 3OFTWARECOSTESTIMATIONANDSIZINGMETHODSISSUESANDGUIDELINES 3HARI,AWRENCE0FLEEGER &ELICIA7U 2OSALIND,EWIS PCM h-' v )NCLUDESBIBLIOGRAPHICALREFERENCES )3".
    [Show full text]