Defense Cost and Resource Center The Defense Cost and Resource Center: An Update

Ronald Lile 40th DoDCAS February 2007 We need good data reporting!!

Acquisition Executives need to make decisions based on “actual” experience

2 DCARC Mission & Objectives

• Mission – To collect historical Major Defense Acquisition Program cost and software resource data and make those data available for use by authorized government analysts to estimate the cost of ongoing and future government programs, particularly DoD weapon systems. • Objectives – Make CSDR reporting as inexpensive and least disruptive as possible for contractors. – Provide wide availability of CSDR data to legitimate government users. – Maintain integrity and accuracy of data collected. – Improve quality of data reported by industry.

Primary Objective: Ensure that DoD cost estimates provided to senior management reflect as accurately as possible DoD’s experience 3 “Actual” Contractor Cost Data

• Provides basis of cost estimates of future systems – Long and painful DoD experience has shown that “actual” cost of producing systems are, by far, the best basis for cost estimates • To be useful for this purpose, definitionally consistent historical data are needed • CCDRs are DoD’s only systematic mechanism for capturing actual data

4 Software Resource Data

• Software is a cost driver on many systems and the DoD has little historical data upon which to base future estimates – There are no data to compare estimated software cost and size with actual program results; no records of software metrics – Tried and failed to match CARDs with CCDR data – Software data routinely reported as “red” at annual DoD Cost Analysis Symposium

5 DCARC Goal

• Collect and make available data that are: – Comparable (i.e., across programs) – Meaningful (i.e., level of detail that provides insight) – Transparent (i.e., well defined) – Accurate – Auditable

6 So…where are we?

7 DoD 5000 Compliance

8 ACAT IC/ID CSDR Status Post-Dropping of Advisory Ratings

Active Programs w/CAIG Active Programs w/CAIG Active Programs w/CAIG Approved CSDR Plans & Approved CSDR Plans & Approved CSDR Plans & Pre-MDAP Programs Not Rated No Issues Minor Issues1 Major Issues3

AARGM AMRAAM ADS 21" MRUUVS AWACS Upgrade AEHF** B-2 RMP C-130 AMP** ACS C-5 AMP AIM-9X CEC CH-47F AIAMD ERM ARH CH-53K COBRA JUDY APS FAB-T ATIRCM/CMWS DDG 1000 (DD(X)) CVN-21 AR/LSB JSTARS BRADLEY UPGD** E-2C REPROD EELV BMTC MEADS** C-130J F/A-18 E/F** F-22 CG(X) RMS C-17A F-35 GLOBAL HAWK** DCGS-A SIAP C-5 RERP FBCB2 GMLRS DCGS-AF VTUAV Fire E/A-18G** FCS H-1 Upgrades** JNN No Action E-2D ADV Hawkeye JLENS JTRS GMR (CL 1) LBSD Programs Waived EFV LPD-17** JTRS HMS (CL 5) NGLRS EXCALIBUR MPS JTRS Joint Waveform ODSI CHEM DEMIL FMTV NAVSTAR GPS** LAND WARRIOR SR HPCM GBS NPOESS LB APACHE BLK III* STS CONNECTOR JPATS HIMARS SSGN** LCS** MIDS LVT JASSM V-22 LHA Replacement AOC-WS NAS JDAM (EMD) WIN-T MH-60R E-10 SSDS JSOW MIDS JTRS ECSS MH-60S MM III PRP JTRS AMF Programs Grandfathered MM III GRP MP RTIP MUOS NAVSTAR GPS UE B-2 EHF CVN77 P-8 (MMA) SM 2 (IV) BAMS DDG51 PATRIOT PAC-3 SSN 774 CSAR-X BMDS SBIRS HIGH T-AKE JCA SDB I JHSV SM-6 (ERAM) JPALS Proper CSDR Planning STRYKER JUCAS TAC TOMAHAWK Early Problems KC-X UH-60M Upgrade LUH VH-71 (VXX) MPF(F) WGS As of January 10, 2007 NAVSTAR GPS III 1- Or Agreed Upon Path Forward NMT 2- Includes Data Issues or Plan Implementation Issues SDB II 3- Includes No Progress on Data Issues or Plan TSAT Implementation Issues *-Initial Assessment **-Withhold program on a get well plan 32 18 25 33 18

TOTAL PROGRAMS 126 9 Analytical Activities • CSDR Plans – Approved Plans CY2002 CY2003 CY2004 CY2005 CY2006 Total Plans IC ID Total IC ID Total IC ID Total IC ID Total IC ID Total IC ID Total Program4263131615173232457925336379110196 Contract 11 1 12 12 18 30 34 27 61 71 97 175 58 118 189 186 261 467 Subcontract00003355762361431852710012968303379 Total 15 3 18 15 34 49 54 101 155 139 285 439 110 251 381 333 674 1,042

– Plans in progress Plans in Progress Plans IC ID Total Program 1 4 5 Contract 10 18 29 Subcontract 5 13 20 Total 16 35 54

As of December 31, 2006 10 CY 2006 CSDR Plan Statistics Distribution of Plans Approved in CY06

250 Program Plans Approved

203 200 Contract Plans Approved Subcontract Plans Approved

150 129

100 79 64 65 62 58

50 37 28 24 23 17 1 4 0 Army Navy Air Force DoD Total 11 CY 2006 CSDR Plan Statistics- Cont’d

CY2006 CSDR Plans Approved

More than 50% of CY2006 Plans were 250 revisions

200

93 150 Revised New 100 67

110 50 44 62 21 0 Program Plans Contract Plans Subcontract Plans 12 Analysis of CY2006 CSDR Plans

NEW 1st Revision(R) 2nd Revision(R1) 3rd Revision(R2) 4th Revision(R3) Count 192 135(70%) 62(32%) 12(6%) 4(2%) Air Force 69 36 15 5 Army 49 57 33 7 4 Navy 71 33 14 DoD 39

Aircraft 62 41 24 1 Air Force 18 11 Army 26 7 An Example CSDR Plan Revisions Navy 17 14 DoD 1 9 Missile 11 18 2 4 3 Army 18 Ordnance 45 1 Revisions Air Force 2 2 1st 2nd 3rd 4th Navy 2 3 M issile 18 Electronics 43 14 5 1 1 Army 18 Air Force 7 3 Javelin Jt Venture, Orlando JAVELIN Army 10 9 Raytheon, Tucson JAVELIN Navy 24 2 LMM&FC, Orlando JAVELIN DoD 2 LMM&FC, Dallas PAC-3, HIMARS GMLRS, HIMARS Space 36 20 5 3 Aerojet GMLRS Air Force 36 18 General Dynamics GMLRS Navy 2 CAUSALITY: Surface Vehicle 6143 - AMCOM Army 12 - EDUCATION & TRAINING Navy 6 2 - DEVIATIONS IN CWBS FROM MIL-HDBK-8881A MISSILE SYSTEM WBS Ship 15 10 2 - MULTIPLE PHASES PER CONTRACT - MULTIPLE VARIANTS PER CONTRACT Navy 15 10 - IMPROPER RFP PACKAGES UAV 11 2 1 - LACK OF PRIME-SUBCONTRACTOR-GOV'T PLANNING & POST-AWARD Air Force 4 2 EXECUTION Navy 7 - WRONG WBSs in RFPs Sys of Sys 411202 - LACK OF PRIME OVERSIGHT ON SUBCONTRACTOR REPORTING- Army 11 13 Why so many Plan Revisions?

• Systemic Problems- What we are doing wrong – Incorrect, Incomplete & Non-Compliant WBS in RFP (e.g. NLOS-LS) – Lack of CAIG-Approved CSDR Plan in RFP (JTRS SRW 1 day notice before contract award) – Program Manager’s & Contracting Officers uninformed/don’t care about CSDR Requirements (e.g. MDA- SM-3) – Prime-Subcontractor Flow-Down Requirement Deficiencies (e.g. failed to flow down requirement to on RF Missile Seeker- PAC-3) – Lack of Discipline & Management Oversight by Gov’t CWIPT leads (F-22 10 Years of neglect lead to Resubmissions by Prime & Subs) – Acquisition Reform Policies (e.g. AMRAAM, JASSM retroactive reporting) • Program Unique- Normal Reasons for Revisions – Restructures – Contract Type Changes – Make/Buy Changes – Updated based on Contractor Responses

14 Analytical Activities

• Validate CSDR Submissions – CCDR Metrics: CY2002 CY2003 CY2004 CY2005 CY2006 Total CCDR Reports Received 164 247 387 1214 1306 3318 CCDR Reports Validated 119 145 332 1239 1407 3242 CCDR Reports Accepted 25 16 101 585 860 1587 CCDR Reports Rejected 94 129 231 654 526 1634 Validation Rate/Yr 73% 59% 86% 100% 100% 98% Acception Rate/Yr 21% 11% 30% 47% 61% 49% – SRDR Metrics: CY2002 CY2003 CY2004 CY2005 CY2006 Total SRDR Reports Received 0 0 9 66 99 174 SRDR Reports Validated 0000111111 SRDR Reports Accepted 00002020 SRDR Reports Rejected 00009191 Validation Rate/Yr N/A N/A N/A 0% 100% 64% Acception Rate/Yr N/A N/A N/A 0% 18% 18%

As of December 31, 2006 15 CSDR Training Statistics

CSDR Training Metrics

600 > 1200 Students Trained 500

400 334 Industry 300 235 Gov't 143 # of Students of # 200

100 197 120 124 27 42 0 CY 2004 CY2005 CY2006 CY2007 16 DACIMS Holdings

17 Reports by Commodity

Mil Handbook Number of Reports Aircraft 13635 Electronic/Automated Software 3507 Missile 9607 Ordnance 517 Ship 222 Space 1140 Surface Vehicle 708 System of System 9 UAV 95 Total 29914

18 Top Aircraft Programs

Program Number of Reports F-14 772 F/A-18 726 A-4 688 EA-6 666 A-6 590 E-2 (Hawkeye) 583 P-3 538 A-7 517 UH-60 489 AH-64 455

19 Current Aircraft Programs

Program Number of Reports F/A-18 726 AH-64 455 CH-47 301 F-22 279 C-17 151 V-22 140 F-35 100 EA-18 50 MH-60 50 H-1 Upgrades 45 E-2 (Advanced Hawkeye) 26 VH-71 (Presidential Helicopter) 20 P-8 (MMA) 8 CH-53K (HLR) 5 RAH-70 (ARH) 3

20 Top Missile Programs

Program Number of Reports AIM-9 (Sidewinder) 1057 AIM-7; RIM-7 (Sparrow; Sea Sparrow) 973 AIM-54 (Phoenix) 707 LGM-118 (Peacekeeper) 597 AGM-84; RGM-84; UGM-84 (Harpoon) 548 AGM-114 (Hellfire) 540 FIM-92 (Stinger Basic) 441 MIM-72 (Chaparral) 434 UGM-96 (Trident I (C-4)) 376 MIM-104 (Patriot) 335

21 Current Missile Programs

Program Number of Reports AIM-9 (Sidewinder) 1057 AIM-120 (AMRAAM) 292 LGM-30 (Minuteman III) 71 GMLRS (Guided Multiple Launch Rocket System) 64 AGM-154 (Joint Standoff Weapon (JSOW)) 55 MIM-104D (Patriot PAC-3) 26 MGM-164 (ATACMS BLOCK II) 24 NLOS-LS (Non-Line-of-Sight Launch System) 14 Extended Range Active Missile (Standard Missile-6) 10 RIM-66 (Standard Missile-2 (SM-2 MR )) 10 AGM-158 (Joint Air-To-Surface Standoff Missile (JASSM)) 8 RIM-156 (Standard Missile 2 (SM-2 ER)) 7 Joint Common Missile 6 R/UGM-109 (Tactical Tomahawk) 5 RIM-67 (Standard Missile 2 (SM-2)) 2 HIMARS (High Mobility Artillery Rocket System) 1

22 Top E/ASS Programs

Program Number of Reports F-14 (Tomcat) 457 AH-64 (Apache) 279 B-1 (Lancer) 209 E-8 (Joint STARS) 185 Future Combat Systems (FCS) 153 F-16 (Fighting Falcon) 145 FBCB2 (FORCE XXI Battle Command Brigade & Below) 129 PLRS/JTIDS (Position Locating Reporting System/Joint Tactical Information Distribution 127 MIDS-LVT (Multifunctional Information Distribution System-Low Volume Terminal) 115 F/A-18 (Hornet) 106

23 Current E/ASS Programs

Program Number of Reports AH-64 (Apache) 279 E-8 (Joint STARS) 185 Future Combat Systems (FCS) 153 FBCB2 (FORCE XXI Battle Command Brigade & Below) 129 MIDS-LVT (Multifunctional Information Distribution System-Low Volume Terminal) 115 CEC (Cooperative Engagement Capability) 63 Warfighter Information Network-Tactical (WIN-T) 62 F-35 (Joint Strike Fighter) 41 P-8 (MMA) 33 E-2 (Advanced Hawkeye) 27 F-22 (Raptor) 24 SSN-21 (Seawolf Submarine Combat System) 24 V-22 (Osprey) 22 JTRS Cluster 1 (Joint Tactical Radio System) 16 UH-60 (Black Hawk) 13 MP-RTIP (Multi-Platform Radar Technology Insertion Program) 12 Land Warrior 11 Cobra Judy Replacement (CJR) 9 GBS () 6 Electromagnetic Aircraft Launch System (EMALS) 4

24 Top Space Programs

Program Number of Reports LS-3 ( (DSP)) 191 SLV-5 ( III) 145 NS-7 (Global Positioning System (NAVSTAR GPS)) 140 ES-8 () 76 Transformational Satellite Communications System (TSAT) 69 WS-1 (Defense Meteorological Satellite Program (DMSP) 59 Advanced Extremely High Frequency (AEHF) Satellite 55 SB-5 (Titan IV) 49 SSB-7 ( (IUS)) 47 SB-3 (Delta II) 42

25 Current Space Programs

Program Number of Reports NS-7 (Global Positioning System (NAVSTAR GPS)) 140 Transformational Satellite Communications System (TSAT) 69 Advanced Extremely High Frequency (AEHF) Satellite 55 LS-3 (SBIRS-HIGH) 38 National Polar-orbiting Operational Environmental Satellite System (NPOESS) 30 Mobile User Objective System (MUOS) 27 EELV (Evolved Expendable Launch Vehicle) 7 ABL (Airborne Laser) 5

26 Surface Vehicle Programs

Program Number of Reports Expeditionary Fighting Vehicle (EFV) 204 M2, M3 (Bradley Fighting Vehicle System (BFVS)/Upgrade) 194 M270 (Multiple Launch Rocket System (MLRS)) 90 M1126-M1135 (Stryker) 75 FMTV (Family of Medium Tactical Vehicles) 42 M1074; M1075 (Palletized Loading System (PLS)) 33 Future Combat Systems (FCS) 21 M1 (Abrams) 15 M6 (Bradley Linebacker) 12 HIMARS (High Mobility Artillery Rocket System) 11

27 Sea System Programs

Program Number of Reports Littoral Combat Ship (LCS) 61 DDG 1000 (Zumwalt Class) 46 T-AKE (ADC(X) Auxiliary Dry Cargo Carrier) 26 LPD-17 (San Antonio Class) 19 FFG 7 (Guided Missile Frigate) 17 CVN-21 12 SSGN (Trident) 11 SSN 21 (Seawolf Class) 6 SSBN 726 (Ohio Class Trident) 4

28 SRDR Holdings

Number of Number of Total Number of Available 2630-2 Available 2630-3 2630 Program Submissions Submissions Submissions B-2 RMP 1 1 C-130J 1 1 CEC 3 1 4 Cobra Judy Replacement 1 1 DDG 1000 1 1 EA-18G 2 2 EFV 314 FBCB2 1 2 3 FCS 20 11 31 H-1 Upgrades 1 1 ISPAN 2 2 4 JSF 3 3 LCS 1 1 MP-RTIP 2 2 THAAD 2 3 5 TSAT 1 1 WIN-T 4 1 5 Total 49 21 70

29 IT Systems

30 Public Web Site Today

IT Systems • http://dcarc.pae.osd.mil Launch Pad

31 Today: User Registration • Access of DCARC Systems requires a DCARC issued certificate. • Register to – Obtain a certificate – Request account to access to DACIMS – Request account to access to eRoom • Limitations – Access to DACIMS is restricted to Govt only. – Access to eRoom areas is by MDAP. – CSDR Submission Only Requires Certificate.

32 Today: User Registration • New User Registration

After you register, DCARC IT Support generates a digital certificate for you. Come back to this page to retrieve it

33 Today: User Registration • Step 1

Request Roles: Submitter, DACIMS Access, or eRoom Access

34 Today: User Registration • Step 2

35 Today: User Registration • Step 3

36 CSDR and EVM Reporting: Current Status

Defense Acquisition Board EVM Metrics CSDR Summary level only EVM

Cost Estimates & OSD/(AT&L) & Analysis Joint Staff

Takes 3-6 months** post “as of date”, summary level only

**- For the Army DoD Cost SERVICES it is 1-2 months Community post “as of date” PMs & CARS/DAMIR CSDR EVM Data: Data Other DoD Request Organizations Manually entered, Summary level only Request CSDR Data Service Joint Program Data ready Program DCARC 15 days after receipt Office Office 75 days post “as of date” Industry Event-driven CSDR, @ PDF 60 days post “as of date”, Word Monthly EVM, detailed WBS Paper 12- 17 days post EDI “as of date”, @ Excel Excel detailed WBS 37 Today: DACIMS 3.0a

38 Today: DACIMS 3.0a Search Button

Weapon System Type (based on military designation)

Program Name

Model/Subsystem

MIL-HDBK Category Organized around Mil Handbook, Type, Program, and Model

39 Today: DACIMS 3.0a • Sample Search: F-22 Info – Opens page with metadata and attachment list

Open – Opens cost report in new window

All “file cabinets” are accepted or historical reports and should contain DCARC memo

Download – Opens box with option to open/save (this will be the fastest way to download data in bulk)

40 Today: DACIMS 3.0a • Data Record Types – Folder Cabinets: typical for newer records to keep the submission components together – Single files: Icon (A) = Accepted identifies file time; (R) = Rejected (P) = Pending title describes (H) = Historical content. Legacy TIF format

41 Today: DACIMS 3.0a • Sample Data Record Detail (part 1)

42 Today: DACIMS 3.0a • Sample Data Record Detail (part 1)

43 Coming Soon…

44 DoD 5000 Compliance

45 Policy Activities

• Submitted for approval – Revised CSDR Manual – Revised CCDR Forms and DIDs – SRDR DIDs • Expect – Approval by March 1, 2007 – Effective April 2, 2007

46 DoD 5000.04 M-1, CSDR Manual

• Transferred to the DCARC public web site: – CSDR Plan (DD2794) format and instructions – Contracting language (RFP & CDRL examples) – SRDR Sample Format 1, Government Report • Streamlined the Manual – Deleted some background & introductory info – Eliminated duplicative material, – Combined Chapters 4, 5, and 6 into one Chapter (4), Processing and Reporting

47 DoD 5000.04 M-1, CSDR Manual

• Added pre-contract or post contract award conference • Revised recurring and nonrecurring definitions • Clarified reporting for undefinitized contractual actions • Updated and added more detail to the CSDR Process Model in Chapter 4

48 CCDR Forms and DIDs

• Changed Progress Curve Report to its own form and DID • Specified cost interrelationships among the three reports • Revised recurring and nonrecurring definitions • Changed the Manufacturing functional category to Manufacturing Operations • Revised definitions of Tooling and Material (including subclasses) • Added definition and reporting instructions for inter- company transfers • Revised instructions for quantity to date • Clarified reporting requirements for Initial, Interim, & Final • Provided for submission of reports in XML format in addition to the Excel compatible formats

49 SRDR DIDs

• Disestablished DD Form 2630-2 & -3 • Added sample formats • Removed sample data dictionary • Provided guidance on minimum acceptable set of information required in the SRDR data dictionary • Added meta data fields to improve the clarity and context of the data – Software functional descriptions – Characterization of software development effort – Software development method(s) used – Cross-references to contract DD 2794 & contract WBS

50 SRDR DIDs

• Added required data fields – to improve reported software size data • Autogenerated code, • ‘Carryover’ code from previous builds, and • Percent re-design, re-code, re-test (to be explained in the data dictionary) – to clarify reporting of sub-contractor developed software code and sub-contractor software development effort • Added optional data fields for: – COTS/GOTS SW integration efforts – Standardized software sizing counts via a code counting tool directed by the Government

51 DACIMS Holdings

52 Future Submissions

Type 2007 2008 2009 2010 2011 2012 2013 2014 2015 2016 2017 Total CCDR Reports 466 315 272 225 135 93 44 19 26 5 4 1604 SRDR Reports 174 88 64 61 53 27 12 6 14 2 1 502 Total 640 403 336 286 188 120 56 25 40 7 5 2106

Commodity CCDR SRDR Total Aircraft 604 67 671 Electronic/Automated Software 192 116 308 Missile 145 4 149 Ordnance 36 2 38 Ship 215 36 251 Space 171 103 274 Surface Vehicle 61 2 63 System of Systems 137 162 299 UAV 43 10 53 Grand Total 1604 502 2106

53 Future Aircraft Submissions

Program 2007 2008 2009 2010 2011 2012 2013 2014 2015 Total AB3 96612 24 ARH 4 5 1 10 C-5 RERP 1 1 1 3 CH-47F761111 17 CH-53K5 24411 320 E-2C REPRO 3 2 2 7 E-2D 24 16 8 8 8 1 2 67 EA-18G 10 5 6 21 F-22 28 14 6 6 6 2 62 FA-18EF 10 7 10 9 9 6 2 53 H-1 Upgrades 8 3 1 11 23 JSF 262626261715184 158 Longbow Apache 3 1 2 3 9 MH-60R9775422 36 MH-60S423232 16 MMA 6511 5 18 T - 45TS 4 3 4 11 UH-60M 18 6 4 9 2 2 2 43 V-22 12 6 12 9 9 48 VH-71 5 5 10 5 25 Total 195 125 107 102 62 46 27 4 3 671

54 Future E/ASS Submissions

Program 2007 2008 2009 2010 2011 2012 2013 2014 2015 2016 2017 Total ADS 3 3 6 AESA 22 4 AOC-WS3131442321 ATIRCM/CMWS333322 16 B-2 EHF 111 3 B-2 RMP 4 2 4 4 2 16 CEC 4212 9 FBCB2 823 1 14 JLENS 67711612 49 JTRS AMF 3 1 4 JTRS GMR 12 12 JTRS HMS 6 4 10 JTRS SRW2422 10 MIDS - JTRS 2 2 MP-RTIP 7 6 6 19 MPS 2723181815531111113 Total 93594048282471341308

55 Future Space Submissions

Program 2007 2008 2009 2010 2011 2012 2013 2014 2015 2016 Total AEHF 33331 13 EELV 2 2 FAB-T 12 4 8 4 16 44 GPS 2819121 60 MUOS 97433 26 NMT 8 16 4 4 32 NPOESS 6 6 6 12 12 42 SBIRS HI 4 4 4 12 TSAT 1422231233234 WGS 21111111 9 Total 885844343623432274

56 Future Ship Submissions

Program 2007 2008 2009 2010 2011 2012 2013 2014 2015 2016 Total CJR 82433 20 CVN 21 3 3 2 8 DDG 1000 28 26 25 34 2 2 117 LCS 1058111111130 LHA(R) 1 1 1 1 4 LPD 1724 244 16 RMS 13 2 3 18 SSGN 3 4 7 SSN 7743 333336 24 T-AKE 3 2 1 1 7 Total 7449444413114 4 7 1251

57 Future Missile Submissions

Program 2007 2008 2009 2010 2011 2012 2013 2014 Total AARGM 3 1 4 AIM-9X 2 1 1 4 AMRAAM 1111116 GMLRS 15 5 4 24 JASSM 411111 9 JSOW 136655 35 MM III GRP 2 2 2 2 8 MM III PRP 4 3 6 13 NLOS 6 6 12 PAC-3 86222 20 SM-2 1111 4 SM-6 1112 5 Tactical Tomahawk 2 1 1 1 5 Total 553327219 2 1 1149

58 Future SV & UAV Submissions

Program 2007 2008 2009 2010 Total Bradley Upgrade 5 5 10 EFV 11510 26 FMTV 113 5 HIMARS 3 1 1 5 Stryker 6 7 4 17 Total 2119185 63

Program200720082009201020112012Total Global Hawk 11 13 2 4 2 32 NUCAS 21 1116 VTUAV 753 15 Total 2019553153

59 IT Systems

60 cPET 1.0 beta

• CSDR Planning & Execution Tool (cPET) • Major functions: – Implemented against Draft XML Schema. – Development of Program / Contract plans • Generates form DD2794 based on the selected program or contract. • Create Responsibility Assignment Matrix. – Creation and validation of 1921 and 1921-1 reports.

61 cPET 1.0 beta

• Basic Information

62 cPET 1.0 beta

• CSDR Planning Major Components

63 cPET 1.0 beta

• Creating the WBS, defaults to Mil Handbook

64 cPET 1.0 beta

• Block 15: Submission Events

65 cPET 1.0 beta

• Create DD2794

66 cPET 1.0 beta

• Responsibility Assignment Matrix

67 cPET 1.0 beta

• Validation Example

68 cPET 1.0 beta

• Validation Example

69 cPET 1.0 beta

• Validation Example

70 CSDR and EVM Reporting: Improved Model

Decisions required: 1. Location of repository Defense Acquisition Board 2. Single format and submission path 3. Business rules regarding data availability EVM Metrics Cost Estimates & Analysis OSD/(AT&L) & DoD Cost Joint Staff Community Data Request SERVICES

PMs & Other DoD Request Organizations DAMIR Data

Summary level EVM Data via Purview Data ready Request 15 days after receipt 75 days post “as of date” Detailed EVM Data ??????? via Virtual Library EVM Data available to PMs Immediately Event-driven CSDR, upon receipt, 12-17 60 days post “as of date”, Industry days post “as of date detailed WBS EVM Data @ Available via preset Monthly EVM, business rules 12- 17 days post @ Excel “as of date”, detailed WBS Format 71 DACIMS 4.0 beta

• Limited beta test to begin Mar 2007

Organized around user functions – Search – CSDR – FPR – DCARC/OSD Review Area – Integrated Web-Upload Function

NOTES: – Will support both certificates and CAC – Submitters will need certificate AND username/password

72 DACIMS 4.0 beta

• Sample CSDR Search

73 DACIMS 4.0 beta

• Sample CSDR Search Results (contract list)

Number of document Organized by contract meeting search criteria

74 DACIMS 4.0 beta

• Sample CSDR Search Results (document list)

75 EVM Central Repository Pilot DACIMS 4.0 Interface, Organized by User Role

76 EVM Central Repository Pilot

• Reviewer Pending Activity

77 EVM Central Repository Pilot

• Document Review Process: Document Detail

78 EVM Central Repository Pilot

• Document Review Process

During review process, PM controls access to submission data. Can invite submitting contractor and support contractors to participate in review and provided comments.

79 Summary

• Actual Experience is the best predictor • DCARC – Tasked with collecting “actual” cost and software data – Activities designed to • Make CSDR reporting as inexpensive and least disruptive as possible for contractors • Provide wide availability of CSDR data to legitimate government users • Maintain integrity and accuracy of data collected • Improve quality of data reported by industry • Need proactive participation by CWIPT members – Leads to collecting the right type of data

80