Familiar Metric Management - The Effort-Time Tradeoff: It s in the Data

Similar documents
Processes for the Intersection

SIDRA INTERSECTION 8.0 UPDATE HISTORY

Estimating. Proportions with Confidence. Chapter 10. Copyright 2006 Brooks/Cole, a division of Thomson Learning, Inc.

PQ-Box 100 Quick Start Instructions

Real-Time Systems Dr. Rajib Mall Department of Computer Science and Engineering Indian Institute of Technology, Kharagpur

Distribution of Data and the Empirical Rule

Considerations for Blending LED Phosphors

Applications. Media Cost x 1,000 Impressions. Avg. Unit Cost Rating (000) CPM = Avg. Unit Cost Rating % Total Schedule Cost CPP = GRPs

China s Overwhelming Contribution to Scientific Publications

Pitch correction on the human voice

CALIFORNIA GERMANY TRAVEL TRADE BAROMETER

Presented at the IPS 2004 Fulldome Standards Summit, Valencia, Spain, 7/8 July 2004 R.S.A. COSMOS

Processor time 9 Used memory 9. Lost video frames 11 Storage buffer 11 Received rate 11

Skip Length and Inter-Starvation Distance as a Combined Metric to Assess the Quality of Transmitted Video

How to Manage Video Frame- Processing Time Deviations in ASIC and SOC Video Processors

Project Summary EPRI Program 1: Power Quality

Computer Coordination With Popular Music: A New Research Agenda 1

NRDC Follow-up Comments to the 12/15/08 CEC Hearing on TV Efficiency Standards

PCM-16 Phase Synchronization Controller Operators Manual

International Workshop, Electrical Enduse Efficiency, 5th March Residential electricity consumption

Relationships Between Quantitative Variables

technical note flicker measurement display & lighting measurement

Station web sites are nearly universal, but we found a TV station -- a Fox affiliate in the Midwest -- that said no, it didn't have one.

Precision testing methods of Event Timer A032-ET

ECG SIGNAL COMPRESSION BASED ON FRACTALS AND RLE

Affected Products: Product Line Category Device Version Machinery Health Management. Data Analysis

Manual Supplement. This supplement contains information necessary to ensure the accuracy of the above manual.

Set-Top-Box Pilot and Market Assessment

GROWING VOICE COMPETITION SPOTLIGHTS URGENCY OF IP TRANSITION By Patrick Brogan, Vice President of Industry Analysis

Frame Processing Time Deviations in Video Processors

DIFFERENTIATE SOMETHING AT THE VERY BEGINNING THE COURSE I'LL ADD YOU QUESTIONS USING THEM. BUT PARTICULAR QUESTIONS AS YOU'LL SEE

Analysis of local and global timing and pitch change in ordinary

Relationships. Between Quantitative Variables. Chapter 5. Copyright 2006 Brooks/Cole, a division of Thomson Learning, Inc.

HEBS: Histogram Equalization for Backlight Scaling

Margin of Error. p(1 p) n 0.2(0.8) 900. Since about 95% of the data will fall within almost two standard deviations, we will use the formula

Automatic Commercial Monitoring for TV Broadcasting Using Audio Fingerprinting

Preface 11 Key Concept 1: Know your machine from a programmer s viewpoint 17

Chapter 27. Inferences for Regression. Remembering Regression. An Example: Body Fat and Waist Size. Remembering Regression (cont.)

COLOUR CHANGING USB LAMP KIT

DEPARTMENT OF THE ARMY TECHNICAL BULLETIN CALIBRATION PROCEDURE FOR TELETYPEWRITER TEST SET TS-800/UGM-l

Analyzing Numerical Data: Using Ratios I.B Student Activity Sheet 4: Ratios in the Media

Audio Metering Measurements, Standards, and Practice (2 nd Edition) Eddy Bøgh Brixen

Blueline, Linefree, Accuracy Ratio, & Moving Absolute Mean Ratio Charts

> f. > œœœœ >œ œ œ œ œ œ œ

Just a T.A.D. (Traffic Analysis Drone)

2018 RTDNA/Hofstra University Newsroom Survey

AT&T Investor Update. 2Q08 Earnings Conference Call July 23, 2008

A variable bandwidth broadcasting protocol for video-on-demand

Using Digital Fault Recorders As Phasor Measurement Unit Devices

2012 Television Pilot Production Report

Chapter 40: MIDI Tool

(Refer Slide Time 1:58)

Analysis of MPEG-2 Video Streams

New Spill Structure Analysis Tools for the VME Based Data Acquisition System ABLASS at GSI

MITOCW max_min_second_der_512kb-mp4

6.UAP Project. FunPlayer: A Real-Time Speed-Adjusting Music Accompaniment System. Daryl Neubieser. May 12, 2016

Application Note 11 - Totalization

John Stankey President and CEO AT&T Operations

45 MW, 22.8 GHz Second-Harmonic Multiplier for High-Gradient Tests*

Rotary Knife Controller

Practice, Practice, Practice Using Prototek Digital Receivers

Video Industry Making Significant Progress on Path to 4K/UHD

4.9 BEAM BLANKING AND PULSING OPTIONS

CNN American Morning March 12, 2010, 7:30 AM CT

The Measurement Tools and What They Do

Performance. Performance

Practicum 3, Fall 2010

Draft revised Energy Label and Ecodesign regulations for displays: Comments by Topten for the CF meeting on December 10 th 2014

NPR Weekend Programs

Connected TV Definitions. A new set of terms for a new type of channel

MATH 214 (NOTES) Math 214 Al Nosedal. Department of Mathematics Indiana University of Pennsylvania. MATH 214 (NOTES) p. 1/3

What is Statistics? 13.1 What is Statistics? Statistics

Fill out the following table: Solid #1 Solid #2 Volume. Number of Peanuts. Ratio

SDS PODCAST EPISODE 96 FIVE MINUTE FRIDAY: THE BAYES THEOREM

A QUANTITATIVE STUDY OF CATALOG USE

Scientific paper writing - Abstract and Extended abstract

Characterization and improvement of unpatterned wafer defect review on SEMs

Sunday Maximum All TV News Big Four Average Saturday

Mosaic 1.1 Progress Report April, 2010

Table of Contents INTRODUCTION 2. SECTION 1: Executive Summary 3-6. SECTION 2: Where do people get news and how?..7-11

Lesson 25: Solving Problems in Two Ways Rates and Algebra

Lecture 1: What we hear when we hear music

Measurement of overtone frequencies of a toy piano and perception of its pitch

Welcome Screen where you will Create an Account, and Log-In to InspectNet thereafter 1

Digital Signal Processing Laboratory 7: IIR Notch Filters Using the TMS320C6711

TROUBLESHOOTING DIGITALLY MODULATED SIGNALS, PART 2 By RON HRANAC

introduction body of the essay conclusion

RF (Wireless) Fundamentals 1- Day Seminar

Iva Petrinska, Dilyan Ivanov, Valchan Georgiev TU Sofia, Bulgaria. September 25, 2017

SPEED DRILL WARM-UP ACTIVITY

Chapter 14 D-A and A-D Conversion

Sources of Error in Time Interval Measurements

R&S FSW-B512R Real-Time Spectrum Analyzer 512 MHz Specifications

White Paper : Achieving synthetic slow-motion in UHDTV. InSync Technology Ltd, UK

Math and Music Developed by Megan Martinez and Alex Barnett in conjunction with Ilene Kanoff

WEB APPENDIX. Managing Innovation Sequences Over Iterated Offerings: Developing and Testing a Relative Innovation, Comfort, and Stimulation

Promo Mojo: Season Eight of 'The Walking Dead' Debuts

Color measurement and calibration of professional display devices

The National Traffic Signal Report Card: Highlights

Optical Engine Reference Design for DLP3010 Digital Micromirror Device

Transcription:

Familiar Metric Management - The Effort-Time Tradeoff: It s in the Data Lawrence H. Putnam Ware Myers Watch me make this baby jump, Ol Leadfoot said, as the light on the boulevard turned green. He pressed down hard on the gas pedal. The convertible jerked forward. The tires screamed. The driver in the curb lane cautiously waited till we were far ahead before he proceeded slowly after us. The motorcycle policeman on the cross street looked up. Ol Leadfoot made it to the next traffic light, still red, a good two blocks ahead of the car that had been beside him at the last light. The motorcycle cop was still a block behind us. He had not had enough time to catch our speed. The companion car reached us just as the light turned green. It inched cautiously past us, pulling over to the curb side of its lane. The cop hung back. Ol Leadfoot pressed the accelerator. Better take it easy, I said. Cop just behind us. We have all known Ol Leadfoot, particularly in our youth. There is no doubt about it. The power of a jackrabbit start makes the blood tingle. As we begin to pay our own bills, we learn that a screeching start burns up gasoline, several times as much as a more sedate start. Some cars provide a dashboard readout that shows instantaneous fuel usage. A jackrabbit start throws the needle over to maximum burnup, minimum economy. We need that software now In creative work, too, the jackrabbit approach throws the effort needle over to maximum. Ready or not, we need that movie now, the Los Angeles Times headlined. The studio philosophy is that if you throw enough people and money at a movie, you can pull it together fast. But there s a big difference between hiring 20 people for 0 days and 0 people for 20 days... Making a movie is not like assembling a machine you need time to reflect. Studios are becoming aware, however, that no one wins if the movie doesn t work, the Los Angeles Times continued. People are realizing that if the picture is good, it doesn t matter when it is released and if it is not good, it also doesn t matter. In spite of much history of the folly of rushing creative work, movies still reach the theaters in the state that Barry Diller, former Fox chairman, characterized as There s a good movie somewhere in there.

Like movies, software development is creative and design-intensive. Rushing it burns up resources and saves you little time. Everybody who ever floored a gas pedal in his teens (or made a movie a few years later) knows that. The software power law When I (Larry Putnam speaking) developed the software equation Size = (Process Productivity) x (Effort)^/3 x (Development Time)^4/3 I had data from the Army Computer Systems Command and several other US Army major commands. Fortunately, the data set was close to homogeneous batch development, lots of assembly language, some Cobol, no incremental compilers, many development inefficiencies. I didn t realize it at the time, but the project data all fell in nearly the same process productivity bin. My key finding was that the product of effort and time produced the software output, as measured in the size of the product. That finding was not surprising. After all, the number of shoes an old-fashioned shoemaker turns out is the product of his effort and time at some productivity rate, shoes per day. The nearly homogeneous data set then led to the two exponents: four thirds on time, one third on effort. Their ratio became the fourth-power law. The software equation worked in practice, but it is not exact, the way Newton s laws are. The reason is: the underlying project data is fuzzy, as Newton s data was not. From time to time I heard criticism of the four thirds to one third relationship. Other estimating formulas, for instance, either ignored the relationship or gave it a weight of one. That is, in going from a nominal development time to one 25 percent shorter, some methods increased effort by about 25 percent. By 984 my company had data on 750 projects, including several hundred commercial projects from the US, Europe, Australia, and Japan. In one study we searched the database for projects that had nearly the same size and the same process productivity, thus duplicating deliberately the pattern of my earlier work. We found three sets. For the three sets the value of the power ratio was 4.66, close to 4.0. The standard deviation was 0.20. However, the number of cases was small, only 23 in all three sets. In a second study we searched the database for data sets with close to the same value of the ratio between size and process productivity. There were five sets, with from 3 to 76 cases in each set. The average value of the ratio was 3.625 with a standard deviation of 0.48, again close to 4.0. We were delighted! With the imprecision in the underlying data, we did not expect to get exactly 4.0. Still, with this approach the probability was 84 percent that the true value of the power ratio lay between 3.5 and 4.5. Alas, several years later, Professor S. D. Conte and his co-authors pointed out that the process productivity we used in this study had itself been determined by the 2

fourth-power law. [] So, working backward from this process productivity to the power ratio, they were not surprised that it came out near 4.0. We had just proved that a circle is round! Still, the fourth-power law was working in practice. This meant strong tradeoff was there even though we could not prove definitively what its exact relationship was. Analysis by number of staff To make a non-circular study, we had to sort the database by some factor that was not tarnished in the first place by calculation from the fourth-power law. We have been advising organizations for years to use smaller development teams. We thought small teams made a difference because of the fourth-power law. Why not sort by team size? It is a definite number, derived from client data, regardless of any estimating equations. It reflects what people actually do. We sorted out two data sets: one using five or fewer people at peak, and one employing 20 or more people at peak. This is clearly shown in Figure C, below. Recent IS Applications (20 person projects vs 5 person projects) A 30 20 PI D 8 6 4 2 MBI 0 00 0 0 00 0 0 000 B 0 00 Phase 3 Duration E 0 00 00 0 Phase 3 Effort C 0 00 00 0 Phase 3 Peak Staff 0 00 Figure. Over a size range extending from 30K to 300K SLOC, squares represent projects with five or fewer people at peak, circles, projects with 20 or more people at peak. The slanting lines represent the mean values of the data points for each set. F 000 00 0 Errors SIT to FOC 3

Effort. Let s look first at the effect of staff size on effort, as shown in Figure E and blown up in Figure 2, below. The circles (20 plus people) are clearly on top much more effort. The trend lines show an average effort in the center of the range (at 0K ESLOC) of 78 personmonths for large-staff projects compared to 24.5 personmonths for small-staff projects. That is a ratio of 7.3. The ratio is somewhat less for smaller projects (fewer ESLOC) and somewhat more for larger projects. Development Effort vs. Developed SLOC 78 PM s Avg for Circles 24.5 PM s Avg for Squares 0 00 000 00 0 Phase 3 Effort Figure 2. Blow up of Figure E. Effort v. Size for two data sets. Well, some one pushed the accelerator down more people and more than half an order-of-magnitude more personmonths. 4

Time. Did that somebody gain much in the form of shortening development time? No, Figure 3 shows the circles and squares intermingled. The trend lines are almost on top of each other, especially at the center of the size range. The trend lines show an average schedule time in the center of 8.9 months for the large-staff projects compared to 9. months for small-staff projects. Somebody gained 0.2 months for all that effort increase. The time a project takes is very little different whether you use a large team or a small team -- you just don t get much schedule compression. Development Schedule vs Developed SLOC 8.92 Months Avg for Circles 9.2 Months Avg for Squares 0 00 0 Phase 3 Duration Figure 3. Blowup of Figure B, project duration vs. size. Note the two trend lines are almost on top of each other. 5

Defects. Well, how about defects? Figure 4 shows that putting all those people on the project generates a lot more defects causing more rework cycles to find and fix them. Not only do the products built by the larger teams cost more effort and dollars, they eat up any schedule benefits in rework. Defect found in System Test vs. Developed SLOC 70 Avg for Circles 33 Avg Squares 0 00 000 00 0 Errors SIT to FOC Figure 4. Blowup of Figure F. Defects vs. Size. 6

Process productivity. In Figure A the trend lines show that projects with 20-plus staffs are substantially less effective (in terms of the process productivity index) than projects with 5-minus staffs. That is, for the same size project, large teams generate a smaller process productivity index than small teams -- they are less efficient. Figure 5, below, shows this pattern. Process Productivity Index vs. Developed SLOC 30 20 PI 0 00 0 Figure 5. Blowup of Figure A. Process productivity parameter vs. Size. Tradeoff still powerful Our 984 study showed that the effort-time tradeoff varied over a range around a power ratio of 4. This variation is the result of fuzzy project data. Our current study, using staff size as the criterion, shows a marked difference in effort, little difference in schedule. This finding verifies that there is a tradeoff and it is large, but it doesn t say whether it is 3 or 5. We do think it is in that range. The fact that it is a lot more than.0 is the important point. You can trade off a little more development time for a lot less effort, cost and defects within limits, of course. Remember, go light on that gas pedal. Or, if you are a film buff, take time to find the good movie out of all that footage you shot!. S. D. Conte, H. E. Dunsmore, and V. Y. Shen, Software Engineering Metrics and Models, The Benjamin/Cummings Publishing Co., Menlo Park, CA 986, 396 pp. 7