DOWNTIME DATA -- ITS COLLECTION, ANALYSIS, AND IMPORTANCE

Proceedings of the 1994 Winter Simulation Conference
eds. J. D. Tew, S. Manivannan, D. A. Sadowski, and A. F. Seila
pages 1040-1043

Edward J. Williams
114-2 Engineering Computer Center, Mail Drop 3
Ford Motor Company
Post Office Box 2053
Dearborn, Michigan 48121-2053, U.S.A.

ABSTRACT
Until the day when plant production personnel and equipment have no downtime, proper collection and analysis of downtime data will be essential to the development of valid, credible simulation models. Methods and techniques helpful to this task within simulation model building are described.

1 INTRODUCTION
Ford Motor Company is steadily increasing its use of simulation to improve the design of production processes, both those still on the drawing board and those currently in operation. To be valid and credible, these simulation models must include expected or actual downtime experience. Since the collection of downtime data represents heavy investments in both time and cost, it is important to recapture these investments via the benefits of using valid and credible simulation models. The following considerations, to be discussed sequentially in the remainder of this paper, all pertain to the valid modeling of downtime:

2 INVALIDITY OF COMMON SIMPLIFYING ASSUMPTIONS
The most brash assumption is to ignore downtime altogether. Unless downtime never occurs (a situation never yet seen in our process-engineering practice), omission of downtime analysis produces an invalid model. Fortunately, such an invalid model also has no credibility, and hence will not be used by management to reach wrong conclusions. Another, more plausible, simplifying assumption is to


This assumption is typically unworkable for two reasons. First, very rarely does the downtime itself pertain to the entire system being modeled. Second, the analysis outlined above applies a downtime "correction" to the throughput statistic only. In practice, performance statistics other than throughput are of concern to the user. For example, a process engineer designing line layout must determine the maximum queue length upstream from a certain operation. Hence, this simplifying assumption is best reserved for rare system-global downtimes. For example, if records show that a certain plant shuts down a given number of scheduled production days per year due to snowstorms, the computation above is well-suited to evaluate the overall productivity of the plant.
A variant of this assumption may be applied to each machine individually. For example, if a machine's cycle time is a constant x and the machine is down a fraction y of total time, this assumption models the machine's cycle time as x/(1-y). This variant likewise tends to estimate global performance metrics such as throughput well, but estimate local performance metrics such as maximum queue lengths poorly.
A third simplifying assumption is "the downtime duration is a constant equal to its mean," and hence replaces a random variable representing downtime duration with that mean value. This assumption typically produces an invalid model which overestimates throughput. Downtimes markedly longer than the mean exhaust downstream buffer stock; once that stock is exhausted, downstream operations suffer unproductive time which can never be recouped. Similarly, upstream operations experience severe backup which the invalid model will fail to represent as high queue-length maxima. Vincent and Law2 describe an analogous pitfall arising from replacing a processing time by its mean. A variant of this assumption models downtime with a uniform or triangular density. These densities are often useful "rough-draft" approximations for model verification. However, the uniform has no unique mode, neither the uniform nor the triangular has inflection points, and both the uniform and the triangular have finite ranges. Therefore, these densities should not remain in the model without validation that these constraints are appropriate to the downtime being modeled.

3 TECHNIQUES OF DOWNTIME DATA COLLECTION
In industrial practice, the model builder visiting the production floor must often work with non-technical personnel unacquainted with simulation analyses; in turn, those employees often have to answer questions based on scanty or disorganized data. We have encountered the following problems and devised the following countermeasures:

4 MODELING CONSIDERATIONS
4.1 Choosing an Appropriate Probability Density

Since downtime (and uptime) durations oughtn't to be replaced with their means, an appropriate probability density must be included in the simulation model. The temptation to use the existing data as an empirical density should usually be avoided, because doing so tacitly assumes that any duration shorter than the sample minimum or larger than the sample maximum is impossible. This assumption is almost always untenable.
That said, the choice of an appropriate theoretical density becomes important. The following steps will assist in choosing one:


For example, a normal density should be avoided if its standard deviation, relative to its mean, is large enough to imply occasional durations less than zero. Also, since the mean, median, and mode of a normal density are all equal, a normal density should be avoided if these equalities conspicuously fail to hold for the sample data.
Similarly, if the sample mean and sample standard deviation are markedly unequal, an exponential density (for which these two quantities are always equal) should be avoided. Likewise, an exponential density should be avoided if the sample mode is well-removed from the sample minimum. A uniform or beta density should be avoided if no upper limit to durations is apparent, because these densities are non-zero over finite ranges.

4.2 Sensitivity Analysis
Sensitivity analysis is a method of assessing how much or how little the observable behavior of the system being modeled varies as its intrinsic properties vary. In the context of studying downtime, sensitivity analysis examines the extent of change in performance metrics such as throughput, downstream utilization, and queue-length maxima in response to changes in downtime properties such as percentage, duration, and variability of duration. For example, of two candidate machines potentially installed at a critical point of a system, the machine with smaller variance of downtime duration may greatly improve system performance even when percent downtime and average duration of downtime are equal for the two machines.
As indicated above, these sensitivity analyses are valuable when no downtime data are available. Comparing system performance under best-case and worst-case scenarios assesses the criticality of downtime performance at a specific point within the system. The greater this criticality, the greater the attention that should be devoted to increasing the accuracy of downtime estimation at that point.
Additionally, sensitivity analyses, in keeping with the "what-if" gaming abilities of simulation, provide accurate assessment of the return on various investments proposed for downtime-performance improvement. Such proposed investments might include capital expenditure for equipment with shorter downtime durations, less variable downtime durations, or longer uptime durations. Competing proposals might involve increasing payroll costs to accommodate hiring additional and/or more highly trained repair crews to improve downtime performance, or increasing outsourcing costs for contracting work externally to the system during its downtime intervals.

4.3 Modeling System Behavior During Downtime Intervals
Modern model-building tools and languages allow the modeler a variety of options for modeling system behavior during downtime. To use these software capabilities effectively in the building of a valid, credible model, the modeler must ask the system experts questions such as these:

After the above questions have been asked and answered, the modeler must, while building a model in the simulation language or package of choice, study its documentation thoroughly to assure an accurate match between the workflow of the system and the corresponding workflow in the model. Achieving accuracy of this match represents the task of model verification -- checking that the model's behavior on the computer matches the modeler's expectations.

5 SUMMARY AND OUTLOOK
During the rapid rise in simulation modeling usage at Ford Motor Company during the past few years, production and process engineers have become increasingly aware that valid downtime modeling is an essential ingredient of valid, credible models. Each of the following is in turn an essential ingredient of valid downtime modeling:

Planned developments include implementation of automated downtime data collection, increased archival and sharing of downtime data among corporate components, and development of spreadsheet macros to smooth the interface between data collection and simulation software.

ACKNOWLEDGMENTS

Drs. Hwa-Sung Na and Sanaa Taraman of Ford Alpha, Ken Lemanski of Ford Product and Manufacturing Systems, and Dr. Onur Ulgen, president of Production Modeling Corporation and professor of Industrial and Manufacturing Engineering at the University of Michigan - Dearborn, all made valuable criticisms toward improving the clarity of this paper.

REFERENCES

Grajo, Eric S. 1992. An Analysis of Test-Repair Loops in Modern Assembly Lines. In Industrial Engineering, 54-55. Production Modeling Corporation, Dearborn, Michigan.
Law, A. M. and W. D. Kelton. 1991.Simulation Modeling and Analysis, 2d ed. New York: McGraw-Hill.
Vincent, Stephen G. and Averill M. Law. 1993. UniFit II: Total Support for Simulation Input Modeling. In Proceedings of the 1993 Winter Simulation Conference, ed. Gerald W. Evans, Mansooreh Mollaghasemi, Edward C. Russell, and William E. Biles, 199-204. Averill M. Law & Associates, Tucson, Arizona.

AUTHOR BIOGRAPHY

EDWARD J. WILLIAMS holds bachelor's and master's degrees in mathematics (Michigan State University, 1967; University of Wisconsin, 1968). From 1969 to 1971, he did statistical programming and analysis of biomedical data at Walter Reed Army Hospital, Washington, D. C. He joined Ford in 1972, where he works as a computer analyst supporting statistical and simulation software. Since 1980, he has taught evening classes at the University of Michigan, including both undergraduate and graduate simulation classes using GPSS/H, SLAM II, or SIMAN.