Abstract
Every simulation study consists of at least 10 steps:
-
1.
Translating the qualitative description of a system into a formal abstraction that explicitly accounts for all logical and mathematical relationships
-
2.
Identifying all parameters of the abstraction that require numerical values as input
-
3.
Identifying all measures of performance whose values require estimation
-
4.
Estimating the values of all unknown input parameters from available data, expert opinion, etc.
-
5.
Designing the set of sampling experiments, at least one for each distinct set of values for the vector of input parameters
-
6.
Converting the logical abstraction to executable code in a simulation programming language
-
7.
Incorporating into the code statements for computing time averages of performance measures during program execution
-
8.
Performing the set of sampling experiments by executing the code repeatedly, at least once for each vector of input-parameter values
-
9.
For each measure of performance, evaluating how well its time average approximates its unknown long-run average
-
10.
Comparing corresponding sample time averages for each performance measure across experiments
While this numbering allows for orderly progress from beginning to end, actual practice may turn out to blur the distinction between steps, rearrange their order, or in some cases, neglect or pay too little attention to them. For example, the availability of point-click-drag-drop simulation software entices a simulationist to merge steps 1 and 6, possibly making her or him oblivious to a wider range of modeling concepts than the selected language implementation may accommodate. Frequently, the choice of long-run time averages to estimate in step 3 depends solely on what the programming language automatically makes available in converting the model to code in step 6. Too often, step 9 is ignored, making any conclusions based on comparisons across experiments difficult to justify statistically.
Keywords
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsPreview
Unable to display preview. Download preview PDF.
References
Cheng, R.C., and J.D. Lamb (1999). Making efficient simulation experiments interactively with a desktop simulation package, University of Kent at Canterbury, United Kingdom.
Daley, D. and L. Servi (1999). Personal communications.
Fishman, G.S. (1973). Concepts and Methods in Discrete Event Digital Simulation, Wiley, New York.
Gordon, G. (1969). System Simulation, Prentice-Hall, Englewood Cliffs, N.J.
Pritsker, A.A.B., J.J. O’Reilly, and D.K. LaVal (1997). Simulation with Visual SLAM and AweSim, Wiley, New York.
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 2001 Springer Science+Business Media New York
About this chapter
Cite this chapter
Fishman, G.S. (2001). Modeling Concepts. In: Discrete-Event Simulation. Springer Series in Operations Research. Springer, New York, NY. https://doi.org/10.1007/978-1-4757-3552-9_2
Download citation
DOI: https://doi.org/10.1007/978-1-4757-3552-9_2
Publisher Name: Springer, New York, NY
Print ISBN: 978-1-4419-2892-4
Online ISBN: 978-1-4757-3552-9
eBook Packages: Springer Book Archive