Validation checks the accuracy of the model's representation of the real system. Model validation is defined to mean "substantiation that a computerized model within its domain of applicability possesses a satisfactory range of accuracy consistent with the intended application of the model".
Face validity A model that has
face validity appears to be a reasonable imitation of a real-world system to people who are knowledgeable of the real world system.
Structural assumptions Assumptions made about how the system operates and how it is physically arranged are structural assumptions. For example, the number of servers in a fast food drive through lane and if there is more than one how are they utilized? Do the servers work in parallel where a customer completes a transaction by visiting a single server or does one server take orders and handle payment while the other prepares and serves the order. Many structural problems in the model come from poor or incorrect assumptions. A requirement is that both the system data and model data be approximately
Normally Independent and Identically Distributed (NIID). The
t-test statistic is used in this technique. If the mean of the model is μm and the mean of system is μs then the difference between the model and the system is D = μm - μs. The hypothesis to be tested is if D is within the acceptable range of accuracy. Let L = the lower limit for accuracy and U = upper limit for accuracy. Then :H0 L ≤ D ≤ U versus :H1 D U is to be tested. The operating characteristic (OC) curve is the probability that the null hypothesis is accepted when it is true. The OC curve characterizes the probabilities of both type I and II errors. Risk curves for model builder's risk and model user's can be developed from the OC curves. Comparing curves with fixed sample size tradeoffs between model builder's risk and model user's risk can be seen easily in the risk curves. If model builder's risk, model user's risk, and the upper and lower limits for the range of accuracy are all specified then the sample size needed can be calculated.
Confidence intervals Confidence intervals can be used to evaluate if a model is "close enough" to a system for some variable of interest. The difference between the known model value, μ0, and the system value, μ, is checked to see if it is less than a value small enough that the model is valid with respect that variable of interest. The value is denoted by the symbol ε. To perform the test a number,
n, statistically independent runs of the model are conducted and a mean or expected value, E(Y) or μ for simulation output variable of interest Y, with a standard deviation
S is produced. A confidence level is selected, 100(1-α). An interval, [a,b], is constructed by : a = E(Y) - t_{a/2,n-1}S/\sqrt{n} \qquad and \qquad b = E(Y) + t_{a/2,n-1}S/\sqrt{n}, where : t_{a/2,n-1} is the critical value from the t-distribution for the given level of significance and n-1 degrees of freedom. : If |a-μ0| > ε and |b-μ0| > ε then the model needs to be calibrated since in both cases the difference is larger than acceptable. : If |a-μ0| 0| 0| 0| > ε or
vice versa then additional runs of the model are needed to shrink the interval.
Graphical comparisons If statistical assumptions cannot be satisfied or there is insufficient data for the system a graphical comparisons of model outputs to system outputs can be used to make a subjective decisions, however other objective tests are preferable. == ASME Standards ==