STEPS TO IMPLEMENT BAYESIAN INPUT DISTRIBUTION SELECTION  
 
  Stephen E. Chick
 
Department of Industrial and Operations Engineering
The University of Michigan
1205 Beal Avenue
Ann Arbor, Michigan 48109-2117, U.S.A.
 
 
ABSTRACT
 
There are known pragmatic and theoretical difficulties associated with some standard approaches for input distribution selection for discrete-event simulations. One difficulty is a systematic underestimate of the variance of the expected simulation output that comes from not knowing the `true' parameter values. Another is a lack of quantification of the probability that a given distribution is best. Bayesian methods have been proposed as an alternative, but acceptance has not yet been achieved, in part because of increased computational demands, as well as challenges posed by the specification of prior distributions. In this paper, we show that responses to questions like those already asked and answered in practice can be used to develop prior distributions for a wide class of models. Further, we illustrate techniques for addressing some computational difficulties thought to be associated with the implementation of Bayesian methodology.
  Return to Wintersim
 
 

 
USING INPUT PROCESS INDICATORS FOR DYNAMIC DECISION MAKING  
 
Michael Freimer
 
School of Operations Research and Industrial Engineering
Cornell University
Ithaca, NY 14853, U.S.A
  Lee Schruben
 
Department of Industrial Engineering and Operations Research
University of California at Berkeley
4135 Etcheverry Hall
Berkeley, CA 94720-1777, U.S.A
 
ABSTRACT
 
In a continually changing environment, a simulation study that integrates the activities of data collection, model analysis, and decision making has some distinct advantages. In this paper we look at simulation project dynamics from a high-level and examine some ways for integrating these activities. From this perspective, the processes used to drive a simulation model are forecasts of environmental changes, and the parameters for models of these processes are viewed as leading indicators. A simple decision-making scenario having some of the characteristics of semiconductor manufacturing is used to illustrate the ideas.
  Return to Wintersim
 
 

 
REGRESSION METAMODELLING IN SIMULATION USING BAYESIAN METHODS  
 
  Russell C. H. Cheng
 
Faculty of Mathematical Studies
University of Southampton
Highfield
Southampton, SO17 1BJ England
 
 
ABSTRACT
 
This paper further develops some of the ideas set out by Cheng (1998) for output analysis using Bayesian Markov Chain Monte Carlo (MCMC) techniques, when a regression metamodel is to be fitted to simulation output. The particular situation addressed by Cheng was where there is uncertainty about the number of parameters needed to specify a model. This arises because there may be uncertainty about the number of terms to be included in the regression model to be fitted. The statistically non-standard nature of the problem means that it requires special handling. In this paper we shall use the derived chain method suggested by Cheng (1998). However, whereas in that paper the distribution of the response output of interest was assumed to be simply normal, it is typically the case, especially in the study of systems working near their capacity limit, that this distribution is skewed, and moreover the distribution has a support that is effectively bounded below - that is the distribution has a threshold. We describe how the derived MCMC method might be applied in this situation and illustrate with a numerical example involving the simulation of a computer PAD network.
  Return to Wintersim
 
 

 
VARIANCE REDUCTION OF MONTE CARLO AND RANDOMIZED QUASI-MONTE CARLO ESTIMATORS FOR STOCHASTIC VOLATILITY MODELS IN FINANCE  
 
Hatem Ben Ameur
 
 
Service de l'
Enseignement des Methodes Quantitatives de Gestion
École des Hautes Etudes Commerciales 3000, chemin de la Cote-Ste-Catherine
Montréal, H3T 2A7, CANADA
  Pierre L'Ecuyer
Christiane Lemieux

 
Département d'Informatique et de Recherche Opérationnelle
Université de Montréal, C.P. 6128, Succ. Centre-Ville
Montréal, H3C 3J7, CANADA
 
ABSTRACT
 
We illustrate by numerical examples how certain variance reduction methods dramatically improve the efficiency of Monte Carlo simulation for option pricing and other estimation problems in finance, in the context of a geometric Brownian motion model with stochastic volatitity. We consider lookback options and partial hedging strategies, with different models for the volatility process. For variance reduction, we use control variates, antithetic variates, conditional Monte Carlo, and randomized lattice rules coupled with a Brownian bridge technique that reduces the effective dimension of the problem. In some of our examples, the variance is reduced by a factor of more than 100 millions without increasing the work. The examples also illustrate how randomized quasi-Monte Carlo can be effective even if the problems considered involve a large number of dimensions.
  Return to Wintersim
 
 

 
EFFICIENCY IMPROVEMENTS FOR PRICING AMERICAN OPTIONS WITH A STOCHASTIC MESH  
 
  Athanassios N. Avramidis
Paul Hyden

 
School of Operations Research and Industrial Engineering
Cornell University
Ithaca, NY 14853, U.S.A.
 
 
ABSTRACT
 
We develop and study general-purpose techniques for improving the efficiency of the stochastic mesh method that was recently developed for pricing American options via Monte Carlo simulation. First, we develop a mesh-based, biased-low estimator. By recursively averaging the low and high estimators at each stage, we obtain a significantly more accurate point estimator at each of the mesh points. Second, we adapt the importance sampling ideas for simulation of European path-dependent options in Glasserman, Heidelberger, and Shahabuddin (1998a) to pricing of American options with a stochastic mesh. Third, we sketch generalizations of the mesh method and we discuss links with other techniques for valuing American options. Our empirical results show that the bias-reduced point estimates are much more accurate than the standard mesh-method point estimates. Importance sampling is found to increase accuracy for a smooth option-payoff functions, while variance increases are possible for non-smooth payoffs.
  Return to Wintersim
 
 

 
STRATIFICATION ISSUES IN ESTIMATING VALUE-AT-RISK  
 
Paul Glasserman
 
Columbia University
New York, NY 10027
Philip Heidelberger
 
IBM T.J. Watson Research Center
Yorktown Heights, NY 10598
Perwez Shahabuddin
 
Columbia University
New York, NY 10027
 
ABSTRACT
 
This paper considers efficient estimation of value-at-risk, which is an important problem in risk management. The value-at-risk is an extreme quantile of the distribution of the loss in portfolio value during a holding period. An effective importance sampling technique is described for this problem. The importance sampling can be further improved by combining it with stratified sampling. In this setting, an effective stratification variable is the likelihood ratio itself. The paper examines issues associated with the allocation of samples to the strata, and compares the effectiveness of the combination of importance sampling and stratified sampling to that of stratified sampling alone.
  Return to Wintersim
 
 

 
AN ASYMPTOTIC ALLOCATION FOR SIMULTANEOUS SIMULATION EXPERIMENTS  
 
Hsiao-Chang Chen
 
 
SynQuest Inc.
Clifton Park, NY 12065, U.S.A.
  Chun-Hung Chen
Jianwu Lin

 
Dept. of Systems Engineering
University of Pennsylvania
Philadelphia, PA 19104, U.S.A.
 
ABSTRACT
 
In this paper, we consider the allocation of a fixed total number of simulation replications among competing design alternatives in order to (i) identify the best simulated design, (ii) intelligently determine the best simulation run lengths for all simulation experiments, and (iii) significantly reduce the total computation cost. An asymptotically optimal allocation rule for maximizing a lower bound of the probability of correct selection is presented. Moreover, we illustrate the efficiency of our method with a series of generic numerical experiments. The simulation cost is significantly reduced with our sequential approach.
  Return to Wintersim
 
 

 
Can the Regenerative Method be Applied to Discrete-Event Simulation  
 
Shane G. Henderson
 
Department of Engineering Science
University of Auckland
Private Bag 92019
Auckland, NEW ZEALAND
  Peter W. Glynn
 
Department of Engineering-Economic Systems and Operations Research
Stanford University
Stanford CA 94309-4023, U.S.A.
 
ABSTRACT
 
The regenerative method enjoys asymptotic properties that make it a highly desirable approach for steady-state simulation output analysis. It has been shown that virtually all discrete-event simulations are regenerative. However, the method is not in widespread use, perhaps primarily because of a difficulty in identifying regeneration times.
 
Our goal in this paper is to highlight the essence of the difficulty in identifying regeneration times in discrete-event simulations. We focus on a very simple example of a discrete-event simulation, and explore its regenerative properties.
 
We show that for our example, it is possible to explicitly determine regeneration times. The ideas that are used to establish this fact might prove useful in identifying regeneration times in more general discrete-event system simulations.
  Return to Wintersim
 
 

 
FAST SIMULATION OF BROADBAND TELECOMMUNICATIONS NETWORKS CARRYING LONG-RANGE DEPENDENT BURSTY TRAFFIC  
 
José R. Gallardo
Dimitrios Makrakis

 
Advanced Communications Engineering Centre
Department of Electrical and Computer Engineering
The University of Western Ontario
London, Ontario N6A 5B9, CANADA
  Luis Orozco-Barbosa
 
 
School of Information Technology and Engineering
Department of Electrical Engineering
University of Ottawa
Ottawa, Ontario K1N 6N5, CANADA
 
ABSTRACT
 
A technique for the fast simulation of broadband communications systems is proposed, which is based on regenerative Importance Sampling techniques. Our algorithm is applicable to estimate the probability of rare events when modeling the offered traffic using Fractional Stable Noise (FSN) processes (including Fractional Brownian Noise as a particular case), which have been recently proved to be able to capture both the long-range dependence and the burstiness of today's aggregate network traffic. An exact description of FSN processes is given, as well as an approximation that allows for the application of Importance Sampling techniques. The results obtained for a simple example are also included.
  Return to Wintersim
 
 

 
SELECTION PROCEDURES WITH STANDARDIZED TIMES SERIES VARIANCE ESTIMATORS  
 
  David Goldsman
William S. Marshall

 
School of Industrial and Systems Engineering
Georgia Institute of Technology
Atlanta, GA 30332, U.S.A.
 
 
ABSTRACT
 
This article studies a modification of Rinott's two-stage procedure for selecting the normal population with the largest (or smallest) mean. The modification, which is appropriate for use in the simulation environment, uses in the procedure's first stage different variance estimators than the usual batch means (BM) variance estimator. In particular, we will use variance estimators arising from the method of standardized time series (STS). On the plus side, certain STS estimators have more degrees of freedom than that of the BM estimator. On the other hand, STS variance estimators tend to require larger sample sizes than the BM estimator in order to converge to their assumed distributions. These considerations result in trade-offs involving the procedure's achieved probability of correct selection as well as the procedure's expected sample size.
  Return to Wintersim
 
 

 
DESIGNING SIMULTANEOUS SIMULATION EXPERIMENTS  
 
Paul Hyden
 
School of Operations Research and Industrial Engineering
Cornell University
206 Rhodes Hall
Ithaca, NY 14853, U.S.A
  Lee Schruben
 
Department of Industrial Engineering and Operations Research
University of California at Berkeley
4135 Etcheverry Hall
Berkeley, CA 94720-1777, U.S.A
 
ABSTRACT
 
Simulation experiments are often designed assuming that a fixed, and known, computing budget is to be allocated sequentially among different alternatives. However, in actual simulation experiments, there may be budget uncertainty or at least flexibility - for example, when there is a soft deadline for obtaining the study results. In such situations, it may be beneficial to allocate resources simultaneously in dynamically changing proportions. In this paper, we will examine optimal resource allocation paths. These paths climb the contour curves of the probability of selecting the best of several alternatives in a manner that insures that the highest probability of correct selection P(CS) is obtained when the study is halted. To gain insight into the complexity of optimal resource allocation paths, simple models exhibiting serial correlation, cross correlation, and trends are studied.
  Return to Wintersim
 
 

 
SIMULTANEOUS SIMULATION EXPERIMENTS AND NESTED PARTITION FOR DISCRETE RESOURCE ALLOCATION IN SUPPLY CHAIN MANAGEMENT  
 
Leyuan Shi
 
Dept. of Industrial Engineering
University of Wisconsin-Madison
Madison, WI 53706
Chun-Hung Chen
 
Dept. of Systems Engineering
University of Pennsylvania
Philadelphia, PA 19104-6315
Enver Yûcesan
 
INSEAD
Technology Management Area
Fontainebleau, France
 
ABSTRACT
 
Discrete resource allocation is a common problem in supply chain management. However, stochastic discrete resource allocation problems are difficult to solve. In this paper, we propose a new algorithm for solving such difficult problems. The algorithm integrates the nested partitions method with an optimal computing budget allocation method. The resulting hybrid algorithm retains the global perspective of the nested partitions method and the efficient simultaneous simulation experiments of the optimal computing budget allocation. Numerical results demonstrate that the hybrid algorithm can be effectively used for a large-scale discrete resource allocation problem.
  Return to Wintersim
 
 

 
Rare Event Simulation of Delay in Packet Switching Networks Using DPR-Based Splitting  
 
Zsolt Haraszti
 
Ericsson Radio Systems AB
S-12625 Stockholm, Sweden
  J. Keith Townsend
 
Center for Advanced Computing and Communication
Department of Electrical & Computer Engineering
North Carolina State University
Raleigh, NC, 27695-7914 U.S.A.
 
ABSTRACT
 
Rare event simulation using splitting has been shown to provide significant speed-up for large classes of problems, especially when queue length distribution is of primary interest. However, choice of the control parameters is much less straightforward in cases where splitting is applied to systems in which the target event is delay, rather than packet loss. In this paper we propose a control strategy for splitting that allows computationally efficient analysis of very low delay threshold probabilities which typically occur in communication networks. A different technique is required for delay because unlike the cell or packet loss case, the target event (delay) and the prerequisite condition that leads to a rare delay event (a full buffer) do not coincide temporally. We demonstrate the technique by using it to measure delay probabilities in three examples: a simple ATM multiplexer, a queueing system with multiple traffic classes, and a tandem queueing network with tagged and background traffic.
  Return to Wintersim
 
 

 
Efficient Simulation of a Tandem Jackson Network  
 
Dirk P. Kroese
 
Teletraffic Research Center
University of Adelaide
South Australia 5005
  Victor F. Nicola
 
Department of Electrical Engineering
University of Twente
Enschede, The Netherlands
 
ABSTRACT
 
In this paper we consider a two-node tandem Jackson network. Starting from a given state, we are interested in estimating the probability that the content of the second buffer exceeds some high level L before it becomes empty. The theory of Markov additive processes is used to determine the asymptotic decay rate of this probability, for large L. Moreover, the optimal exponential change of measure to be used in importance sampling is derived and used for efficient estimation of the rare event probability of interest.
 
Unlike changes of measures proposed and studied in recent literature, the one derived here is a function of the content of the first buffer, and yields asymptotically efficient simulation for any set of arrival and service rates. The relative error is bounded independent of the level L, except when the first server is the bottleneck and its buffer is infinite, in which case the relative error is bounded linearly in L.
  Return to Wintersim
 
 

 
SIMULATING HEAVY TAILED PROCESSES USING DELAYED HAZARD RATE TWISTING  
 
Sandeep Juneja
 
Indian Insititute of Technology
Delhi 110016, India.
Perwez Shahabuddin
 
Columbia University
New York, NY 10027, USA.
Anurag Chandra
 
Massachusetts Institute of Technology
Cambridge, MA 02139, USA.
 
ABSTRACT
 
Consider the problem of estimating the small probability that the maximum of a random walk exceeds a large threshold, when the process has a negative drift and the underlying random variables may have heavy tailed distributions. We consider one class of such problems that has applications in estimating the ruin probability associated with insurance claim processes with subexponentially distributed claim sizes, and in estimating the probability of large delays in single server queues with subexponentially distributed service times. Significant work has been done on analogous problems for the light tailed case (when the moment generating function exists in a neighborhood around zero, so that the tail decreases at an exponential rate or faster) involving importance sampling methods that use exponential twisting. However, for the subexponential case, moment generating functions do not exist in the pertinent regions making exponential twisting infeasible. In this paper we introduce importance sampling techniques where the new probability measure is using the change of measure obtained by twisting the hazard rate of the original distribution. For subexponential distributions this amounts to twisting at a subexponential rate. We also introduce the technique of "delaying" the change of measure and show that the combination of the two techniques produces asymptotically optimal estimates of the small probabilities mentioned above for a large class of subexponential distributions.
  Return to Wintersim
 
 

 
SIMULATION-BASED ESTIMATION OF QUANTILES  
 
  E. Jack Chen
W. David Kelton

 
Department of Quantitative Analysis and Operations Management
University of Cincinnati
Cincinnati, Ohio 45221, U.S.A.
 
 
ABSTRACT
 
This paper discusses implementation of a sequential quantile-estimation algorithm for highly correlated steady-state simulation output. Our primary focus is on issues related to computational and storage requirements of order statistics. The algorithm can compute exact sample quantiles and process sample sizes up to several billion without storing and sorting the whole sequence. The algorithm dynamically increases the sample size so that the quantile estimated satisfies a pre-specified precision requirement.
  Return to Wintersim
 
 

 
On the Implementation of Smoothed Perturbation Analysis Estimator for a Single-Server Queue with Multiple Vacations  
 
Takayuki Takagi
 
Department of Systems Science
Kyoto University
Kyoto 606-8501, JAPAN
  Naoto Miyoshi
 
Department of Mathematical and Computing Sciences
Tokyo Institute of Technology
Tokyo 152-8552, JAPAN
 
ABSTRACT
 
We consider to apply the perturbation analysis~(PA) to a single-server queue with multiple server vacations. A major difficulty in the implementation of PA estimator for such queueing systems is that the introduced perturbations are propagated and accumulated continuously without any resetting. This fact may lead to the divergence of PA estimates even if the limiting distribution exists. We show that it is possible to construct a sequence of points on the observed sample path such that the perturbations are accumulated only between the two adjacent points. The key idea lies in constructing a perturbed path which is not on the same sample as the observed nominal path but is identical in probability law.
  Return to Wintersim
 
 

 
IMPROVED BATCHING FOR CONFIDENCE INTERVAL CONSTRUCTION IN STEADY-STATE SIMULATION  
 
Natalie M. Steiger
 
Department of Information Systems and Operations Management
University of North Carolina at Greensboro
Greensboro, NC 27402-6165, U.S.A.
  James R. Wilson
 
Department of Industrial Engineering
North Carolina State University
2401 Stinson Drive
Raleigh, NC 27695-7906, U.S.A.
 
ABSTRACT
 
We describe an improved batch-means procedure for building a confidence interval on a steady-state expected simulation response that is centered on the sample mean of a portion of the corresponding simulation-generated time series and satisfies a user-specified absolute or relative precision requirement. The theory supporting the new algorithm merely requires the output process to be weakly dependent (phi-mix\-ing) so that for a sufficiently large batch size, the batch means are approximately multivariate normal but not necessarily uncorrelated. A variant of the method of nonoverlapping batch means (NOBM), the Automated Simulation Analysis Procedure (ASAP) operates as follows: the batch size is progressively increased until either (a) the batch means pass the von Neumann test for independence, and then ASAP delivers a classical NOBM confidence interval; or (b) the batch means pass the Shapiro-Wilk test for multivariate normality, and then ASAP delivers a corrected confidence interval. The latter correction is based on an inverted Cornish-Fisher expansion for the classical NOBM ratio, where the terms of the expansion are estimated via an autoregressive--moving average time series model of the batch means. An experimental performance evaluation demonstrates the advantages of ASAP versus other widely used batch-means procedures.
  Return to Wintersim
 
 

 
SENSITIVITY OF OUTPUT PERFORMANCE MEASURES TO INPUT DISTRIBUTION SHAPE IN MODELING QUEUES - 3: REAL DATA SCENARIO  
 
  Donald Gross
 
Department of Systems Engineering and Operations Research
George Mason University
Fairfax, VA 22030, U.S.A
 
 
ABSTRACT
 
This paper follows-on papers presented at the two previous WSC conferences on sensitivity of output measures to input distribution selection in queueing modeling. Here, a real situation is studied, where data on input distributions are utilized and distributions selected by two fitting packages, Arena Input Analyzer and ExpertFit. Empirical distributions made from histograms of the raw data itself , as well as the first two choices from Arena and ExpertFit are compared for this small bank queueing network model, showing that an output measure such as mean wait in queue is quite sensitive to input distribution choice.
  Return to Wintersim
 
 

 
SIMULATING A NONSTATIONARY POISSON PROCESS USING BIVARIATE THINNING: THE CASE OF "TYPICAL WEEKDAY" ARRIVALS AT A CONSUMER ELECTRONICS STORE  
 
  K. Preston White, Jr.
 
Department of Systems Engineering
University of Virginia
Charlottesville, VA 22903-2442, U.S.A.
 
 
ABSTRACT
 
We present a case study in which thinning is applied to simulate time-varying arrivals at a consumer electronics store. The underlying simulation was developed to support an analysis of new staffing schedules for retail sales associates, given proposed changes in store layout and operating procedures. A principal challenge was developing a modeling approach for customer arrivals, where it was understood that the arrival rate varied by time-of-day and by day-of-the-week, as well as seasonally. An analysis of arrival data supported a conjectured "typical weekday" as one basic arrival model. For this model, arrivals were assumed to be nonstationary Poisson, with a piecewise-linear arrival rate independently modulated by hour and by day. Arrival data were filtered and independent hourly and daily thinning factors computed. In the simulation, potential arrivals were generated with a mean equal to the minimum average interarrival rate, determined from the average arrival count for the hour/day time block with unit thinning factors. Candidate arrivals were then thinned using a bivariate acceptance probability equal to the product of the corresponding hourly and daily thinning factors.
  Return to Wintersim
 
 

 
Sensitivity Analysis of Simulation Output To Parameters of Nonhomogeneous Poisson Processes  
 
Michael E. Kuhl
 
Department of Industrial and Manufacturing Systems Engineering
Louisiana State University
Baton Rouge, LA 70803-6409, U.S.A.
  Sun Ewe Lim
 
Department of Industrial and Manufacturing Systems Engineering
Louisiana State University
Baton Rouge, LA 70803-6409, U.S.A.
 
ABSTRACT
 
Nonhomogeneous Poisson processes (NHPPs) are frequently used in stochastic simulations to model nonstationary point processes. These NHPP models are often constructed by estimating the parameters of the rate function from one or more observed realizations of the process. This paper focuses on the degree of accuracy to which the rate function parameters of the NHPP need to be estimated such that the simulation output performance measures are not significantly different from performance measures that would be obtained for the underlying (true) process.
  Return to Wintersim
 
 

 
OPTIMIZATION OVER DISCRETE SETS VIA SPSA  
 
László Gerencsér
 
Computer and Automation Institute
Hungarian Academy of Sciences
Kende 13-17, Budapest, 1111, HUNGARY
Stacy D. Hill
 
Applied Physics Laboratory
John Hopkins University
Laurel, MD 20723-6099, U.S.A.
Zsuzsanna Váagó
 
Computer and Automation Institute
Hungarian Academy of Sciences
Kende 13-17
Budapest, 1111, HUNGARY
 
ABSTRACT
 
A fixed gain version of the SPSA (simultaneous perturbation stochastic approximation) method for function minimization is developed and the error process is characterized. The new procedure is applicable to optimization problems over, the grid of points in with integer components. Simulation results and a closely related application, a resource allocation problem, is shortly described.
  Return to Wintersim
 
 

 
Solving Stochastic Optimization Problems with Stochastic Constraints: An Application in Network Design  
 
Gül Gürkan
 
CentER for Economic Research
Tilburg University
5000 LE Tilburg
The Netherlands
A. Yonca
 
GE Corporate R & D
One Research Circle
Niskayuna, NY 12309 USA
Stephen M. Robinson
 
Department of Industrial Engineering
University of Wisconsin-Madison
1513 University Avenue
Madison, WI 53706 USA
 
ABSTRACT
 
Recently sample-path methods have been successfully used in solving challenging simulation optimization and stochastic equilibrium problems. In this paper we deal with a variant of these methods to solve stochastic optimization problems with stochastic constraints. Using optimality conditions, we convert the problem to a stochastic variational inequality. We outline a set of sufficient conditions for the almost-sure convergence of the method. We also illustrate an application by using the method to solve a network design problem. We find optimal arc capacities for a stochastic network (in which the demand and supply at each node is random) that minimize the sum of the capacity allocation cost and a measure of the expected shortfall in capacity.
  Return to Wintersim
 
 

 
ITERATIVE RANKING-AND-SELECTION FOR LARGE-SCALE OPTIMIZATION  
 
  Sigurdur Ólafsson
 
Department of Industrial and Manufacturing Systems Engineering
Iowa State University
205 Engineering Annex, Ames, IA 50010
 
 
ABSTRACT
 
We develop a new algorithm for simulation-based optimization where the number of alternatives is finite but very large. Our approach draws on recent work in adaptive random search and from ranking-and-selection. In particular, it combines the nested partitions method for global optimization and Rinott's two-stage ranking-and-selection procedure. We prove asymptotic convergence of the new algorithm under fairly mild conditions.
  Return to Wintersim
 
 

 
STATISTICAL METHODS FOR SENSITIVITY AND PERFORMANCE ANALYSIS IN COMPUTER EXPERIMENTS  
 
Leslie M. Moore
 
Statistical Sciences Group, TSA-1, MS F600
Los Alamos National Laboratory
Los Alamos, NM 87545-0600, USA
  Bonnie K. Ray
 
Dept. of Mathematical Sciences
New Jersey Institute of Technology
Newark, NJ 07102, USA
 
ABSTRACT
 
We describe statistical methods for sensitivity and performance analysis of complex computer simulation experiments. Graphical methods, such as trellis plots, are suggested for exploratory analysis of individual or aggregate performance metrics conditional on different experiment inputs. More formal statistical methods, such as analysis of variance-based methods and regression tree analysis, are used to determine variables having substantive influence on the experimental results and to investigate the structure of the underlying relationship between inputs and outputs. The methods are discussed in relation to a supply chain model of the textile manufacturing process having many possible input and output variables of interest and for a computer model used to describe the flow of material in an ecosystem.
  Return to Wintersim
 
 

 
A SURVEY OF RANKING, SELECTION, AND MULTIPLE COMPARISON PROCEDURES FOR DISCRETE-EVENT SIMULATION  
 
James R. Swisher
 
Lotus Biochemical Corporation
7335 Lee Highway
Radford, Virginia 24141 U.S.A.
  Sheldon H. Jacobson
 
Department of Mechanical and Industrial Engineering
University of Illinois at Urbana-Champaign
Urbana, Illinois 61801 U.S.A.
 
ABSTRACT
 
Discrete-event simulation models are often constructed so that an analyst may compare two or more competing design alternatives. This paper presents a survey of the literature for two widely-used statistical methods for selecting the best design from among a finite set of k alternatives: ranking and selection (R&S) and multiple comparison procedures (MCPs). A comprehensive survey of each topic is presented along with a summary of recent unified R&S-MCP approaches. In addition, an example of the application of Nelson and Matejcik's (1995) combined R&S-MCP procedure is given.
  Return to Wintersim
 
 

 
The Main Issues in Nonlinear Simulation Metamodel Estimation  
 
Isabel Reis dos Santos
 
Departamento de Matemática
Instituto Superior Técnico
Av. Rovisco Pais, 1049 Lisboa, PORTUGAL
  Acácio M. O. Porta Nova
 
Seccào Autónoma de Economia e Gestào
Instituto Superior Técnico
Av. Rovisco Pais, 1049 Lisboa, PORTUGAL
 
ABSTRACT
 
In this paper, we investigate and discuss some of the main issues concerning the estimation of nonlinear simulation metamodels. We propose a methodology for identifying a tentative functional relationship, estimating the metamodel coefficients and validating the simulation metamodel. This approach is illustrated with a simple queueing system. Finally, we draw some conclusions and identify topics for further work in this area.
  Return to Wintersim
 
 

 
Copyright © 1999, Wintersim 99 <mailto:vlopez@linknet.net> ... All rights reserved.