Академический Документы
Профессиональный Документы
Культура Документы
1. Introduction
Discrete-event simulation has been widely used to compare
alternative system designs or operating policies. When
evaluating k alternative system designs, we select one or
more systems as the best and control the probability that the
selected systems really are the best. Let mi denote the
expected response of system i. Our goal is to nd the system
with the smallest expected response m* min1pipkmi. If the
system with the biggest expected response is desired, just
replace min with max in the formula. We achieve this goal by
using a class of ranking and selection (R&S) procedures.
Most R&S procedures, for example Dudewicz and Dalal
(1975) or Rinotts (1978) indifference-zone selection procedures, have focused on identifying the best system. Nevertheless, Koenig and Law (1985) have developed a two-stage
indifference-zone procedure to select a subset of size m
containing the v best of k systems; where (1pvpmok).
A restricted subset attempts to exclude populations that
deviate more than d* from the smallest mean. Extending the
work of Gupta and Santner (1973), Sullivan and Wilson
(1989) have developed a two-stage restricted-subset-selection
procedure, denoted VE, that determines a subset of maximum size m that contains at least one system that is within
a pre-specied amount of the best. In this paper, we derive
the probability of correct selection of determining a restricted subset based on the distribution of order statistics in
a clear and concise manner. Furthermore, the proposed
restricted-subset-selection procedures have the capability to
determine subsets of maximum size m that contain only and
Correspondence: E Jack Chen, BASF Corporation, 333 Mount Hope
Avenue, Rockaway, NJ 07866, USA.
2. Background
First, some notation:
Xij:
Ni:
mi:
i:
X
s2i :
S2i (Ni):
S2i n0
j1
1
Xij Xi 2
n0 1
for i 1, 2, y, k.
2. Compute the required sample sizes
l
m
i maxn0 1; h1 Si n0 =d 2 ;
for i 1; 2; . . . ; k
where Jzn is the smallest integer that is greater than or
equal to the real number z.
3. Simulate additional Nin0 samples.
4. Compute the second-stage sample means
2
Xi
Ni
X
1
Xij
Ni n0 jn 1
0
where
Ga b a1
bx; a; b
x 1 xb1
GaGb
is the beta distribution with shape parameters a and b and
G(a) is the gamma function. Note that G(a) (a1)! for any
positive integer a. Furthermore, the cdf of the uth order
statistics
Gm;u yu
Zyu
gm;u tdt
1
Let X[c] be the cth smallest weighted sample mean from X~il
for l 1, 2,y, n and let m[c] be its unknown true mean. Let
X[u] be the uth (u mc 1) smallest weighted sample mean
from X~il for l n 1,n 2,y, k and let m[u] be its unknown
true mean. To simply the notation, we use Xc, Xu, mc, and mu
instead of X[c], X[u], m[c], and m[u] in remainder of this section.
We can write
~ u
~ c oX
PCS PX
m m
P Tc oTu u c
d =h
Z1
X
Gv;c tu hdGkv;mc 1 tu
1
1
Z1
3. Methodologies
In this section, we extend the subset selection procedures to
select a restricted subset. As with most selection procedures,
the proposed selection procedures require the input data to
be independent and identically distributed (iid) normal.
However, the variance can be different across systems. Many
performance measures of interest are taken over some
average of a sample path or a batch of samples. Thus, many
applications tend to have a normally distributed simulation
output. If the non-normality of the samples is a concern,
users can use batch means to manufacture samples that
appear to be iid normal, as determined by the tests
of independence and normality (see, for example, Chen
and Kelton, 2007). In the selection procedures described
below, the sampling operations can be carried out independently across systems. Hence, one can deploy the selection
procedures in a parallel and distributed environment.
1
when mi pm0
when m0 d omi
q
s2i =Ni s20 =N0
q
s2i =Ni s20 =N0
pFz1a 1 a
The inequality holds because Z has a standard normal
distribution when mim0 d*.
To control Type I errors, which occur when we do not
accept the null hypothesis when it is true, and Type II errors,
we use the test statistic
Xi X0
Z q
s2i =Ni s20 =N0
q
s2i =Ni s20 =N0
d
B
C
bpF@z1a qA
2
2
si =Ni s0 =N0
0
d
z1a q pzb
s2i =Ni s20 =N0
PX~b1 d =2oX~u
XPX~c d =2oX~u
XPX~c wcu d oX~u
XPX~c X~u wcu omc mu
P
Hence,
PX~u pX~b1 d =2p1 P
There is no more than 1P* probability that system u (ie
the system whose weighted sample means is the uth smallest
among system il for l n 1, n 2, y, k) will be included in
the restricted subset.
Under the conguration that mil mi1 d for
l 2, 3, y, k (ie the LFC and v 1) and vom, the
probability of X~bm 1 pX~b1 d =2 will be less than
1P*. However, the probability of X~bl pX~b1 d =2 for
l 2, 3, y, m will be greater than 1P*. Note that under
3. Calculate
the required
l
m sample size Nbl ;t 1 max
n0 ; 2hSbl Nbl ;t =dbl 2 , for l 1, 2, y, k. Here dbl is
computed according to (3).
4. If Ni,t 1pNi,t, for i 1, 2, y, k go to step 6.
5. Simulate additional J(Ni,t 1Ni,t) /2n samples for
system i. Set t t 1. go to step 2.
6. Select system bl iff Xbl pminXbm ; Xb1 d =2.
i, instead of the weighted
Note that sample means X
sample means Xi, are used to determine the subset. This is
because there are more than two stages of sample means and
we can no longer use the approach of Dudewicz and Dalal
i for
(1975) to compute the weighted sample means. While X
i 1, 2, y, k are still t-distributed, they have different df.
Hence, this sequential procedure is a heuristic. Nevertheless,
our empirical studies indicate that this procedure performs
well in terms of P(CS) and sample sizes. The critical value h
depends on c, v, m, k, n0, and P*. Even though the sample
sizes for each system change at each iteration, we use the
initial value of h through all iterations, thereby simplifying
the programming effort and providing conservative estimates of the sample sizes.
Let us consider the steps for taking additional samples,
that is steps 25, as one iteration. We can reduce the number
of iterations with a larger incremental sample size for system
i at the tth iteration, but we run the risk of allocating more
samples than necessary to non-promising systems. For
example, two-stage procedures allocate all the required
samples at the second stage based on the information
obtained at the rst stage. At the other extreme, some
selection procedures, for example Andradottir and Kim
(2007) take only one additional sample from each system
that is still under consideration at each iteration. While these
procedures generally require smaller sample sizes, they
require many iterations and incur the associated overhead.
We propose to use the sample size allocation strategy of
Chen and Kelton (2005) to compute the incremental sample
size dynamically with all the information obtained up to the
current iteration. The additional sample size for alternative i
at iteration t 1 is
di;t 1 Ni;t 1 Ni;t =2
2i )r/(r1) to
Since we use the equation S2i (r) (Srj X2ij/rX
compute the variance estimator, we are only required to
PNi;t
PNi;t 2
Xit ; t1
Xit instead of the entire
store the triple Ni;t ; t1
sequence Xi1 ; Xi2 ; . . . ; XiNi;t .
see Law and Kelton (2000, p 559) for details. Since fij
will not, in general, be an integer, interpolation will probably be necessary. The procedure will eliminate system j
I(1) wij for some feasible system i.
j(1)4X
such that X
q
Here wij tP;fij S2i 1; Ni =Ni S2j 1; Nj =Nj is the one1/k
) CI half width. The SWC procedure is
tailed P( (P*)
o
as follows.
4. Selection-with-constraints procedure
1. Let Ni,t be the sample size allocated for system i and
i,t[g] be the sample mean of the gth performance
X
measure of system i at the tth iteration. Simulate n0
samples for all systems. Set the iteration number t 0,
and N1,t N2,t y Nk,t n0. Specify the value of
the indifference amount d*[g] for g 1, 2, y, o, the
soft constraints m0[g] for g 2, 3, y, o, and the
required precision P*. Let M be the set of systems that
are still under consideration and is initialized to include
all k systems. Compute P*
o 1(1P*)/o and
1/k
.
The
critical
constant
h1 is obtained with
P (P*)
o
k, n0, and P*.
o
2. Calculate sample means and sample variances for each
performance measure of each system. Obtain the index
i[1]), that is the system having the
b1( arg miniAMX
smallest sample means of the primary performance
measure.
3. Calculate the new sample size Ni,t 1[1] max
(n0, J(h1Si(1,Ni,t)/di[1])2n), for iAM and Ni,t 1[g]
max(n0,J(2tP,Ni,t1Si(g,Ni,t)/di[g])2n), for iAM and
g 2, 3, y, o. Here di[1] is computed according
to (7), di[g], for g 2, 3, y, o, is computed according
to (5), and S2i (g,Ni,t) is the sample variance of the
gth performance measure of system i with sample
size Ni,t.
p
4. Let wi0 g tP;Ni;t 1 Si g; Ni;t = Ni;t . For each system
i[g]4m0[g] wi0[g] for some g 2, 3, y, o,
iAM, if X
then remove system i from M. For each system iAM, if
i[g]om0[g]wi0[g] for some g 1, 2, y, o, then set
X
Ni,t 1[g] Ni,t[g].
5. For each iAM having Ni,t 1[g] Ni,t[g], for g
2, 3, y, o, then for each jAM compute wij
q
i[1]
j[1]4X
tP;fi;j S2 1; Ni;t =Ni;t S2j 1; Nj;t =Nj;t . If X
i
0
9. Simulate additional min(Ni,t
1,J(Ni,t 1Ni,t) /2n)
samples for systems iAM. Set t t 1. Go to step 2.
i[1].
10. Select system b1 miniAMX
In the case that there is only one secondary performance
measure, a tighter lower bound can be achieved. Let b1 be
the probability that an unacceptable system is included in the
subset and let b2 be the probability that the best feasible
system is not selected when compared to other feasible
systems in isolation. When b1 b2 b, Andradottir and
Kim (2007) show that
PCSX21 bk1=2 1 b P
(ie Xis).
We tested the following two congurations:
Setting 1: m1 0, and mj d*, for j 2, 3, y, k.
Setting 2: mi 0, for i 1, 2, y, 5, and mj d*, for
j 6, 7, y, k.
jai
5. Empirical experiments
In this section, we present some empirical results of
performing restricted subset selection and SWC.
P* 0.95
v1
v5
v1
v5
RSS
PCS1
PCS2
m
T
std(T)
0.9813
0.4633
2.12
2967
303
1.0
0.9580
1.06
20947
2163
0.9973
0.5979
1.80
4986
514
1.0
0.9795
1.03
26214
2675
SRSS
PCS1
PCS2
m
T
std(T)
Iter
0.9866
0.4886
2.10
2664
310
10
1.0
0.9753
1.03
20864
248
13
0.9992
0.6314
1.74
4749
351
11
1.0
0.9912
1.01
26215
261
14
P* 0.95
v1
v5
v1
v5
RSS
PCS1
PCS2
m
T
std(T)
0.9331
0.1254
3.11
2967
303
1.0
0.9132
4.89
20947
2163
0.9761
0.2888
3.70
4986
514
1.0
0.9570
4.95
26214
2675
SRSS
PCS1
PCS2
m
T
std(T)
Iter
0.9456
0.1335
3.18
2570
265
10
1.0
0.9323
4.92
20850
264
13
0.9845
0.3128
3.78
4633
321
11
1.0
0.9749
4.97
26206
266
14
b 12
b 10
b7
b3
b1
a0
a1
a3
a6
a 10
a 12
SWC
PCS
T
std(T)
Iter
0.954
4971
445
13
0.955
4881
434
12
0.953
4702
416
12
0.959
4429
380
12
0.965
4058
351
11
0.972
3869
347
9
AK
PCS
T
0.973
4063
0.973
4109
0.973
4184
0.974
4319
0.975
4502
0.975
4604
AK
PCS
T
d
Iter
0.960
3976
140
0.962
3749
130
0.963
3726
130
0.963
3686
128
0.966
3615
125
0.968
3581
124
d row under AK
different random number streams. The Iter
lists the lower bound of the number of iterations. The
observed P(CS)s are all greater than the specied nominal
value. Procedure AK is a two-phase procedure; it performs the feasibility check on the secondary performance
measure before it performs the optimization on the primary
performance measure. On the other hand, AK and SWC
perform the feasibility check and optimization simultaneously. Even though the allocated sample sizes of SWC are
greater than those of AK , SWC requires far fewer
iterations and may take less runtime. For example, when
b 1 SWC terminates, on average, with nine iterations. On
the other hand, AK terminates, on average, with no less
than 124 (ie 3581/2520) iterations.
6. Conclusions
We have presented two restricted-subset-selection procedures that provide effective means for screening a large set of
systems. These procedures are versatile, easy to apply, and
Appendix
Given v, m, k, n0, and P*, let h be the value such that
Z1
1
#
X~j mj X~u mu
mu mj
o
Pj P
d =2h d =2h
d =2h
Z1
1
1
01
1
Z1
Z
Ft 2hgk1;1 tdtA
@ Ft 2hgk1;m tdt
1
Z1
1
Z1
1
1
Z1
1
1
Z1
Ft 3hgk1;1 tdt
1
Pe X
Pb PX~i1 d =2oX~b2
"
#
X~i1 mi1 X~b2 mb2
mb2 mi1
d =2
P
o
d =2h
d =2h
d =2h d =2h
mb2 mi1
P Ti1 oTb2
h
d =2h
Z1
mb2 mi1
h gk1;1 tdt
F t
d =2h
1
1
Z1
1
Z1
Z1
1
1
References
Ft hgk1;1 tdt
1
Chen EJ (2008). Restricted subset selection. In: Mason SJ, Hill RR,
Moench L, Rose O, Jefferson T and Fowler JW (eds).
Proceedings of the 2008 Winter Simulation Conference, IEEE;
Piscataway, NJ, pp 281289.
Chen EJ (2009). Subset selection procedures. J Simulation. To
Appear.
Chen EJ and Kelton WD (2005). Sequential selection procedures:
Using sample means to improve efciency. Eur J Opl Res 166(1):
133153.
Chen EJ and Kelton WD (2007). A procedure for generating batchmeans condence interval for simulation: Checking independence and normality. Simulation: Trans of the Soc for the
Modeling and Simul International 83: 683694.
Chen EJ and Lee LH (2009). A multi-objective selection procedure
of determining a Pareto set. Vienna, Vol. 1., pp. 478486.
Comput Opns Res 36: 18721879.
Dudewicz EJ and Dalal SR (1975). Allocation of observations
in ranking and selection with unequal variances. Sankhya B37:
2878.
Gupta SS and Santner TJ (1973). On selection and ranking
proceduresA restricted subset selection rule. Proceedings of the
39th Session of the International Statistical Institute, Vol. 1.,
Vienna, pp 478486.