Вы находитесь на странице: 1из 11

Approximation of Pareto Optima in Multiple-Objective, Shortest-Path Problems

Author(s): Arthur Warburton


Source: Operations Research, Vol. 35, No. 1 (Jan. - Feb., 1987), pp. 70-79
Published by: INFORMS
Stable URL: http://www.jstor.org/stable/170911 .
Accessed: 20/03/2014 04:28
Your use of the JSTOR archive indicates your acceptance of the Terms & Conditions of Use, available at .
http://www.jstor.org/page/info/about/policies/terms.jsp
.
JSTOR is a not-for-profit service that helps scholars, researchers, and students discover, use, and build upon a wide range of
content in a trusted digital archive. We use information technology and tools to increase productivity and facilitate new forms
of scholarship. For more information about JSTOR, please contact support@jstor.org.
.
INFORMS is collaborating with JSTOR to digitize, preserve and extend access to Operations Research.
http://www.jstor.org
This content downloaded from 136.152.209.151 on Thu, 20 Mar 2014 04:28:44 AM
All use subject to JSTOR Terms and Conditions
APPROXIMATION OF PARETO OPTIMA
IN MULTIPLE-OBJECTIVE, SHORTEST-PATH PROBLEMS
ARTHUR WARBURTON
University of Ottawa, Ontario, Canada
(Received December 1983; revisions received September 1985, February, May 1986; accepted June 1986)
We study methods for approximating the set of Pareto optimal paths in multiple-objective, shortest-path problems.
Known generalizations of standard shortest-path methods will compute this set, but can suffer from rapidly increasing
computational and storage demands as problem size increases. In an effort to avoid such difficulties, we develop
approximation methods that can estimate the Pareto optima to any required degree of accuracy. The approximation
methods are "fully polynomial"; that is, they operate in time and space bounded by a polynomial in problem size and
accuracy of approximation-the greater the accuracy, the more time required to reach a solution. We show how
approximation methods may be applied to yield fully polynomial approximation schemes for a variety of NP-complete,
single-objective problems.
Given the remarkable variety of problems that
have been successfully formulated and solved
by shortest-path methods, and the many problems
that involve more than one objective function, it is
natural to consider multiple-objective, shortest-path
(MOSP) models with several objective functions as-
sociated with each path. Normally, no path is optimal
with respect to all objective functions simultaneously;
one tries, instead, to determine efficient (Pareto opti-
mal, nondominated, minimal) paths of interest. The
ability to identify such paths has obvious relevance to
decision problems involving conflicting goals. Further,
a variety of interesting single-objective optimization
problems can be solved by identifying optimal solu-
tions with efficient paths in a related MOSP problem.
Typical examples include
(i) the solution of shortest-path problems in which
the set of allowable paths is restricted in various
ways; for example, Cox (1984), Aneja, Aggarwal
and Nair (1983), Warburton (1982), Lawler
(1976) and Yucaoglu (1973);
(ii) the solution of nonlinear knapsack problems
(Morin and Marsten 1976);
(iii) computation of utility-maximizing paths in
graphs with stochastic arc lengths (Loui 1983).
Methods for determining efficient paths, which have
been described in various contexts by a number of
authors (see Section 1), form part of a very extensive
literature on multiple-objective optimization. These
methods have features common to Pareto optima-
generating methods in general. First, the efficient set
tends to be poorly structured and, to identify efficient
alternatives "of interest," it may be necessary to gen-
erate the entire efficient set or a substantial part of it.
Second, the efficient set may be very large (possibly
exponential in size) and computational and storage
requirements tend to increase rapidly with increases
in problem size and number of objective functions.
For shortest-path problems, at least, these difficul-
ties are intrinsic, as evidenced by the fact that even
the following two-criteria decision problem is NP-
complete (Garey and Johnson 1979):
(Q) Given a directed graph G = (N, A), positive
integer lengths v(a) and weights w(a) for each arc
a E A, specified nodes s, t E N, and positive
integers V, W, does there exist a simple (s, t)-path
in G with total weight W or less and total length
V or less?
Given the enormous-but unsuccessful-effort that
researchers have devoted to solving NP-complete
problems in polynomial time, it is unlikely that it will
ever be possible to design an algorithm that answers
(Q) in polynomial time. Further, it is easy to construct
examples for which the list of efficient (s, t)-path
length vectors (v, w) has a length that is exponential
in the number of nodes of G.
In view of the rather negative results just presented,
we investigate the possibility of using approximation
methods to reduce the heavy computational demands
inherent in MOSP problems. Previous approaches
along this line have been restricted to approximations
obtained by considering weighted linear combinations
Subject classification: 483 multiple-objective shortest paths, 651 approximation of Pareto optima.
Operations Research
0030-364X/87/3501-0070 $01.25
Vol. 35, No. 1, January-February 1987 70 ?
1987 Operations Research Society of America
This content downloaded from 136.152.209.151 on Thu, 20 Mar 2014 04:28:44 AM
All use subject to JSTOR Terms and Conditions
Multiple-Objective, Shortest-Path Problems / 71
of objective functions and solving a parametric family
of single-objective, shortest-path problems (Henig
1982, White 1982c). The approximating set so ob-
tained is a subset of the efficient set and can be
determined by standard shortest-path methods. Un-
fortunately, such an approximation has a serious dis-
advantage-large sections of the efficient set may be
"missed" completely (and we do not know which
sections). Consequently there is no precise notion of
how well the points that are computed actually rep-
resent the efficient set.
To avoid this problem, we develop a different type
of approximation-rather than computing a subset
of the efficient paths, we consider paths that are, in
a well-defined way, "approximately" efficient. We
determine conditions under which it is possible to
approximate the set of efficient paths to any desired
degree of precision, e. For the cases in which the
conditions are met, we present fully polynomial
approximation schemes (FPAS) for computing the
approximating sets. Specifically, we show that an
FPAS exists for problems having any fixed number of
criteria. Thus, from the point of view of computa-
tional complexity theory, MOSP problems are among
the most tractable of NP-hard discrete optimization
problems (Garey and Johnson). On the other hand,
our results also indicate that practical e-approximate
algorithms are generally limited either to problems
having 2 or 3 criteria, or to problems requiring the
e-approximation of only certain restricted sets of
efficient paths. For such problems, our methods can
be employed for the relatively rapid approximate
generation of trade-off curves and surfaces.
We illustrate the application of approximate
MOSP procedures to single-objective optimization
problems by presenting e-optimal algorithms both for
the problem of finding shortest paths subject to a side
condition, and for the problem of finding paths that
minimize the maximum of several linear path length
functions. Special cases include a variety of knapsack
problems and multiprocessor scheduling problems for
which FPAS have previously been developed (Ibarra
and Kim 1975, Horowitz and Sahni 1976, Sahni 1976,
Lawler 1979).
The remainder of this paper is organized as follows.
In Section 1 we formally define the MOSP problem,
present a typical algorithm for solving it exactly, and
discuss the essential features of exact MOSP algo-
rithms. We use the algorithm for illustrative purposes
throughout the paper. In Section 2, we show how
exact MOSP algorithms can have exponentially in-
creasing computational demands. This discussion
leads to a consideration of approximation methods in
Section 3, where we combine scaling and rounding
techniques (e.g., Sahni and Lawler) with an exact
MOSP algorithm and obtain, for positive data, an
FPAS. In Section 4 we extend our methods to prob-
lems in which the data are not necessarily positive. In
the final section, we discuss applications to single-
objective optimization problems.
1. Exact Computation of Efficient Sets
We now present some definitions and notation. Sup-
pose that x =
(xi,
.. . , x,)
and
y
=
(yi,
. . ., Yr) are
real r-vectors. If xk
<
Yk for 1
<
k
<
r, then x dominates
y, and we write x
-
y. Let X be a finite set of r-vectors,
and call x E X efficient if there is no y E X, y $ x,
that dominates x. X* denotes the set of efficient
vectors of X. Let G =
(N, A) be a directed graph with
node set N =
I
1, ..., n
I
and arc set A. Associate an
integer arc length vector c(a)
=
(c'(a), . .. , cr(a)) with
each arc a E A, and extend c to subsets S C A be
defining Ck(S)
=
XaeS ck(a) for 1
<
k < r, and c(S) =
(c'(S), . . . c cr(S)). For collections 5 of sets of arcs,
define c(Y) = {c(S) I S E 51. S E Y is efficient
if S E 5* =
IS
I
c(S) E (c(Y))*}. Call node 1 the
source, and node n the destination. For 1
<
t <
n, t
denotes the set of all simple (1, t)-paths in G, and
Ct
=
c(9t) is the set of (1, t)-path-length vectors (we
set 9, = 1}, C,
=
10}).
"0" denotes the zero vector.
We single out the (1, n)-paths for special attention,
and write s =
n, C= Cn .
The MOSP problem can now be formally defined
as that of identifying C*. We might also require *,
or a subset M of ?*, satisfying c(S) = C*.
Algorithms that solve the MOSP problem can be
constructed in a variety of ways. We will choose only
one algorithm for detailed consideration and will refer
to it throughout the remainder of this paper. Subject
to a certain restriction (to be described later), any
other algorithm that solves the MOSP problem could
be substituted in the discussion.
The most direct way of obtaining an algorithm that
solves the MOSP problem is to invoke a "multiple-
objective" version of the dynamic programming op-
timality principle that dates at least to Brown and
Strauch (1965). In our context, the principle states
that at each stage we retain the set of presently non-
dominated path length vectors (rather than the current
shortest-path lengths). The principle permits an al-
most direct transcription of standard shortest-path
algorithms.
For simplicity, suppose that G is a complete acyclic
This content downloaded from 136.152.209.151 on Thu, 20 Mar 2014 04:28:44 AM
All use subject to JSTOR Terms and Conditions
72 / WARBURTON
graph. Then we can assume that the nodes of G
are numbered so that an arc (i, j) from node i to
node j exists if and only if i <
j.
Let
cij
=
(Cl, ..., cij) denote the corresponding arc length
vector. If r = 1, the problem is a regular shortest-path
problem, and the length u* of a shortest (1, n)-path in
G can be found in 0(n2) time, using the well-known
recursion
U,*= 0,
u* =min
Iui
+ c!-b, 2
-
j
< n.
i<j
The corresponding MOSP algorithm can be written:
Algorithm ACYCLIC
C,*=
10},
C*
= MIN
IC
+
c1j,
2 j
-
n.
i<j
In this statement, C0 +
cij
consists of all vectors of the
form ci +
cij
with ci E C0*. "MIN" means that, at each
stage j, we retain only the efficient vectors in the set
E1= U (C* + c). (1)
i<j
Note that, to compute C*, we require all the sets
Cj*
for j
< n. Using these sets, we can recover a path
P E
*
corresponding to a given c E C* = C* in
O(rn) operations, using a simple labeling procedure
suggested by Hu (1969).
The amount of work required to implement a re-
cursion such as ACYCLIC clearly depends upon the size
of the sets
Cj*
and the manner in which "MIN" is
performed. Let
WjY
=
I for 2 <
j
<
n. At step
j, the set
Ej
of (1) contains Wj r-vectors (assuming
there are no ties) and requires rWj additions for its
generation. Cj*
=
MIN(Ej)
must then be found. This
computation may be accomplished in a variety of
ways. Consider first the case r = 2 (i.e., we seek a
trade-off "curve," rather than some higher dimen-
sional set). We can maintain each set C0 as a list of
vectors sorted by the first coordinate, and merge the
sets C0 +
cij by pairs to form
Ej. Cj*
can then be found
in a single pass through
Ej,
implying an
0(WjVlog j)
computation. If r
>
3, finding Cj*
is more computa-
tionally demanding. Most easily programmed is the
naive method:
Ej
is generated element by element,
beginning with C* +
c,j
=
fclj}
and an initial non-
dominated list I =
Vj
. As each vector x is generated,
it is compared with the vectors in the list, I, of cur-
rently nondominated vectors, yielding an updated list
(I U {x})*. Thus, computation of
Cj*
requires rWj
additions and at most 0(rWJ) comparisons. The
rWJ terms may be improved to
O(Wj(log j)r-2)
using a divide and conquer algorithm of Kung, Luccio
and Preparata (1975), which provides the best-known
worst-case bound, O(v(log V)r2), for finding the effi-
cient vectors in a set of v r-vectors. For simplicity, for
r > 3 we will assume in our complexity estimates that
the naive method is used to compute efficient vectors.
Improved bounds can be obtained by using the Kung,
Luccio and Preparata algorithm.
Since the computation of C* requires all C* for
<j, we obtain the following upper bounds on the
overall time complexity of ACYCLIC:
r=2, O((2W2+ W21og 2) + ... + (2 W + Wnlog n));
rB 3, O((rW2+
rW2)
+ +
(rWn
+
rW2)).
Space requirements are
O(rW,), excluding problem
data.
These bounds are very dependent upon the cardi-
nality of the sets C*. To illustrate the dependency, let
Vmax
=maxi I
C1 |, , I C* }. (2)
Utilizing the fact that
Wj
<
i Vmax for
j
, 1, we obtain
from the bounds for ACYCLIC the following very pes-
simistic performance guarantees:
r = 2, O(n2Vmaxlog n);
r 3, O(rn3V2ax).
Space requirements are bounded by O(rn Vmax), r > 2.
The complexities derived for ACYCLIC are typical
for MOSP algorithms in general. Complexity is
bounded by polynomials in n, r and Vmax, and the
algorithms are reasonably efficient provided Vmax re-
mains small. If G is not a complete graph, implemen-
tations taking advantage of the sparsity of G may be
possible. For example, the
0(I
A I) implementation
of ACYCLIC for r = 1 can be generalized to r > 2 by
using an appropriate sparse representation of G and
noting that in (1) we need compute only one set
C* +
cij
for each arc (i, j) terminating at node j.
Daellenbach and de Kluyver (1980) and Henig
have both described ACYCLIC. Warburton (1983) and
Loui generalize the Bellman-Ford (1957) method;
Loui generalizes Spira's (1973) algorithm, and
Warburton (1982) describes a MOSP version of
Floyd's (1962) algorithm.
Another approach for generating C* (or certain
subsets or supersets of it) is to treat r - 1 of the
objective functions as extra state variables and to solve
a family of restricted shortest-path problems via a
standard dynamic programming recursion. See Henig,
and Lawler (1976, p. 92) for a discussion of the case
when r =
2. See White (1982a), Cox et al. (1980),
Hartley (1979), Thuente (1979), and Randolph and
This content downloaded from 136.152.209.151 on Thu, 20 Mar 2014 04:28:44 AM
All use subject to JSTOR Terms and Conditions
Multiple-Objective, Shortest-Path Problems / 73
Ringeisen (1975) for further discussion of MOSP
methods.
Vector Maximization. If the object is to maximize
one or more components of c(P), MOSP algorithms
can be applied to compute efficient paths only if G is
acyclic. If G contains cycles and r = 1, computing C*
solves the NP-hard longest-path problem; if r = 2,
we can easily show that identifying C* is at least
as difficult as finding a shortest Hamiltonian path
between node 1 and node n (a strongly NP-hard
problem).
Generation of Subsets of Efficient Vectors. Let
d(a) = (dl(a), ..., dr(a)) for a E A be nonnegative
integer arc length vectors defined on the arc set A,
and define
Dj
= {d(P)IPE Aj
Ifor
1 j
<
n. The
approximation procedures will require a method for
calculating D*, where
Dj
for 1
<
j < n is defined
as the intersection of
Dj
with the set Z of vectors
(xl, . .. , xr) satisfying
0 <
Xk
<
Mk (3)
for given constants Mk for 1 < k
<
r - 1. Nonnegativity
of the problem data implies that the following modi-
fication of ACYCLIC yields the required set. It differs
from ACYCLIC only in that it eliminates vectors not in
Z whenever they are encountered.
Algorithm RESTRICT
f=
MIN{Zn(D*+dij)b,
2<j n.
<IJ
At step j, vectors not in Z can be eliminated in at
most
Wj
= ID* I + * + ID *I
I
vector compari-
sons. Hence time and space bounds for RESTRICT
may be obtained from the bounds for ACYCLIC
simply by replacing Wj
and Vmax by Wj and Vmax
=
max{ I
1, 1.., I D*
I
}, respectively. Further, inte-
grality and nonnegativity of the data together with
the definition of Z imply
Vmax S (Ml
+ 1) - (Mr-I
+ 1). (4)
Remark. The RESTRICT version of a MOSP algorithm
is normally applicable provided problem data are
nonnegative. For example, MOSP analogues of the
Bellman-Ford and Floyd algorithms are valid for prob-
lems containing no negative cycles with respect to
any objective function (Warburton 1983), and the
RESTRICT versions are valid for those instances in
which all problem data are nonnegative. Again, we
merely delete vectors not in Z whenever they are
encountered.
2. Size of the Efficient Set
In the previous section we showed how computational
requirements for a MOSP algorithm depend upon the
cardinality of Vmax. We now consider possibilities for
the growth of Vmax.
First, note that an upper bound on Vmax is easily
obtained as a function of r and the length of longest
paths in G. In fact, if the arc length vectors are
nonnegative, then integrality implies the very conserv-
ative bound Vmax < B"1, where B ? Ck(P) + 1 for all
P E
j
for 2
-
k < n. Since we can always choose
B = (n - 1)cmax + 1, where cmax is the largest data
element, ACYCLIC operates in time and space bounded
by a polynomial in n and cmax. Thus, for fixed r,
ACYCLIC is pseudopolynomial (Garey and Johnson). It
does not qualify as a true polynomial algorithm since
running time and storage requirements are not
bounded by a polynomial in the length of an encoding
of the problem data. For example,
0(f2(log(cnaj))r)
would be a polynomial bound, but not O(n2CMax). If
B or r is large, the pseudopolynomiality of our pro-
cedure is not very encouraging, especially since con-
version of rational data to integer data can lead to
large values of B. In fact, by allowing integer arc
lengths as large as 2n1, we can construct examples for
any r > 2 with IC* =
I9,I
= 2n', even for an
acyclic graph with n nodes and 2n arcs. Further,
even if all data are selected from a small data range,
C*
I
can increase
exponentially
with the number of
criteria. It is easy to construct an r-criteria example
with n + 1 nodes, nr arcs, and having all data elements
either 0 or 1, for which I Cn I nI
=
(n+?')
3. Approximation of Efficient Sets
The previous discussion established the possibility of
heavy, and even exponential, computational demands
for MOSP algorithms. We now explore the possibility
of reducing the workload by generating sets that are
merely approximately efficient. In this section, all arc
length vectors are assumed to be positive and integral.
First we require a notion of approximate efficiency
(Orlin 1982, White 1982b). Let x
=
(x,, .. ., Xr) and
Y = (Yi,
.
.., Yr) be positive r-vectors and suppose an
"'accuracy" e
>
0 is specified. Then x e-dominates y if
Xk
-<
(1 +
) yk
for 1
<
k
<
r. For e
=
0, this concept
reduces to the usual notion of vector dominance. Now
let X be a set of positive r-vectors, and suppose
Y C X. Then Y(E) c X is E-efficient for Y, provided
that
(i) each vector in Y is E-dominated by some vector
in Y(E), and
This content downloaded from 136.152.209.151 on Thu, 20 Mar 2014 04:28:44 AM
All use subject to JSTOR Terms and Conditions
74 / WARBURTON
(ii) no vector in Y(E) dominates any other; that is,
Y(E) = (Y(E))*.
Condition (ii) ensures that the E-efficient set is not too
large. In particular, if X = Y and E = 0 then X* is the
unique 0-efficient set for X. Note that we do not
require Y(E) C Y if Y $ X.
An approximation scheme for computing an E-
efficient set for Y C C is an algorithm that receives
two inputs: an encoding of the problem (i.e., a suitable
representation of G and an encoding of the arc length
vectors), and a number e > 0 prescribing the degree of
accuracy required. The algorithm determines a set
Y(E) C C that is E-efficient for Y. If it operates in time
and space bounded by a polynomial in
I/E
and the
length, W, of the encoded input, the algorithm is an
FPAS.
Consideration of an E-efficient set C(E) for C offers
immediate advantages. First, whereas I C* I may grow
exponentially with W, we will show that C(E) can be
determined by an FPAS, and so may be computable
much more quickly than C* itself. Second, we will see
in Section 5 that appropriate modeling techniques
enable us to obtain fully polynomial approximation
schemes for certain single objective optimization
problems by identifying E-optimal solutions as ele-
ments of E-efficient sets for corresponding MOSP
problems.
A Basic Approximation. Suppose we are given a pos-
itive integer vector L = (L1, . .. , L,_ ) and a positive
vector 0 = (61, ...,
6tR,1),
with Ok > 1 for 1 < k <
r - 1. Let CL C C be the image under c of the set,
5, of all P 8 SD satisfying
Lk
<1
Ck(p) <1 OkLk,
I <, k <, r-
1.(5)
Choose E > 0. We first determine an E-efficient set
for CL.
For each arc a E A, define a scaled, nonnegative
integer arc length vector d(a)
=
(d'(a), . . ., dr(a)) by
dr(a) = cr(a), (6)
dk(a)
cLka) 1 <k <r- 1,
(7)
where
k
ELk_
k l r-1 (8)
T
n- 1
and ltJ denotes the integer part of t. Because
ck(a) k
c(a
Tk T1
d (a)
<Tk
k< r- 1, (9)
we have
Ck(P) - (n - 1) Tk
<
Tk dk(P)< Ck(P),
1<,k,r-1, (10)
for any P E8-
Now consider the set ?L of all P E SD for which
d k(p) '< (n
-
1 )
Ok
1
<
k
<
r- 1.
Let ML C ?-L be in one-to-one correspondence with
(d(gL9))* in the following way: for each u E
(d(5L))*9
there is exactly one P E ML satisfying d(P) = u. We
claim that C(WL) is e-efficient for CL. To see this, first
note that
(d (L))
=
d(qL)
=
(dG(9L))* (12)
The condition (12) and the definitions of ML and d
imply that C(WL) = (C(WL))*. It remains to show that
for each v E CL there is a u E C(ML) satisfying
Uk -<(- +
E)Vk,
I < k < r.
(13)
To demonstrate ( 13), first observe that by (5), (8) and
(10) we have
S>L C ?'L. (14)
Hence, if Q E 5L, there exists P E 5?L such that
d(P) E (dQ(L))*
and
dk(P) < d
k(Q), 1 < k < r.
(15)
Further, by (12), we may assume P E ?L. Then (6)
gives
cr(P) <
cr(Q), (17)
and (10), (8), and (5) imply
Ck(P) <
ck(Q) + (n
-
1)Tk,
= Ck(Q) + ELk,
< (1 + E)ck(Q), I < k < r - 1. (18)
Letting u =
c(P) and v
=
c(Q) we see that (17) and
(18) imply (13), as required.
Computing C(?L). According to the previous discus-
sion, the E-efficient set C(ML) for CL may be found by
first determining (d(57L))*, then finding a correspond-
ing set of paths WL, and finally calculating C(JL).
The computation of (dG(9L))* is accomplished
merely by applying RESTRICT to the nonnegative data
set defined by (6) and (7) and taking
Mk
=
(n
-
1)-, 1 < k
<
r - 1,
E
This content downloaded from 136.152.209.151 on Thu, 20 Mar 2014 04:28:44 AM
All use subject to JSTOR Terms and Conditions
Multiple-Objective, Shortest-Path Problems / 75
in (3). At the final step we have
(
))
=
D*
6RL can then be found by applying the path-finding
procedure to each element of Dn*. However, it is not
necessary to compute 3L in order to find C(G4L). We
save all remainders ck(a)/Tk- Lck(a)/TkJ, and accu-
mulate them in a computation parallel to the com-
putation of D5*. The elements of C(S?L) are found
by adding the accumulated remainders to vectors in
Dn* and then multiplying by the appropriate scale
factors. Thus, except for a constant factor, C(WL) can
be computed within the same time bounds as Dn*.
According to (4), and our choice of Mk, we have
Vmax <f(n, E. 0),
where 0
=
(6k, . 0, or-)
and
f(n,
E,
0)
= (l + (n-
l)-
I
)...(l + (n -1) I (9)
The pessimistic complexity bounds for RESTRICT then
imply that C(WL) is computable in O(n301E-1log n)
time and O(n201E-1) space if r = 2. For r > 3, we
obtain worst-case guarantees of O(rn3(f(n,
E9 0))2)
time and O(rnf(n, E, 0)) space. The time and space
bounds are now independent of the arc lengths and,
for fixed r and 0, are polynomial in n and 1/E. Thus,
for fixed r and 0, we have a FPAS for computing
E-efficient sets for CL. (Any other pseudopolynomial
algorithm for solving the MOSP problem could be
substituted for ACYCLIC, provided its RESTRICT version
is valid.)
When to Scale. Note that scaling in (8) is advanta-
geous only if Lk is sufficiently large. In fact, there is
no point in scaling any component k for which
Lk S
n
(20)
In such a case, (8) implies that the scaling constant Tk
is less than unity, and (11) holds automatically for
component k. If (20) holds for 1 s k ? r - 1, then
computing the 0-efficient set C* defined by C* =
C* n R(lX,
9 Xr) I Xk
1<
OkLkg
I <
k s r
-
1 } with
RESTRICT already constitutes an FPAS. On the other
hand, computing the approximation C(GWL) using scal-
ing techniques is never of a higher order of difficulty
than the direct computation of C*, since scaling and
rounding can only reduce the size of efficient sets.
Remark. The E-approximation technique can easily
be modified to deal with those situations in which E =
(El, ... Er)
is a vector of
"approximation
factors."
Then x E-dominates y if Xk < (1 + Ek)Yk for 1 <
k < r. The kth component of vectors in X* is closely
estimated if Ek iS small, and may be effectively ignored
by choosing Ek sufficiently large. Thus, in some sense,
the Ek represent criterion "weights."
The General Method for Positive Data. An FPAS for
C* can be obtained by computing E-efficient sets C(ML)
for each member of an appropriate family F of sets
c(5'L) of the form (5), chosen so that U {5L I S>L E
F)
= 6P. To develop such a family, suppose that B is large
enough to ensure that
I C k(p) <B, I1 sk <r,
for any simple path P in G. Let m be the smallest
integer satisfying 2m > B; that is, m = Flog B1 where
Ft] is the integer "round up" of t. Define j% = 1,
Oj
=
20j-l
= 2i for 1 <
j
<
m,
and let I denote the
(r - 1)-fold Cartesian product of 10, ..., m - 1}.
For each (ml, ...,
m(rl))
E I, consider the vector
L =
(13m1,
.
-
13f(r-i))'
and note that 93 is the union of sets 5'L of the form
(5) with each 0k = 2, where the union is taken over all
vectors L indexed by L For each such 5L, we consider
CL
and
compute
the
corresponding
E-efficient set
C(WL) using the method previously described. Since
each C(WL) is E-efficient for CL, C(E) = U CL contains
an e-efficient set for C where, again, the union is taken
over vectors L indexed by I. (C(e))* is an e-efficient
set for C.
The computation of C(E) requires mrl = [log B 1
applications of the basic approximation. For r = 2,
the resulting worst case time requirement is
O(n3E-'log nFlog B1). If r > 3, we obtain the time
bound O(rn3f 2rlog B lr-l) where, by (19),
f=f(n, E, (2, ..., 2)) = + 2(n - 1))r
Since we may have to retain an e-efficient set for each
CL, C(e)
may
contain as
many
as
tr
=
frlog
B]r-I
r-vectors. Thus for r > 2 storage requirements are
0(tr + rnf ), excluding problem data. For r =
2, (C(E))*
can be found in O(t2log log B) time by maintaining
each C(2?L) as a list sorted by first coordinate, merging
O(log B) lists by pairs, and then making a single pass
through the merged list. If r >
3, (C(e))* can be
determined in O(rt2) time. For fixed r, all time and
space bounds are polynomial in n, log B, and 1/E.
Since we can always choose B
=
(n -1 )Cmax we have
an FPAS for computing E-efficient sets for C.
The FPAS is limited by the fact that it requires
Flog Blr-1 applications of the basic approximation.
This content downloaded from 136.152.209.151 on Thu, 20 Mar 2014 04:28:44 AM
All use subject to JSTOR Terms and Conditions
76 / WARBURTON
For example, if B = 32, r 5, we already require
54
=
625 applications! The Flog B] r-1 factor can
easily be reduced by performing shortest-path cal-
culations to bound vectors in C*, and further
improvements may well be possible. However, com-
putational complexity theory shows that, if r is not
fixed, a FPAS for the MOSP problem probably does
not exist. This negative conclusion is implied by the
fact that "MULTIPROCESSOR SCHEDULING"
with a variable number of processors, a strongly
NP-complete problem (Garey and Johnson) can be
reduced to a special case of a MOSP problem.
Practical c-approximation, in general, appears to be
confined to problems that require only a few applica-
tions of the basic approximation and such that each
application is of "reasonable" complexity. For exam-
ple, if r = 2 or 3, and both log B and n/E are moderate,
then an e-approximation of the entire set C* can be
considered. On the other hand, a single application of
the basic approximation can require that time and
space increase exponentially with r for fixed n/E, thus
limiting e-approximation even for the sets CL given
by (5).
The advantages of scaling are most apparent when
B is large. For example, suppose r
=
2. An exact
computation of C* is O(n2B log n), and an E-efficient
set can be obtained in time 0(n3'-'log nrlog B 1). The
ratio of the bounds is of the order EB/n log B, implying
that for any fixed n and c, the approximate method
could be exponentially faster for large B.
Finally, we note that fully polynomial approxima-
tion schemes can be constructed for maximization
versions of MOSP problems, provided G is acyclic.
On the other hand, if G contains cycles, we have
already noted that the strongly NP-hard problem of
finding a shortest Hamiltonian path is a special case,
implying the unlikelihood of finding an FPAS even
for r = 2.
4. Extension to More General Data Sets
By appropriately transforming the arc length vectors
and shifting the origin, we may extend the methods
of Section 3 to problems whose data are not necessar-
ily positive. Let
u,
= (u , .. ., ur ) be the "ideal" vector
of shortest (1, n)-path lengths (Zeleny 1976). Further,
define v =(v'..., Vr) by v =
(u
-
1,..., u - 1).
Then 1 c ck(P)
-
vk for all P E R. That is, all simple
(l, n)-paths have strictly positive length with respect
to v as the origin. Call a set Y(E, v) C C (e, v)-efficient
for Y c C provided, for each y E Y, there is an
x E Y(E, v) with (Xk - Vk)
<
( 1 + Y)(yk
- Vk) for
1
-
k < r. We also reouire Y(i- W)
=
Y*(e 1). We
wish to compute an (e, v)-efficient set for C. Do-
ing so requires a transformation of the arc length
vectors. Let Uk
=
(Uk ..., Uk) be the vectors of
shortest (1, k)-path lengths, 2 < k
<
n. Set ul -
(Ul, Ur)
=
(OS . 0).
For each arc (i, j) E A, define a new nonnegative
arc length vector c = c11 + ui -
uj. Then, for any
c(P)
= c(P) - Un , 0. (21)
The nonnegativity of all c and the strict positivity of
the Cjk(P) + 1 enable us to compute (, v)-efficient sets
by a slight alteration of the methods of Section 3.
First, consider the basic approximation. By (21) we
can choose constants 0k 1 for 1 < k < r - 1, a
positive integer vector L =
(LI,
. .. , Lr1),
and define
the set 5?L of all simple (1, n)-paths Q satisfying
Lk ck(Q) + I <1
OkLk,
I -,< k < r- 1
Defining new nonnegative arc length vectors
dij by
scaling c'- exactly as in (6) to (8), and considering the
set YL of all simple (1, n)-paths P satisfying
dk(P) (n- 1)
k
I <k r- 1, (22)
C
we can then show, almost as we did in Section 3, that
c (L) is (E, v)-efficient for C(SL), where &WL C L
satisfies
d(L)-(d(L))*. The set (d( L))* may be
calculated using RESTRICT, saving only vectors d that
satisfy (22). Hence the computational complexity is
the same as that of Section 3 once the Uk and
cij
have
been determined by shortest-path computations. A set
that is (e, v)-efficient for c(9) can be obtained, as
before, by repeated approximation over a set of grid
points.
5. Applications to Single-Objective Problems
A variety of optimization problems may be solved,
either exactly or approximately, by identifying optimal
solutions with effilcient path-length vectors. In this
section we provide several examples.
Min-Max Shortest Path. For a set of arcs S C A, let
f(S)
=
maxlc'(S),
c2(S)}. The min-max shortest-path
problem is to findf* and P* E 3
so that
f*
=
f(P*)
=
min{ f(P) I
P
E }.
Note that we can find f* by searching C* since f is
nondecreasing in c' and c2. Suppose R is a shortest
(1, n)-path with respect to c' + c2 and that cl(R) +
c2(R) > 0. Then R E Y* also, and we can assume
without loss of generality that f(R)
=
c'(R)
>
c2(R).
This content downloaded from 136.152.209.151 on Thu, 20 Mar 2014 04:28:44 AM
All use subject to JSTOR Terms and Conditions
Multiple-Objective, Shortest-Path Problems / 77
(tO,t) (?,t2 (tO,
Figure 1. Scheduling on parallel processors.
Let L = (c'(R) + c2(R))/2. Consequently,
L <f(P*)
<
c1(R), (23)
so P* belongs to one of the sets
5i
for i =
1, 2 defined
as ={P E 9 I L <
ci(P)
<
OL}, where 0 =
c'(R)/L
< 2. Calculation of c-efficient sets for c(5S) and c(52)
using the basic approximation yields a simple path-
length vector P that is c-optimal for the min-max
problem. For example, if P* e SN, then we have c2(P)
< c2(P*) and c'(P) < (1 + E)cl(P*), so that f(P) <
(1 + E)f(P*). Space and time requirements with
ACYCLIC are O(n2'-') and 0(n3G-'log n), respectively.
Iff is the maximum of r > 2 path lengths, we obtain
an FPAS by replacing (23) with
cl(R) + ... + cr(R)
f(P*) < c'(R).
r
The problem of nonpreemptively scheduling n jobs
on r parallel processors to minimize makespan may
be viewed as a special case of the min-max shortest-
path problem. Suppose, for simplicity, that r
=
2, and
that
tj
is the processing time of job j
on machine k,
for 1 <
j
<
n and k = 1, 2. Consider the multigraph
G of Figure 1. A min-max (0, n)-path in G solves the
problem, and a shortest (0, n)-path R with respect to
c' + c2 is a shortest-processing-time (SPT) schedule.
The special structure of G yields an
O(n2/A) time and
space calculation. The resulting procedure is essen-
tially that described in Sahni and in Horowitz and
Sahni.
Restricted Shortest-Path Problems. Consider the re-
stricted shortest-path problem
cl(P*)
=
minIc'(P) I P E , c2(P) < K}.
Assuming nonnegative data, RESTRICT solves the prob-
lem by saving only those vectors whose second com-
ponent does not exceed K. This implementation leads
to an O(n2K log n) calculation. For a related approach,
see Lawler (1976, p. 92). Handler and Zang (1980)
solve the problem using Lagrangian relaxation and
kth-shortest paths. Aneja, Aggarwal and Nair com-
bine branch-and-bound with a Dijkstra-type labeling
scheme.
If K is fixed, the problem is solvable in polynomial
time. However, if K is not fixed, the associated deci-
sion problem is NP-complete. We obtain a FPAS by
computing an e-efficient set for C, and so find a path
P
E
-9 satisfying C2(P) < K and c'(P)
<
(1 + E)c2(P*).
If both upper and lower bounds on c'(P*) are
known, it may be possible to obtain an FPAS that
requires a single application of the basic approxima-
tion. Further, a maximization may be considered if G
is acyclic. For example, the 0- 1 knapsack problem
n
c*= max E
c'xi
j=1
n
subject to E
c.2x1
< K
j=l
xj
E [0 I],
I 6,n,
may be modeled as a multiple-objective, shortest-path
problem in which the objective is to maximize cl
while minimizing c2. Referring to Figure 2, we see
that an optimal solution to the knapsack problem may
be found among those vectors in C* whose second
component does not exceed K. Lawler (1979) provides
an easily computed bound L, L <
c* S 2L, that can
be used to obtain an E-optimal solution in
O(n2/A)
time and space by computing an c-efficient set for CL
of (5) with r = 2, 0 = 2. His procedure obtains a path
)
C,' (c 1c'
)(C2 ,Ct
-1
--
0.| _ r0 0 2 (~~~~~~~~~~
0000
n-)
(0.0) (0,0)
(0,0)
Figure 2. The 0-1 knapsack problem.
This content downloaded from 136.152.209.151 on Thu, 20 Mar 2014 04:28:44 AM
All use subject to JSTOR Terms and Conditions
78 / WARBURTON
P with c2(P)
<
K, c'(P) 2 (1 - g)cl(P*). He also
discusses several refinements that considerably im-
prove the efficiency of the method, and shows how to
extend the approximation procedure to a variety of
other knapsack problems.
In the present context, variations on the knapsack
problem may be incorporated in a straightforward
way by adding to Figure 2 arcs, nodes, objective
functions, or a combination. For example, consider a
0-1 multiple-choice knapsack problem with the addi-
tional restriction that no more than L items from
some special set S can be chosen. The multiple choice
constraints are modeled by additional arcs, and an
additional objective function can be used to count
items in the special set. A fully polynomial approxi-
mation scheme is obtained by scaling the profit coef-
ficients. Variations on the scheduling problem can be
handled in a similar way.
Acknowledgment
This research was partially supported by National
Sciences and Research Engineering Council of Canada
Grant A5543. The author gratefully acknowledges the
comments of the referees and Associate Editor.
References
ANEJA, Y. P., V. AGGARWAL AND K. P. K. NAIR. 1983.
Shortest Chain Subject to Side Constraints. Net-
works 13, 295-302.
BELLMAN, R. 1957. Dynamic Programming. Princeton
University Press, Princeton, N.J.
BROWN, T. A., AND R. E. STRAUCH. 1965. Dynamic
Programming in Multiplicative Lattices. J. Math.
Anal. Appl. 12, 364-376.
COX, R., J. THOMAS, C. REVELLE AND J. COHON. 1980.
Methodologies for Multiple Objective Shortest Path
Problems, paper presented at TIMS/ORSA Meeting,
Colorado Springs.
Cox, R. G. 1984. Routing and Scheduling of Hazardous
Materials Shipments: Algorithmic Approaches to
Managing Spent Nuclear Fuel Transport. Ph.D. dis-
sertation, Civil Engineering, Cornell University.
DAELLENBACH, H. G., AND C. A. DE KLUYVER. 1980.
Note on Multiple Objective Dynamic Programming.
J. Opnl. Res. Soc. 31, 591-594.
FLOYD, R. W. 1962. Algorithm 97: Shortest Path. Comm.
Assoc. Comput. Mach. 5, 362.
GAREY, M., AND D. JOHNSON. 1979. Computers and
Intractability: A Guide to the Theory of NP Com-
pleteness. W. H. Freeman, San Francisco.
HANDLER, G., AND I. ZANG. 1980. A Dual Algorithm for
the Constrained Shortest Path Problem. Networks
10, 293-310.
HARTLEY, R. 1979. Dynamic Programming in Vector
Networks. Notes in Decision Theory, No. 86, Man-
chester University.
HENIG, M. 1982. The Shortest Path Problem with Two
Objective Functions. Working Paper 743/82, Fac-
ulty of Management, Tel Aviv University.
HOROWITZ, E., AND S. SAHNI. 1976. Exact and Approx-
imate Algorithms for Scheduling Nonidentical Proc-
essors. J. Assoc. Comput. Mach. 23, 317-327.
Hu, T. C. 1969. Integer Programming and Network
Flows. Addison-Wesley, Reading, Mass.
IBARRA, O., AND C. KIM. 1975. Fast Approximation
Algorithms for the Knapsack and Sum of Subset
Problems. J. Assoc. Comput. Mach. 22, 463-468.
KUNG, H. T., F. LuCCIO AND F. P. PREPARATA. 1975.
On Finding the Maxima of a Set of Vectors. J. Assoc.
Comput. Mach. 22, 469-476.
LAWLER, E. 1976. Combinatorial Optimization: Net-
works and Matroids. Holt, Rinehart, & Winston,
New York.
LAWLER, E. 1979. Fast Approximation Algorithms for
Knapsack Problems. Math. Opns. Res. 4, 339-356.
LoUi, R. 1983. Optimal Paths in Graphs with Stochastic
or Multidimensional Weights. Comm. Assoc.
Comput. Mach. 26, 670-676.
MORIN, T. L., AND R. E. MARSTEN. 1976. An Algorithm
for Nonlinear Knapsack Problems. Mgmt. Sci. 22,
1147-1158.
ORLIN, J. 1982. Fast Approximation Schemes for Com-
binatorial Optimization Problems. Paper presented
at TIMS/ORSA Meeting, San Diego.
RANDOLPH, P., AND R. RINGEISEN. 1975. Shortest Paths
through Multiple Criteria Networks. ERI 76073,
Iowa State University.
SAHNI, S. 1976. Algorithms for Scheduling Independent
Tasks. J. Assoc. Comput. Mach. 23, 116-127.
SPIRA, P. M. 1973. A New Algorithm for Finding All
Shortest Paths in a Graph of Positive Arcs in Average
Time O(n210g2n). SIAM J. Comput. 2, 28-32.
THUENTE, D. 1979. Two Algorithms for Shortest Paths
through Multiple Criteria Networks. Mathematics
Department, Purdue University, Fort Wayne, Ind.
WARBURTON, A. 1982. Solving Restricted Shortest Path
Problems Using a Multiple Objective Version of
Floyd's Algorithm. Proceedings of the 1982 ASAC
Conference, University of British Columbia, Van-
couver, Canada.
WARBURTON, A. 1983. Exact and Approximate Solution
of Multiple Objective Shortest Path Problems.
Working Paper 83-74, Faculty of Administration,
University of Ottawa, Ottawa, Canada.
WHITE, D. 1982a. Optimality and Efficiency. John Wiley
& Sons, New York.
WHITE, D. 1982b. Epsilon Efficiency, Notes in Decision
This content downloaded from 136.152.209.151 on Thu, 20 Mar 2014 04:28:44 AM
All use subject to JSTOR Terms and Conditions
Multiple-Objective, Shortest-Path Problems / 79
Theory, No. 134. Department of Decision Theory,
Manchester University.
WHITE, D. 1982c. The Set of Efficient Solutions for
Multiple Objective Shortest Path Problems. Comput.
Opns. Res. 9, 101-107.
YUCAOGLU, E. 1973. Solution Algorithms for Resource
and Route Constrained Shortest Path Problems in
Time-Dependent Transportation Networks. Ph.D.
thesis, Industrial Engineering, Stanford University.
ZELENY, M. 1976. A Concept of Compromise Solutions
and the Method of the Displaced Ideal. Comput.
Opns. Res. 1, 479-496.
This content downloaded from 136.152.209.151 on Thu, 20 Mar 2014 04:28:44 AM
All use subject to JSTOR Terms and Conditions

Вам также может понравиться