Вы находитесь на странице: 1из 15

Articles

Using Game Theory for


Los Angeles Airport Security

James Pita, Manish Jain, Fernando Ordóñez,


Christopher Portway, Milind Tambe, Craig Western,
Praveen Paruchuri, and Sarit Kraus

P
n Security at major locations of economic or
political importance is a key concern around
the world, particularly given the threat of ter- rotecting national infrastructure such as airports, historical
rorism. Limited security resources prevent full landmarks, or a location of political or economic importance is
security coverage at all times, which allows a challenging task for police and security agencies around the
adversaries to observe and exploit patterns in world, a challenge that is exacerbated by the threat of terrorism.
selective patrolling or monitoring; for exam-
Such protection of important locations includes tasks such as
ple, they can plan an attack avoiding existing
monitoring all entrances or inbound roads and checking
patrols. Hence, randomized patrolling or
monitoring is important, but randomization inbound traffic. However, limited resources imply that it is typ-
must provide distinct weights to different ically impossible to provide full security coverage at all times.
actions based on their complex costs and Furthermore, adversaries can observe security arrangements
benefits. To this end, this article describes a over time and exploit any predictable patterns to their advan-
promising transition of the latest in multia- tage. Randomizing schedules for patrolling, checking, or moni-
gent algorithms into a deployed application. toring is thus an important tool in the police arsenal to avoid
In particular, it describes a software assistant the vulnerability that comes with predictability. Even beyond
agent called ARMOR (assistant for random- protecting infrastructure, randomized patrolling is important in
ized monitoring over routes) that casts this
tasks ranging from security on university campuses to normal
patrolling and monitoring problem as a
Bayesian Stackelberg game, allowing the
police beats to border or maritime security (Billante 2003,
agent to appropriately weigh the different Paruchuri et al. 2007, Ruan et al. 2005).
actions in randomization, as well as uncer- This article focuses on a deployed software assistant agent
tainty over adversary types. ARMOR com- that can aid police or other security agencies in randomizing
bines two key features. It uses the fastest their security schedules. We face at least three key challenges in
known solver for Bayesian Stackelberg games building such a software assistant. First, the assistant must pro-
called DOBSS, where the dominant mixed vide quality guarantees in randomization by appropriately
strategies enable randomization; and its weighing the costs and benefits of the different options avail-
mixed-initiative-based interface allows users able. For example, if an attack on one part of an infrastructure
occasionally to adjust or override the auto-
will cause economic damage while an attack on another could
mated schedule based on their local con-
straints. ARMOR has been successfully
potentially cost human lives, we must weigh the two options
deployed since August 2007 at the Los Ange- differently—giving higher weight (probability) to guarding the
les International Airport (LAX) to randomize latter. Second, the assistant must address the uncertainty in
checkpoints on the roadways entering the air- information that security forces have about the adversary.
port and canine patrol routes within the air- Third, the assistant must enable a mixed-initiative interaction
port terminals. This article examines the with potential users rather than dictate a schedule; the assistant
information, design choices, challenges, and may be unaware of users’ real-world constraints, and hence
evaluation that went into designing ARMOR. users must be able to shape the schedule development.
We have addressed these challenges in a software assistant

Copyright © 2009, Association for the Advancement of Artificial Intelligence. All rights reserved. ISSN 0738-4602 SPRING 2009 43
Articles

agent called ARMOR (assistant for randomized six-month trial period of ARMOR deployment at
monitoring over routes). Based on game-theoretic LAX. The feedback from police at the end of this
principles, ARMOR combines three key features to six-month period was extremely positive; ARMOR
address each of the challenges outlined above. will continue to be deployed at LAX and expand to
Game theory is a well-established foundational other police activities at LAX.
principle within multiagent systems to reason
about multiple agents, each pursuing its own inter-
ests (Fudenberg and Tirole 1991). We build on Security Domain Description
these game-theoretic foundations to reason about We will now describe the specific challenges in the
two agents—the police force and its adversary—in security problems faced by the LAWA police. LAX1
providing a method of randomization. In particu- is the fifth busiest airport in the United States and
lar, the main contribution of our article is mapping the largest destination airport in the United States,
the problem of security scheduling as a Bayesian serving 60–70 million passengers per year (Stevens
Stackelberg game (Conitzer and Sandholm 2006) et al. 2006). LAX is unfortunately also suspected to
and solving it through the fastest optimal algo- be a prime terrorist target on the West Coast of the
rithm for such games (Paruchuri et al. 2008), United States, with multiple arrests of plotters
addressing the first two challenges. The algorithm attempting to attack LAX (Stevens et al. 2006). To
used builds on several years of research regarding protect LAX, LAWA police have designed a securi-
multiagent systems and security (Paruchuri et al. ty system that utilizes multiple rings of protection.
2005, 2006, 2007). In particular, ARMOR relies on As is evident to anyone traveling through an air-
an optimal algorithm called DOBSS (decomposed port, these rings include such things as vehicular
optimal Bayesian Stackelberg solver) (Paruchuri et checkpoints, police units patrolling the roads to
al. 2008). the terminals and inside the terminals (with dogs),
While a Bayesian game allows us to address and security screening and bag checks for passen-
uncertainty over adversary types, by optimally gers. There are unfortunately not enough resources
solving such Bayesian Stackelberg games (which (police officers) to monitor every single event at
yield optimal randomized strategies as solutions), the airport; given its size and the number of pas-
ARMOR provides quality guarantees on the sched- sengers served, such a level of screening would
ules generated. These quality guarantees obviously require considerably more personnel and cause
do not imply that ARMOR provides perfect securi- greater delays to travelers. Thus, assuming that all
ty; instead, ARMOR guarantees optimality in the checkpoints and terminals are not being moni-
utilization of fixed security resources (number of tored at all times, setting up available checkpoints,
police or canine units) assuming the rewards are canine units, or other patrols on deterministic
accurately modeled. In other words, given a schedules allows adversaries to learn the schedules
specific number of security resources and areas to and plot an attack that avoids the police check-
protect, ARMOR creates a schedule that random- points and patrols, which makes deterministic
izes over the possible deployment of those schedules ineffective.
resources in a fashion that optimizes the expected Randomization offers a solution here. In partic-
reward obtained in protecting LAX. ular, from among all the security measures to
The third challenge is addressed by ARMOR’s use which randomization could be applied, LAWA
of a mixed-initiative-based interface, where users police have so far posed two crucial problems to
are allowed to graphically enter different con- us. First, given that there are many roads leading
straints to shape the schedule generated. ARMOR is into LAX, they want to know where and when
thus a collaborative assistant that iterates over gen- they should set up checkpoints to check cars driv-
erated schedules rather than a rigid one-shot ing into LAX. For example, figure 1 shows a vehic-
scheduler. ARMOR also alerts users in case over- ular checkpoint set up on a road inbound towards
rides may potentially deteriorate schedule quality. LAX. Police officers examine cars that drive by,
ARMOR thus represents a very promising transi- and if any car appears suspicious, they do a more
tion of multiagent research into a deployed appli- detailed inspection of that car. LAWA police
cation. ARMOR has been successfully deployed wished to obtain a randomized schedule for such
since August 2007 at the Los Angeles Internation- checkpoints for a particular time frame. For exam-
al Airport (LAX) to assist the Los Angeles World ple, if we are to set up two checkpoints, and the
Airport (LAWA) police in randomized scheduling timeframe of interest is 8 AM to 11 AM, then a can-
of checkpoints and since November 2007 for gen- didate schedule may suggest to the police that on
erating randomized patrolling schedules for canine Monday, checkpoints should be placed on route 1
units. In particular, it assists police in determining and route 2, whereas on Tuesday during the same
where to randomly set up checkpoints and where time slot, they should be on route 1 and 3, and so
to randomly allocate canines to terminals. Indeed, on. Second, LAWA police wished to obtain an
February 2008 marked the successful end of the assignment of canines to patrol routes through the

44 AI MAGAZINE
Articles

terminals inside LAX. For example, if there are


three canine units available, a possible assignment
may be to place canines on terminals 1, 3, and 6
on the first day, but on terminals 2, 4, and 6 on
another day, and so on based on the available in-
formation. Figure 2 illustrates a canine unit on
patrol at LAX.
Given these problems, our analysis revealed
three key challenges: first, potential attackers can
observe security forces’ schedules over time and
then choose their attack strategy—this fact makes
deterministic schedules highly susceptible to
attack; second, there is unknown and uncertain in-
formation regarding the types of adversary we may
face; and third, although randomization helps
eliminate deterministic patterns, it must also
account for the different costs and benefits associ-
ated with particular targets.
In summarizing the domain requirements, we
Figure 1. LAX Checkpoint.
emphasize the following key points. First, it is
LAWA police, as domain experts, who expressed a
requirement for randomization, leading us to
design ARMOR. Second, there exist different rings
of security (including canines and checkpoints
that ARMOR schedules), which are not static and
therefore may change independently of the other
rings different times. The end result of such shift-
ing randomized security rings is that adversary
costs and uncertainty increase, particularly for
well-planned attacks, which in turn may help deter
and prevent attacks.

Approach
We modeled the decisions of setting checkpoints
or canine patrol routes at the LAX airport as
Bayesian Stackelberg games. These games allow us
to accomplish three important tasks: they model
the fact that an adversary acts with knowledge of
security forces’ schedules, and thus randomize
schedules appropriately; they allow us to define Figure 2. LAX Canine Patrol.
multiple adversary types, meeting the challenge of
our uncertain information about our adversaries;
and they enable us to weigh the significance of dif- leader. For example, given our security domain, the
ferent targets differently. Since Bayesian Stackel- police force (leader) must first commit to a mixed
berg games address the challenges posed by our strategy for placing checkpoints on roads in order
domain, they are at the heart of generating mean- to be unpredictable to the adversaries (followers),
ingfully randomized schedules. From this point we where a mixed strategy implies a probability distri-
will explain what a Bayesian Stackelberg game con- bution over the actions of setting checkpoints. The
sists of, how an LAX security problem can be adversaries, after observing check- points over
mapped onto Bayesian Stackelberg games, some of time, can then choose their own strategy of attack-
the previous methods for solving Bayesian Stackel- ing a specific road. To see the advantage of being
berg games, and how we use DOBSS to optimally the leader in a Stackelberg game, consider a simple
solve the problem at hand. game with the payoff table as shown in figure 3.
The leader is the row player and the follower is the
Bayesian Stackelberg Games column player. Given a simultaneous move game,
In a Stackelberg game, a leader commits to a strat- that is, the leader and follower now act at the same
egy first, and then a follower selfishly optimizes its time, the only pure-strategy Nash equilibrium for
reward, considering the action chosen by the this game is when the leader plays a and the fol-

SPRING 2009 45
Articles

lower. We also assume knowing a priori probabili-


ty pl, where l represents the type of adversary (1, 2,
and so on), of the different follower types (that is l
c d ∈ L). Our goal is to find the optimal mixed strate-
a 2, 1 4,0 gy for the leader to commit to, given that the fol-
lower may know the leader’s mixed strategy when
b 1,0 3,2
choosing its strategy and that the leader will not
know the follower’s type in advance.

Techniques for Solving


Stackelberg Games
Figure 3. Payoff Table for
In previous work it has been shown that finding an
Example Normal Form Game.
optimal solution to a Bayesian Stackelberg game
with multiple follower types is NP-hard (Conitzer
and Sandholm 2006). Researchers in the past have
identified an approach, which we will refer to as
Follower Type 1 the multiple-LPs method, to solve Stackelberg
games (Conitzer and Sandholm 2006), and this can
c d be used to solve Bayesian Stackelberg games. This
a 2, 1 4,0 approach, however, requires transforming a
Bayesian game into a normal form game using the
b 1,0 3,2
Harsanyi transformation (Harsanyi and Selten
1972). Similarly one may apply efficient algo-
rithms for finding Nash equilibria (Sandholm,
Follower Type 2 Gilpin, and Conitzer 2005), but they require the
c' d' same Harsanyi transformation. Since our research
crucially differs in its nonuse of the Harsanyi trans-
a 1, 1 2,0 formation, it is important to understand this trans-
b 1,0 3,2 formation and its impact.
Harsanyi Transformation. The first step in solv-
ing Bayesian games for previous methods is to
apply the Harsanyi transformation (Harsanyi and
Selten 1972) that converts the incomplete infor-
Figure 4. Security Agent Versus Followers 1 and 2. mation game into a normal form game. Given that
the Harsanyi transformation is a standard concept
lower plays c, which gives the leader a payoff of 2. in game theory, we explain it briefly through a
However, if the leader commits to a uniform mixed simple example without introducing the mathe-
strategy of playing a and b with equal (0.5) proba- matical formulations. Consider the case of the two
bility, then the follower will play d in order to max- follower types 1 and 2 as shown in figure 4. Fol-
imize its payoff, leading to a payoff for the leader lower type 1 will be active with probability α, and
of 3.5. Thus, by committing to a mixed strategy follower type 2 will be active with probability 1 –
first, the leader is able to obtain a higher payoff α. Performing the Harsanyi transformation in-
than could be obtained in a simultaneous move volves introducing a chance node that determines
situation. the follower’s type, thus transforming the leader’s
The Bayesian form of such a game, then, implies incomplete information regarding the follower
that each agent must be of a given set of types. For into an imperfect information game. The trans-
our security domain, we have two agents, the formed, normal form game is shown in figure 5. In
police force and the adversary. While there is only the transformed game, the leader still has two
one police force type, there are many different strategies while there is a single follower type with
adversary types, such as serious terrorists, drug four (2 * 2) strategies. For example, consider the sit-
smugglers, and petty criminals, denoted by L. Dur- uation in the transformed game where the leader
ing the game, the adversary knows its type, but the takes action a and the follower takes action cc⬘. The
police do not know the adversary’s type; this is an leader’s payoff in the new game is calculated as a
incomplete information game. For each agent (the weighted sum of its payoffs from the two tables in
police force and the adversary) i, there is a set of figure 4, that is, α times payoff of leader when fol-
strategies σi and a utility function ui : L ⫻ σ1 × σ2 lower type 1 takes action c plus 1 – α times payoff
→ ᑬ. Figure 4 shows a Bayesian Stackelberg game of leader when follower type 2 takes action c⬘. All
with two follower types. Notice that follower type the other entries in the new table, both for the
2 changes the payoff of both the leader and the fol- leader and the follower, are derived in a similar

46 AI MAGAZINE
Articles

cc' cd' dc' dd'


a 2α + (1 – α), 1 2, α 4α + (1 – α), (1 – α) 4α + 2(1 – α), 0
b α, (1 – α) α + 3(1 – α), 2(1 – α) 3α, 2α + (1 – α) 3, 2

Figure 5. Harsanyi Transformed Payoff Table.

fashion. In general, for n follower types with k


strategies per follower type, the transformation
results in a game with kn strategies for the follow-
er, thus causing an exponential blowup losing max x ,q ,a ∑∑∑ p R x q
i∈ X l∈ L j∈Q
l l
ij i
l
j
compactness.
Methods such as those described in Conitzer
and Sandholm (2006) and Sandholm, Gilpin, and
s.t. ∑ i∈ X xi = 1
Conitzer (2005) must use this Harsanyi transfor- ∑ j∈Q qlij = 1
mation, which implies the game loses its compact
structure. Nonetheless, the solutions their meth- 0 ≤ (al − ∑ i∈X Cijl xi ) ≤ (1 − qlj ) M
ods obtain can be transformed back into the origi- xi ∈ [ 0 …1]
nal game.
qlj ∈ {0,1
1}
DOBSS
a∈ℜ
One key advantage of the DOBSS approach is that
it operates directly on the Bayesian representation,
without requiring the Harsanyi transformation. In
particular, DOBSS obtains a decomposition scheme
by exploiting the property that follower types are
independent of each other. The key to the DOBSS Problem 1.
decomposition is the observation that evaluating
the leader strategy against a Harsanyi-transformed
game matrix is equivalent to evaluating against probabilities pl, with l ∈ L, of facing each follower
each of the game matrices for the individual fol- type, the leader solves problem 1.
lower types. Here for a set of leader’s actions x and actions for
We first present DOBSS in its most intuitive form each follower ql, the objective represents the
as a mixed-integer quadratic program (MIQP); we expected reward for the agent considering the a
then illustrate how it may be transformed into a
priori distribution over different follower types pl.
linearized equivalent mixed-integer linear program
Constraints with free indices mean they are repeat-
(MILP). While a more detailed discussion of the
ed for all values of the index. For example, the
MILP is available in Paruchuri et al. (2008), the cur-
fourth constraint means xi ∈ [0 ... 1] for all i ∈ X.
rent section may at least serve to explain at a high
The first and the fourth constraints define the set
level the key idea of the decomposition used in this
of feasible solutions x as a probability distribution
MILP.
The model we propose explicitly represents the over the set of actions X. The second and fifth con-
actions by the leader and the optimal actions for straints limit the vector of actions of follower type
the follower types in the problem solved by the l, ql to be a pure distribution over the set Q (that is
agent. We denote by x the leader’s policy (mixed each ql has exactly one coordinate equal to one
strategy), which consists of a vector of probability and the rest equal to zero). Note that we need to
distributions over the leader’s pure strategies. consider only the reward-maximizing pure strate-
Hence, the value xi is the proportion of times in gies of the follower types, since for a given fixed
which pure strategy i is used in the policy. We mixed strategy x of the leader, each follower type
denote by ql the vector of strategies of follower type faces a problem with fixed linear rewards. If a
l ∈ L. We also denote by X and Q the index sets of mixed strategy is optimal for the follower, then so
leader and follower l’s pure strategies, respectively. are all the pure strategies in support of that mixed
We also index the payoff matrices of the leader and strategy.
each of the follower types l by the matrices Rl and The two inequalities in the third constraint
Cl. Let M be a large positive number. Given a priori ensure that q1j = 1 only for a strategy j that is opti-

SPRING 2009 47
Articles

mal for follower type l. Indeed this is a linearized adversary type observes the LAWA police check-
form of the optimality conditions for the linear pro- point policy and then decides where to attack.
gramming problem solved by each follower type. Since adversaries can observe the LAWA police pol-
We explain these constraints as follows: note that icy before deciding their actions, this can be mod-
the leftmost inequality ensures that for all j ∈ Q, eled through a Stackelberg game with the police as
the leader.
al ≥ ∑ i∈XCijl xi
. In this setting, the set X of possible actions for
This means that given the leader’s vector x, al is an LAWA police is the set of possible checkpoint com-
upper bound on follower type l’s reward for any binations. If, for instance, LAWA police were set-
action. The rightmost inequality is inactive for ting up one checkpoint then X = {1, ..., k}. If LAWA
every action where q1j = 0, since M is a large positive police were setting up a combination of two check-
quantity. For the action that has q1j = 1 this inequal- points, then X = {(1, 2), (1, 3)...(k − 1, k)}, that is, all
ity states that the adversary’s payoff for this action combinations of two checkpoints. Each adversary
must be ≥ al, which combined with the previous type l ∈ L = {1, ..., m} can decide to attack one of the
inequality shows that this action must be optimal k roads or maybe not attack at all (none), so its set
for follower type l. Notice that problem one is a of actions is Q = {1, …, k, none}. If LAWA police
decomposed MIQP in the sense that it does not uti- select road i to place a checkpoint on and adver-
lize a full-blown Harsanyi transformation; instead sary type l ∈ L selects road j to attack then the
it solves multiple smaller problems using individ- police receive a reward Rijl and the adversary
ual adversaries’ payoffs (indexed by l). Further- receives a reward Cijl . These reward values vary
more, this decomposition does not cause any sub- based on three considerations: the chance that the
optimality (Paruchuri et al. 2008). LAWA police checkpoint will catch the adversary
We can linearize the quadratic programming on a particular inbound road; the damage the
problem 1 through the change of variables zlij = adversary will cause if it attacks by means of a par-
xiqlj. The substitution of this one variable allows us ticular inbound road; and the type of adversary,
to create an MILP. The details of this transforma- that is, adversary capability. If LAWA police catch
tion and its equivalence to problem 1 are present- the adversary when i = j, we make Rijl a large posi-
ed in Paruchuri et al. (2008). DOBSS refers to this tive value and Cijl a large negative value. However,
equivalent mixed-integer linear program, which the probability of catching the adversary at a
can be solved with efficient integer programming checkpoint is based on the volume of traffic
packages. Although DOBSS still remains as an through the checkpoint (significant traffic will
exponential solution to solving Bayesian Stackel- increase the difficulty of catching the adversary),
berg games, by avoiding the Harsanyi transforma- which is an input to the system. If the LAWA police
tion it obtains significant speedups over the previ- are unable to catch the adversary, then the adver-
ous approaches as shown in the experimental sary may succeed; that is, we make Rijl a large nega-
results and proofs in Paruchuri et al. (2008). tive value and Cijl a large positive value. Certainly,
if the adversary attacks from an inbound road
Bayesian Stackelberg Game for the Los where no checkpoint was set up, there is no chance
Angeles International Airport that the police will catch the adversary. The mag-
We now illustrate how the security problems set nitude of Rijl and Cijl vary based on the adversary’s
forth by LAWA police can be cast in terms of a potential target, given the road from which the
Bayesian Stackelberg game. We focus on the check- adversary attacks. Some roads lead to higher val-
point problem for illustration, but the case of the ued targets for the adversary than others. The game
canine problem is similar. Given the checkpoint is not a zero sum game, however, as even if the
problem, our game consists of two players, the adversary is caught, the adversary may benefit due
LAWA police (the leader) and the adversary (the to publicity.
follower), in a situation consisting of a specific The reason we consider a Bayesian Stackelberg
number of inbound roads on which to set up game is because LAWA police face multiple adver-
checkpoints, say roads 1 through k. LAWA police’s sary types. Thus, differing values of the reward
set of pure strategies consists of a particular subset matrices across the different adversary types l ∈ L
of those roads to place checkpoints on prior to represent the different objectives and valuations of
adversaries selecting which roads to attack. LAWA the different attackers (for example, smugglers,
police can choose a mixed strategy so that the criminals, terrorists). For example, a hard-core,
adversary will be unsure of exactly where the well-financed adversary could inflict significant
checkpoints may be set up, but the adversary will damage on LAX; thus, the negative rewards to the
know the mixed strategy LAWA police have cho- LAWA police are much higher in magnitude than
sen. We assume that there are m different types of an amateur attacker who may not have sufficient
adversaries, each with different attack capabilities, resources to carry out a large-scale attack. If these
planning constraints, and financial ability. Each are the only two types of adversaries faced, then a

48 AI MAGAZINE
Articles

External Information

Armor System

Bayesian Stackelberg Game


User Input

Interface

DOBSS

S1=30%
S2=10%
S3=60%

Finalized Schedule Suggested Schedule Mixed Strategy

Front End Back End

Figure 6. ARMOR System Flow Diagram.

20-80 split of probability implies that while there System Architecture


is a 20 percent chance that the LAWA police face
the former type of adversary, there is an 80 percent There are two separate versions of ARMOR,
chance that they face an amateur attacker. Our ex- ARMOR-checkpoint and ARMOR-canine. While in
perimental data provides initial results about the the following we focus on ARMOR-checkpoint for
sensitivity of our algorithms to the probability dis- illustration, both these versions use the same
tributions over these two different adversary types. underlying architecture with different inputs. As
While the number of adversary types has varied shown in figure 6, this architecture consists of a
based on inputs from LAWA police, for any one front end and a back end, integrating four key
adversary type the largest game that has been con- components: a front-end interface for user interac-
structed, which was done for canine deployment, tion; a method for creating Bayesian Stackelberg
consisted of 784 actions for the LAWA police game matrices; an implementation of DOBSS; and
(when multiple canine units were active) for the a method for producing suggested schedules for
eight possible terminals within the airport and 8 the user. They also contain two major forms of
actions per adversary type (one for a possible attack external input. First, they allow for direct user
on each terminal). input into the system through the interface. Sec-

SPRING 2009 49
Articles

Figure 7. ARMOR Interface.

ond, they allow for file input of relevant informa- inbound roads and the number of checkpoints al-
tion for checkpoints or canines, such as traffic/pas- lowed during that time slot determines the avail-
senger volume by time of day, which can greatly able actions for the LAWA police, whereas the
affect the security measures taken and the values action space of the adversary is determined by the
of certain actions. At this point we will discuss in number of inbound roads. Thus, the system can set
detail what each component consists of and how up the foundation for the Bayesian Stackelberg
they interact with each other. game by providing all the actions possible in the
game. Once the action space has been generated, it
Interface can be sent to the back end to be set up as a
The ARMOR interface, seen in figure 7, consists of Bayesian Stackelberg game, solved, and returned as
a file menu, options for local constraints, options a suggested schedule, which is displayed to the
to alter the action space, a monthly calendar and a user by means of a spreadsheet.
main spreadsheet to view any day(s) from the cal- There are three options that serve to restrict cer-
endar. Together these components create a work- tain actions in the generated schedule: forced
ing interface that meets all the key requirements checkpoint; forbidden checkpoint; at least one
set forth by LAWA officers for checkpoint and checkpoint. These constraints are intended to be
canine deployment at LAX. used sparingly to accommodate situations where a
The base of the interface is designed around six user, faced with exceptional circumstances and
possible adjustable options; three of them alter the extra knowledge, wishes to modify the output of
action space and three impose local constraints. the game. The user may impose these restrictions
The three options to alter the action space are the by forcing specific actions in the schedule. In par-
number of checkpoints allowed during a particular ticular, the forced checkpoint option schedules a
time slot; the time interval of each time slot; and checkpoint at a specific time on a specific day. The
the number of days to schedule over. For each giv- forbidden checkpoint option designates a specific
en time slot, the system constructs a new game. time on a specific day when a checkpoint should
As discussed previously, the total number of not be scheduled. Finally, the at least one check-

50 AI MAGAZINE
Articles

point option designates a set of time slots and between the combinations with the given proba-
ensures that a checkpoint is scheduled in at least bilities. Each time a selection is made, that combi-
one of the slots. nation is sent to the user interface to be reviewed
The spreadsheet in the interface serves as the by the user as necessary. So if, for instance, combi-
main mechanism for viewing, altering, and con- nation one was chosen, the user would see check-
straining schedules. The columns correspond to point A and B as scheduled for the given time slot.
the possible checkpoints, and the rows correspond In rare cases, as mentioned previously, a user
to the time frames in which to schedule them. Up may have forbidden a checkpoint or required a
to a full week can be viewed at a single time as seen checkpoint. ARMOR accommodates such user
in figure 7. Once a particular day is in view, the directives when creating its schedule; for example,
user can assign to that day any desired constraints. if checkpoint C is forbidden, then all the probabil-
Each constraint is represented by a specific color ity in our example shifts to the combination A and
within the spreadsheet, namely green, red, and yel- B. Unfortunately, by constraining the schedule fre-
low for the forced, forbidden, and at least con- quently, a user can completely alter the mixed
straints, respectively. strategy produced as the output of DOBSS, defeat-
ing DOBSS’s guarantee of optimality. To avoid such
Matrix Generation and DOBSS a possibility, ARMOR incorporates certain alerts
Given the submitted user information, the system (warnings) to encourage noninterference in its
must create a meaningful Bayesian Stackelberg schedule generation. For example, if a combina-
game matrix. Previously we illustrated the genera- tion has zero or very low probability of being cho-
tion of the action space in this game. Based on the sen and the user has forced that checkpoint com-
prespecified rewards as discussed earlier, we can bination to occur, ARMOR will alert the user.
provide the rewards for the LAWA police and the Similarly, if a combination has a very high likeli-
adversaries to generate a game matrix for each hood and the user has forbidden that event,
adversary type. After the final game matrices are ARMOR will again alert the user. However, ARMOR
constructed for each adversary type, they are sent only alerts the user; it does not autonomously
to the DOBSS implementation, which calculates remove the user’s constraints. Resolving more sub-
the optimal mixed strategy over the current action tle interactions between the user’s imposed con-
space. straints and DOBSS’s output strategy remains an
To demonstrate the process, assume there are issue for future work.
three possible checkpoint locations (A, B, C), one When a schedule is presented to the user with
possible time slot to schedule over, and two check- alerts, the user may alter the schedule by altering
points available for scheduling. Given this sce- the forbidden or required checkpoints, or possibly
nario, the unique combinations possible include by directly altering the schedule. Both possibilities
scheduling checkpoints A and B, A and C, and B are accommodated in ARMOR. If the user simply
and C, over the given time frame. We will assume adds or removes constraints, ARMOR can create a
that checkpoints A and B are highly valuable while new schedule. Once the schedule is finalized, it can
C, although not completely devoid of value, has a be saved for actual use, thus completing the system
very low value. Based on this information, a likely cycle. This full process was designed specifically to
mixed strategy generated by DOBSS would be to meet the requirements at LAX for checkpoint and
assign a high probability to choosing action A and canine allocation.
B, say 70 percent, and a low probability to both the
other actions, say 15 percent each. Whatever the
mixed strategy actually comes out to be, it is the
Design Challenges
optimal strategy a user could take to maximize Designing and deploying the ARMOR software on
security based on the given information. This a trial basis at LAX posed numerous challenges and
mixed strategy is then stored and used for the actu- problems to our research group. Some key lessons
al schedule generation. learned during the design and deployment of
ARMOR include the importance of tools for ran-
Mixed Strategy and Schedule Generation domization, the importance of manual schedule
Once an optimal mixed strategy has been chosen overrides, and the importance of providing police
by DOBSS and stored within the system, a particu- officers with operational flexibility.
lar combination of actions must be chosen to be
displayed to the user. Consider our example from Importance of Tools for Randomization
the previous section involving three possibilities There is a critical need for randomization in secu-
(checkpoints A and B, A and C, B and C) and their rity operations. Security officials are aware that
probabilities of 70 percent, 15 percent, and 15 per- requiring humans to generate randomized sched-
cent. Knowing this probability distribution, the ules is unsatisfactory because, as psychological
system can formulate a method to randomly select studies have often shown (Wagenaar 1972),

SPRING 2009 51
Articles

humans have difficulty randomizing. Instead, Our run-time analysis adds to the above results,
mathematical randomization that appropriately focusing specifically on the current security
weighs the costs and benefits of different actions domain for which this work has been applied. For
and randomizes accordingly leads to improved this reason we compare the run-time results of
results. Security officials were hence extremely DOBSS versus multiple LPs, described previously,
enthusiastic in their reception of our research and given the specific domain used for canine deploy-
eager to apply it to their domain. In addition, these ment at LAX. MIP-Nash (Sandholm, Gilpin, and
officials have indicated that obtaining schedules Conitzer 2005) has not been included in this
automatically reduces their burden of having to analysis of run times as it only provides the best
construct such schedules manually taking all the Bayes-Nash equilibrium as opposed to the optimal
relevant factors into account. mixed strategies provided by the multiple-LPs
method and the DOBSS method. The aim of this
Importance of analysis is to show that DOBSS is indeed the most
Manual Schedule Overrides suitable procedure for application to real domains
While ARMOR incorporates all the knowledge that such as the LAX canine and checkpoint allocation.
we could obtain from LAWA police and provides the To that end, we used the data from a full week of
best output possible, it may not be aware of dynam- canine deployment to analyze the time necessary
ic developments on the ground. For example, police to generate a schedule given the DOBSS method
officers may have very specific intelligence for and the multiple-LPs method. For completeness we
requiring a checkpoint on a particular inbound road. show the results given one to four adversary types
Hence, it was crucial to allow LAWA police officers where four adversary types is the minimum
(in rare instances when it is necessary) to manually amount LAWA has set forth as necessary.
selectively override the schedule provided. In figure 8 we summarize the run-time results for
our Bayesian games using DOBSS and multiple LPs.
Importance of Providing Police Officers We tested our results on the Bayesian games pro-
with Operational Flexibility vided from the canine domain with number of
When initially generating schedules for canine adversary types varying between one to four. Each
patrols, the system created a very detailed sched- game between LAWA and one adversary type is
ule, micromanaging the patrols. This did not get as modeled as a normal form game. Thus, there are
positive a reception from the officers. Instead, an four normal form games designed for the game
abstract schedule that afforded the officers some between LAWA and the various adversary types for
flexibility to respond to dynamic situations on the the base case. The size of each of these normal
ground was better received. form games is (784, 8), corresponding to 784
strategies for LAWA and 8 for the adversary. We
then used the seven generated instances, taken
Experimental Results from an arbitrary week of canine deployment, of
Our experimental results explore the run-time effi- this base case to obtain averaged results.
ciency of DOBSS and evaluate the solution quality The x-axis in figure 8 shows the number of fol-
and implementation of the ARMOR system. lower types the leader faces starting, and the y-axis
of the graph shows the run time in seconds. All the
Run-Time Analysis experiments that were not concluded in 20 min-
It has been shown in Paruchuri et al. (2008) that utes (1,200 seconds) were cut off. From the graph
DOBSS significantly outperforms its competitors, we summarize that DOBSS outperforms the multi-
which include MIP-Nash (Sandholm, Gilpin, and ple-LPs method by a significant margin given our
Conitzer 2005) and multiple LPs (Conitzer and real canine domain. In the graph, while multiple
Sandholm 2006) in an experimental domain that LPs could solve the problem only for up to two
involves a security agent patrolling a world consist- adversary types, DOBSS could solve for all four
ing of m houses, 1 … m and a robber trying to rob adversary types within 80 seconds.
these houses. These results show that even for a Hence we see that the DOBSS method is faster
world as small as three houses, MIP-Nash and mul- than the multiple-LPs method. Consequently, we
tiple LPs are unable to converge on a solution with- conclude that DOBSS is the algorithm of choice for
in the allowed time of 30 minutes when there are 8 Bayesian Stackelberg games (Paruchuri et al. 2008),
or more adversary types. DOBSS, however, is able to especially given the particular games created by
achieve a solution in less than 10 seconds for up to real security domains such as the canine patrolling
14 adversary types. Also, as the number of houses problem presented in this article.
increases up to five, Paruchuri and colleagues find
that MIP-Nash and multiple LPs are unable to con- Evaluation of ARMOR
verge on a solution within 30 minutes for even low We now evaluate the solution quality obtained
numbers of adversary types (Paruchuri et al. 2008). when DOBSS is applied to the LAX security

52 AI MAGAZINE
Articles

DOBSS Multiple LPs

1400
1200
Runtime (in secs)

1000
800
600
400
200
0
1 2 3 4
Number of Adversary Types

Figure 8. Run Times: DOBSS and Multiple-LPs Methods.

domain. We offer three types of evaluation. While two checkpoints are placed. Here also the reward
our first evaluation is in the lab, ARMOR is a in the case of the DOBSS strategy is greater than
deployed assistant, and hence our remaining two the reward of the uniform random strategy. When
evaluations are of its deployment in the field. With we have two checkpoints, the type 2 adversary
respect to our first evaluation, we conducted four chooses the action none (to not attack). This leads
experiments. The first three compared ARMOR’s to the observation that the reward of the DOBSS
randomization with a uniform randomization strategy and the reward of the uniform strategy are
technique that does not use ARMOR’s weights in the same when only the type 2 adversary is pres-
randomization. ent. Figure 9c presents the case of three check-
The results of the first experiment are shown in points. Here the reward values obtained by DOBSS
figures 9a, 9b, and 9c. The x-axis represents the are always positive—this is because the chances of
probabilities of occurrence of the two adversary catching the adversary of type 1 improve signifi-
types we chose to focus on. Since the actual num- cantly with three checkpoints. This also leads to
ber of adversary types used for LAX is secure infor- the reward of DOBSS decreasing with the decrease
mation, we use two adversary types for simplicity in the probability of occurrence of the adversary of
in this analysis. The x-axis shows the probability p type 1. Note that the type 2 adversary, as with the
of adversary type 2 (the probability of adversary case of two checkpoints, decides none and hence
type 1 is then obtained on 1 – p). The y-axis repre- the reward of the DOBSS strategy and the uni-
sents the reward obtained by LAWA. This reward formly random strategy are the same when only
represents the expected reward LAWA would type 2 adversary is present.
obtain given the optimal adversary response to the The three experiments reported allow us to con-
strategy adopted by LAWA. Figure 9a shows the clude that DOBSS weighted randomization pro-
comparison when one checkpoint is placed. For vides significant improvements over uniform ran-
example, when adversary of type 1 occurs with a domization in the same domain, thus illustrating
probability of 0.1 and type 2 occurs with a proba- the utility of our algorithms. We continue these
bility of 0.9, the reward obtained by the DOBSS results in the following fourth experiment, focus-
strategy is −1.72 whereas the reward obtained by a ing now on canine units. Figure 9d shows the com-
uniform random strategy is −2.112. It is important parison of the reward obtained between schedul-
to note that the reward of the DOBSS strategy is ing canine units with DOBSS and scheduling them
strictly greater than the reward of the uniform ran- with a uniform random strategy (denoted URS). In
dom strategy for all probabilities of occurrence of the uniform random strategy, canines are random-
the adversary types. ly assigned to terminals with equal probability. The
Figure 9b also has the probability distribution on x-axis represents the weekday and the y-axis repre-
the x-axis and the reward obtained on the y-axis. It sents the reward obtained. We can see that DOBSS
shows the difference in the obtained reward when performs better even with three canine units as

SPRING 2009 53
Articles

0 1.5

-0.5 1
DOBSS Strategy Uniform Random Strategy

-1
0.5
Reward

Reward
-1.5
0

-2
-0.5

-2.5
-1

-3 DOBBS Strategy Uniform Random Strategy


-1.5

-3.5
0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 -2
0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
Probability of Adversaries (type2) Probability of Adversaries (type2)

a b

1.6 6
DOBSS Strategy Uniform Random Strategy DOBSS (5 Canines) DOBSS (6 Canines)
1.4 5 DOBSS (3 Canines) URS (6 Canines)

1.2 4

1 3
Reward

Reward

0.8 2

0.6 1

0.4 0

0.2 -1

0 -2
0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Mon Tue Wed Thu Fri Sat Sun

Probability of Adversaries (type2) Days

c d

Figure 9. DOBSS Strategy Versus Uniformly Random Strategy.


a. One checkpoint. b. Two checkpoints. c. Three checkpoints. d. Canines.

compared to six canine units being scheduled The numbers 1 to 5 in the table denote the check-
using the uniform random strategy. For example, point number (we have assigned arbitrary identi-
on Friday, the reward of a uniformly random strat- fication numbers to all checkpoints for the purpose
egy with six canine units is −1.47, whereas the of this experiment) and the values of the table
reward of three, five, and six canine units with show the percentage of times this checkpoint was
DOBSS is 1.37, 3.50, and 4.50 respectively. These used. For example, in week 1, checkpoint 2 was
results show that DOBSS weighted randomization used just less than 5 percent of times, while check-
with even three canine units provides better results point 2 was used about 25 percent of the times in
against uniform randomization in the same week 2. We can make two observations from these
domain with six canine units. Thus our algorithm two weeks: First, we do not appear to have uniform
provides better rewards and can help in reducing use of these checkpoints; that is, there is great vari-
the cost of resources needed. ance in the percentage of times checkpoints are
In the next evaluation, we examine ARMOR’s deployed. Second, the checkpoint deployment
setting of checkpoints at LAX. The first experiment varies from week to week; for example, checkpoint
examines the change in checkpoint deployment 4 was not used in week 1, but it was used 15 per-
during a fixed shift (that is, keeping the time fixed) cent of the times in week 2.
over two weeks. The results are shown in table 1. The goal of the next experiment was to provide

54 AI MAGAZINE
Articles

Checkpoint Number 1 2 3 4 5
Week 1 33.33 4.76 33.33 0 28.57
Week 2 19.04 23.80 23.80 14.28 19.05

Table 1. Variation in Usage Percentage.

results on the sensitivity analysis, specifically, how Related Work and Summary
the probabilities of different actions will change if
we change the proportion of adversary types. Fig- The patrolling problem itself has received signifi-
ure 10 shows the variation in strategy for placing cant attention in multiagent literature due to its
two checkpoints together when the probability of wide variety of applications ranging from robot
occurrence of the adversary changes. The x-axis patrol to border patrolling of large areas (Ruan et
shows the variation in the probability of occur- al. 2005, Billante 2003). The key idea behind the
rence of the adversary types, whereas the y-axis policies provided by these techniques is random-
shows the variation in the probabilities in the ization, which decreases the amount of informa-
DOBSS strategy. For example, when adversary of tion given to an adversary. However, no specific
type 1 occurs with a probability of 1, the probabil- algorithm/procedure has been provided for the
ity of placing both checkpoints 1 and 4 is 0.353; generation of randomized policies; hence, they can
when adversaries 1 and 2 occur with probabilities lead to highly suboptimal policies. Two exceptions
0.4 and 0.6, respectively, then the probability of are Praveen Paruchuri and colleagues and their ear-
placing checkpoints 3 and 4 is 0.127. We can ly work (Paruchuri et al. 2006), which provides
observe that there is no variation in the probabili- algorithms for analyzing randomization-reward
ties in the DOBSS strategies when the probabilities trade-offs, and Noa Agmon and colleagues and
of occurrence of the two adversary types vary from their recent work (Agmon, Kraus, and Kamink
.1 to .9. This indicates that our results are not par- 2008), which provides algorithms for reducing the
ticularly sensitive to variations in probabilities of probability of penetration. However, unlike our
opponents except at the extremes. work, neither model any adversaries or adversary
Our final evaluation is a more informal evalua- types.
tion based on feedback from the LAWA police. First, Finally, the sequence from Koller and Pfeffer
they have provided very positive feedback about (1997) provides an alternative compact representa-
the deployment. They suggest that the technique tion to normal form representation. However, rep-
they had previously used was not one of random- resenting commitment to a mixed strategy, as
ization, but one of alternating checkpoints; such a required in our Stackelberg games, is difficult in
routine can bring about determinism in the sched- this representation, making its use difficult. Fur-
uling, which we have avoided. Second, ARMOR has thermore, Koller and Pfeffer (1997) have not
reduced routine work in scheduling, which allows focused on computing optimal response in Stack-
LAWA police to focus on more important tasks. elberg games, but rather in only finding equilibria.
Third, several arrests have been made at check- While ARMOR is a game-theoretic security
points scheduled by ARMOR. Typically these scheduler, there are many other competing non-
involved cars attempting to carry weapons into game-theoretic tools in use for related applications.
LAX. Finally, Director James Butts of LAX police has For example, the “Hypercube Queuing Model”
commented that the new random placement (Larson 1974) based on queuing theory depicts the
“makes travelers safer” and even gives them “a detailed spatial operation of urban police depart-
greater feeling of police presence” by making the ments and emergency medical services and has
police appear more numerous (Murr 2007). Also found application in police beat design, allocation
Chief Erroll Southers in testimony to a congres- of patrolling time, and so on. However, this mod-
sional committee has commented, “LAX is safer el does not take specific adversary models into
today than it was 18 months ago,” citing ARMOR account; ARMOR, on the other hand, tailors poli-
as one of the key factors.2 This does not necessarily cies to combat various potential adversaries.
suggest that ARMOR’s schedule was responsible Two different approaches have been presented
because this is not a controlled experiment per se. previously to find solutions to Bayesian Stackel-
Nonetheless, it illustrates that the first line of berg games efficiently. One of the approaches,
defense at the outer airport perimeter is helping named ASAP (Paruchuri et al. 2007), is able to
alleviate the threat of violence at the airport. operate on the Bayesian form of Stackelberg

SPRING 2009 55
Articles

0.45

0.4

0.35

Probability of Action
0.3
C1,C3 C1,C4
0.25

0.2

0.15 C2,C3 C3,C4

0.1

0.05

0
0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1

Probabilities of Adversaries (type2)

Figure 10. Sensitivity Analysis.

games, but it provides an approximate solution. 2008) for the past two years into the real world.
The second approach, the multiple-LPs method,
requires a Bayesian game to be transformed into a Acknowledgements
normal form game using the Harsanyi transfor- ARMOR’s deployment at LAX has only been possible
mation (Harsanyi and Selten 1972). DOBSS is due to the exceptional effort by LAWA police to
superior to ASAP in that it provides exact solu- strike a collaboration. This research was supported
tions, and as shown, it also outperforms the mul- by the United States Department of Homeland Secu-
tiple-LPs method for our domain of interest. rity through the Center for Risk and Economic
In summary, establishing security around infra- Analysis of Terrorism Events (CREATE) under grant
structure of economic or political importance is a number 2007-ST-061-000001. However, any opin-
challenge that is faced today by police forces ions, findings, and conclusions or recommendations
around the world. While randomized monitoring in this document are those of the authors and do not
is important—as adversaries can observe and necessarily reflect views of the United States Depart-
exploit any predictability—randomization must ment of Homeland Security. We would also like to
use dif ferent weighing functions to reflect the thank the National Science Foundation for their
complex costs and benefits of different police contributions under grant number IS0705587.
actions. This article describes a deployed agent
assistant called ARMOR that casts the monitoring Notes
problem as a Bayesian Stackelberg game, where 1. www.lawa.org/lax/justTheFact.cfm
randomized schedule generation can appropriate- 2. homeland.house.gov/SiteDocuments/20080506102233
ly weigh the costs and benefits as well as uncer- -79198.pdf
tainty over adver sary types. ARMOR combines
two key features: first, it uses the fastest known References
solver for Bayesian Stackelberg games called DOB- Agmon, N.; Kraus, S.; and Kamink, G. A. 2008. Multi-
SS, where the dominant mixed strategies provide Robot Perimeter Patrol in Adversarial Settings. In Proceed-
ings of the 2008 IEEE International Conference on Robotics
schedule randomization; and second, its mixed-
and Automation. Piscataway, NJ: Institute of Electrical and
initiative-based interface allows users to occa-
Electronics Engineers.
sionally adjust or override the automated sched-
Billante, N. 2003. The Beat Goes On: Policing for Crime
ule based on their local constraints. ARMOR has
Prevention. Leonards, NSW, Australia: Center for Inde-
been successfully deployed at the Los Angeles In- pendent Studies (www.cis.org.au/issueanalysis/IA38/IA38
ternational Airport, randomizing allocation of .HTM).
checkpoints since August 2007 and canine Conitzer, V., and Sandholm, T. 2006. Computing the Op-
deployment since November 2007. ARMOR thus timal Strategy to Commit To. In Proceedings of the Seventh
represents a successful transition of multiagent ACM Conference on Electronic Commerce (ACM-EC). New
algorithmic advances (Paruchuri et al. 2006, 2007, York: Association for Computing Machinery.

56 AI MAGAZINE
Articles

Fudenberg, D., and Tirole, J. 1991. Game Theory. Cam-


bridge, MA: The MIT Press.
Harsanyi, J. C., and Selten, R. 1972. A Generalized Nash
Solution for Two-Person Bargaining Games with Incom-
plete Information. Management Science 18(5): 80–106. Theoretical Aspects of Rationality
Koller, D., and Pfeffer, A. 1997. Representations and Solu-
tions for Game-Theoretic Problems. Artificial Intelligence
and Knowledge (TARK XII)
94(1–2): 167–215. Stanford University
Larson, R. C. 1974. A Hypercube Queuing Model for Fa- July 6 - 8, 2009
cility Location and Redistricting in Urban Emergency Ser-
vices. Computer and Operations Research 1(1): 67–95.
TARK brings together researchers from a wide va-
Murr, A. 2007. The Element of Surprise. Newsweek Nation-
al News (September 28) (www.newsweek.com/id/41845).
riety of fields, including Artificial Intelligence, Dis-
Paruchuri, P.; Pearce, J. P.; Marecki, J.; Tambe, M.;
tributed Computing, Cryptography, Economics and
Ordóñez, F.; and Kraus, S. 2008. Playing Games for Secu- Game Theory, Linguistics, Philosophy, and Psychol-
rity: An Efficient Exact Algorithm for Solving Bayesian ogy, in order to further our understanding of interdis-
Stackelberg Games. In Proceedings of the 2008 Internation-
ciplinary issues involving reasoning about rationality
al Conference on Autonomous Agents and Multiagent Sys-
tems. Richland, SC: International Foundation for and knowledge. For more information see:
Autonomous Agents and Multiagent Systems.
Paruchuri, P.; Pearce, J. P.; Tambe, M.; Ordóñez, F.; and ai.stanford.edu/∼epacuit/tark09
Kraus, S. 2007. An Efficient Heuristic Approach for Secu-
rity against Multiple Adversaries. In Proceedings of the Key Dates
2007 International Conference on Autonomous Agents and Paper submission deadline: March 2, 2009
Multiagent Systems. Richland, SC: International Founda-
Notification of authors: April 27, 2009
tion for Autonomous Agents and Multiagent Systems.
Paruchuri, P.; Tambe, M.; Ordóñez, F.; and Kraus, S. 2005.
Camera ready version due: June 10, 2009
Safety in Multiagent Systems by Policy Randomization.
Paper presented at the Second International Workshop Organizers:
on Safety and Security, Utrecht, the Netherlands, 26 July.
Joseph Halpern (Conference Chair)
Paruchuri, P.; Tambe, M.; Ordóñez, F.; and Kraus, S. 2006.
Security in Multiagent Systems by Policy Randomization.
Aviad Heifetz (PC Chair)
In Proceedings of the 2006 International Conference on Eric Pacuit (Local Organizer)
Autonomous Agents and Multiagent Systems. Richland, SC:
International Foundation for Autonomous Agents and
Multiagent Systems.
Ruan, S.; Meirina, C.; Yu, F.; Pattipati, K. R.; and Popp, R.
L. 2005. Patrolling in a Stochastic Environment. In Pro-
ceedings of the Tenth International Command and Control
Research and Technology Symposium. Washington, DC:
Department of Defense.
Sandholm, T.; Gilpin, A.; and Conitzer, V. 2005. Mixed-
Integer Programming Methods for Finding Nash Equilib-
ria. In Proceedings of the Twentieth National Conference on appointments in the Department of Industrial and Sys-
Artificial Intelligence. Menlo Park, CA: AAAI Press. tems Engineering and the Department of Computer Sci-
Stevens, D.; Hamilton, T.; Schaffer, M.; Dunham-Scott, ence at the Viterbi School of Engineering, University of
D.; Medby, J. J.; Chan, E.; Gibson, J.; Eisman, M.; Mesic, Southern California.
R.; Kelley, C.; Kim, J.; LaTourrette, T.; and Riley, K. J. 2006.
Implementing Security Improvement Options at Los Angeles Christopher Portway is a student in the Computer Sci-
International Airport. Rand Documented Briefing. Santa ence Department at the Viterbi School of Engineering,
Monica, CA: The Rand Corporation. University of Southern California.
Wagenaar, W. A. 1972. Generation of Random Sequences Milind Tambe is a professor in the Computer Science
by Human Subjects: A Critical Survey of Literature. Psy- Department at the Viterbi School of Engineering, Uni-
chological Bulletin 77(1): 65–72. versity of Southern California.

James Pita is a graduate student in the Computer Science Craig Western is a student in the Computer Science
Department at the Viterbi School of Engineering, Uni- Department at the Viterbi School of Engineering, Uni-
versity of Southern California. versity of Southern California.

Manish Jain is a graduate student in the Computer Sci- Praveen Paruchuri is a research scientist with Intelligent
ence Department at the Viterbi School of Engineering, Automation, Inc., Rockville, Maryland.
University of Southern California.
Saris Kraus is a professor in the Computer Science
Fernando Ordóñez is an associate professor with joint Department at Bar-Ilan University.

SPRING 2009 57

Вам также может понравиться