Title: Effort Games and the Price of Myopia
1Effort Games and the Price of Myopia
Joint work with Yoram Bachrach and Jeff
Rosenschein
2Agenda
- What are Effort Games
- The complexity of incentives in unanimous effort
games - The price of myopia (PoM) in unanimous effort
games - The PoM in SPGs simulation results
- Rewards and the Banzhaf power index
3Effort Games - informally
- Multi-agent environment
- A common project depends on various tasks
- Winning and losing subsets of the tasks
- The probability of carrying out a task is higher
when the agent in charge of it exerts effort - There is a certain cost for exerting effort
- The principal tries to incentivize agents to
exert effort - The principal can only reward agents based on
success of the entire project
4Effort Games - Example
- A communication network
- A source node and a target node
- Each agent is in charge of maintenance tasks for
some link between the nodes - If no maintenance is expended on a link, it has
probability a of functioning
5Effort Games example (2)
- If maintenance effort is made
- for the link, it has probability ß a of
functioning - A mechanism is in charge of sending some
information between source and target - The mechanism only knows whether it succeeded to
send the information, but does not know which
link failed, or whether it failed due to lack of
maintenance
6Example 2
- Voting domain
- Only the result is published, and not the votes
of the participants - Some outsider has a certain desired outcome x of
the voting - The outsider uses lobbying agents, each agent
able to convince a certain voter to vote for the
desired outcome
7Example 2 (2)
- When the agent exerts effort, his voter votes for
x with high probability ß, otherwise the voter
votes for x with probability a ß - The outsider only knows whether x was chosen or
not, he doesnt know what agents exerted effort,
or even who voted for x.
8Related Work
- E. Winter. Incentives and discrimination, 2004.
- ß 1, the only winning coalition is the grand
coalition, a is the same for all agents,
iterative elimination of dominated strategies
implementation, focuses on economic question of
discrimination. - M. Babaioff, M. Feldman and N. Nisan.
Combinatorial agency, 2006 - Focuses on Nash Equilibrium
- Very different results
9Preliminaries
- An n-player normal form game is given by
- A set of agents (players) I 1,,n
- For each agent i a set of pure strategies Si,
- A utility (payoff) function Fi(s1,, sn).
- We denote the set of strategy profiles
- Denote items in S as .
- We also denote
, and - denote
10Dominant Strategy
- Given a normal form game G, we say agent is
strategy strictly dominates
if for any incomplete strategy profile
, - .
- We say agent is strategy sx is is dominant
strategy if it dominates all other strategies
.
11Dominant Strategy Equilibrium
- Given a normal form game G, we say a strategy
profile is a
dominant strategy equilibrium if for any agent i,
strategy si is a dominant strategy for i.
12Iterated Elimination of Dominated Strategies
- In iterated dominance, strictly dominated
strategies are removed from the game, and have no
longer effect on future dominance relations. - Well-known fact iterated strict dominance is
path-independent.
p1\p2 s1 s2
s1
s2
13Simple Coalitional Game
- A simple coalitional game is a domain that
consists of a set of tasks, T, and a
characteristic function . - A coalition wins if ,
and it loses if .
14Weighted Voting Game
- A weighted voting game is a simple coalitional
game with tasks T t1,,tn, a vector of
weights w (w1,,wn), and a threshold q. - We say ti has the weight wi.
- The weight of coalition is
- The coalition C wins the game if ,
and it loses if .
15The Banzhaf power index
- The Banzhaf power index depends on the number of
coalitions in which an agent is critical, out of
all possible coalitions. - It is given by ,
where -
16Effort Game domain
- A set I1,n of n agents
- A set T t1,,tn of n tasks
- A simple coalitional game G with task set T and
with the value function - A set of success probability pairs (a1,ß1),,
(an,ßn), such that , and - A set of effort exertion costs
, such that ci gt 0.
17Effort Game domain interpretation
- A joint project depends on completion of certain
tasks. - Achieving some subsets of the tasks completes the
project successfully, and some fail, as
determined by the game G. - Agent i is responsible of task ti.
- i can exert effort on achieving the task, which
gives it probability of ßi to be completed - Or i can shirk (do not exert effort), and the
task will be completed with probability ai. - The exertion of effort costs the agent ci.
18Observations about the Model
- Given a coalition C of agents that contribute
effort, we define the probability that a certain
task is completed by - Given a subset of tasks T and a coalition of
agents that contribute effort C, we can calculate
the probability that exactly the tasks in T are
achieved
if
otherwise
19Observations about the Model (2)
- We can calculate the probability that any winning
subset of tasks is achieved - Given the reward vector r(r1,,rn), and given
that the agents that exert effort are
, is expected reward is
20Effort Game - definition
- An Effort Game is the normal form game
defined on the above domain, a simple
coalitional game G and a reward vector r (r1,,
rn), as follows. - In i has two strategies Si exert,
shirk - Given a strategy profile
, we denote the coalition of agents that
exert effort in s by - The payoff function of each agent
if si exert
if si shirk
21Incentive-Inducing Schemes
- Given an effort game domain Ge, and a coalition
of agents C that the principal wants to exert
effort, we define - A Dominant Strategy Incentive-Inducing Scheme for
C is a reward vector r (r1,,rn), such that
for any , exerting effort is a
dominant strategy for i. - An Iterated Elimination of Dominated Strategies
Incentive-Inducing Scheme for C is a reward
vector r (r1,,rn), such that in the effort
game , after any sequence of eliminating
dominated strategies, for any , the
only remaining strategy for i is to exert effort.
22The Complexity of Incentives
- The following problems concern a reward vector r
(r1,,rn), the effort game , and a
target agent i - DSE (DOMINANT STRATEGY EXERT) Given ,
is exert a dominant strategy for i ? - IEE (ITERATED ELIMINATION EXERT) Given
, is exert the only remaining strategy for i,
after iterated elimination of dominated
strategies ?
23The Complexity of Incentives (2)
- The following problems concern the effort game
domain D, and a coalition C - MD-INI (MINIMUM DOMINANT INDUCING INCENTIVES)
Given D, compute the dominant strategy incentive
inducing scheme r (r1,,rn) for C that
minimizes the sum of payments, . - MIE-INI (MINIMUM ITERATED ELIMINATION INDUCING
INCENTIVES) Given D, compute the iterated
elimination of dominated strategies
incentive-inducing scheme, that minimizes the sum
of payments, .
24Unanimous Effort Games
- The underlying coalitional game G has task set T
t1,,tn - v(T) 1, and for all C ? T v(C) 0.
- Success probability pairs are identical for all
the tasks (a1 a, ß1 ß), , (an a, ßn ß) - a lt ß
- Agent i is in charge of the task ti, and has an
effort exertion cost of ci.
25Results MD-INI
- Lemma if and only if
exerting effort is a dominant strategy for i. - Corollary MD-INI is in P for the unanimous
effort games. The reward vector -
is the dominant strategy incentive-inducing
scheme that minimizes the sum of payments.
26Results IE-INI
- Lemma Let i and j be two agents, and ri and rj
be their rewards such that
and . - Then under iterated elimination of dominated
strategies, the only remaining strategy for both
i and j is to exert effort. - Theorem IE-INI is in P for the effort game in
the above-mentioned weighted voting domain. For
any reordering of the agents , a
reward vector rp where the following holds for
any agent p(i), is an iterated elimination
incentive-inducing scheme
27Results MIE-INI
- Theorem MIE-INI is in P for the effort games in
the above unanimous weighted voting domain. The
reward vector rp that minimizes the sum of
payments is achieved by sorting the agents by
their weights wi ci, from the smallest to the
largest.
28The Price of Myopia (PoM)
- Denote by RDS the minimum sum of rewards ( )
such that for all i, ri is a dominant strategy
incentive-inducing scheme. - Denote by RIE the minimum sum of rewards ( )
such that for all i, ri is an iterated
elimination of dominated strategies
incentive-inducing scheme. - Clearly RDS RIE
- Since in RIE we are taking into account only a
subset of strategies - How large can the ratio be ?
- We call this ratio the price of myopia.
29PoM in Unanimous Effort Games
- In the underlying weighted voting game G a
coalition wins if it contains all the tasks, and
loses otherwise. - All the agents have the same cost for exerting
effort, c. - Success probability pairs are identical for all
tasks, (a, ß). - Theorem in the above setting ,
where n 2 is the number of players.
30Series-Parallel Graphs (SPGs)
- SPGs have two distinguished vertices, s and t,
called source and sink, respectively. - Start with a set of copies of single-edge graph
K2.
31Effort Games over SPGs
- An SPG G (V,E) representing a communication
network. - For each edge there is an agent i,
responsible for maintenance tasks for ei. - i can exert an effort at the certain cost ci, to
maintain the link ei, and then it will have
probability ß of functioning
32Effort Games over SPGs (2)
- Otherwise (if the agent does not exert effort on
maintaining the link), it will have probability a
of functioning - The winning coalitions are those containing a
path from source to sink - This setting generalizes the unanimous weighted
voting games
33Simulation Setting
- All our simulations were carried
- out on SPGs with 7 edges,
- whose composition is described
- by the following tree
- The leaves of the tree represent the edges of the
SPG. - The inner node of the tree represents SPG that is
series or parallel composition of its children.
34Simulation Setting (2)
- We have sampled uniformly at random SPGs giving
probability of ½ for series composition and
probability of ½ for parallel composition at each
inner node of the above tree. - We have computed the PoM for a 0.1, 0.2,,0.8
and ß a 0.1, a 0.2,,0.9 - The costs ci have been sampled uniformly at
random in 0.0, 100). - For each pair (a, ß) 500 experiments have been
made in order to find the average PoM, and the
standard deviation of the PoM.
35Simulation Results
0.9 0.8 0.7 0.6 0.5 0.4 0.3 0.2 a \ ß
177..85 94.45 58.04 38.04 25.27 16.26 9.43 4.22 0.1
37.64 22.64 14.98 10.12 6.70 4.15 2.27 0.2
14.12 9.08 6.21 4.25 2.82 1.76 0.3
6.67 4.55 3.22 2.26 1.54 0.4
3.71 2.67 1.95 1.41 0.5
2.34 1.76 1.33 0.6
1.64 1.28 0.7
1.24 0.8
Average PoM
36Simulation Results (2)
0.9 0.8 0.7 0.6 0.5 0.4 0.3 0.2 a \ ß
472.22 221.13 119.72 69.40 39.99 21.85 10.23 3.05 0.1
65.05 34.17 19.68 11.43 6.30 3.01 1.00 0.2
19.52 10.39 5.92 3.27 1.60 0.57 0.3
7.28 3.91 2.14 1.07 0.39 0.4
3.02 1.60 0.80 0.30 0.5
1.32 0.65 0.25 0.6
0.56 0.22 0.7
0.20 0.8
Standard deviation of the PoM
37Simulation Results Interpretation
- As one can see from the first table, the higher
the distance between a and ß, the higher the PoM
is. - When there are large differences in the
probabilities, the structure of the graph is more
important - Fits our expectations
38SPG with Parallel Composition
- Theorem Let G be SPG which is obtained by
parallel composition of a set of copies of
single-edge graphs K2. As before, each agent is
responsible for a single edge, and a winning
coalition is one which contains a path (an edge)
connecting the source and target vertices. Then
we have RDS RIE.
39Rewards and the Banzhaf power index
- Theorem Let D be an effort game domain where for
i we have ai 0 and ßi 1, and for all j ? i we
have aj ßj ½, and let r (r1,,rn) be a
reward vector. Exerting effort is a dominant
strategy for i in Ge(r) if and only if
(where ßi(v) is the Banzhaf power index of
ti in the underlying coalitional game G, with the
value function v).
40DSE (DOMINANT STRATEGY EXERT) hardness result
- Theorem DSE is as hard computationally as
calculating the Banzhaf power index of its
underlying coalitional game G. - Corollary It is NP-hard to test whether exerting
effort is dominant strategy in effort game where
the underlying coalitional game is weighted
voting game.
41Conclusions
- We defined the effort games
- Found how to compute optimal rewards schemes in
unanimous effort games - Defined the PoM
- Provided results about PoM in unanimous weighted
voting games - Gave simulation results about PoM in SPGs
- Connected the complexity of computing incentives
and complexity of finding the Banzhaf power
index.
42Future Work
- Complexity of incentives for other classes of
underlying games - Approximation of incentive-inducing schemes
- The PoM in various classes of effort games