Effort Games and the Price of Myopia - PowerPoint PPT Presentation

About This Presentation
Title:

Effort Games and the Price of Myopia

Description:

Effort Games and the Price of Myopia. Michael Zuckerman ... There is a certain cost for exerting effort. The principal tries to incentivize agents to exert effort ... – PowerPoint PPT presentation

Number of Views:24
Avg rating:3.0/5.0
Slides: 43
Provided by: Han82
Category:
Tags: effort | games | myopia | price

less

Transcript and Presenter's Notes

Title: Effort Games and the Price of Myopia


1
Effort Games and the Price of Myopia
  • Michael Zuckerman

Joint work with Yoram Bachrach and Jeff
Rosenschein
2
Agenda
  • What are Effort Games
  • The complexity of incentives in unanimous effort
    games
  • The price of myopia (PoM) in unanimous effort
    games
  • The PoM in SPGs simulation results
  • Rewards and the Banzhaf power index

3
Effort Games - informally
  • Multi-agent environment
  • A common project depends on various tasks
  • Winning and losing subsets of the tasks
  • The probability of carrying out a task is higher
    when the agent in charge of it exerts effort
  • There is a certain cost for exerting effort
  • The principal tries to incentivize agents to
    exert effort
  • The principal can only reward agents based on
    success of the entire project

4
Effort Games - Example
  • A communication network
  • A source node and a target node
  • Each agent is in charge of maintenance tasks for
    some link between the nodes
  • If no maintenance is expended on a link, it has
    probability a of functioning

5
Effort Games example (2)
  • If maintenance effort is made
  • for the link, it has probability ß a of
    functioning
  • A mechanism is in charge of sending some
    information between source and target
  • The mechanism only knows whether it succeeded to
    send the information, but does not know which
    link failed, or whether it failed due to lack of
    maintenance

6
Example 2
  • Voting domain
  • Only the result is published, and not the votes
    of the participants
  • Some outsider has a certain desired outcome x of
    the voting
  • The outsider uses lobbying agents, each agent
    able to convince a certain voter to vote for the
    desired outcome

7
Example 2 (2)
  • When the agent exerts effort, his voter votes for
    x with high probability ß, otherwise the voter
    votes for x with probability a ß
  • The outsider only knows whether x was chosen or
    not, he doesnt know what agents exerted effort,
    or even who voted for x.

8
Related Work
  • E. Winter. Incentives and discrimination, 2004.
  • ß 1, the only winning coalition is the grand
    coalition, a is the same for all agents,
    iterative elimination of dominated strategies
    implementation, focuses on economic question of
    discrimination.
  • M. Babaioff, M. Feldman and N. Nisan.
    Combinatorial agency, 2006
  • Focuses on Nash Equilibrium
  • Very different results

9
Preliminaries
  • An n-player normal form game is given by
  • A set of agents (players) I 1,,n
  • For each agent i a set of pure strategies Si,
  • A utility (payoff) function Fi(s1,, sn).
  • We denote the set of strategy profiles
  • Denote items in S as .
  • We also denote
    , and
  • denote

10
Dominant Strategy
  • Given a normal form game G, we say agent is
    strategy strictly dominates
    if for any incomplete strategy profile
    ,
  • .
  • We say agent is strategy sx is is dominant
    strategy if it dominates all other strategies
    .

11
Dominant Strategy Equilibrium
  • Given a normal form game G, we say a strategy
    profile is a
    dominant strategy equilibrium if for any agent i,
    strategy si is a dominant strategy for i.

12
Iterated Elimination of Dominated Strategies
  • In iterated dominance, strictly dominated
    strategies are removed from the game, and have no
    longer effect on future dominance relations.
  • Well-known fact iterated strict dominance is
    path-independent.

p1\p2 s1 s2
s1
s2
13
Simple Coalitional Game
  • A simple coalitional game is a domain that
    consists of a set of tasks, T, and a
    characteristic function .
  • A coalition wins if ,
    and it loses if .

14
Weighted Voting Game
  • A weighted voting game is a simple coalitional
    game with tasks T t1,,tn, a vector of
    weights w (w1,,wn), and a threshold q.
  • We say ti has the weight wi.
  • The weight of coalition is
  • The coalition C wins the game if ,
    and it loses if .

15
The Banzhaf power index
  • The Banzhaf power index depends on the number of
    coalitions in which an agent is critical, out of
    all possible coalitions.
  • It is given by ,
    where

16
Effort Game domain
  • A set I1,n of n agents
  • A set T t1,,tn of n tasks
  • A simple coalitional game G with task set T and
    with the value function
  • A set of success probability pairs (a1,ß1),,
    (an,ßn), such that , and
  • A set of effort exertion costs
    , such that ci gt 0.

17
Effort Game domain interpretation
  • A joint project depends on completion of certain
    tasks.
  • Achieving some subsets of the tasks completes the
    project successfully, and some fail, as
    determined by the game G.
  • Agent i is responsible of task ti.
  • i can exert effort on achieving the task, which
    gives it probability of ßi to be completed
  • Or i can shirk (do not exert effort), and the
    task will be completed with probability ai.
  • The exertion of effort costs the agent ci.

18
Observations about the Model
  • Given a coalition C of agents that contribute
    effort, we define the probability that a certain
    task is completed by
  • Given a subset of tasks T and a coalition of
    agents that contribute effort C, we can calculate
    the probability that exactly the tasks in T are
    achieved

if
otherwise
19
Observations about the Model (2)
  • We can calculate the probability that any winning
    subset of tasks is achieved
  • Given the reward vector r(r1,,rn), and given
    that the agents that exert effort are
    , is expected reward is

20
Effort Game - definition
  • An Effort Game is the normal form game
    defined on the above domain, a simple
    coalitional game G and a reward vector r (r1,,
    rn), as follows.
  • In i has two strategies Si exert,
    shirk
  • Given a strategy profile
    , we denote the coalition of agents that
    exert effort in s by
  • The payoff function of each agent

if si exert
if si shirk
21
Incentive-Inducing Schemes
  • Given an effort game domain Ge, and a coalition
    of agents C that the principal wants to exert
    effort, we define
  • A Dominant Strategy Incentive-Inducing Scheme for
    C is a reward vector r (r1,,rn), such that
    for any , exerting effort is a
    dominant strategy for i.
  • An Iterated Elimination of Dominated Strategies
    Incentive-Inducing Scheme for C is a reward
    vector r (r1,,rn), such that in the effort
    game , after any sequence of eliminating
    dominated strategies, for any , the
    only remaining strategy for i is to exert effort.

22
The Complexity of Incentives
  • The following problems concern a reward vector r
    (r1,,rn), the effort game , and a
    target agent i
  • DSE (DOMINANT STRATEGY EXERT) Given ,
    is exert a dominant strategy for i ?
  • IEE (ITERATED ELIMINATION EXERT) Given
    , is exert the only remaining strategy for i,
    after iterated elimination of dominated
    strategies ?

23
The Complexity of Incentives (2)
  • The following problems concern the effort game
    domain D, and a coalition C
  • MD-INI (MINIMUM DOMINANT INDUCING INCENTIVES)
    Given D, compute the dominant strategy incentive
    inducing scheme r (r1,,rn) for C that
    minimizes the sum of payments, .
  • MIE-INI (MINIMUM ITERATED ELIMINATION INDUCING
    INCENTIVES) Given D, compute the iterated
    elimination of dominated strategies
    incentive-inducing scheme, that minimizes the sum
    of payments, .

24
Unanimous Effort Games
  • The underlying coalitional game G has task set T
    t1,,tn
  • v(T) 1, and for all C ? T v(C) 0.
  • Success probability pairs are identical for all
    the tasks (a1 a, ß1 ß), , (an a, ßn ß)
  • a lt ß
  • Agent i is in charge of the task ti, and has an
    effort exertion cost of ci.

25
Results MD-INI
  • Lemma if and only if
    exerting effort is a dominant strategy for i.
  • Corollary MD-INI is in P for the unanimous
    effort games. The reward vector

  • is the dominant strategy incentive-inducing
    scheme that minimizes the sum of payments.

26
Results IE-INI
  • Lemma Let i and j be two agents, and ri and rj
    be their rewards such that
    and .
  • Then under iterated elimination of dominated
    strategies, the only remaining strategy for both
    i and j is to exert effort.
  • Theorem IE-INI is in P for the effort game in
    the above-mentioned weighted voting domain. For
    any reordering of the agents , a
    reward vector rp where the following holds for
    any agent p(i), is an iterated elimination
    incentive-inducing scheme

27
Results MIE-INI
  • Theorem MIE-INI is in P for the effort games in
    the above unanimous weighted voting domain. The
    reward vector rp that minimizes the sum of
    payments is achieved by sorting the agents by
    their weights wi ci, from the smallest to the
    largest.

28
The Price of Myopia (PoM)
  • Denote by RDS the minimum sum of rewards ( )
    such that for all i, ri is a dominant strategy
    incentive-inducing scheme.
  • Denote by RIE the minimum sum of rewards ( )
    such that for all i, ri is an iterated
    elimination of dominated strategies
    incentive-inducing scheme.
  • Clearly RDS RIE
  • Since in RIE we are taking into account only a
    subset of strategies
  • How large can the ratio be ?
  • We call this ratio the price of myopia.

29
PoM in Unanimous Effort Games
  • In the underlying weighted voting game G a
    coalition wins if it contains all the tasks, and
    loses otherwise.
  • All the agents have the same cost for exerting
    effort, c.
  • Success probability pairs are identical for all
    tasks, (a, ß).
  • Theorem in the above setting ,
    where n 2 is the number of players.

30
Series-Parallel Graphs (SPGs)
  • SPGs have two distinguished vertices, s and t,
    called source and sink, respectively.
  • Start with a set of copies of single-edge graph
    K2.

31
Effort Games over SPGs
  • An SPG G (V,E) representing a communication
    network.
  • For each edge there is an agent i,
    responsible for maintenance tasks for ei.
  • i can exert an effort at the certain cost ci, to
    maintain the link ei, and then it will have
    probability ß of functioning

32
Effort Games over SPGs (2)
  • Otherwise (if the agent does not exert effort on
    maintaining the link), it will have probability a
    of functioning
  • The winning coalitions are those containing a
    path from source to sink
  • This setting generalizes the unanimous weighted
    voting games

33
Simulation Setting
  • All our simulations were carried
  • out on SPGs with 7 edges,
  • whose composition is described
  • by the following tree
  • The leaves of the tree represent the edges of the
    SPG.
  • The inner node of the tree represents SPG that is
    series or parallel composition of its children.

34
Simulation Setting (2)
  • We have sampled uniformly at random SPGs giving
    probability of ½ for series composition and
    probability of ½ for parallel composition at each
    inner node of the above tree.
  • We have computed the PoM for a 0.1, 0.2,,0.8
    and ß a 0.1, a 0.2,,0.9
  • The costs ci have been sampled uniformly at
    random in 0.0, 100).
  • For each pair (a, ß) 500 experiments have been
    made in order to find the average PoM, and the
    standard deviation of the PoM.

35
Simulation Results
0.9 0.8 0.7 0.6 0.5 0.4 0.3 0.2 a \ ß
177..85 94.45 58.04 38.04 25.27 16.26 9.43 4.22 0.1
37.64 22.64 14.98 10.12 6.70 4.15 2.27 0.2
14.12 9.08 6.21 4.25 2.82 1.76 0.3
6.67 4.55 3.22 2.26 1.54 0.4
3.71 2.67 1.95 1.41 0.5
2.34 1.76 1.33 0.6
1.64 1.28 0.7
1.24 0.8
Average PoM
36
Simulation Results (2)
0.9 0.8 0.7 0.6 0.5 0.4 0.3 0.2 a \ ß
472.22 221.13 119.72 69.40 39.99 21.85 10.23 3.05 0.1
65.05 34.17 19.68 11.43 6.30 3.01 1.00 0.2
19.52 10.39 5.92 3.27 1.60 0.57 0.3
7.28 3.91 2.14 1.07 0.39 0.4
3.02 1.60 0.80 0.30 0.5
1.32 0.65 0.25 0.6
0.56 0.22 0.7
0.20 0.8
Standard deviation of the PoM
37
Simulation Results Interpretation
  • As one can see from the first table, the higher
    the distance between a and ß, the higher the PoM
    is.
  • When there are large differences in the
    probabilities, the structure of the graph is more
    important
  • Fits our expectations

38
SPG with Parallel Composition
  • Theorem Let G be SPG which is obtained by
    parallel composition of a set of copies of
    single-edge graphs K2. As before, each agent is
    responsible for a single edge, and a winning
    coalition is one which contains a path (an edge)
    connecting the source and target vertices. Then
    we have RDS RIE.

39
Rewards and the Banzhaf power index
  • Theorem Let D be an effort game domain where for
    i we have ai 0 and ßi 1, and for all j ? i we
    have aj ßj ½, and let r (r1,,rn) be a
    reward vector. Exerting effort is a dominant
    strategy for i in Ge(r) if and only if
    (where ßi(v) is the Banzhaf power index of
    ti in the underlying coalitional game G, with the
    value function v).

40
DSE (DOMINANT STRATEGY EXERT) hardness result
  • Theorem DSE is as hard computationally as
    calculating the Banzhaf power index of its
    underlying coalitional game G.
  • Corollary It is NP-hard to test whether exerting
    effort is dominant strategy in effort game where
    the underlying coalitional game is weighted
    voting game.

41
Conclusions
  • We defined the effort games
  • Found how to compute optimal rewards schemes in
    unanimous effort games
  • Defined the PoM
  • Provided results about PoM in unanimous weighted
    voting games
  • Gave simulation results about PoM in SPGs
  • Connected the complexity of computing incentives
    and complexity of finding the Banzhaf power
    index.

42
Future Work
  • Complexity of incentives for other classes of
    underlying games
  • Approximation of incentive-inducing schemes
  • The PoM in various classes of effort games
Write a Comment
User Comments (0)
About PowerShow.com