Document 10965323

advertisement
It’s Now or Never:
Deadlines and cooperation
Kaz Miyagiwa
Emory University
Abstract: Cooperation among opportunistic agents often breaks down when agents cannot
observe one another’s actions. The standard remedy for such a problem is a two-mode
approach pioneered by Green and Porter (1984), where agents switch back and forth
between cooperation and punishment mode. Here, I consider use of a deadline as an
alternative. I find that, under certain conditions, imposing a deadline can induce
cooperation with unobservable actions, and that the optimal deadline can payoffdominates the optimal two-mode strategy.
Correspondence: Kaz Miyagiwa, Department of Economics, Emory University, Atlanta,
GA 30322, U.S.A.; E-mail: kmiyagi@emory.edu; Telephone: 404-727-6363; Fax: 404727-4639
Acknowledgments: I have benefited from comments by seminar participants at Emory,
Florida International, Kobe, and Osaka Universities. Errors are my responsibility.
1. Introduction
Today most of us academic economists do research with someone else. Having a
co-author makes research more enjoyable than working alone, and improves the chance
of solving tough problems, such as proving theorems and identifying next research
agendas. However, it can also give rise to the temptation to free-ride on one’s co-author’s
effort. Such temptations are greater especially when co-authors are in separate locations
and cannot observe each other’s actions. If both researchers succumb to such temptations,
of course, co-authorship breaks down.
How to induce cooperation among opportunistic agents has long been a subject of
intensive research among social scientists, possibly dating back to Rousseau’s problem
over stag hunting.1 Game theory has taught us that, if agents interact repeatedly over
time, cooperation can be induced with threats of punishment.2 However, in stochastic
environments where bad outcomes occur due to bad luck as well as due to lack of effort,
it is impossible to mete out punishment if agents cannot observe one another’s actions.
The standard remedy for this type of problem is a two-mode strategy pioneered by Green
and Porter (1984), in which agents switch back and forth between cooperation and
punishment mode over time. Although all agents are aware in equilibrium that no one has
ever shirked, occasional self-flagellation is necessary in the wake of poor outcomes so as
to make agents toe the line in cooperation mode.
In this paper I consider use of a deadline as an alternative solution to this type of
1
As retold in Fudenburg and Tirole (1991, p. 3), each hunter of a group on a stag-hunting expedition has a
hard time resisting the temptation to catch a rabbit that happens cross his path and to call it a day instead of
hunting a stag that requires cooperation and coordination.
2
A contract-theoretic approach has recently been applied to study how a principal can induce cooperation
among agents. The present paper concerns the desirability to cooperate outside principal-agent
relationships.
2
problem. The setting I consider is as follows. I suppose that agents are engaged in some
‘search activity’ and exert efforts until there is success. The probability of success
depends on total efforts of agents. Agents however cannot observe each other’s action, so
it is impossible to disentangle failures due to shirking from those due to bad luck.
Therefore, if shirking goes undetected an agent who shirks faces the same continuation
payoff as an agent who exerts effort.
Suppose that in such circumstances all agents succumb to the temptation to shirk
so cooperation cannot be maintained. Now consider the effect of a deadline, which
specifies the number of periods during which agents agree to exert efforts. When the
deadline arrives, agents know that it is the last chance to cooperate. Since there is no
future cooperation, agents know that failure leads to a smaller continuation payoff than
when there is no deadline. This decrease in continuation payoff motivates agents to exert
efforts to succeed. Once agents have the incentive to exert efforts at the deadline without
observing each other’s action, I can move back to check their incentive to shirk in earlier
periods. My analysis shows that if the deadline is set too much in the future agents cannot
cooperate. In fact, I find that there is the unique optimal deadline, yielding the maximum
expected payoff to each agent. I then compare the expected payoff from the optimal
deadline with that from the two-mode strategy a la Green-Porter (1984) and show that the
former can payoff-dominate the latter.
Now that I have outlined my analysis it is time to turn to the relevant literature.
As already mentioned, research on cooperation with unobserved actions in stochastic
environment was pioneered by Green and Porter (1984) Focusing on collusion
enforcement in oligopoly under stochastic demand, these authors proposed a two-mode
3
strategy in the class of trigger strategies, whereby firms start out in cooperation mode but
switch to punishment mode when the price falls below a certain level. Abreu, Milgrom
and Pearce (1991) advanced this approach and applied it in a situation similar to the one
considered here. However, their paper differs considerably from the present work. Their
main interest lies with the folk theorem, i.e., the effect of changing discount factors on
the stability of cooperation under the two-mode strategy, whereas here focus is on the
cooperation-inducing effect of a deadline for a given discount factor.
The effect of a deadline has also been examined in the recent bargaining
literature.3 The principal result in this literature is that, when there is a deadline, agents
stall out, making no attempt to reach an agreement until the deadline arrives. Thus, in
bargaining situations a deadline causes inefficient delays in reaching an agreement. In
contrast, in the present paper a deadline has a positive outlook.
The remainder of the paper is organized in six sections. Section 2 describes the
basic environment and presents the benchmark case, in which cooperation fails without
observable actions. Section 3, the main section, introduces a deadline and examines its
property. Section 4 compares the expected payoffs between the optimal deadline and the
optimal two-mode strategy a la Green-Porter (1984) and shows that the former may
dominate the latter. Section 5 extends the analysis to the case of asymmetric agents in
success probability, while Section 6 considers another extension, in which agents never
leave the game when there is success.
See. e.g., Hendricks, Weiss and Wilson (1988), Fershtman and Seidmann (1993), Ponsati (1995), and
Damiano et al. (2010)
3
4
Section 7 concludes the discussion and suggests possible applications of the
analysis to three specific cases; joint research projects in industrial economics, law
enforcement efforts as regards statute of limitations and commitment in military
campaigns.
2. Setup
2.A. Going it alone
Consider a single agent in an infinite time horizon. Time is discrete and all actions
take place at t = 1, 2,…., which are called ‘dates.’ At each t, conditional on not having
succeeded to date, an agent decides whether to exert effort or not. Exerting effort costs
the disutility e, but leads to success at t + 1 with (conditional) probability (1 – p) > 0,
where p > 0 is constant over time. Success entitles an agent to the perpetual flow of
benefits at the rate b, suming up to π = b/(1 - δ), where δ ∈(0, 1) is the discount factor.
If there is failure, a probability p event, an agent faces exactly the same
continuation payoff as at date t due to the stationary environment. Therefore, v, the
expected payoff satisfies this recursive structure:
v = - e + (1 – p)δπ + pδv.
Collecting terms yields
v=
(1 − p)δ b − e
.
1−δ p
5
On the other hand, if an agent chooses not to exert effort, there is possibility of success,
yielding the normalized utility of zero to an agent. I assume v ≥ 0 so that an agent exerts
effort.4
2.B. Cooperation with observable actions
Now suppose that m (≥ 2) identical agents agree to independently and
simultaneously exert efforts and to share the reward yielded by success. If probability of
failure per agent remains constant at p, cooperation reduces the team’s probability of
failure to pm < p. Thus, cooperation functions as insurance here.
Cooperation can also change the rewards. Let B denote the benefit per-period to
each agent and let Π = B/(1 - δ). B need not be equal to b.
In the remainder of this subsection, I assume actions are observable and seek the
condition for cooperation under the standard trigger strategy: “At t = 1, agree to
cooperate and exert effort. At any t ≥ 2, exert effort as long as all agents exerted efforts at
all dates up to t – 1; otherwise go it alone.”
If each agent plays the above strategy, by an argument similar to the one given for
an individual agent, the expected payoff V to each agent must satisfy this recursive
equation:
(1)
V = – e + (1 − pm)δΠ + δpmV
Collecting terms yields
4
This is just for convenience’s sake. If the inequality is reversed, an agent’s reservation utility is till zero,
and the rest of the analysis is unaffected.
6
(1 − p m )δΠ − e
V=
.
1 − δ pm
I assume that cooperation is desirable; i. e., V > v.
The above trigger strategy is an equilibrium strategy if no agent has the unilateral
incentive to shirk. Shirking saves the disutility e, but increases the probability of failure
from pm to pm-1. It also triggers punishment mode, reducing the continuation payoff from
V to v. Thus, an agent who shirks would get the expected payoff
Vd = (1 – pm-1)δΠ + δpm-1v
Therefore, no one shirks if and only if V ≥ Vd, which I assume. This condition can be
expressed, after substitution and rearrangement, as
pm-1(1 − p)δ(Π - V) + pm-1δ(V - v) ≥ e.
Assumption 1: When actions are observable, agents can cooperate as a team in the sense
that V ≥ Vd.
2.C. Unobservable actions
Suppose next that agents cannot observe one another’s actions. If all agents exert
efforts, each agent faces the expected payoff V defined above. However, shirking goes
undetected. If an agent can shirk with impunity, he faces the continuation payoff V insead
of v. Since shirking also increases the probability of failure to pm-1, an agent sho shirks
expects the payoff
(2)
Wd = (1 - pm-1)δΠ + pm-1δV.
7
I assume that
V – Wd = pm-1(1 − p)δ(Π - V) – e < 0.
Assumption 2: With unobserved actions, cooperation is impossible, i. e.,V < Wd.
3. Deadlines
In this section I focus my attention on the cases in which both assumptions 1 and
2 hold; i.e., Wd > V ≥ Vd. In other words, agents prefer to cooperate but cannot because
they cannot observe each other’s actions. Using the defining expressions, these
conditions can be expressed, after manipulation, as
pm-1(1 − p)δ(Π - V) + pm-1δ(V - v) ≥ e > pm-1(1 − p)δ(Π - V).
Since V > v, there is a range of e satisfying the above conditions.
The standard remedy to this problem is a two-mode strategy a la Green-Porter
(1984). Translated into the present context, the Green-Porter (1984) approach has all
agents exert efforts at t = 1. If there is failure, agents go solo for a given number of
periods, after which they return to cooperation mode, and so forth. By temporarily halting
cooperation, this strategy reduces the continuation payoff and induces cooperation. The
objective of the present section is to present a deadline as an alternative strategy for
inducing efforts.
Let me being by explaining what I mean by deadlines. First, define the sequence
of strategies Cn; n = 1, 2,… , which says “At dates t (where 1 ≤ t ≤ n) cooperate and exert
effort. At t > n, go it alone.” That is, agents cooperate and exert efforts only up to date t =
8
n and go it alone from t = n + 1 on. In this case, I say that there is a deadline at t = n, and
call Cn the deadline-n strategy.
Let U(n) denote the expected payoff to each agent at t = 1, given that all agents
play Cn, and let Ud(n) the corresponding expected payoff to an agent who shirks at t = 1.
With C1, these are expressed as
U(1) = – e (1 – pm)δΠ + pmδv
Ud(1) = (1 – pm-1)δΠ + pm-1δv.
No agent shirks if and only if
U(1) – Ud(1) = pm-1(1 - p)δ(Π – v) – e ≥ 0.
Since V > v, there is a range of e satisfying:
(3)
pm-1(1 - p)δ(Π – v) ≥ e ≥ pm-1(1 − p)δ(Π - V).
The first inequality ensures that U(1) > Ud(1) while the second equality implies V < Wd.
It is easy to check that an e satisfying (3) also satisfies assumption 1. We have proved
Proposition 1: If an e satisfies (3), it satisfies assumptions 1 and 2. For such an e, all
agents exert effort at date 1 under strategy C1 without observing one another’s action.
Proposition 1 has the following intuitive explanation. Shirking always increases the
probability of failure from pm to pm-1. Further, while the continuation is V without the
deadline, it is now v < V with the deadline. Thus, the deadline reduces the continuation
payoff, prompting agents to exert efforts, as does the Green-Porter strategy.
9
If condition (3) holds so that agents cooperate for at least one period, then I
consider C2. If all agents play C2, the expected payoff at t = 1 is
U(2) = – e + (1 – pm)δΠ + pmδU(1),
whereas shirking yields
Ud(2) = (1 – pm-1)δΠ + pm-1δU(1).
There is no shirking if
U(2) – Ud(2) = pm-1(1 – p)δ(Π - U(1)) – e ≥ 0.
If this inequality holds, agents cooperate at least for the first two periods, in which case I
move to C3 and so on.
This process of extending the deadline eventually comes to an end because the
deadline extended indefinitely is no deadline at all so that by assumption 2 agents shirk.
To show this formally, assume that Cn-1 induces cooperation (for the first n – 1 periods).
Then, the expected payoff under the strategy Cn is:
(4)
U(n) = – e + (1 – pm)δΠ + pmδU(n – 1).
This is a first-order difference equation, with the solution
U(n) = (v – V) (pmδ)n + V.
Since v < V, U(n) increases monotonically; that is, the longer agents can cooperate, the
larger the expected payoff.
Shirking yields the expected payoff:
(5)
Ud(n) = pm-1(1 – q)δΠ + pm-1δU(n - 1)
10
so there is no incentive to shirk under Cn if
U(n) – Ud(n) = – e + pm-1(1 – p)δ(Π - U(n - 1)) ≥ 0.
(6)
Because U(n) is monotone increasing, (6) implies that U(n) – Ud(n) is monotone
decreasing.
I now show that there is an n* ≥ 2 such that U(n) – Ud(n) < 0 for all n > n*.
Substituting from (1), I rewrite (4) as
U(n) = V – pmδ(V – U(n – 1)).
Similarly, substituting from (2) into (5) yields
Ud(n) = Wd – pm-1δ(V – U(n - 1)).
Therefore,
U(n) – Ud(n) = V – Wd + δpm-1(1 – p)(V – U(n - 1)).
Now, as n → ∞, U(n – 1) → V since a deadline set at infinity is no deadline at all,
and hence U(n) – Ud(n) → (V – Wd), which is negative by assumption 2.
I have proved
Proposition 2: If U(1) – Ud(1) ≥ 0, there exists the unique integer n* ≥ 1 such that
U(n* + 1) – Ud(n* + 1) < 0 ≤ U(n*) – Ud(n*).
The uniqueness of n* comes from the monotonicity of U(n) proved above. Proposition 2
has the following explanation. When the deadline t = n is reached, an agent faces the
11
continuation payoff v because there will no more be cooperation. Moving back one
period, agents at t = n – 1 faces the continuation payoff U(1) because there is one more
period of cooperation left even if there is failure. Since U(1) > v, each agent has a greater
incentive to shirk at t = n – 1. The incentive to shirk is even greater at t = n – 2, when the
continuation payoff is U(2) > U(1). Since the continuation payoff U(n) is increasing,
moving back in time, agents eventually reach the period n* so that U(n* + 1) < Ud(n* +
1).
If the deadline cannot be extended beyond n* without destorying cooperation, the
monotonicity of U(n) implies that the deadline n* yields the maximal welfare to each
agent; i.e., n* is the optimal deadline. The next proposition gives further characterizations
of the optimal deadline n* (see Appendix A for the proof).
Proposition 3.
(A) The greater the reward Π of cooperation, the greater n* tends to be.
(B) Given Π, the greater the number of cooperating agents m, the smaller n* tends to be.
Note that to derive part B of proposition 3 I assumed constancy of Π with respect to the
number of agents. In applications, Π may fall with the number of agents involved,
especially if success yields the fixed reward to be divided among agents. In such cases,
12
depending the rate at which Π changes, part B may not hold locally, although it must at n
sufficiently large.5
4. A comparison with the two-mode strategy a la Green-Porter (1984)
Although the original work of Green and Porter (1984) analyzed a situation
drastically different from the one considered here, their methodology is fully applicable
to the present model. Therefore, it would be of considerable interest to compare the two
approaches.
Translated into our context, the two-mode strategy a la Green-Porter (henceforth
the GP strategy) has agents exerting efforts in the first period, only to switch to
punishment mode if there is failure. Once in punishment mode, agents go it alone for a
fixed number of periods, after which they return to cooperation mode. In equilibrium this
pattern repeats itself over time. Inclusion of punishment mode in the strategy lowers the
continuation payoff and makes shirking less appealing to agents. The main difference is
that the two-mode strategy spreads out cooperation periods over the entire horizon
whereas the deadline frontloads them.
Now I derive the optimal two-mode strategy in the present context. Begin by
letting Gc and Gp denote, respectively, the cooperation-model and the punishment-mode
expected payoffs. Since failure triggers punishment mode, Gc can be expressed as
Gc = – e + (1 − pm)δΠ + δpmGp.
5
With n approaching infinity, cheating by one agent will have no effect on success probability but a cheater
saves the disutility e and hence deviates.
13
Substituting from the definition of V in (1), I can rewrite this as
(7)
Gc = (1 - δpm)V + δpmGp.
Similarly, if agents stay in punishment model for τ – 1 periods, Gp can be expressed, after
some manipulation, as
(8)
Gp = [1 – (δp)τ]v + (δp)τGc.
Solving (7) and (8) simultaneously yields
(9)
[1 − (δ p)τ ]δ p m v + (1 − δ p m )V
Gc =
[1 − (δ p)τ ]δ p
Gp =
[1 − (δ p)τ ]v + (δ p)τ (1 − δ p m )V
.
[1 − (δ p)τ ]δ p
Now, let Gd denote the payoff to a cheater when agents are in cooperation mode,
i. e.,
Gd = (1 – pm- 1)δΠ + pm-1δGp.
There would not be cheating if
(10)
Gc – Gd = – e + pm-1(1 − p)δ(Π - Gp) ≥ 0.
All these payoffs depend on τ so write Gc(τ), Gp(τ) and Gd(τ).
It is easy to check that Gp(τ) is decreasing, because the longer agents stay in
punishment mode, the smaller the expected payoff. Thus, the optimal GP strategy
minimizes Gp(τ) with respect to τ subject to the no-deviation constraint (10) and the
integer constraint. Let τ* be the optimizer of this program.
14
Since the above is a well-structured optimization problem, one may be
tempted to conclude that the optimal GP strategy payoff-dominates the optimal
deadline strategy. However, the next proposition shows that the converse can be
true.
Proposition 4: If Gp(τ*) ≤ U(τ*), then the optimal deadline strategy payoff-dominates
the GP strategy; that is, U(n*) > Gc(τ*).
Proof: U(n) is increasing and Gp(τ) is decreasing. Further, U(0) = Gp(∞) = v and U(∞) =
Gp(0) = V. Thus, there is a unique positive integer, say, τo, so that U(τ) ≥ Gp(τ) for all τ
≥ τo. Now, consider:
(11)
0 ≤ – e + pm-1(1 − p)δ(Π - Gp(τ*))
≤ – e + pm-1(1 − p)δ(Π - U(τ*)) = U(τ* + 1) – Ud(τ* + 1).
The first inequality holds because τ* is the integer satisfying constraint (10). The second
is due to the assumption of proposition 4. The equality follows because of (6).
(11) imply that the deadline strategy Cτ* + 1 is incentive-compatible, meaning that
the optimal deadline strategy yields the expected payoff at least as large as U(τ* + 1).
The proof is now complete because
U(τ* + 1) = – e + (1 − pm)δΠ + δpmU(τ*)
15
≥ – e + (1 − pm)δΠ + δpmGp(τ*) = Gc(τ*),
where the first equality follows from the recursive definition of U(n), while the second
equality follows from the definition of Gc(τ) given in (9).

6. Asymmetric agents
In this section I relax the assumption that agents are symmetric, and study the
effect from the differences in probability of success among agents. To lighten notation, I
consider the case with two agents, whom I call Senior (s) and Junior (j). Senior has a
better chance of success, or a lower probability of failure, than Junior; i.e., ps < pj. They
are identical in all other respects agents. In particular, they have the same discount factor,
incur the same disutility from effort and enjoy the same benefits b and B from individual
and team success, respectively. The analysis follows closely that of the preceding section,
so details are omitted.
Going it alone yields the expected payoff vi to agent i = s, j, where:
(12)
vi = [(1 – pi)δπ – e]/(1 – δpi).
The difference in failure probability implies that vj < vs; Junior’s expected payoff is
smaller because he is more likely to fail than Senior. Assume vj ≥ 0 so both agents find
individual search worthwhile undertaking.
If agents agree to cooperate and exert effort, the probability of failure is pspj, so
the expected payoff V to each agent from cooperation satisfies
V = – e + (1 - pjps)δΠ + δpjpsV,
16
and hence is given by:
V = [(1 - pjps)δΠ – e]/(1 – δpjps).
If agents cannot observe each other’s actions, the expected payoff from shirking is
Vid = (1 - pk)δΠ + δpkV;
i, k = j, s (i ≠ k).
Therefore, there is no cooperation if
(13)
V – Vid = – e + pk(1 − pi)δ(Π - V) < 0.
Since Π > V and
ps(1 − pj) – pj(1 − ps) = ps – pj < 0,
(13) implies that V – Vjd < V – Vsd; i.e., Senior has more of an incentive to cooperate
than Junior. Intuitively, the team’s success depends more on Senior’s effort. Put
differently, Junior’s shirking does not decrease the team’s success probability as much as
Senior’s and yet both agents face the same disutility of effort and receive the same
reward. Therefore, Junior has a greater temptation to free-ride on Senior’s effort. Thus,
this asymmetry in incentives may be corrected with a side payment from Senior to Junior,
provided that V – Vjd < 0 < V – Vsd. However, it can be checked that a side payment
reduces V – Vsd more than it increases V – Vsd, so such an arrangement may not always
be possible.6
by p (1 – p )(1 - δ)ΔΠ/(1 - δp p ) and V – V by p (1 – p )(1 sd
j
s
j s
sd
s
j
δ)ΔΠ/(1 - δp p ). Since p > p , the latter is smaller (in absolute value) than the former.
j s
j
s
6
Transfer of ΔΠ decreases V – V
17
For the rest of the analysis I assume that cooperation is impossible to maintain,
and consider the effect of deadlines. With the deadline n = 1, the expected payoff to agent
i is
Ui(1) = – e + (1 - pjps)δΠ + δpjpsvi.
The expected payoffs from cheating are
Uid(1) = (1 - pk)δΠ + δpkvi.
Hence, there is no cheating if
Uj(1) - Ujd(1) = – e + ps(1 − pj)δ(Π - vj) ≥ 0.
(14)
for Junior and
(15)
Us(1) – Usd(1) = – e + pj(1 − ps)δ(Π - vs) ≥ 0.
for Senior. The second terms on the right of (14) and (15) cannot be ranked a priori
because ps(1 − pj) < pj(1 − ps) = ps – pj < 0 whereas Π – vj > Π – vs. However,
substituting from (12) shows, after manipulation, that if
(16)
e + (1 – δpj)(1 – δps)Π < (1 – pjps)δπ,
then Us(1) – Usd(1) < Uj(1) – Ujd(1), implying that Senior has less of the incentive to
cooperate than Junior. Thus, imposing a deadline can result in an incentive reversal.
An incentive reversal occurs because with the deadline Junior has a lower
continuation payoff than Senior (vj < vs) and is harmed more from failure at the deadline.
Thus, while Junior’s shirking does not impact the probability of success as much as
Senior’s, success now means more to Junior than to Senior. When (16) holds, Junior is
more concerned with success than saving the disutility, and has a greater incentive to
18
cooperate than Senior. In that case, if Us(1) – Usd(1) < 0 < Uj(1) – Ujd(1), Junior may
offer a side payment to Senior to ensure cooperation. Such a payment increases Us(1) –
Usd(1) more than it decreases Uj(1) – Ujd(1), so that such a side payment can always be
arranged provided that Us(1) + Uj(1) > Usd(1) + Ujd(1). The next proposition summarizes
the findings.
Proposition 5: With unobserved actions, Junior has a greater incentive to shirk without a
deadline. If(16) holds, the deadline n = 1 results in an incentive reversal; Senior now has
a greater incentive to shirk.
If both agents have the incentive to cooperate with the deadline n = 1, the rest of
the analysis is similarly to the preceding analysis. The following result, proved in the
appendix, is analogue to proposition 4.
Proposition 6. If (14) and (15) are satisfied, a unique optimal deadline exists.
7. Cooperation with repeated search
In the preceding analyses, agents always exit the game when there is success. This
section, motivated by Rouseau’s parable concerning stag hunters (see footnote 1),
considers the case in which the rewards are transitory so that agents must return to search
when the rewards run out. To keep the analysis simple, I assume that the rewards last just
one period, although the qualitative results obtained here are valid for any finite number.
19
When an agent is going it alone, the expected payoff, v, now satisfies this
recursive structure
v = - e + (1 – p)(δb + δ2v) + pδv.
which differs only in the reward from success from the previous analysis. While success
brings b in perpetuity in the preceding analysis, here an agent enjoys the benefit b for one
period (at date t + 1) and returns to search the next date (t + 2), at which he faces the
same expected payoff v. That is, the total reward from success is b + δv instead of π.
Collecting terms yields
v=
(1 − p)δ b − e
.
1 − δ p − (1 − p)δ 2
I again assume that search is worthwhile for each agent, i. e., v ≥ 0.
Turning to cooperation, assume there are just two agents to ease exposition. Given
the reward B per agent, the expected payoff V satisfies
V = - e + (1 – p2)(δB + δ2V) + p2δV
and hence
(1 − p 2 )δ B − e
V=
.
1 − δ p 2 − (1 − p 2 )δ 2
Assume that V > v or cooperation is desirable.
If agents cannot observe each other’s action so that shirking goes unpunished, the
expected payoff to an agent who shirks is
Wd = (1 – p)(δB + δ2V) + pδV.
I assume
20
(17)
V – Wd = - e + p(1 – p)δB – p(1 – p)(1– δ)δV < 0
so that cooperation is impossible to maintain.
Assuming that (15) holds, I now consider the effect of a deadline. As before, let
U(n) denote the expected payoff at t = 1 if both agents play the strategy Cn, and let Ud(n)
be the corresponding expected payoff from shirking at t = 1. For C1 these are written as:
U(1) = - e + (1 – p2)(δB + δ2v) + p2δv
and
Ud(1) = (1 – p)(δB + δ2v) + pδv.
There is no shirking if and only if
(18)
U(1) – Ud(1) = - e + p(1 – p)δ[B – (1 – δ)v].
Substituting from(17) and (18) yields
(V – Wd) – (U(1) – Ud(1)) = pδ(1 – p)(1 – δ)(v – V) < 0.
This indicates that there is a range of e such that imposition of a deadline induces agents
to cooperate for at least one period, while they cannot without a deadline.
The remainder of the analysis is similar to the one in section 4. I show, in
Appendix C, that U(n) is monotone increasing and that U(n) – Ud(n) becomes negative
for n sufficiently large. This leads to the next conclusion (see Appendix C for the proof).
Proposition 7: Suppose that U(1) – Ud(1) > 0. Then there is the optimal deadline n* ≥ 1,
which yields the maximum welfare U(n*) per agent.
21
8. Summary and suggestions for applications and extensions
In this paper I examine stochastic environments in which cooperation among
opportunistic agents is impossible to maintain when they cannot observe one another’s
actions. The main result has been that in such circumstances imposing a deadline can
induce agents to cooperate if the deadline is not set too far in time. In fact, there exists the
unique optimal deadline, which allows cooperation for the maximum duration, yielding
the greatest payoff to each agent.
Since the standard remedy for effort inducement in such circumstance is a twomode approach a la Green and Porter (1984), I then proceed to compare the two
approaches, finding that neither approach can strictly payoff-dominate the other. In the
text, I furnish a sufficient condition in which the optimal deadline strategy yields a
greater expected payoff than the optimal two-mode strategy.
I also consider two extensions to the basic model. In one I examine the effect of
asymmetry in probability of success between two agents. Here I find that a deadline can
result in an incentive reversal; without a deadline an agent having a greater chance of
success has less of an incentive to shirk, but without an deadline he may more likely shirk
than his partner. In the second extension, I examined the case in which agents do not exit
the game with success. In both cases I find that a deadline can induce cooperation when
agents cannot cooperate without one.
Besides co-authorship, the present theoretical framework may illuminate the role
of a deadline for inducing cooperation in a number of diverse situations. I make three
suggestions. A first is in industrial organization. Specifically, for pharmaceutical and biomed firms trying jointly to discover a cure for diseases, joint research projects often
22
depend crucially on government and private grants. Practically all such grants come with
time limits. Our analysis suggests that research grants with time limits can help firms
exert effort. In the cases of more general research joint ventures, the durations are usually
predetermined and often overseen by government agencies wary of antitrust actions.7 For
example, the Advanced Technology Program (ATP), a program within the U.S.
Department of Commerce, requires firms to specify the duration of an RJV on its
application form submitted to secure exemptions from antitrust laws. In Europe, firms
participating Eureka projects must commit to the duration of joint ventures at the time of
applications for research grants. Although these requirements are intended to keep track
of RJV activities, our analysis shows that they may also have the effect of facilitating
cooperation among firms.
Our analysis can also be applied to an area in law and economics. In most
countries including the U.S., crimes that are considered exceptionally heinous by society
have no statute of limitations. In 2006, following the footsteps of other industrial
countries, Japan also finally abolished its statute of limitations for murders. However,
according to the present analysis, statute of limitations may serve as a deadline, inducing
cooperating among law enforcement officers or agencies. The implication is that
abolishment of statute of limitations may have a negative impact on arrest records in
murder cases, which can be empirically investigated. The Japanese case may present a
natural experiment to test the deadline effect.
As a third application, the present analysis can throw a positive spin in
commitment in military strategies. In particular, President Obama drew heavy criticisms
7
Miyagiwa (2009) examines the relationship between research joint ventures and product market collusion.
23
when he announced his commitment to withdraw American troops from Iran by certain
dates. The criticism was based on the fear that the enemies would then just wait out until
withdrawals are complete and then occupy the vacated lands unopposed. However, the
present analysis furnishes a more positive note to the President’s exit plan. Our analysis
suggests that the commitment to withdraw troops by a certain date can serve as a
deadline, inducing the military to exert more efforts in a coordinated manner among all
its branches, making a final victory over enemies more likely before troop withdrawals.
All these applications involve empirical investigation of the deadline effect,
requiring separate treatments, as the present framework must be re-specified to fit the
separate cases. I look forward to working together with empirical economists in these
investigations in the near future, if they agree to a deadline.
24
Appendices
Appendix A: Proof of propositions 3 and 4
To prove proposition 3, differentiate both sides of (10) with respect to Π, noting that
sgn {d(U(n) – Ud(n))/dΠ} = sgn {d(Π - U(n - 1))/dΠ > 0,
The inequality holds since U(n – 1) contains Π only with positive probability. Thus, an
increase in benefit intensifies the incentive to cooperate. Thus, if U(n*+ 1) – Ud(n* + 1) >
0 in (10), then cooperation can be maintained at least for the first n* + 1 periods instead
of the first n* periods. This establishes proposition 3.
Turning to proposition 4, I show that an increase in m increases the incentive to
deviate, given n. The first things to show is that U(n) increases with m. To do so, write
U(n) = U(n; m). Suppose U(n; m) is positive. Then I can show that U(n; m) < U(n; m + 1)
by induction on n. It is easy to check using (3) that the inequality holds for n = 1.
Suppose that U(n; m) < U(n; m + 1) for n ≥ 1. Then by (7)
U(n + 1; m + 1) - U(n + 1; m)
= (pm – pm+1)δΠ + pm+1δU(n; m + 1) - pmδU(n; m)
= δpm{(Π – U(n; m)) –p(Π – U(n; m + 1))} > 0.
However, the incentive to deviate increases in m even faster. To see it, let Ud(n) = Ud(n;
m) and observe that by (8)
[U(n; m + 1) – Ud(n; m + 1)] – [U(n; m) – Ud(n; m)]
= pm(1 – p)δ(Π - U(n – 1; m + 1)) – pm-1(1 – p)δ(Π - U(n – 1; m))
= δ(1 – p)pm{p[Π – U(n – 1; m + 1)] – [Π – U(n – 1; m)]} < 0.
25
The inequality follows because U(n – 1; m + 1) > U(n – 1; m).
Appendix B: proof of proposition 6. For the deadline n ≥ 2, assuming that both exert
effort between t = 1 and t = n, I can write Ui(n) to agent i as
Ui(n) = – e + (1 - pjps)δΠ + δpjpsUi(n – 1).
Ui(n) is monotone increasing for each agent. There will be no incentive for agent i to
cheat if
Ui(n) - Uid(n) = – e + pk(1 − pi)δ(Π - Ui(n-1)) ≥ 0; i, k = j, s (i ≠ k).
As before, it can be shown that the right-hand side of this equation converges to V – Vid
< 0 as n tends to infinity. Thus, there is ni* satisfying that
Ui(n) - Uid(n) ≥ 0 > Ui(n + 1) - Uid(n + 1).
Let n* = min {ns*, nj*}.
Appendix C: Proof of Proposition 7. I first prove that U(n) is monotone increasing by
induction. Define ΔU(n) = U(n) – U(n – 1), with ΔU(1) = U(1) – v > 0. Given that ΔU(1)
> 0, a calculation shows that
(1b)
ΔU(2) = p2δΔU(1) > 0
Now assume that ΔU(n – 2) and ΔU(n – 1) are both positive. Then
(2b)
ΔU(n) = (1 – p2)δ2ΔU(n – 2) + p2δΔU(n – 1)
which is positive under the assumption. Thus, U(n + 1) > U(n), for all n ≥ 1.
26
Next I prove the existence of the optimal deadline n*, given that U(1) – Ud(1) > 0.
Note that U(n) approaches V as n tends to infinity. The expected payoff from cheating is
given by
Ud(n) = (1 – pm-1)(δB + δ2U(n – 2)) + pm-1δU(n – 1)
= Wd + pmδ(U(n – 2) – V)
which approaches Wd as n tends to infinity. Thus, as n tends to infinity, U(n) – Ud(n)
approaches V – Wd < 0. Thus the deadline cannot be extended infinitely while
maintaining cooperation.
27
References
Abreu, D., P. Milgrom, and D. Pearce, 1991, Information and timing in repeated
partnerships, Econometrica 59, 1713-733.
Damiano, E., H. Li, and W. Suen, 2010, Optimal deadlines for agreement, unpublished
paper
Fershtman, C. and D. Seidmann, 1993, Deadline effects and inefficient delay in
bargaining with endogenous commitment, Journal of Economic Theory 60, 306321.
Fudenberg, D., and J. Tirole, 1991, Game theory (The MIT Press: Cambridge,
Massachusetts).
Green, E.J., and R. H. Porter, 1984, Noncooperative collusion under imperfect price
formation, Econometrica 52, 87-100.
Hendricks, K., A Weiss, and C. Wilson, 1988, The war of attrition in continuous time and
complete information, International Economic Review 29, 663-680.
Miyagiwa, K., 2009, Collusion and research joint ventures, Journal of Industrial
Economics 63, 768-784.
Ponsati, C., 1995, Deadline effect: a theoretical note, Economics Letters 48, 281-285.
Download