From Equals to Despots: The Dynamics of Repeated Decision Taking in Partnerships with Private Information Vinicius Carrascoy

William Fuchsz

Abstract This paper considers the problem faced by two agents who repeatedly have to take a joint action, cannot resort to side payments, and each period are privately informed about their favorite actions. We study the properties of the optimal contract in this environment. We establish that …rst best values can be arbitrarily approximated (but not achieved) when the players are extremely patient. Also, we show that the provision of intertemporal incentives necessarily leads to a dictatorial mechanism: in the long run the optimal scheme converges to the adoption of one player’s favorite action.

1

Introduction

There are many situations in which, repeatedly, two agents have to take a common action, cannot resort to side payments, and each period are privately informed about their favorite actions. Examples include many supranational organizations such as a Monetary Union or a Common Market. In the former, monetary policy must be jointly taken and, in the latter, a common tari¤ with the outside world must be adopted each period. At the national level, political coalitions which must jointly decide on policy issues are also a good example. In this paper we study the properties of the optimal contract for environments with such features. We …rst show that e¢ ciency can be arbitrarily approximated, but never attained, when players’ are su¢ ciently patient. The intuition goes as follows. In a repeated setting, the promise of continuation (equilibrium) values can play a similar role to the one side payments play in static mechanism design problems. The di¤erence between side payments and continuation values is that the latter can only imperfectly transfer utility across players. In particular, to transfer continuation utility from one player to anothe in any period t, allocations (decisions) for periods

> t must be altered. When players’ are su¢ ciently patient

(i.e., their discount rate ( ) is close to one), their current payo¤, which is weighted by (1

), becomes

insigni…cant relative to the promised continuation values. Hence, in order to guarantee truth-telling in the current period, continuation values have to vary only minimally. Since, the (Incentive Compatible) Utility Possibilities Frontier is locally linear, this implies that the associated e¢ ciency losses from the variation in continuation values are arbitrarily small in the limit. The attainment of full e¢ ciency, however, would call We have bene…ted from conversations with Simon Board, Sergio Firpo, Roger Myerson, Alessandro Pavan, Phil Reny, Leo Rezende, Yuliy Sannikov, Andy Skrzypacz, Nancy Stokey, and Balazs Szentes. Ezra Ober…eld provided excellent research assistance. We are particularly thankful to Ferdinando Monte for careful comments on an early draft. All errors are ours. y Department of Economics, PUC-Rio. z UC Berkeley

1

for no variation whatsoever in continuation values. This is clearly at odds with the provision of incentives needed for an e¢ cient action to be taken. Hence, full e¢ ciency is not attainable. Although the limiting e¢ ciency result is of interest, our main focus is on understanding the dynamic properties of the optimal allocation rule for the case in which (although potentially large) the discount factor is strictly smaller than 1. In order to understand these properties, it is useful to keep in mind, as a benchmark, what the …rst best allocation would entail. The …rst best would call for a constant weighted average of the players’types (favorite actions). The problem with this allocation when types are private is that the agents away from the center have an incentive to exaggerate their positions. If they expect the other types to be to the left (right) of them they would have an incentive to claim to be far right (far left) and in that way bring the chosen allocation closer to their preferred point. As a result, absent any additional tools to provide incentives, the only way one has to prevent the agents from lying is by making the allocation more insensitive to their reports.1 When decisions are taken repeatedly, the optimal allocation can be made more sensitive to extreme announcements of preferences. This is the case because, when the agents care about the future, they can trade decision power in the current allocation for decision power in the future. More extreme types are given more weight in the current decision but they pay for it by having less in‡uence in future allocations. As known from static mechanism design, once incentive compatibility constraints are taken into account, the agents’ utilities have to be adjusted to incorporate the rents derived from their private information. Following Myerson (1981), the adjusted utility is referred to as virtual utility. In our repeated setting, virtual utilities also play a key role. In fact, we show that the dynamics of decision taking are fully determined by: (i) a decision rule that, at each period, maximizes the weighted sum of the agents’(instantaneous) virtual utilities and (ii) a process that governs the evolution of the weights given to the agents’virtual utilities on decisions. The dynamics of the decision taking process lead to our second and most interesting result. Continuation values vary from period to period re‡ecting the agents’weights in the allocation rule.2 Indeed, continuation values increase (higher future decision power) for the agent that reports a less extreme preferred action, and to decrease (lower future decision power) for a player that reports to prefer an extreme action. Such dynamics eventually lead to one player becoming a dictator. Put di¤erently, in the limit, only the preferences of one agent are taken into account to determine current and future allocations. Our approximate e¢ ciency result can be contrasted with the one obtained by Sonnenschein and Jackson (2007). They study a "budgeting mechanism" which allows the agents to report each possible type (they have a discrete type space) a …xed number of times.3 The number of times they can report a given type is given by the frequency with which that type should be realized. They prove (Corollary 2 in their paper) that, for any > 0, their "budgeting mechanism" is, for a …nite (but large) number of periods of interaction, less than

ine¢ cient relative to the …rst best if players are patient. The sources of the ine¢ ciency in their

mechanism and in our scheme are quite di¤erent though. In their setting, when the last periods get close, agents may not be able to report truthfully, as they might have run out of their budgeted reports for a particular type. Instead, in our setting, the ine¢ ciency arises from the fact that the weights each agent has 1 See

Carrasco and Fuchs (2009) for a complete analysis of the static problem. approach to the analysis of the optimal allocation rule in the repeated game relies on the factorization results of Abreu, Pearce and Stacchetti (1990), which show that the agent’s payo¤ can be split into a current value and a continuation value. 3 Although not necessarily e¢ cient for given , their mechanism has the nice feature that it is robust i.e. the planner does not need to know the exact functional form of the agents preferences. 2 Our

2

on the choice of the allocation must vary over time. More remarkably, we slowly drift towards one of the agents becoming a dictator. Indeed, we show that the optimal way to link decisions over time necessarily leads to a dictatorship ex-post. In a somewhat simpler environment in which there is a binary choice each period and agents can have either have weak or strong preferences for either option, Casella (2005) studies a mechanism in which agents are given a vote every period which they can use over time.4 The possibility of shifting votes intertemporally allows agents to concentrate their votes when preferences are more intense. Therefore, if one agent has a long string of strong preferences and the other doesn’t, the other agent will accumulate a lot of votes and will be able to outvote him in the future. In a voting setting with two players, two binary issues, a continuum of preference intensities, and where votes across issues are cast simultaneously, Hortala-Vallve (2007) shows that if players are allowed to freely distribute a given number of votes across the two issues, the ex-ante e¢ cient decision can be attained. In our setting the continuation values play the same role as the number of remaining votes in Casella’s mechanism. However, by considering the optimal mechanism instead of a particular scheme, we don’t restrict the accrual of votes to one per period and we allow agents to borrow votes from the future. In our case one of the agents will eventually "run out of votes" and the other will dictate the allocation in the future. In contrast to Hortala-Vallve (2007), decisions are made sequentially in our model. Future decisions are used as an instrument to provide incentives for current decisions. This, in turn, leads to some ine¢ ciency ex-ante and to a dictatorship ex-post. Our paper also relates to a vast literature, pioneered by the seminal work of Fundenberg, Levine and Maskin (1994), that shows that continuation values are close substitutes to side payments in repeated settings. Within this literature, the closest paper to ours is Athey and Bagwell (2001), who analyze an in…nitely repeated Bertrand duopoly, and establish that, for a …nite discount factor, monopoly pro…ts can be exactly attained if …rms make use of asymmetric continuation values. The di¤erence from our setting is that, for some states that occur with positive probability, …rms in their paper can transfer pro…ts perfectly. This allows them to reconcile the variation in promised continuation values –which, in our setting, leads to a dictatorial mechanism –with those values being provided at the region of the Frontier that is linear. In dynamic insurance problems with one-sided private information such as Thomas and Worral (1990), the privately informed agent’s marginal utility is driven to 5

1, so that his consumption goes to zero. This is

known as the immiseration result. This is related to our dictatorship result. Indeed, an agent who reports to have an extreme type in a given period is like an agent that reports to have a low income realization. The optimal mechanism will respond by giving that agent more weight in the current allocation decision (similarly

a higher transfer today). Incentive compatibility then calls for the agent to "pay" for this by forgoing future weight in the allocation decision (future consumption). In their setting, a Principal designing an optimal insurance policy trades-o¤ risk-sharing (which calls for a constant consumption stream) and the provision of incentives –through varying continuation values. No matter if there is one agent or a continuum, this leads to inmisseration in the limit. In our model, in contrast, privacy of information would not pose a problem if there were just one agent, nor it is a problem in the case in which there is a continuum of agents. In the one agent case, there is no incentive problems for the agent because his report will not be weighted with any other reports. In the 4 Skrzypacz

and Hopenhayn (2004) use a similar chip mechanism to sustain collusion in a repeated auctions environment. and Lucas (1990) establish a similar result for an economy with a continuum of agents. They show that the income distribution fans out and in the limit almost all agents are impoverished. 5 Atkeson

3

continuum of agents case, any report has no e¤ect on the allocation and hence, there is no incentive to lie either. In our environment, incentives are harder to provide when there is a small number of agents. Despite the similarities with the dynamic insurance literature, we …nd it remarkable that an optimal incentive scheme among two ex-ante identical agents leads to the granting of all bargain power to a single player. Also, note that as opposed to immiseration, dictatorship is an absorbing state: once an agent is granted all the decision rights incentive constraints will not bind any longer and the continuation values will be constant rather than constantly drifting towards

1.

The paper is organized as follows. We introduce the model in section 2. The optimal mechanism and its properties are characterized in Section 3. All proofs are relegated to the Appendix.

2

The Model

There are two ex-ante symmetric players, i = 1; 2, who, at every period, must take a new joint action, a 2

[0; 1] : At each period t 2 f0; 1; :::g ; they receive privately preference shocks

i

2 [0; 1] : The preference shocks

are i.i.d. over time and across players, and are drawn from a distribution F (:) ; with density f ( i ) > 0; which is symmetric around 21 : Their instantaneous utility is a twice continuously di¤erentiable function6 u (a; i ) ; with u ( i; i)

u (a; i ) f or all a;

and ua; i (a; i ) > 0 > uaa (a; i ) : Put in words, their preferences are single peaked, with

i

representing their favorite action.

We additionally assume that preferences are symmetric around 21 ; for all a; u (a; i ) = u (1

a; 1

i

2 [0; 1] ;

7 i) :

Symmetry of preferences and the distribution of types around

1 2

makes the problem itself symmetric

around that point. Therefore, it is natural to measure extreme preferences in terms of how distant they are from 12 : After the players observe their preference shocks, they make reports ^i ; i = 1; 2. Letting the initial history h0 be the empty set, a public history at time t > 0, ht ; is a sequence of (i) past announcements of all players, and (ii) past realized actions: 0 0 t 1 t 1 ht = f?; ^1 ; ^2 ; a0 ; :::; ^1 ; ^2 ; at

1

g:

Given the reports and the history of the game, a history dependent allocation is determined according to an enforceable contract, which is a sequence of functions of the form n o1 t t 2 3t at ^1 ; ^2 ; ht : [0; 1] [0; 1] ! [0; 1] : t=0

6 All

the results extend to the case in which the individual players’utility function depend on their identities, ui (a; i ). 7 Note that, in particular, this holds for preferences such that an agent with type i is indi¤erent between any two actions a and b that are equidistant from i :

4

This contract is chosen a priori before the agents learn their preference shocks. Let H t be the set of all public histories ht . A public strategy for player i is a sequence of functions t f^i (:; :)gt ,

where

^t : H t [0; 1] ! [0; 1] : i n t o n t o ^ (:) ; ^ (:) Each pro…le of strategies ^ = de…nes a probability distribution over public histories. 1 2

Letting

t

t

2 [0; 1) denote the (common) discount factor, player i’s discounted expected payo¤ is given by: " # 1 X t t t t E (1 ) u at (^ ; h ); i : t=0

We analyze this game using the recursive methods developed by Abreu, Pearce and Stacchetti (1990). <2 be the set of Public Pure Strategy Equilibria (PPSE) payo¤s for the players, we can decompose the payo¤s into a current utility u (a; i ) and a continuation value vi (^) 2 W:

More speci…cally, letting W

i=1;2

E [(1

)u(a ^ ; i ) + vi (^i ; ^ i )];

In other words, any PPSE can be summarized by the actions to be taken in the current period and equilibrium continuation values as a function of the announcements.8

3

Properties of the Optimal Allocation Rules

3.1

Approximate E¢ ciency:

In the setting we consider, the …rst best allocation would be a weighted average of the agents’types, with the weights being constant over time. The di¢ culty of implementing such an allocation when types are private information is that, whenever his preference shock is di¤erent than 21 ; the agent would have an incentive to exaggerate his report towards the extremes. In particular, for the case in which

= 0, the only way one

has to prevent the agents from lying is by making the allocation more insensitive to the their reports. As a result, the allocation is biased towards the center.9 When

> 0, continuation values can be used as an additional instrument to get agents to report truthfully.

Now, an agent that reports an extreme type can be allowed to have a larger impact on the allocation in an incentive compatible way. The key is to present the agents with a trade-o¤ between the bene…t of a larger in‡uence in the current allocation vs. the loss they will incur in future continuation values. This allows the mechanism to take into account the intensity of the agents’ preferences, which, in turn, leads to e¢ ciency gains when compared to a static decision taking problem. Continuation values play a similar role to side payments in standard static incentive problems. The di¤erence between side payments and continuation values is that the latter can only imperfectly transfer utility across players. In particular, to transfer continuation utility from player i to player j in any period t; allocations for periods

> t must be altered. This, together with the lack of observability, implies that

one cannot attain exact e¢ ciency as an equilibrium outcome. Indeed, exact e¢ ciency would call for an equilibrium in which for all histories future allocations would not respond to current announcements. Hence, 8 This 9 See

is known as the Factorization result. Carrasco and Fuchs (2009) for a complete analysis of the static problem.

5

truthtelling would have to be a static best response for the players, and this cannot be attained with an e¢ cient allocation. Although e¢ ciency cannot be attained, one can arbitrarily approximate it as the players become patient. In fact, when

! 1; the utility value of the current period, which is weighted by (1

) ; becomes insignif-

icant relative to the continuation values. Hence, in order to guarantee truth-telling in the current period continuation values have to vary only minimally. Since V (v) is locally linear, the associated losses from the variation in continuation values become negligible. Indeed, letting a ( ) = arg max a

2 X

u (a; i ) ;

i=1

be the (ex-ante) symmetric Pareto e¢ cient allocation, and v

FB

=E

" 2 X

u (a ( ) ; i ) , we have10

i=1

Proposition 1 (Approximate E¢ ciency) Given

> 0; there exists

sum of players PPE payo¤ s at an optimum are within

#

< 1 such that, for all

>

the

of v F B .

This result can be contrasted with the one obtained by Jackson and Sonnenschein (2007). They study a "budgeting mechanism" which allows the agents to report each possible type (they have a discrete type space) a …xed number of times. The number of times they can report a given type is given by the frequency with which that type should statistically be realized. They prove (Corollary 2 in their paper) that, for any > 0, their "budgeting mechanism" is less than ine¢ cient relative to the …rst best if players are patient and face su¢ ciently many similar problems. Although it appears to operate very di¤erently from the storable votes mechanism proposed by Casella (2005) or our own mechanism, in essence, the budgeting mechanism also presents the players with a trade-o¤ between current allocation and continuation values. The way continuation values vary in Jackson and Sonnenschein (2007) with the current reports is not e¢ cient, but, for

close to 1, they are su¢ cient to sustain incentive compatibility and the ine¢ ciency becomes negligible.

The sources of the e¢ ciency losses in their mechanism and in our scheme are quite di¤erent though. In their setting, when the last periods get close, agents may not be able to report truthfully, as they might have run out of their budgeted reports for a particular type. Instead, in our setting, the ine¢ ciency arises from the fact that the weight each agent has on the choice of the allocation must vary over time.

3.2

The Dynamics of Decision Taking

To derive some properties of the dynamics of decision taking in our environment, it is convenient to use the payo¤ decomposition implied by APS (1990) to write the Bellman equation that characterizes the frontier of equilibrium values that can be attained. At a given period, let v denote the expected future value promised to player 2 by the optimal contract. De…ne V (v) as the highest value to player 1 given that player 2’s expected values are v: Let v be lowest value the designer can assign to a agent 2, and de…ne v as the players’payo¤ when their preferred action is always taken, v = E [u ( i ; i )].11 1 0 Using virtually the same proof (with a proper adjustment of the relevant promised values), one can easily extend the approximate e¢ ciency result to the case in which the welfare criterion is not utilitarian. 1 1 We assume v E [u ( j ; i )] for i 6= j.

6

Letting

= ( 1;

2) ;

we can write V (v) as: V (v) = max E [(1 a( );w( )

) u (a ( ) ;

1)

+ V (w ( ))]

s.t. E [(1

E

E

2

[(1 1

[(1

) u (a ( ) ;

) u (a ( ) ;

1)

) u (a ( ) ;

2 ) + w ( )]

E

+ V (w ( ))]

2)

h (1 1

E

2

h

(1

+ w ( )] = v

(Promise Keeping 2)

b 1; 2 ;

)u a

) u a b1 ;

2

;

2

+ w

1

b 1; 2

i

+ V w b1 ;

f or all

^ 2 [0; 1]

2; 2

(IC2)

i

2

f or all

w ( ) 2 [v; v] f or all :

^ 2 [0; 1]

1; 1

(IC1)

(Feasibility)

In the appendix, we derive the following …rst order condition for the optimal current allocation a (:): h h i 2 i +

where

@u(a( ); 1 ) h @a @u(a( ); 2 ) @a

( f(

1)

@u(a( ); 1 ) @a ( 2 ) @u(a( ); 2 ) f ( 2) @a 1)

(1) ( 1 ) @ u(a( ); 1 ) f ( 1) i @ 1 @a i (1) ( 2 ) @ 2 u(a( ); 2 ) f ( 2) @ 2 @a

h

(:) is a multiplier on the …rst order counterpart of the IC constraints,

= 0:

( i) =

(FOC1)

d ( i) d i ,

and , which

plays the role of a Pareto weight, is the multiplier on the Promise Keeping constraint. As suggested by Myerson (1984), it is convenient to think about the Lagrangian that yields this …rst order condition as representing the weighted sum of the agents’ virtual utilities. Indeed, letting agent i0 s instantaneous virtual utility be u (a ( ) ; i ) = u (a ( ) ; i )

( i) u (a ( ) ; i ) f ( i)

[ (1) ( i )] @u (a ( ) ; i ) ; f ( i) @ i

it can be readily seen from the …rst order condition for a (:) ; that the optimal mechanism maximizes the weighted sum of the agents’virtual instantaneous utilities, with the weight given to agent 1 being equal to one, and the weight given to agent 2 being equal to .12 In solving for the optimal continuation value for player 2, we obtain the following condition when, although smaller than one,

is large: V 0 (v) = E Q [V 0 (w ( ))] :

By the Envelope Theorem, V 0 (v) = : 1 2 See

Myerson’s notes on virtual utility at http://home.uchicago.edu/~rmyerson/research/virtual.pdf

7

(Martingale)

so that the weight ( ) agent 2’s virtual utility is given when the action is taken is a martingale process with respect to a distribution Q.

It follows that, for large ; the dynamics of decision taking is fully determined by (i) a decision rule that,

at each period, maximizes the weighted sum of the agents’instantaneous virtual utilities, and (ii) the process that governs the evolution of the weights the agents’virtual utilities are given on decisions. The distribution Q di¤ers from the true distribution of the players’types by an explicit account –through

the multipliers and their derivatives – of the incentive compatibility constraints. Except for the change of measure, similar martingale properties for marginal values also hold in many dynamic insurance models.13

Our model di¤ers from them in that, in the dynamic insurance models, the problem is that agents have an incentive to claim to be poorer than they actually are. Instead, we face a situation were agents have an incentive to claim their type is more extreme than it actually is. In fact, the symmetry of the problem around type

1 2;

along with the tilting of the optimal allocation toward the middle to curb the players’ incentives

to exaggerate their preferences, implies that relevant direction in which the IC constraints bind depends on whether the players’s favorite action is above or below 21 : The relevant constraints for players whose favorite action is above

1 2

are those that ensure they don’t want to lie upwards. Conversely, for the case in which the

players favorite action is below

1 2;

the relevant constraints are the local downward constraints. Therefore,

the multipliers on the First Order Condition counterparts of inequalities IC1 and IC2 change sign at 21 . The change of measure needed for the property to hold in our setting follows from this point. Moreover, the dynamic insurance models either deal with the case in which there is a single agent or there is a continuum of them. In our setting, privacy of information would not pose a problem if there were just one agent, nor it is a problem in the case in which there is a continuum of agents. In the one agent case, there is no incentive problems for the agent because his report will not be weighted with any other reports. In the continuum of agents case, any report has no e¤ect on the allocation and hence, there is no incentive to lie either. In our environment, incentives are harder to provide since there is a …nite number of agents. It can be shown (this is done in the appendix) that, under the optimal mechanism, for any given w in (v; v) – i.e., whenever 0 <

< 1 – continuation values vary from period to period for agent 2. That is,

there is positive probability of both

w( ) > v and w ( ) < v:

Unlike the insurance models, variation in the continuation values is not necessary in order to provide (some) insurance. For example, in Thomas and Worral (1990), when

= 0; there is no way the Principal can

provide any insurance to the agent that gets a low income realization. In our model, instead, since agents don’t know how aligned their interests are it is possible even in the static case to have an allocation which depends on the Agents’types (Carrasco and Fuchs (2008)). Nonetheless, in an optimal scheme it will always be e¢ cient to have continuation values varying over time. The intuition for this is similar to the insurance models. Continuation values allow for agents with an extreme type in the current period (poor agents in the insurance models) to get more weight in the current allocation choice (higher current consumption) in exchange for forgoing decision rights (consumption) in the future. 1 3 For

examples those studied by Green (1987), Thomas and Worral (1990), and Atkenson and Lucas (1993).

8

The proof of Theorem 1 (see below) follows similar arguments to those in Thomas and Worral (1990). We …rst note that the Martingale Convergence Theorem implies that, for all V 0 (v) must converge to a random variable. Then, we show by contradiction that this random variable cannot have positive density for any value in ( 1; 0).14 Therefore, eventually, the action taken will place weight only to one of the players.

Alternatively, eventually, either

w ( ) ! v f or some; or V (w ( )) ! v with probability 1. Theorem 1 (Dictatorship in the limit) Assume is large but smaller than one ( << 1). The provision of intertemporal incentives necessarily leads to a dictatorial mechanism: In the limit as t ! 1, either v converges to v almost surely; or V (v) converge to v almost surely:

Whenever an agent is promised continuation values of v; it must be the case that his favorite action is taken from then on. In other words, v is an absorbing state. Therefore, eventually, a single player will be given all bargain power over decisions to be taken. Dictatorship is an ex-post consequence of an optimal mechanism in repeated decision taking settings. It is worth pointing out that although Sonnenschein and Jackson’s (2007) budgeting mechanism does not have this long run implication, it can lead to even lower values in long run. This happens when the set of reports left to an agent is very di¤erent from the distribution of types. In environments with endogenous participation constraints, such as Fuchs and Lippi (2006), the threat of abandoning the partnership puts a limit on the extent to which one of the agents can dominate the decision process. We believe that incorporating these considerations is an interesting avenue for future research.

References [1] Abreu, D., Pearce, D. and Stacchetti, E., 1990. "Toward a Theory of Discounted Repeated Games with Imperfect Monitoring", Econometrica, Vol. 58(5), pp. 1041-1063. [2] Arrow, K., 1979. "The Property Rights Doctrine and Demand Revelation under Incomplete Information", Economics and Human Welfare. Academic Press. [3] Athey, S., and Bagwell, K., 2001, "Optimal Collusion with Private Information", Rand Journal of Economics, 32 (3): 428-465. [4] Atkeson, A., and Lucas, R.E., 1993, "On E¢ cient Distribution with Private Information", Review of Economic Studies, 59, pp.427-453. [5] Casella, A., 2005. "Storable Votes", Games and Economic Behavior, 51, pp. 391-419. [6] Carrasco, V., and Fuchs, W., 2008, "Dividing and Discarding: A Procedure for taking decisions with non-transferable utility", mimeo. [7] d’Apresmont, C. and Gerard-Varet, L., 1979. "Incentives and Incomplete Information", Journal of Public Economics 11, pp. 25-45. 1 4 Convergence

to a value would violate the property that continuation values must vary to provide incentives.

9

[8] Dobb, J., 1953. Stochastic Processes, John Wiley and Sons, Inc., New York, N. Y. [9] Fuchs, W., and Lippi, F., 2006. "Monetary Union with Voluntary Participation", Review of Economic Studies, 73, pp. 437-457. [10] Fundenberg, D., Levine, D., and Maskin, E., 1994. "The FolkTheorem with Imperfect Public Information", Econometrica, 62, pp. 997-1040. [11] Green, E. "Lending and the smoothing of uninsurable income", in Contractual Arrangements for Intertemporal Trade (E. C. Prescott and N. Wallace, Eds.), Press, Minnesota, 1987. [12] Jackson, M. and Sonnenschein, H., 2007. "Overcoming Incentive Constraints by Linking Decisions", Econometrica, Vol. 75(1), pp. 241-258. [13] Kolmogorov, A., and Fomin, S., Introductory Real Analysis. New York, Dover Publications, 1970. [14] Milgrom, P, and I. Segal, 2002, "Envelope Theorems for Arbitrary Choice Sets", Econometrica, 70 (2) March, 583-601. [15] Myerson, R., 1981. “Optimal Auction Design”, Mathematics of Operations Research, Vol. 6(1) , pp. 58-73. [16] Myerson, R., 1984. “Cooperative games with imcomplete information”, International Journal of Game Theory, Vol. 13 (2) , pp. 69-96. [17] Luenberger, D., 1969, Optimization by Vector Space Methods, John Wiley and Sons, Inc. [18] Skrzypacz, A. and Hopenhayn, H., 2004. “Tacit Collusion in Repeated Auctions.”Journal of Economic Theory 114 (1), pp. 153-169. [19] Stokey, N., Lucas, R.E., and Prescott, E., 1989, Recursive Methods in Economic Dynamics, Harvard University Press. [20] Thomas, J. and Worrall, T., 1990. "Income Fluctuations and Asymmetric Information: An Example of a Repeated Principal-Agent Problem", Journal of Economic Theory, 51, pp. 367-390.

4

APPENDIX:

4.1

The Approximate E¢ ciency Result

Proof of Approximate E¢ ciency. We de…ne the ex-ante e¢ cient allocation as " 2 # X a ( ) = arg max E ui (a; i ) : a

We prove that, for any > 0; there exists payo¤s is within

i=1

2 (0; 1) such that for

> ; the sum of the players’equilibrium

of the payo¤ associated with v

FB

=E

" 2 X

#

ui (a ( ) ; i ) :

i=1

10

We do so by constructing continuation values that replicate as closely as possible the expected payments of the expected externality mechanism proposed by Arrow (1979), and d’Aspremont and Gerard-Varet (1979), that guarantee e¢ ciency in a standard (static) Mechanism Design problem. De…ne i

( i) = E

i

[u

i

(a ( ) ;

i )] ;

and consider, for i = 1; 2; the following candidates for continuation values 1

E

vi ( ) = where

i

[u

i

(a ( ) ;

i )]

i

(

!

i)

( )

+E [ui (a ( ) ; i )]

;

( ) is given by ( ) = (1

)

max E

i

[u

i

(a ( ) ;

i )]

min E i [ui (a ( ) ;

i

(1

i )]

) d;

i

where d is a …nite number. Note that

( ) is strictly positive and just depends on : It is chosen so to guarantee that, for all ;

fvi ( )gi are feasible values. Note, moreover, that 2 X

i

(

i)

i=1

2 X

=

i ( i) =

2 X

E

i

[u

i

(a ( ) ;

i )] :

i=1

i=1

Hence, upon inducing truthfulness from the players –so that a ( ) can be implemented in the …rst period in an incentive compatible way –, 2 X

vi ( ) =

i=1

2 X

!

E [ui (a ( ) ; i )]

i=1

2 ( )

so that the sum of the players’expected payo¤s when these continuation values are used is v F B

2 ( )

:

We now proceed by showing that, with these continuation values, one can implement a ( ) in an incentive compatible way. We then show that we can make

( ) arbitrarily small as

! 1:

Note that, if the other player is being truthful, player i’s problem, if a is implemented and he faces vi ( ) as a continuation value, is max (1 bi

)E

i

u a

bi ;

i

;

i

+ E

which has the same solution as the one associated with the program15 h h max (1 ) E i ui a bi ; i ; i + (1 ) E i u

Since

bi

a ( ) = arg max a

1 5 All

other terms do not not a¤ect incentives.

11

2 X i=1

ui (a; i ) ;

i

h

i

i

vi (bi ; a

i)

bi ;

i

;

i

ii

:

the announcement bi = Now, pick

i

is optimal.

> 0: Consider the

that solves 2

=

2 1

d

2d 2d +

< 1:

= :

It is easy to see that = Moreover, for

> ; the sum of the players’equilibrium payo¤ is within of v F B ; when one implements a ( ) 2

with continuation values fvi ( )gi=1 : Since the optimal contract can not do worse than fa ( ) ; vi ( )gi , we establish that v F B can be arbitrarily approximated.

The fact that v F B cannot be exactly attained follows because that would call for the decision a ( ) being taken in every period and every history, so that .future allocations would not respond to current announcements. Hence, truthtelling would have to be a static best response for the players, and this cannot be attained with an e¢ cient allocation. It is worth pointing out that, using virtually the same proof (with a proper adjustment of the relevant promised values), one can easily extend the approximate e¢ ciency result to the case in which the welfare criterion is not utilitarian. We will use this fact in Lemma 3 below.

4.2 4.2.1

The Dictatorship Result Convexity of the Program and Concavity of the Frontier

In order to convexify our problem and apply Lagrangean methods, we allow for general stochastic mechanisms throughout this Appendix. De…nition 1 A stochastic Recursive Mechanism is a mapping that takes announcements ( 1 ;

2)

into a

probability distribution over actions a and continuation values for agent 2, w : P (a; wj ) ; A Stochastic Recursive Mechanims is Incentive Compatible if Z Z E 1 E 1 [(1 ) u (a; 2 ) + w] dP (a; wj 2 ; 1 ) [(1 ) u (a; E

and Z

2

[(1

) u (a;

1 ) + V (w)] dP (a; wj 1 ;

2)

E

2

Z

[(1

2)

) u (a;

+ w] dP a; wjb2 ; 1)

Now, consider the problem: E

max

fP (a;wj )g

2[0;1]2

Z

[(1

subject to Z E 1 [(1 ) u (a; 2 ) + w] dP (a; wj 2 ; Z E 2 [(1 ) u (a; 1 ) + V (w)] dP (a; wj 1 ;

) u (a;

1)

E

2)

E

12

1)

1

2

Z

; 8 2 ; b2 ;

+ V (w)] dP a; wjb1 ;

+ V (w)] dP (a; wj )

Z

1

[(1

) u (a;

2)

[(1

) u (a;

1)

1

; 8 1 ; b1

(Optimal)

+ w] dP a; wjb2 ;

1

+ V (w)] dP a; wjb1 ;

; 8 2 ; b(ICs) 2; 1

; 8 1 ; b1

E We have the following result:

Z

[(1

) u (a; i ) + w] dP (a; wj ) = v;

(PK)

Proposition 2 There exists a solution to (Optimal). Proof. The objective function is continuous as it is a bounded linear functional. By the Alouglu‘s Theorem (see Luenberger, 1969), the set of all possible distributions is compact in the weak-* topology The set of contraints is formed by an equality (the promise keeping constraint) and weak inequalities (the IC constraints). It follows that, since the payo¤s are linear in the mechanism, the set of distributions that satisfy these constraints is a (weak-*) closed subset of the set of all distributions. Hence, the constraint set is compact in the weak-* topology. The result then follows. Denote by V (v) the Value of (Optimal), we have Proposition 3 The program (Optimal) is convex. and V (:) is concave Proof. Since the set of contraints is formed by an equality (the promise keeping constraint) and weak inequalities (the IC constraints).(i) the constraints are de…ned by equalities and weak inequalities, and (ii) the payo¤s are linear in fP (a; wj )g ; the constraint set is convex. Noting that the objective is linear, the

result follows. In fact, if fP (a; wj ; vk )g is the solution of (Optimal) when PK constraints are indexed by

vk ; one has that

f P (a; wj ; v1 ) + (1

) P (a; wj ; v2 )g

is feasible when the Promise Keeping constraints (equations PK) are indexed by

v1 + (1

) v2 : It then

follows that V ( v1 + (1 ) v2 ) Z E [(1 ) u (a; n ) + V (w)] d [ P (a; wj ; v1 ) + (1 ) P (a; wj ; v2 )] Z 0 1 E [(1 ) u (a; n ) + V (w)] dP (a; wj ; v1 ) B C Z = @ A + (1 )E [(1 ) u (a; n ) + V (w)] d [P (a; wj ; v2 )] =

V (v1 ) + (1

) V (v2 ) :

where the …rst inequality follows from the de…nition of V (:) Proposition 4 V (:) is continuously di¤ erentiable over (v; v). Proof. Since V (:) is a concave value function, this follows from Corollary 2 in Milgrom and Segal (2002). Combining Propositions 3 and 4, one has the following result Proposition 5 V (:) is strictly concave Proof. Since V (:) is continuously di¤erentiable over (v; v) ; if V (:) were not strictly concave, it would have to be be linear. Our ine…ciency result rules this possibility out. 13

4.2.2

Incentive Compatible Representation of Preferences

De…ne U1 ( 1 ) = E

Z

2

and U2 ( 2 ) = E

1

[(1 Z

) u (a;

[(1

1)

) u (a;

+ V (w)] dP (a; wj )

2)

+ w] dP (a; wj ) :

We have the following result, which is standard im Mechanism Design problems in which the agents’ preferences satisfy a single-crossing condition: Lemma 1 A Recursive Stochastic Mechanism is IC if, and only if, Zi Ui ( i ) = Ui (0) + E

i

0

and E 4.2.3

i

Z

[(1

Z

[(1

) u i (a; )] dP (a; wj ;

) u i (a; i )] dP (a; wj ;

i)

i)

d

is non-decreasing in :

The Lagrangian Representation and the Result

In what follows, we ignore the monotonicity constraints. Note that, given that u (a; i ) satisfy a singlecrossing condition, the monotonicity constraint will be satis…ed for large enough

(which is the case of

interest for us); since, by Proposition 1, the optimal allocation will get closer to the e¢ cient one.(which satis…es a stronger monotonicity condition given that

@ 2 u(a; i ) @u@ i

> 016 ). Also, our dictatorship result will be

derived from the FOC for the continuation values, and not the current decision (which is the object that the monotonicity constraint restricts). De…ne the Lagrangean as follows

L (P (a; wj:) j ; ) 2 2 6 2 Z1 X = 6 4 4Ui ( i )

E Ui (0)

i=1 0

[U1 ( 1 )] + [E 2 [U2 ( 2 )] v] Zi Z E i [(1 ) u i (a; )] dP (a; wj ; 1

0

3

3

7 7 5 d i ( i) 5 i) d

We have the following result, which is an application of the results in Luenberger (1969, Theorem 1 Section 8.2 and Theorem 1, Section 8.3) to our setting: Lemma 2 (Optimality) Assume that the monotonicity constraints slack at an optimum.17 An incentive compatible Recursive Mechanism fP (a; wj )g is optimal if, and only if, there exists a non-decreasing

f

i

( i )g

i 2[0;1];i=1;2

and a non-negative number

for which

L Pe (a; wj:) j ;

L (P (a; wj:) j ; ) : 1 6 By

Topkis (1998), a ( i;

is non-decreasing in

i

for all

i)

= arg max u (a;

i)

+ u (a;

i)

i:

2 1 7 Since @ u(a; i ) @a@ i

> 0;this will be true whenever ; although smaller than one, is large as, in such case (for arbitrary Pareto weights) one can arbitrarily approximate the e¢ cient allocation, which is strictly increasing in the agents’(announced) types.

14

for all Incentive Compatible Recursive Mechanisms

n o Pe (a; wj )

2[0;1]2

:

We note that, since preferences and the density f ( i ) are symmetric around symmetric around s 2 0;

1 2

1 2:

1 2;

the solution will be

This, in turn, implies that the multipliers will be symmetric around 12 ; taht is, for all

; i

1 +s 2

=

i

1 2

s :

Also, over time, as decisions take place, agents will not look alike: agent 2 is promised continuation values v and the mechanism must deliver such values. Hence, as time goes by, the mechanism must give di¤erent weights to the agents. These di¤erent weights are captured by the multiplier

on the Promise Keeping

constraint. Obviously, the relative weights the agents are given by the mechanism over time also a¤ect the IC constraints and, consequently, their multipliers. We, therefoe, take that 2

Last, we take that

i

( 2) =

1

( 2) :

( i ) is absolutely continuous. We denote its (Radon-Nykodyn) derivative by

i

( i) :

After some integration by parts, the Lagrangean can be read as

L (P (a; wj:) j ; ) h h ii h h h 2 1( 1) E 1 U1 ( 1 ) 1 + E 2 U2 ( 2 ) 1 f ( 1) 6 6 +U1 (0) [ 1 (1) 1 (0)] + U1 (0) [ 1 (1) 6 Z h 6 = 6 +E E 2 [(1 ) u 1 (a; 1 )] dP (a; wj ) 6 1 6 Z h 4 + E 2 E 1 [(1 ) u 2 (a; 2 )] dP (a; wj )

1( 1) f ( 1)

1

ii

(0)]

1 (1)

f( 1 (1)

f(

1( 1) 1)

v i

1( 2) 2)

i 3

i

7 7 7 7 7: 7 7 5

Since the Lagrangean is linear in P (a; wj ) ; if (a ( ) ; w ( )) are in the support of P (a; wj ), one must have that (a ( ) ; w ( )) maximize the Lagrangean when the agents types are : Hence, the following FOC for the current decision a (:) must be satis…ed: i h ii 3 2 h h 1 (1) 1( 1) 1( 1) + u (a ( ) ; ) ua (a ( ) ; 1 ) 1 a 1 h h f ( 1) i 1 h f ( 1) ii 5 = 0 4 1( 2) 1 (1) 1( 2) + ua (a ( ) ; 2 ) 1 + u (a ( ) ; ) a 2 1 f ( 2) f ( 2)

(FOCa)

This is the condition in the text.

As for continuation values, the following First Order Condition must be satis…ed: V 0 (w ( )) 1

( 1) f ( 1) f ( 2) + f ( 1) 1

1

( 1) f ( 1 ) f ( 2 ) = 0: f ( 1) 1

(FOCw)

We can now show: Lemma 3 (Martingale Lemma) For

large but smaller than 1, there exists a measure Q such that Player

1´ s marginal value follows a martingale, i.e.

E Q [V 0 (w ( ))] = V 0 (v)

15

Proof. As we showed above, the FOC of the Lagrangean with respect to w (:) is V 0 (w ( )) 1

1 ( 1) f ( 1) f ( 2) = f ( 1)

1

1 ( 1) f ( 1) f ( 2) = 0 f ( 1)

By the Envelope Theorem, V 0 (w) =

;

so one can re-write the above First Order Condition as: V 0 (w ( )) 1

1 ( 1) f ( 1 ) f ( 2 ) = V 0 (w) 1 f ( 1)

1 ( 2) f ( 1) f ( 2) : f ( 2)

The e¢ cient outcome can be approximated when agents are su¢ ciently patient whatever the welfare criterion used. Hence, as and

1

! 1; the Incentive Compatibility constraints will slack.; so that, as

( i ) approach zero uniformly.

18

It follows that, for a su¢ ciently large ; 1 ( i) > 0 for all f ( i)

1

! 1; both

1

i

Now, dividing both sides of equation (FOCw) by Z Z 1

and integrating over ( 1 ;

2) ;

1

( 2) f ( 1) f ( 2) d 1d f ( 2) 1

0

2

2

one has E Q [V 0 (w ( ))] = V 0 (w) ;

where Q is an absolutely continuous distribution with density h i 1( 1) 1 f ( 1) f ( 1) f ( 2) q( ) = Z Z h i 1( 2) 1 f ( 2) f ( 1) f ( 2) d 1d 1

0: 2

2

Lemma 4 (Spreading of Values) Assume that v 2 (v; v). Then, there is positive probability of both w( ) > v and w ( ) < v: Proof. Assume toward a contradiction that w( ) for almost all

(the other case is analogous): Since V is concave, V 0 (w ( ))

1 8 Note

that the

v

v

in the proof of Proposition 1 only depends on ; and not on any realization of (

16

1 ; 2 ).

( i)

for almost all : Using Lemma Martingale, V (v) = E Q [V 0 (w ( ))] ; this implies that V 0 (w ( )) = V 0 (v) for almost all : Since V (:) is strictly concave, one must have w ( ) = v for almost all : Plugging this in the …rst order conditions for w (:) ; we get ( 1) 1 ( 2) f ( 1 ) f ( 2 ) = V 0 (v) 1 f ( 1) f ( 2) ; f ( 1) f ( 2) 1 ( 2) 1 ( 1) = for almost all ( 1 ; 2 ) f ( 1) f ( 2) 1

V 0 (v) 1

Since the left hand side just depends on all ( 1 ;

1

( 1) =

Since for all s 2 0; 21 i

the fact that all

1

and the right hand side on

2;

the above can hold for almost

2 ) only if

1

( 1) =

1

1

( 2 ) = 0 for almost all ( 1 ; 1 2

s

=

i

( 2 ) = 0 for almost all ( 1 ;

2)

2) :

1 +s ; 2 implies that

i

( i ) must also be zero for almost

i:

Plugging

i

( i) =

i

( i ) = 0 in the First Order Condition for a; one gets: @u (a ( ) ; @a

1)

+

@u (a ( ) ; @a

2)

= 0:

It is easy to see that the a (:) implicitly de…ned by the above equation is not IC when continuation values, w ( ) ; are constant, unless contradicts w 2 (v; v) :

= 0, or

= 1; that is, unless the dictatorship holds: Dictatorship, however,

We are now able to show: Theorem 1 The provision of intertemporal incentives necessarily leads to a dictatorial mechanism: In

the limit as t ! 1, either v converges to v almost surely, or V (v) converge to v almost surely: Proof. Since V 0 (v) is a non-positive martingale, by Dobb’s convergence Theorem (see Dobb (1953)), it converges almost surely to some random variable, R. Next we show by contradiction that R cannot have any positive likelihood for values in (0; 1) : Hence, all the probability is concentrated where R = 0 or R =

1:

Hence, one has that either V (v) goes to v, or v converges to v, i.e., one of the players becomes a dictator in the limit.

In search of a contradiction, suppose there existed a positive probability of …nding a path V 0 (v) with the property that limt!1 V 0 (v) = C; where 0 < C < 1: Since V 0 (v) is continuous for any v 2 (v; v) ; the sequence vt converges. Denote its limit by limt!1 vt = v 0 2 (v; v). Let W (w; ) denote the next period’s

continuation value given the current promised value w and reported state . For wt to converge it must be that W (w0 ; ) = w0 for all : This however contradicts Lemma (4) : 17

From Equals to Despots: The Dynamics of Repeated ...

study the properties of the optimal contract in this environment. We establish that ... small in the limit. The attainment of full effi ciency, however, would call ... that the agents away from the center have an incentive to exaggerate their positions.

192KB Sizes 2 Downloads 169 Views

Recommend Documents

Repeated Signaling and Firm Dynamics
We thank the seminar participants at Columbia University, Dartmouth College, New York University, London. Business School .... In our model, firms with positive information choose higher capital stocks and credibly ...... in simulated data, the corre

Repeated Signaling and Firm Dynamics
Mar 17, 2010 - University of California, Berkeley ... firms with negative private information have negative leverage, issue equity, and overin- vest. ... We thank the seminar participants at Columbia University, Dartmouth College, New York University

What is Unequal among the Equals? Ranking Equivalent Rules from ...
Oct 26, 2010 - Based on these interestingness measures, an incremental Apriori-like algorithm is designed to select more interesting rules from the lower ...

The Left Hand of Equals - Research at Google
of loading Smalltalk-80 off the Apple-branded floppy disks onto a Lisa. So much romance! Meanwhile, down in the basement machine room, Simula had been ...

repeated thoracic discharges from a stun device
2 Department of General Surgery, Rush University Medical Center, Chicago, ... lack of objective laboratory data describing the physiological effects and safety of.

Efficient Repeated Implementation ... - Faculty of Economics
Consider regime ̂R defined in Section 4.2 of LS. ..... Define ¯g = (M,ψ) as the following mechanism: Mi = .... strategy bi except that at (ht,θt) it reports z + 1.

Supplement to "Efficient Repeated Implementation"
the definition of ψ of ˆg) but induces regime Dj in which, by (A.1), j obtains vj j > πθ(t) θt j . But this is a contradiction. Q.E.D. ... Next define ρ ≡ maxi θ a a [ui(a θ)−ui(a θ)] and ¯δ ≡ ρ ρ+ε . Mechanism ˜g = (M ψ) is def

From Time-Dependent Dynamics to Many-Body ...
the fundamental Schrodinger equation appear. The author has included topics that are not usually covered in standard textbooks and has written the book in such a way that every topic contains varying layers of difficulty, so that the instructor can d

A Solution to the Repeated Cross Sectional Design
Jul 28, 2011 - unemployment rate). • Box-Jenkins and fractional differencing techniques can control for autocorrelation at level-2. (Box and Jenkins. 1976; Box-Steffensmeier and Smith 1996, 1998; Lebo,. Walker and Clarke 2000; Clarke and Lebo 2003)

An Effective Approach to the Repeated Cross& ... - Wiley Online Library
The PCSTS tool kit does not provide an appropriate solution, and we offer one here: double filtering with ARFIMA methods to account for autocorrelation in longer. RCS followed by the use of multilevel modeling to estimate both aggregate- and individu

Dynamics of giant volcanic ash clouds from supervolcanic ... - CiteSeerX
Sep 13, 2005 - volcanic cloud by assuming a localised constant flux source at ground level .... data in Figure 3a are only indicative, but are sufficient to support the .... currents: Two-layer shallow-water and numerical solutions, J. Fluid. Mech.

Dynamics of giant volcanic ash clouds from supervolcanic ... - CiteSeerX
Sep 13, 2005 - [1] The largest explosive volcanic eruptions that have occurred on Earth generated giant ash clouds from rising plumes that spread in the ...

Global Inequality Dynamics: New Findings from WID.world
We start with a brief history of the. WID.world project. We then present selected findings on income inequality, private ... and COnICET (e-mail: [email protected]); Chancel: Paris School of Economics,. 48 Boulevard Jourdan, 75014 Paris, and Iddri

Efficient Repeated Implementation: Supplementary Material
strategy bi except that at (ht,θt) it reports z + 1. Note from the definition of mechanism g∗ and the transition rules of R∗ that such a deviation at (ht,θt) does not ...

Efficient Repeated Implementation
‡Faculty of Economics, Cambridge, CB3 9DD, United Kingdom; Hamid. ... A number of applications naturally fit this description. In repeated voting or .... behind the virtual implementation literature to demonstrate that, in a continuous time,.

Efficient Repeated Implementation
[email protected] ..... at,θ ∈ A is the outcome implemented in period t and state θ. Let A∞ denote the set ... t=1Ht. A typical history of mechanisms and.

Introduction to Repeated Games with Private Monitoring
Stony Brook 1996 and Cowles Foundation Conference on Repeated Games with Private. Monitoring 2000. ..... actions; we call such strategies private). Hence ... players.9 Recent paper by Aoyagi [4] demonstrated an alternative way to. 9 In the ...

Repeated proximity games
If S is a. ®nite set, h S will denote the set of probability distributions on S. A pure strategy for player i in the repeated game is thus an element si si t t 1, where for ..... random variable of the action played by player i at stage T and hi. T

Introduction to Repeated Games with Private Monitoring
our knowledge about repeated games with imperfect private monitoring is quite limited. However, in the ... Note that the existing models of repeated games with.

The Dynamics of Mutational Effects
Apr 3, 2007 - We use mutation accumulation experiments and molecular data from experimental evolution to show ... manipulation of organisms in the laboratory (i.e., mutation ... experimental evolution of large populations in the laboratory.