Computing dynamic optimal mechanisms when hidden types are Markov and controlled by hidden actions Kenichi Fukushima

Yuichiro Waki

University of Wisconsin, Madison

University of Queensland

September 25, 2011

This note documents how the main theoretical results in Fukushima and Waki (2011) extend to a richer setting where the agent can inuence the evolution of his hidden type

θt

through a hidden action

yt .

An example of such a setting is one where

hidden stock of wealth or human capital and

yt

θt

represents a

is a hidden investment. The notation follows

Fukushima and Waki (2011) unless otherwise indicated.

rt ∈ Θ to the planner. The planner chooses an outcome xt ∈ X and recommends an action yt ∈ Y given the agent's 0 history of reports. The agent then chooses an action yt ∈ Y which may or may not equal yt . We assume Y is a nite set with cardinality M . If the agent's current type is θt and he chooses action yt , his next period type θt+1 is drawn from the density π(·|θt , yt ) > 0. The initial distribution is π(·|θ−1 , y−1 ) where (θ−1 , y−1 ) is ∞ t+1 publicly known. We let Y denote the set of function sequences y = {yt }t=0 , yt : Θ →Y for each t, and write In each period the agent draws a type

Pr(θ

t

θt ∈ Θ

and sends a report

|θ−1 , y−1 , y) = π(θt |θt−1 , yt−1 (θt−1 )) × · · · × π(θ1 |θ0 , y0 (θ0 )) × π(θ0 |θ−1 , y−1 ). y|θt−1 = {yt+s (θt−1 , ·)}∞ s=0

t−1 denote the continuation of y after θ . ∞ t+1 → X and yt : An allocation is then a sequence (x, y) = {xt , yt }t=0 , where xt : Θ Θt+1 → Y for each t. We do not introduce randomizations to keep the notation simple, We also let

although doing so is quite straightforward and useful for computations (as it helps obtain convexity). takes place, that is, if after each shock history θ t occurs and the agent chooses yt (θ ), the agent obtains lifetime utility: If allocation

(x, y)

U (x, y; θ−1 , y−1 ) =

∞ X X t=0

t

the outcome

β t u(xt (θt ), yt (θt ); θt )Pr(θt |θ−1 , y−1 , y)

θt

and the planner incurs cost:

C(x, y; θ−1 , y−1 ) =

∞ X X t=0

q t c(xt (θt ))Pr(θt |θ−1 , y−1 , y).

θt

1

xt (θt )

An allocation

(x, y)

is therefore incentive compatible if

U (x, y; θ−1 , y−1 ) ≥ U (x ◦ r, y0 ; θ−1 , y−1 ),

∀(r, y0 ) ∈ R × Y

(1)

and satises promise keeping if

U (x, y; θ−1 , y−1 ) ≥ U0 . The planning problem starting from of

(x, y)

(θ−1 , y−1 , U0 )

(2)

is to minimize

C(x, y; θ−1 , y−1 )

by choice

subject to incentive compatibility and promise keeping.

We have the following analog of Lemma 1: Lemma A1.

An allocation (x, y) is incentive compatible if and only if

u(xt (θt ), yt (θt ); θt ) + βUt+1 (θt ; θt , yt (θt )) ≥ u(xt (θt−1 , θt0 ), yt0 ; θt ) + βUt+1 (θt−1 , θt0 ; θt , yt0 )

(3)

for all t, θt−1 , θt , θt0 , and yt0 , where Ut (θ

t−1

; θ− , y− ) =

∞ X X s=t

β s−t u(xs (θt−1 , θts ), ys (θt−1 , θts ); θs ) Pr(θts |θ− , y− , y|θt−1 ).

θts

Proof.

0 The only if part is clear. So let (x, y) satisfy (3) and x (r, y ) ∈ R × Y . For s s s 0 0 t s t 0 t t s each t, dene r| and y | by (r|s (θ ), y |s (θ )) = (rs (θ ), ys (θ )) for all s ≤ t and θ , and s )) for all s ≥ t + 1 and θs . I.e., (r|t , y0 |t ) follows (r, y0 ) (r|ts (θs ), y 0 |ts (θs )) = (θs, y s (rt (θt ), θt+1 until period t and then reverts back to truth-telling and obedience from t + 1. Applying (3) 0 0 0 1 0 1 inductively we have U (x, y; θ−1 , y−1 ) ≥ U (x ◦ r| , y | ; θ−1 , y−1 ) ≥ U (x ◦ r| , y | ; θ−1 , y−1 ) ≥ · · · ≥ U (x ◦ r|t , y0 |t ; θ−1 , y−1 ) for any t. Since u is bounded and β ∈ (0, 1), this implies:

U (x, y; θ−1 , y−1 ) ≥ lim U (x ◦ r|t , y0 |t ; θ−1 , y−1 ) = U (x ◦ r, y0 ; θ−1 , y−1 ). t→∞

Hence

(x, y)

is incentive compatible.

Notice here that the continuation utility prole

Θ×Y,

Ut (θt−1 ; ·, ·)

is a function of

(θ− , y− ) ∈ N ×M

so a recursive formulation in the spirit of Fernandes and Phelan (2000) has

continuous state variables. We say that

π

has an order

K

mixture representation if we can write:

π(θ|θ− , y− ) =

K X

pk (θ)wk (θ− , y− ),

k=1 K where p : Θ → R+ and w : Θ × Y → PK k=1 wk (θ− , y− ) = 1 for each (θ− , y− ). Under this representation, we can dene

RK +

satisfy

P

θ∈Θ

pk (θ) = 1

k

and

p(θt )

(4)

for each

( at (θ

t−1

)=

X

u(xt (θt ), yt (θt ); θt )

θt



∞ X X

s β s−t−1 u(xs (θs ), ys (θs ); θs ) Pr(θt+1 |θt , yt (θt ), y|θt )

s s=t+1 θt+1

  

2

and write

Ut (θ

t−1

; ·, ·) =

K X

akt (θt−1 )wk (·, ·).

k=1

at

This suggests that, by using

instead of

Ut

as an endogenous state variable, it should be

possible to reduce the dimensionality from N × M to K . t−1 Let us now write at (θ ; x, y) to describe the mapping from (x, y) to at (θt−1 ) dened by −1 (4). Let us also write a0 (x, y) = a0 (θ ; x, y), as this is independent of θ−1 . We then dene the auxiliary planning problem starting from to minimize

C(x, y; θ−1 , y−1 )

(θ−1 , y−1 , a0 )

as the problem of choosing

(x, y)

subject to incentive compatibility (1) and

a0 (x, y) = a0 . We let

A∗ ⊂ V K

(5)

a0 's for which the constraint set of this problem is non(θ−1 , y−1 )) and let J ∗ : Θ × Y × A∗ → R denote the optimal

denote the set of

empty (which is independent of value function. If

a∗0 ∈ arg min∗ J ∗ (θ−1 , y−1 , a0 ) a0 ∈A

s.t.

a0 · w(θ−1 , y−1 ) ≥ U0 ,

then a solution to the auxiliary planning problem starting from the planning problem starting from

(θ−1 , y−1 , a∗0 ) is a solution to

(θ−1 , y−1 , U0 ).

The analog of Lemma 2 is:

An allocation (x, y) satises the constraints of the auxiliary planning problem t ∗ (1) and (5) if and only if there exists a = {at }∞ t=0 , at : Θ → A , such that (x, y, a) satises Lemma A2.

u(xt (θt ), yt (θt ); θt ) + βat+1 (θt ) · w(θt , yt (θt )) ≥ u(xt (θt−1 , θt0 ), yt0 ; θt ) + βat+1 (θt−1 , θt0 ) · w(θt , yt0 ) at (θt−1 ) =

X

u(xt (θt ), yt (θt ); θt ) + βat+1 (θt ) · w(θt , yt (θt )) p(θt )

(7)

θt

for all t, θ , t

Proof.

θt0

,

yt0

, and a0 (θ−1 ) = a0 .

Virtually identical to that of Lemma 2.

The analog of the

B

operator therefore maps

A⊂VK

into

B(A) = {a ∈ V K |∃(x, y, a+ ) ∈ F (a; A)} where

F (a; A)

is the set of function triples

(x, y, a+ ) : Θ → X × Y × A

satisfying:

u(x(θ), y(θ); θ) + βa+ (θ) · w(θ, y(θ)) ≥ u(x(θ0 ), y 0 ; θ) + βa+ (θ0 ) · w(θ, y 0 ), X a= u(x(θ), y(θ); θ) + βa+ (θ) · w(θ, y(θ)) p(θ). θ

3

(6)

∀θ, θ0 , y 0

At this point it is useful to construct a particular incentive compatible allocation as follows. First pick any

x¯ ∈ X

and let

W (θ) = max y∈Y

For each for each

Ut (θ

 

W :Θ→R

u(¯ x, y; θ) + β



X θ+

; θ− , y− ) =

∞ X X s=t

So for each

solve the Bellman equation:

  W (θ+ )π(θ+ |θ, y) . 

θ let y¯(θ) solve the right hand side problem. t and θt . We then have for any t and θt−1 :

t−1

¯) (¯ x, y

Then set

x¯t (θt ) = x¯ and y¯t (θt ) = y¯(θt )

β s−t u(¯ x, y¯(θs ); θs )Pr(θts |θ− , y− , y|θt−1 ) =

θts

X

W (θ)π(θ|θ− , y− ).

θ

t, θt−1 , θt , θt0 , yt0 :

u(¯ x, y¯(θt ); θt ) + βUt+1 (θt ; θt , y¯(θt )) ≥ u(¯ x, yt0 ; θt ) + βUt+1 (θt−1 , θt0 ; θt , yt0 ). It follows from Lemma A1 that

¯) (¯ x, y

is incentive compatible.

We have the following analog of Proposition 3:

A∗ is a non-empty and compact set, and is the largest xed point of B . If A0 ⊂ V K is a compact set satisfying A0 ⊃ B(A0 ) ⊃ A∗ (one example being A0 = V K ) then n ∗ K satises A∗ ⊃ B(A0 ) ⊃ A0 B n (A0 ) is decreasing in n and ∩∞ n=0 B (A0 ) = A . If A0 ⊂ V n n (one example being A0 = {a0 (¯x, y¯ )}), then B (A0 ) is increasing in n and cl(∪∞ n=0 B (A0 )) = A∗ . Proposition A3.

Proof. A ⊂ B(A)

The analogs of Lemmas 5-8 follow from virtually identical arguments. Thus: (i) V K , A ⊂ B(A) =⇒ B(A) ⊂ A∗ , (ii) B(A∗ ) = A∗ , (iii) A ⊂ A0 ⊂ V K =⇒ ⊂ B(A0 ), and (iv) A is compact =⇒ B(A) is compact. Similarly for the rst two

parts of the proposition.

∗ To prove the nal part of the proposition, suppose A0 ⊂ B(A0 ) ⊂ A . Then from (ii), (iii), n ∗ ∗ n ∞ and the compactness of A , we know that B (A0 ) is increasing and cl(∪n=0 B (A0 )) ⊂ A . ∗ ∞ n ∗ ∞ n To prove A ⊂ cl(∪n=0 B (A0 )), pick any a ∈ A . We construct a sequence in ∪n=0 B (A0 ) 0 ∗ ∗ that converges to a. For this, rst pick another a ∈ A0 (⊂ A ). By the denition of A 0 0 there exist incentive compatible allocations (x, y) and (x , y ) such that a = a0 (x, y) and 0 0 0 a = a0 (x , y ). Next for each n ≥ 1, do the following. Dene xn = {xnt }∞ t=0 by truncating x 0 after n periods and appending x . Thus for t > n: t n+1 )). (xn0 (θ0 ), ..., xnt (θt )) = (x0 (θ0 ), ..., xn (θn ), x00 (θn+1 ), ..., x0t−n−1 (θn+1 And let

(rn , yn ) ∈ arg

max

ˇ ). U (xn ◦ ˇr, y

(ˇ r,ˇ y)∈R×Y 0 0 Here, since (x , y ) is incentive compatible, we can assume without loss that for t > n, n t n t 0 t ˆ n ) = (xn ◦ rn , yn ). By construction, rt (θ ) = θt and yt (θ ) = yt−n−1 (θn+1 ). Finally, let (ˆ xn , y ˆ n ) is incentive compatible, a0 (ˆ ˆ n ) ≥ a0 (xn , y), and an+1 (θn ; x ˆn, y ˆ n ) = a0 for all θn . (ˆ xn , y xn , y

4

n ˆ n ) ∈ ∪∞ We next show a0 (ˆ xn , y n=0 B (A0 ) for all n. From the incentive compatibility of ˆ n ) and an+1 (θn ; x ˆn, y ˆ n ) = a0 we obtain by induction a0 (ˆ ˆ n ) ∈ B n+1 ({a0 }). This, (ˆ xn , y xn , y n (iii), and the fact that B (A0 ) is increasing in n then imply the result. 0 0 n n ˆ n )}∞ ˆ ) → a as n → ∞, we pick an arbitrary subsequence {a0 (ˆ To verify a0 (ˆ x ,y xn , y n0 =1 00 00 ˆ n )}∞ and show that it has a further subsequence {a0 (ˆ xn , y that converges to a . Applying 00 n =1 n n n to (r , y ) the argument we applied to r in the proof of Proposition 3, we obtain a subindex 00 00 00 ˜ ). Also for each t we have xnt = xt for n00 along which (rn , yn ) converges to some (˜r, y 00 00 00 00 00 ˆ n ) = a0 (xn ◦ rn , yn ) → n00 ≥ t. This together with the boundedness of u implies a0 (ˆ xn , y 00 00 ˜ ). Combining this with a0 (ˆ ˆ n ) ≥ a0 (xn , y) and a0 (xn , y) → a, we obtain a0 (x ◦ ˜r, y xn , y

˜ ) ≥ a. But the incentive compatibility of (x, y) implies a0 (x ◦ ˜r, y ˜ ) ≤ a0 (x, y) = a, a0 (x ◦ ˜r, y ˜ ) = a. so a0 (x ◦ ˜ r, y ¯ )}. From the incentive compatibility of (¯ ¯ ), (ii), and (iii), Now let A0 = {a0 (¯ x, y x, y ∗ + we have B(A0 ) ⊂ A . To see A0 ⊂ B(A0 ), observe that if we set (x(θ), y(θ), a (θ)) = + ¯ )) for each θ we have (x, y, a ) ∈ F (a0 (¯ ¯ ); A0 ). (¯ x, y¯(θ), a0 (¯ x, y x, y The analog of the

T

operator maps

J : Θ × Y × A∗ → R

into

T J : Θ × Y × A∗ → R,

dened as:

T J(θ− , y− , a) =

inf +

(x,y,a )∈F (a;A∗ )

X

c(x(θ)) + qJ(θ, y(θ), a+ (θ)) π(θ|θ− , y− ).

(8)

θ

The analog of Proposition 4 is therefore:

J ∗ is a bounded lower semicontinuous function, and ||T n J − J ∗ || → 0 as n → ∞ for any bounded J : Θ × Y × A∗ → R. There exists a function g ∗ : Θ × Y × A∗ → (X × Y × A∗ )Θ which attains the inmum on the right hand side of (8) when J = J ∗ , and for any such g ∗ the allocation (x∗ , y∗ ) dened recursively by (x∗t (θt ), yt∗ (θt ), a∗t+1 (θt )) = ∗ (θt−1 ), a∗t (θt−1 ))(θt ) solves the auxiliary planning problem starting from (θ−1 , y−1 , g ∗ (θt−1 , yt−1 a∗0 (θ−1 )). Proposition A4.

Proof.

Virtually identical to that of Proposition 4.

References Fernandes, A.,

and

C. Phelan (2000): A Recursive Formulation for Repeated Agency with

History Dependence, Fukushima, K.,

and

Journal of Economic Theory, 91(2), 223247.

Y. Waki (2011):

Computing Dynamic Optimal Mechanisms When

Hidden Types Are Markov, Working paper.

5

Computing dynamic optimal mechanisms when hidden ...

Sep 25, 2011 - University of Wisconsin, Madison. Yuichiro Waki. University of Queensland ..... with the boundedness of u implies a0(xn , yn ) = a0(xn ◦rn ,yn ) →.

211KB Sizes 1 Downloads 197 Views

Recommend Documents

Computing Dynamic Optimal Mechanisms When ...
Sep 25, 2011 - than 15 dimensions, it is not good enough in many computational ... the problem in which non-local incentive constraints are ignored, and then goes on to check ..... rank K. One can always write π in this way with K = N,3 but it is al

Computing Dynamic Optimal Mechanisms When ...
Sep 25, 2011 - University of Wisconsin, Madison. Yuichiro ..... If the environment is convex, B maps convex sets into convex ...... Princeton University Press.

Optimal Allocation Mechanisms with Single ... - Semantic Scholar
Oct 18, 2010 - [25] Milgrom, P. (1996): “Procuring Universal Service: Putting Auction Theory to Work,” Lecture at the Royal Academy of Sciences. [26] Myerson ...

Optimal Mechanisms for Selling Information
Store. Pop Art. Supplies. Their value depends who is the user behind the click. .... Expected revenue: • How big does s (size of signaling space) need to be ?

Optimal Allocation Mechanisms with Single ... - Semantic Scholar
Oct 18, 2010 - We study revenue-maximizing allocation mechanisms for multiple heterogeneous objects when buyers care about the entire ..... i (ci,cLi)], where (pLi)z denotes the probability assigned to allocation z by pLi. The timing is as follows: S

Optimal Dynamic Hedging of Cliquets - Semantic Scholar
May 1, 2008 - Kapoor, V., L. Cheung, C. Howley, Equity securitization; Risk & Value, Special Report, Structured. Finance, Standard & Poor's, (2003). Laurent, J.-P., H. Pham, Dynamic Programming and mean-variance hedging, Finance and Stochastics, 3, 8

Optimal Dynamic Hedging of Cliquets - Semantic Scholar
May 1, 2008 - some presumed mid price of vanillas results in a replicating strategy for the exotic. Risk management departments (also called Risk Control) are charged with feeding the pricing model sensitivity outputs into a VaR model, and generally

Optimal Dynamic Actuator Location in Distributed ... - CiteSeerX
Center for Self-Organizing and Intelligent Systems (CSOIS). Dept. of Electrical and ..... We call the tessellation defined by (3) a Centroidal Voronoi. Tessellation if ...

An Optimal Lower Bound for Anonymous Scheduling Mechanisms
An easy observation made by [13] is that the well-known VCG mechanism ... of at most m times the optimal makespan, while making it the dominant strategy ..... Figure 1: An illustration of an instance that is a ({1, ..., j},i)-projection of t, as in D

Optimal Allocation Mechanisms with Single-Dimensional ... - DII UChile
Oct 18, 2010 - the allocation of “sponsored-link” positions on a search engine: Each .... virtual surpluses, the solution derived via pointwise optimization is not ...

An Optimal Lower Bound for Anonymous Scheduling Mechanisms
Mu'alem and Schapira [12] ...... each job independently using some non-affine-maximizer mechanism for single-dimensional domains. (those are abundant).

An Optimal Lower Bound for Anonymous Scheduling Mechanisms
scheduling algorithms are anonymous, and all state-of-the-art mechanisms ..... Figure 1: An illustration of an instance that is a ({1, ..., j},i)-projection of t, as in Def.

Optimal mechanisms for heterogeneous multi-cell ...
NW (1809T), Washington, DC; phone: 202-566-2265; fax: 202-566-2373; [email protected]. ‡Water Center of the Earth Institute, Columbia University, New ...

Efficiency of (Revenue-)Optimal Mechanisms - Research at Google
within a small additive constant of 5.7. ... Auctions are bid-based mechanisms for buying and selling ... benefit of the business in mind, they might want to keep.

Computing Dynamic Heterogeneous-Agent Economies
Jun 24, 2011 - Feasibility is achieved by using Smolyak's (1963) sparse-grid ... to introduce a method for computing equilibrium in these models by ...... This is made clear in Figure 2 where a scatter plot of today vs ... today's shocks and capital

Computing Dynamic Heterogeneous-Agent Economies ...
Jun 24, 2011 - Keywords: Numerical Solutions, Heterogeneous Agents, Projection Methods. JEL Codes: ..... analytic solution is also given. The OLG model is ...

Optimal Allocation Mechanisms with Single-Dimensional ... - DII UChile
Oct 18, 2010 - the allocation of “sponsored-link” positions on a search engine: Each .... virtual surpluses, the solution derived via pointwise optimization is not incentive-compatible. ... surplus of an allocation: For types between the critical

Points mechanisms and dynamic contracting with ...
Mar 31, 2017 - in a points program and to obtain the good through it with some ... billion, and concluded that the loyalty program, if it operated as a stand-alone business, ...... the mechanism is optimal for the seller, because both points owners .

Dynamic Resource Allocation Mechanisms for Grid ...
applications. Keywords: Grid Computing, Resource Allocation,. Resource Management, E-learning, Learning Object,. Grid Learning Object, Ontology, Semantic ...

Robot Mechanisms and Mechanical Devices Illustrated - Dynamic DNS
It is my sincere hope that this book will decrease the time it takes to pro- duce a working robot, ... a success. Building, designing, and working with practical mobile robots ..... scan the traced areas of resin to solidify the model's first cross s

Robot Mechanisms and Mechanical Devices Illustrated - Dynamic DNS
cost and labor of manual or machine finishing by sanding or polishing. ...... the host machine's mounting surface for their installation. Figure 1-14 This single-axis ...