Dilation Bootstrap: a method for constructing confidence regions for partially identified models Alfred Galichon and Marc Henry ´ Ecole polytechnique Universit´e de Montr´eal, CIRANO and CIREQ First draft: May 27, 2006 This draft1 : December 31, 2007

Abstract We propose a methodology for constructing confidence regions with partially identified models of general form. The region is obtained by inverting a test of internal consistency of the econometric structure. We develop a dilation bootstrap methodology to deal with sampling uncertainty without reference to the hypothesized economic structure, and apply a duality principle to reduce the dimensionality of the remaining deterministic problem. As a result, the confidence region becomes easily computable, and the methodology can be applied to the estimation of models with sample selection, censored observations and to games with multiple equilibria.

JEL Classification: C10, C12, C13, C15, C34, C52, C61. Keywords: Partial identification, incomplete specification test, duality, dilation bootstrap, nearest neighbours.

1

We thank Christian Bontemps, Gary Chamberlain, Victor Chernozhukov, Pierre-Andr´e Chiappori,

Ivar Ekeland, Rustam Ibragimov, Guido Imbens, Thierry Magnac, Francesca Molinari, Geert Ridder, Bernard Salani´e, participants at the ”Semiparametric and Nonparametric Methods in Econometrics” con´ ference in Oberwolfach and seminar participants at BU, Brown, CalTech, Chicago, Ecole polytechnique, Harvard, MIT, Northwestern, Toulouse, UCLA, UCSD and Yale for helpful comments (with the usual disclaimer). Financial support from NSF grant SES 0532398 is gratefully acknowledged by both authors.

1

Introduction In several rapidly expanding areas of economic research, the identification problem is steadily becoming more acute. In policy and program evaluation (Manski (1990)) and more general contexts with censored or missing data (Shaikh and Vytlacil (2005), Magnac and Maurin (2005)) and measurement error (Chen, Hong, and Tamer (2005)), ad hoc imputation rules lead to fragile inference. In demand estimation based on revealed preference (Blundell, Browning, and Crawford (2003)) the data is generically insufficient for identification. In the analysis of social interactions (Brock and Durlauf (2002), Manski (2004)), complex strategies to reduce the large dimensionality of the correlation structure are needed. In the estimation of models with complex strategic interactions and multiple equilibria (Tamer (2003), Andrews, Berry, and Jia (2003), Pakes, Porter, Ho, and Ishii (2004)), assumptions on equilibrium selection mechanisms may not be available or acceptable. More generally, in all areas of investigation with structural data insufficiencies or incompletely specified economic mechanisms, the hypothesized structure fails to identify a unique possible data generating mechanism for the data that is actually observed. Hence, when the structure depends on unknown parameters, and even if a unique value of the parameter can still be construed as the true value in some well defined way, it does not correspond in a one-to-one mapping with a probability measure for the observed variables. In other words, even if we abstract from sampling uncertainty and assume the distribution of the observable variables is perfectly known, no unique parameter but a whole set of parameter values (hereafter called identified set) will be compatible with it. In such cases, many traditional estimation and testing techniques become inapplicable and a framework for inference in incomplete models is developing, with an initial focus on estimation. A question of particular relevance in applied work is how to construct valid confidence regions for identified sets of parameters. Formal methodological proposals include Chernozhukov, Hong, and Tamer (2007), Imbens and Manski (2004), Andrews, Berry, and Jia (2003), Pakes, Porter, Ho, and Ishii (2004), Romano and Shaikh (2005), Beresteanu and Molinari (2006), Galichon and Henry (2006). Some of these contributions are reviewed in section 1.4. The picture is far from complete and there is space for alternative proposals, possibly computationally more efficient, more general or less conservative. In the present work, we propose a methodology that contributes to a better understanding of inference in incomplete models by clearly distinguishing how to deal with sampling un-

2

certainty on the one hand, and model uncertainty on the other. The key to this separation is to deal with sampling variability without any reference to the hypothesized structure, using a methodology we call dilation bootstrap. This consists in dilating each point in the space of observable variables in such a way that the empirical probability (which is known) of dilated sets dominates the true probability (which is unknown) of the original sets. The unknown true probability (i.e. the true data generating mechanism) is then removed from the analysis, and we can proceed as if the problem were purely deterministic. This deterministic problem is that of determining whether a parameter value is compatible with a known distribution, and the resulting confidence region is the set of such parameters. In great generality (i.e. in the case of an economic structure defined by a correspondence between observable and unobservable variables, and moment conditions on the latter), we show that this requirement of compatibility can be formulated as a programming problem, the dual of which provides a function of the parameters that takes value zero for compatible parameters, and is non-negative everywhere. Hence, the confidence region we propose is the set of zeros of a function that is straightforward to compute with convex optimization techniques. Section 1 defines the economic structure, the identified set and the confidence region. Section 2 defines the test statistic. Section 2.3 discusses the power of the test and the corresponding tightness of the confidence regions. Section 2.4 explains the dilation bootstrap procedure. The last section concludes.

1 1.1

General setup Econometric structure specification

We define the economic structure as in Jovanovic (1989), who pioneered the study of empirical implications of multiple equilibria. Variables under consideration are divided into two groups. Latent variables, u ∈ U ⊆ Rdu , are typically not observed by the analyst, but some of their components may be observed by the economic actors. Observable variables, y ∈ Y ⊆ N × Rdy , are observed by the analyst and the economic actors. P is the true data generating process for the observable variables, and ν a hypothesized data generating process for the latent variables. The econometric structure under consideration is given by a relation between observable and latent variables, i.e. a subset of Y × U, which we shall write as a correspondence from Y to U denoted by Γθ : Y ⇒ U, where θ ∈ Θ ⊂ Rdθ 3

is a vector of unknown parameters. The distribution ν of the unobservable variables U is assumed either to belong to a parametric family νθ , θ ∈ Θ, or to satisfy a set of moment conditions, namely Eν (mi (U ; θ)) = 0, mi : U → R, i = 1, . . . , dm

(1)

and we denote by Vθ the set of distributions that satisfy (1). In addition, denote M(P, νθ ) the set of probabilities on (Y × U ) with marginals P and νθ , and M(P, Vθ ) the set of probabilities on (Y × U ) with marginal P on Y and marginal on U in Vθ . Finally, we call M(θ, νθ ) the structure defined by the correspondence Γθ and the hypothesized distribution of unobservables νθ , and we call M(θ, Vθ ) the structure defined by the correspondence Γθ and the moment conditions underlying the definition of Vθ . We make the additional technical assumptions that Γθ is non-empty and closed-valued and measurable, i.e. for each open set O ⊆ U , Γ−1 θ (O) = {y ∈ Y | Γθ (y) ∩ O 6= ∅} ∈ BY , with BY and BU denoting the Borel σ-algebras of Y and U respectively. Observed Sample

DGP (Y, U ) ∼ π

Y U

{Y1 , . . . , Yn }

Restriction U ∈ Γθ (Y )

Latent Variable U Parametric νθ Or Emθ (U ) = 0

Figure 1: Summary of the structure. DGP stands for data generating process, i.e. a joint law generating the pairs (Yi , Ui ), i = 1, . . . , n, the first component of which is observed.

Example 1. Games with multiple equilibria. A simple class of examples is that of models defined by a set of rationality constraints. Suppose the payoff function for player j, j = 1, . . . , J is given by Πj (Sj , S−j , Xj , Uj ; θ), where Sj is player j’s strategy and S−j is the opponents’ strategies. Xj is a vector of observable characteristics of player j and Uj a vector of unobservable determinants of the payoff whose distributions are only known to satisfy a set of moment conditions as in (1). Finally θ is a vector of parameters. Pure strategy Nash equilibrium conditions Πj (Sj , S−j , Xj , Uj ; θ) ≥ Πj (S, S−j , Xj , Uj ; θ), for all S define a correspondence Γθ from unobservable variables U to observable variables Y = (S 0 , X 0 )0 . Example 2. Model defined by moment inequalities. A special case of the specification above is provided by models defined by moment inequalities. E(ϕi (Y ; θ)) ≤ 0, ϕi : Y → R, i = 1, . . . , dϕ , 4

(2)

where Y denotes the whole vector of observable variables, including explanatory variables. This is a special case of our general structure, where Γθ (y) = {u ∈ U : ui ≥ ϕi (y; θ), i = 1, . . . , du }, and mi (u; θ) = u, i = 1, . . . , dϕ , with du = dϕ . Example 3. Model defined by conditional moment inequalities. E(ϕi (Y ; θ)|X) ≤ 0, ϕi : Y → R, i = 1, . . . , dϕ ,

(3)

where X is a sub-vector of Y . Bierens (1990) shows that this model can be equivalently rephrased as E(ϕi (Y ; θ)1{t1 ≤ X ≤ t2 }) ≤ 0, ϕi : Y → R, i = 1, . . . , dϕ ,

(4)

for all pairs (t1 , t2 ) ∈ R2dx (the inequality is understood element by element). Conditionally on the observed sample, this can be reduced to a finite set of moment inequalities by limiting the class of pairs (t1 , t2 ) to observed pairs (Xi , Xj ), Xi < Xj . Hence this fits into the framework of example 2. Example 4. Unobserved random censoring (also known as accelerated failure time) model. A continuous variable Z = µ(X, θ) + ², where µ is known up to a vector of unknown parameters θ, is censored by a random variable C. The only observable variables are X, V = min(Z, C) and D = 1{Z < C}. The error term ² is supposed to have zero conditional median P (² < 0|X) = 0. Khan and Tamer (2006) show that this model can be equivalently rephrased in terms of unconditional moment inequalities. ·µ ¶ ¸ 1 E 1{V ≥ µ(X, θ)} − 1{t1 ≤ X ≤ t2 } ≤ 0 2 ·µ ¶ ¸ 1 E − D × 1{V ≤ µ(X, θ)} 1{t1 ≤ X ≤ t2 } ≥ 0 2 for all pairs (t1 , t2 ) ∈ R2dx (the inequality is understood element by element). Hence this fits into the framework of example 3. Finally we turn to an example of binary response, and a simple unconditional inequalities example, which we shall use as a pilot examples to illustrate each step of our procedure. Pilot Example 1. A Binary Response Model: The observed variables Y and X are related by Z = 1{Xθ + ε ≤ 0}, under the conditional median restriction Pr(ε ≤ 0|X) = η for a known η. In our framework the vector of observed variables is Y = (Z, X)0 , and to deal 5

with the conditioning, we take the vector U to also include X, i.e. U = (X, ε)0 . To simplify exposition, suppose X only takes values in {−1, 1}, so that Y = {0, 1} × {−1, 1} and U = {−1, 1} × [−2, 2], where the restriction on the domain of ε is to ensure compactness only. The multi-valued correspondence defining the model is Γθ : Y ⇒ U characterized by Γθ (1, x) = {x} × (−2, −xθ] and Γθ (0, x) = {x} × (−xθ, 2]. The two moment restrictions are m± (x, ε) = (1{ε ≤ 0} − η)(1 ± x). Pilot Example 2. A System of Moment Inequalities Model: We consider the set of moment conditions: Eϕ(Y, θ) ≤ 0 with θ0 = (θ1 , θ2 ), ϕ0 = (ϕ1 , ϕ2 , ϕ3 ) and ϕ1 (y, θ) = θ1 + θ2 y − 1, ϕ2 (y, θ) = −θ1 y, ϕ3 (y, θ) = −θ2 y. So Γθ (u) = {u ∈ R3 : ui ≥ ϕ(y, θ)}, and m(u, θ) = u.

1.2

Definition of the identified set

We argue that the structure is internally consistent for some value of the parameter if and only if the correspondence Γθ it defines is almost surely respected, i.e. if U ∈ Γθ (Y ) with probability one for some underlying probability structure. It captures the idea that the true data-generating process is compatible with the hypothesized structure. This statement is made precise in the following definition: Definition 1. For a given θ, model M(θ, Vθ ) (respectively M(θ, νθ )) is internally consistent if and only if there exists a joint probability π ∈ M(P, Vθ ) (respectively M(P, νθ )) such that π({(u, y) ∈ Y × U : u ∈ / Γθ (y)}) = 0. This prompts the natural definition of the identified set as the set of parameters such that the econometric structure is internally consistent. Call M(θ) the structure M(θ, νθ ) when the distribution of unobservable variables is parameterized, and M(θ, Vθ ), when the distribution of unobservable variables satisfies the moment conditions. Definition 2. The identified set is ΘI = { θ ∈ Θ : M(θ) is internally consistent }. Pilot example 1 continued One has Pr(Z = 1|X) = Pr(ε ≤ −Xθ). Suppose θ > 0. Then Pr(Z = 1|X = 1) = Pr(ε ≤ −θ|X = 1) ≤ Pr(ε ≤ 0|X = 1) = η, and similarly, Pr(Z = 1|X = −1) ≥ η. Symmetrical results hold for θ < 0. The resulting identified regions for θ are summarized in table 1. This illustrates the fact that θ is set-identified : in this example, only a set of θ is identifiable, depending on features of distribution of (X, Z).

6

Pr (Z = 1|X = −1)

Pr (Z = 1|X = 1)











{θ > 0}

{θ > 0}



{θ < 0}

{θ ∈ R} {θ > 0}



{θ < 0}

{θ < 0}



Table 1: Identified regions for the binary response pilot example.

Example 5. In the case of a model defined by moment inequalities as in example 2, the identified set is simply the set of θ such that the inequalities are satisfied, i.e. ΘI = {θ ∈ Θ : E(ϕi (Y ; θ)) ≤ 0, ϕi : Y → Rdu , i = 1, . . . , dϕ }. So, in particular: Pilot example 2 continued In this case, the identified region has the graphical representation given in figure 2.

θ1 = 0 1/E(Y )

θ

1

+

θ

2E

(Y

)

=

ΘI

1

1 0

θ2 = 0

Figure 2: Identified region for the inequalities pilot example for E(Y ) > 0.

1.3

Confidence region for the identified set

Given a sample (Y1 , . . . , Yn ) of independently and identically distributed realizations of Y , our objective is to construct a sequence of random sets Θαn such that for all θ ∈ ΘI , limn→∞ Pr (θ ∈ Θαn ) ≥ 1 − α. In other words, we are concerned with constructing a region 7

˜ that covers the Θαn that covers each value of the identified set, as opposed to a region Θ ˜ ≥ 1 − α. We do so (as in Chernozhukov, identified set uniformly, i.e. such that Pr(ΘI ⊆ Θ) Hong, and Tamer (2007), Imbens and Manski (2004) and others) by including in Θαn all the values of θ such that we fail to reject a test of internal consistency of M(θ) with asymptotic level at least 1 − α. We shall demonstrate the construction of a test statistic Tnα (θ) such that, conditionally on the structure M(θ) being internally consistent, the probability that Tnα (θ) = 0 is at least 1 − α asymptotically, i.e. lim Pr (Tnα (θ) = 0 | M(θ) is internally consistent) ≥ 1 − α.

n→∞

(5)

Hence we define our confidence region in the following way. Definition 3. The (1 − α) confidence region for ΘI is Θαn = {θ ∈ Θ : Tnα (θ) = 0}. The full procedure is summarized in table 2. It is clear from equation 5 and the above definition that our confidence region covers each element of the identified set with probability at least 1 − α asymptotically. Hence, after a section devoted to discussing in detail our contribution within the literature on the topic, the remainder of this paper will be concerned with the construction of the statistic Tnα (θ) with the required property 5.

1.4

Review of the literature

The first general purpose method to construct confidence regions with partially identified models was proposed by Chernozhukov, Hong, and Tamer (2007). The idea is to extend Mestimation to cases where the criterion function Q(θ) is null on a set of parameter values (as opposed to a unique parameter value as in identified models) and non negative everywhere. The identified set ΘI is simply defined as the set of parameter values for which Q(θ) is zero. The procedure consists in approximating the limiting distribution of the suitably normalized version of supθ∈ΘI Qn (θ), where Qn is the empirical criterion function, using an initial estimate of ΘI and either a bootstrap, a sub-sampling or a simulation procedure (based on the asymptotic distribution), and constructing the 1 − α confidence region for ΘI using the α quantiles obtained with the approximation. Although this approach is very elegant and very general in scope, it gives no theoretical guidance as to the choice of criterion function in a given situation, in contrast with the M-estimation literature which is motivated by a need for robust alternatives to maximum 8

Table 2: SUMMARY OF THE PROCEDURE 1. Draw B bootstrap samples (Y1b , . . . , Ynb ), and for each b and for each l, define Wlb as the matrix with (i, j)-th entry equal to 1 if Yjb is among the l nearest neighbours of Yi in Euclidian distance, and zero otherwise. Call lb the smallest value of l such that min{trace(Wlb Π) : Π permutation} = 0. Order the lb and call lnα the [Bα] largest, and take Jnα to be the correspondence that to each sample point Yj associates the lnα nearest neighbours of Yj within the initial sample (Y1 , . . . , Yn ). 2. Distinguish parametric and semiparametric cases: 2.1. If the distribution of unobservable variables is assumed to satisfy moment conditions. For each j, minimize 1{u ∈ / Γθ (Jnα (Yj ))} − λ0 m(u, θ) over u, keeping λ fixed. Call the minimum hλ (Yj ). MaxP imize supλ n1 nj=1 hλ (Yj ) over λ. 2.2. If the distribution of unobservable variables is parameterized, maximize Pn (A) − νθ (Γθ ◦ Jnα (A)) where A ranges over the quadrants {(−∞, Yj ], (Yj , ∞) : j = 1, . . . , n}. 3. Define the confidence region Θαn as the set of θ’s for which the previous maximization returns zero.

9

likelihood. Chernozhukov, Hong, and Tamer (2007) do propose a choice of Q in several leading examples, but it is difficult to assess the influence of this choice on the size of the confidence regions obtained within their framework. One proposal to fill this gap appears in Galichon and Henry (2006). Within the framework presented in the present paper, except for the fact that the distribution of the unobservable variables is parametric, viz. νθ instead of ν ∈ Vθ here, they show that the null hypothesis of internal consistency of the econometric structure for any given parameter value θ is equivalent to P (B) ≤ νθ Γθ (B) for all sets B in a class of sets C that they call coredetermining (meaning that P (B) ≤ νθ Γθ (B) for all B ∈ C implies P (B) ≤ νθ Γθ (B) for all measurable sets B). Hence the inf B∈C [νθ (Γθ (B)) − P (B)] is a classical choice of criterion function Q, with empirical version obtained by replacing P by Pn in the latter expression, thereby obtaining a generalized Kolmogorov-Smirnov statistic. Galichon and Henry (2006) show that the latter statistic can also be used directly to derive confidence regions that cover each element of ΘI with a prescribed probability. Indeed they derive the asymptotic distribution of inf B∈C [νθ (Γθ (B)) − Pn (B)] and propose to form the 1 − α confidence region with all the values of the parameter such that the hypothesis of internal consistency of the structure is not rejected at the 1 − α level of significance. Andrews, Berry, and Jia (2003) also propose a heuristic for testing the inequality P (B) ≤ νθ Γθ (B), and hence were the first to apply a Kolmogorov-Smirnov specification testing type of approach to problems of partial identification. However, they restrict attention to Y finite (as in the case of games with discrete strategies) and singleton B’s. In other words, they test the sufficient conditions P ({y}) ≤ νθ Γθ ({y}) for each observable value y in the finite set Y. Although it is natural to restrict attention to Y finite, since structures arising from games with discrete strategies are more likely to give rise to identification failures, restricting attention to the singletons leads to conservative inference, since νθ Γθ (.) is an alternating capacity (i.e. a set function that satisfies νθ (Γθ (B∪C)) ≤ νθ (Γθ (B))+νθ (Γθ (C)) for all B ∩ C = ∅, see for instance definition 1.8 page 7 of Molchanov (2005)). The present paper extends the results of Galichon and Henry (2006) to the case where no parametric assumption is made on the unobservable variables U , thereby including as a special case structures defined by moment inequalities. It also improves on the Galichon and Henry (2006) proposal significantly in terms of implementability. Indeed, the asymptotic distribution of the test statistic in the latter is not distribution free, which means that quantiles must be approximated via the simulation of a Brownian bridge or via a subsampling procedure, both of which are computationally costly and require some ad-hoc choices of deterministic sequences that are unpalatable to practitioners and always leave 10

open the way for data-snooping. To avoid these drawbacks, we propose a (computationally economical) dilation bootstrap procedure which controls sample variability without any reference to the hypothesized econometric structure under scrutiny. Credit also goes to the pioneering insight of Andrews, Berry, and Jia (2003), who propose a procedure in a similar spirit. However, since they restrict attention to singletons, theirs is a simple bootstrap, whereas the dilation bootstrap presented here is a genuine statistical innovation. After the dilation bootstrap is performed (once and for all, since it is independent of θ), one can proceed as if everything was known, so the remainder of the construction of the confidence region is a deterministic optimization problem. In addition to the obvious practical advantages, we believe this approach to be the most natural in cases where the hypothesized structure incompletely specifies the true data generating process. Note also that since the dilation bootstrap operates without reference to the parameterized structure, it is free of the drawbacks documented in Andrews (2000). A significantly different approach is advocated in Beresteanu and Molinari (2006). Their proposal can be seen as a non structural alternative to our fully structural approach. They propose to construct from the observed variables a sequence of random sets with expectation equal to the identified set. As a result of that construction, which they explicit in the case of the linear model with interval censored endogenous variables, they can apply laws of large numbers and central limit theorems for random sets to test hypotheses on the identified set, and invert such tests to obtain confidence regions. Other notable works on the subject include Andrews and Guggenberger (2007), Andrews and Soares (2007), Bugni (2007), Rosen (2006) and Santos (2007).

2

Internal consistency test statistic

This section is concerned with constructing the statistic Tnα (θ) that satisfies requirement 5. In other words, calling M(P, θ) either M(P, Vθ ) or M(P, νθ ) depending on the case considered, we set out to construct a conservative test statistic for the hypothesis of internal consistency H0 (θ) :

∃π ∈ M(P, θ) : π({(u, y) ∈ Y × U : u ∈ / Γθ (y)}) = 0,

with rejection region R\{0}. The definition of alternatives will be discussed in section 2.3. We define our test statistic and present our formal results before proceeding to a heuris11

tic discussion of the principles involved. In all that follows, Pn refers to the empirical distribution of the sample (Y1 , . . . , Yn ), which gives mass 1/n to each of the sample points.

2.1

Internal consistency as an optimization problem

The null hypothesis of internal consistency H0 (θ) is equivalent to the optimization problem min

π∈M(P,θ)

: π({(u, y) ∈ Y × U : u ∈ / Γθ (y)}) = 0.

This program is not computationally workable as such as it requires optimizing over an infinite-dimensional space. This problem has the following dual formulations sup (P (A) − νθ (Γθ (A))) = 0

A∈BY

in the case of a parametric assumption on the distribution of the unobservable variables, and sup EP [gλ,θ (Y )] = 0, where gλ,θ (y) = inf (1{u ∈ / Γθ (y)} − λ0 m(u; θ)) , u

λ∈Rdm

in the case of where unobservable variables are constrained by moment inequalities. We use these dual formulations to construct feasible test statistics as explained in the next sections. Pilot example 1 continued Here, we have λ = (λ1 , λ2 ) ∈ R2 and gλ,θ (x, 0) = min( inf {−λ0 m(ε, x)}; inf {1 − λ0 m(ε, x)}), ε≥−xβ

ε≤−xβ

0

gλ,θ (x, 1) = min( inf {−λ m(ε, x)}; inf {1 − λ0 m(ε, x)}). ε≤−xβ

ε≥−xβ

Since we have seen that only the sign of θ can be identified, we assume that θ ∈ Θ = {−1, 0, 1}. We show at the end of the appendix that supλ∈R2 EP (hλ,θ=−1 (Z)) = 0 if and only if η ≥ PrZ|X (1| − 1) and η ≤ PrZ|X (1|1).

2.2

Definition of the test statistics

To define our test statistic, we need an auxiliary construction called a dilation. Call Jnα a sequence of correspondences Jnα : Y ⇒ Y that satisfies Pr (∀A ∈ BY : P (A) ≤ Pn (Jnα (A))) → 1 − α, where BY denotes the class of (Borel) measurable subsets of Y. 12

(6)

Y

U

Γ(y)

α Γ ◦ Jn (y)

Γ(y)

α Γ ◦ Jn (y)

y

Y

U

y α Jn (y)

Figure 3: The shaded area shows the graph of the original structure in both figures. The dotted line in top figure shows the dilated structure, and the bottom figure shows how the dilated structure is constructed.

13

The idea of the dilation is to make the structure slightly more permissive, as illustrated in figure 3, so that we can check its compatibility with the empirical distribution in place of the true unknown distribution of the observed variables. We define the test statistics as follows: • In case the distribution of the unobservable variables is parametric: Tnα (θ, νθ ) = sup(Pn (A) − νθ (Γθ ◦ Jnα (A))),

(7)

A∈C

where C = {(−∞, Yj ], (Yj , ∞) : j = 1, . . . , n}. • In case the unobservable variables satisfy moment conditions: Ã n ! X 1 Tnα (θ, Vθ ) = sup hλ (Yj ) , n j=1 λ

(8)

where hλ (y) = inf u (1{u ∈ / Γθ ◦ Jnα (y)} − λ0 m(u, θ)). Pilot example 2 continued In that case, Jnα (y) = [y − δnα (y), y + δnα (y)], where for each j, δnα (Yj ) is the distance between Yj and its neighbour of rank lnα . The latter is determined with the algorithm of section 2.4.2 based on B bootstrap replications. Defining H(y) := Γθ (Jnα (y)) = {(u1 , u2 , u3 )0 :

u1 ≥ −θ1 (y + δnα (y)) u2 ≥ −θ2 (y + δnα (y)); u3 ≥ −1 + θ1 + θ2 (y − δnα (y))},

we can write

µ hλ (y) = min

¶ 0

0

inf −λ u, inf (1 − λ u) ,

u∈H(y)

u∈H(y) /

so that the inner optimization, i.e. the computation of hλ , is a simple linear programming P problem, and the outer optimization, i.e. the maximization of n1 nj=1 hλ (Yj ), is a nonsmooth convex optimization problem1 .

We can now state the theorem that validates the construction of our confidence region. 1

For simulation purposes in Appendices 2 and 3, we performed the outer optimization with the Hybrid

Algorithm for Non-Smooth Optimization, written by Michael Overton

14

Theorem 1. Suppose the following assumptions hold: (A1) The restriction of P to Rdy is absolutely continuous with respect to Lebesgue measure and (A2) Jnα satisfies requirement (6). Then limn→∞ Pr(Tnα (θ) = 0) ≥ 1 − α for all θ ∈ ΘI , where Tnα (θ) = Tnα (θ, Vθ ) or Tnα (θ, νθ ). Hence, limn→∞ Pr(θ ∈ Θαn ) ≥ 1 − α for all θ ∈ ΘI . Remark 1. Note that assumption (A1) does not rule out discrete observable variables. It only states that the observable variables are a combination of purely discrete variables and absolutely continuous ones. The formal proof is in appendix 3. We give a detailed heuristic to explain the principle of the test. The hypothesis of internal consistency of the structure is that U ∈ Γθ (Y ) almost surely, where U is the unobservable variable, and Y is the observable variable. A suitable choice of Jnα ensures that, with probability larger than (1 − α) asymptotically, we have Y ∈ Jnα (Y ∗ ), where Y ∗ is a bootstrap variable (i.e. a variable with distribution Pn ), hence that U ∈ Γθ (Jnα (Y ∗ )). Call the latter statement (Snα ). Since (Snα ) does not depend on the unknown data generating process P , there only remains to determine whether (Snα ) is true or not. Formally, (Snα ) states the existence of a joint law π ∈ M(Pn , θ) such that π({(u, v) : u ∈ / Γθ (Jnα (v))}) = 0. Hence (Snα ) can be written as the solution of a programming problem minπ∈M(Pn ,θ) π({(u, v) : u ∈ / Γθ (Jnα (v))}) = 0. Since Tnα (θ) is the dual formulation of this programming problem, the level of the test of the internal consistency hypothesis based on the rejection region {Tnα (θ) 6= 0} is larger than 1 − α asymptotically.

2.3

Power of the test against fixed alternatives

Remember that for a fixed value of the parameter vector θ, our null hypothesis of internal consistency is the following: H0 (θ) :

∃π ∈ M(P, θ) : π({(u, y) ∈ Y × U : u ∈ / Γθ (y)}) = 0

which is equivalent to min

π∈M(P,θ)

: π({(u, y) ∈ Y × U : u ∈ / Γθ (y)}) = 0

We need to distinguish two kinds of fixed alternatives, the • Quasi-consistent alternatives HQC (θ) :

inf

π∈M(P,θ)

: π({(u, y) ∈ Y × U : u ∈ / Γθ (y)}) = 0 15

but the infimum is not attained, • Inconsistent alternatives HIC (θ) :

2.3.1

inf

π∈M(P,θ)

: π({(u, y) ∈ Y × U : u ∈ / Γθ (y)}) > 0.

Quasi-consistent alternatives

Since our procedure is insensitive to the fact that the infimum is attained or not, we need to make sure quasi-consistent alternatives are ruled out. Such occurrences are ruled out by Theorem 1 of Galichon and Henry (2006) for the case of a parametric distribution of unobservable variables, so we concentrate on the semiparametric case where the distribution of unobservable variables satisfies a set of moment conditions. The following example shows that the infimum is not always attained. Example 6. Let P = N (0, 1), U = Y = R, Vθ = {ν : Eν (U ) = 0}, and Γθ (y) = {1} for all y ∈ Y, and consider the distribution πm = P ⊗ νm such that νm ({1}) = 1 − 1/m, and νm ({1 − m}) = 1/m. The πm probability of U ∈ / Γθ (y) is 1/m which indeed tends to zero as m → ∞, but it is clear that there exists no distribution ν which puts all mass on {1} and has expectation 0. It is clear from example 6 that we need to make some form of assumption to avoid letting masses drift off to infinity. The theorem below gives formal conditions under which quasiconsistent alternatives are ruled out. It says essentially that the moment functions m(u, θ) need to be bounded. In all that follows, we assume without loss of generality that U ⊂ Rdu is actually the domain of U . £ ¤ Assumption 1. For all θ ∈ Θ, limM →∞ supν∈Vθ ν k m (U, θ) k 1{km(U,θ)k>M } = 0. Assumption 2. For all θ ∈ Θ, for every K ≥ 0, the set {u :k m (u, θ) k≤ K} is included in a compact set. Assumption 3. The graph of Γθ , i.e. {(u, y) ∈ U × Y : u ∈ Γθ } is closed. Example 7. In example 2, by Theorem 1.6 page 9 of Rockafellar and Wets (1998), we know that assumption 3 is satisfied when the moment functions ϕj , j = 1, . . . , dϕ are lower semi-continuous. We can now state the result: 16

Theorem 2. Under assumptions 1, 2 and 3, H0 implies HIC , ie. HQC is ruled out. Remark 2. Assumption 1 is an assumption of uniform integrability. It is immediate to note that assumptions 1 and 2 are satisfied when the moment functions m(u, θ) are bounded and U is compact. Pilot example 1 continued Here, U=[-2,2] is compact, and m± (ε, x) = [1{ε≤0} − η](1 ± x) are bounded. Thus the admissible set of measures M(P, Vθ ) is uniformly tight, and we can replace the “min” by an “inf” in the expression of H0 .

2.3.2

Inconsistent alternatives

Now that we have ruled out quasi-consistent alternatives (that our procedure is insensitive to), we give conditions for our procedure to have power against inconsistent alternatives, i.e. that our test rejects inconsistent alternatives with probability tending to one. First, this requires that the dilation used for the confidence region shrinks to the identity mapping at a suitable rate to ensure that µ ¶ α Pr min π{(v, u) : u ∈ / Γθ ◦ Jn (v)} = 0 | HIC → 0 π∈M(Pn ,θ)

(9)

for almost all samples and for all θ ∈ Θ\ΘI . Second, we need to ensure that the dual formulation used in the test statistic is not smaller than the primal program. Semiparametric case In the case of unobservable distributions defined by moment restrictions, we need an additional assumption (generally called a Slater condition in the optimization literature): Assumption 4. For almost all samples (Y1 , . . . , Yn ), there exists a measurable function g and a vector λ and ² > 0 such that for all (u, v) ∈ U × {Y1 , . . . , Yn }, g(y) + λ0 m(u, θ) < 1{u ∈ / Γθ ◦ Jnα (y)} − ². Remark 3. This condition is an interior condition, i.e. it ensures there exists a feasible solution to the optimization problem in the interior of the constraints. Theorem 3. Under 9 and assumptions 1, 2, 3 and 4, Pr(Tnα (θ, Vθ ) = 0) → 0 for each θ ∈ Θ\ΘI . 17

Remark 4. As described in the appendix, this is ensured by the fact that there is no duality gap, i.e. that the statistic obtained by duality is indeed positive when the primal is. Parametric case In the case of parametric assumptions on the distribution of unobservable variables, we know from Theorem 1 of Galichon and Henry (2006) that minπ∈M(Pn ,θ) π{(v, u) : u ∈ / Γθ ◦ Jnα (v)} is equal to its dual supA∈BY (Pn (A) − νθ (Γθ ◦ Jnα (A))). However, the latter is infeasible, so the problem of power is equivalent to the problem of determining when it is equal to the proposed statistic (7). This problem is equivalent to showing that the class of sets C = {(−∞, Yj ], (Yj , ∞) : j = 1, . . . , n} is “core determining” for the Choquet capacity functional A 7→ νθ (Γθ ◦ Jnα (A)). A definition of core determining classes is given in Galichon and Henry (2006), together with conditions on the structure that ensure that the class C = {(−∞, Yj ], (Yj , ∞) : j = 1, . . . , n} is core determining, and hence that the statistic supA∈C (Pn (A) − νθ (Γθ ◦ Jnα (A))) has power against inconsistent alternatives.

2.4 2.4.1

Dilation Bootstrap Dilation

We now turn to the question of how to construct the dilation Jnα that satisfies requirement 6. The idea is to dilate each point in space in such a way that we account for all sampling variability before introducing any reference to the hypothesized structure M(θ). In one dimension, this boils down to the problem of constructing uniform bands for the quantile based on the law of iterated logarithm as illustrated on figure 4.

In general, the suitable dilation is determined with an appeal to the empirical bootstrap principle. We use the discrepancy between the empirical distribution Pn and the bootstrap distribution Pn∗ (the empirical distribution of a sample of n independent and identically distributed variables with distribution Pn ) to approximate the discrepancy between the true distribution of observable variables P and the empirical distribution Pn . Hence we choose Jnα such that: Pr (∀A ∈ BY : Pn∗ (Jnα (A)) ≥ Pn (A)) → 1 − α in probability conditionally on the original sample (Y1 , . . . , Yn ). 18

(10)

F −1 (v) Y = F −1 (V )

Fn−1 (v)

α Jn (F −1 (v))

α F −1 (v) − δn (v)

v V ∼ U [0, 1]

Figure 4: Uniform quantile bands: F and Fn are the true and empirical distributions of the observed variables Y . Jnα is the dilation.

2.4.2

Dilation Bootstrap Procedure

Consider B iid samples (Y1b , . . . , Ynb ), b = 1, . . . , B drawn from Pn and call Pnb the empirical distribution of sample b.

1. For each bootstrap replication b and for each l, define: b Wi,j (l) = 1{Yjb is not among the l nearest neighbours of Yi (Yi included)}

cb (l) = min{trace(Wlb Π) : Π is a permutation matrix} lb = min{l : cb (l) = 0} 2. Let lnα be the [Bα] largest among the lb , b = 1, . . . , B, and take Jnα to be the correspondence that to each sample point Yj associates the lnα nearest neighbours (in Euclidian distance) of Yj (itself included) within the initial sample (Y1 , . . . , Yn ). Discussion: b (l))ni,j=1 is a matrix of 0-1 weights which penalizes the pairs (Yi , Yjb ) The matrix Wlb = (Wi,j

where Yi and Yjb are too far apart in nearest neighbour ranking. The quantity trace(Wlb Π) is the number of pairs that are more than l neighbours apart in the one-to-one matching of initial sample points Y1 . . . , Yn with bootstrap sample points Y1b , . . . , Ynb defined by the 19

permutation matrix Π. The cost function cb (l) is the minimum matching cost among all possible matches (i.e. all possible permutation matrices). A cost cb (l) = 0 means that there is a matching where the pairs (Yi , Yjb ) are never more than l neighbours apart. Hence, if we define Jn,l as the correspondence y ⇒ {y and its l − 1 nearest neighbours}, we have for any F ⊆ Y, Pnb (Jn,l (F )) ≥ Pn (F ) since by construction Jn,l (F ) contains at least as many bootstrap draws as F contains original draws. As a result, by choosing Jnα as in step 2 of the bootstrap procedure above, we ensure that for a proportion 1−α of the bootstrap samples, we have Pnb (Jnα (F )) ≥ Pn (F ) for all F ⊆ Y. Remark 5. Although the problem of finding the minimum cost matching (called the assignment problem or marriage problem) is very familiar to economists, as far as we know, its application within a bootstrap procedure is unprecedented. The best known algorithms to find the optimal assignment require O(n3 ) operations, which is computationally the most demanding element in our bootstrap procedure2 (whose formal complexity is in O(n3 ln n)).

2.4.3

Bootstrap Quantile

In the case where Y is the real line, the optimal assignment can be obtained in the following way (known as the greedy algorithm): for each order statistic Y(i) of the original sample, starting with the smallest Y(1) , find the closest bootstrap draw that has not been matched yet, and match it with Y(i) . Lemma 4 in the appendix justifies this construction and this optimal match has the simple graphical representation given in figure 5. As shown in lemma 4 and seen in figure 5, where the Y(j) ’s j = 1 . . . , n are the order statisb tics of the original sample, and the Y(j) ’s j = 1 . . . , n are the order statistics of the b-th boot-

strap sample, the matching that minimizes the maximum index distance between matched points is the bootstrap quantile function. Hence, by (5.3.1) page 163 of Cs¨org˝o and R´ev´esz (1981), the number of nearest neighbours lb is bounded above by (2n ln ln n)1/2 , which bounds the rate at which the correspondence Jnα shrinks to the identity: viz. (ln ln n/n)1/2 . 2

We propose a Matlab implementation of the whole bootstrap procedure, where the assignment pro-

cedure is borrowed from Markus Buehren’s implementation of the Kuhn-Munkr´es algorithm (also known as the Hungarian method, see Papadimitriou and Steiglitz (1982)). One bootstrap iteration takes about 15 minutes for n = 1000 on an AMD Opteron (tm) Processor 250 with 4G of RAM. Since the dilation bootstrap needs to be performed only once on a given data set, this seems very reasonable, especially if one uses parallel processing.

20

Y(1)

Y(2)

Y(n)

b Y(1)

b Y(n)

Figure 5: Optimal matching. The maximum number of neighbours is lb = 4 since the longest arrow visible on the graph links a point with its third nearest neighbour in Euclidian distance.

In higher dimensions, the identification with the quantile process is no longer available, but the limit of the distribution of the bootstrap neighbour process lb can be derived using results in combinatorial probability theory. The results are collected in lemma 1, and complemented by a simulation study described in appendix 2. Lemma 1. If Y is uniformly distributed on [0, 1]d , then for almost all samples (Y1 , . . . , Yn ), there exist constants c, C, c(d), C(d), K such that the diameter δn of Jnα satisfies: • For d = 1,



2nδn (ln ln n)−1/2 → 1

∗ −a.s.

√ ¡ ¢ • For d = 2, P r∗ c(ln n)3/4 n−1/2 ≤ δn ≤ C(ln n)3/4 n−1/2 ≥ 1 − n−K ln n

• For d ≥ 3, c(d) ≤ lim inf

δn n−1/d (ln n)1/d

δn ≤ lim sup n−1/d (ln ≤ C(d) n)1/d

∗ −a.s.

Remark 6. This result implies that in the uniform case (and it can presumably be extended to a large class of distributions), the rate at which Jnα tends to the identity is bounded above by (ln ln n/n)1/2 when d = 1 (so that lb is bounded above by (2n ln ln n)1/2 ), n−1/2 (ln n)3/4 when d = 2 (so that lb has rate bounded by (ln n)3/2 ) and n−1/d (ln n)1/d when d ≥ 3 (so that lb has rate bounded by ln n).

2.4.4

Randomization of Discrete Variables

The dilation bootstrap procedure outlined above relies on Euclidian geometry for the nearest neighbour matching. Hence it is not directly applicable to the case where the observable 21

variable Y has a discrete component, as in our pilot example 1. However, we can embed the discrete case within the continuous case using an extraneous randomization device, as we now explain. Let Y = (Y1 , Y20 )0 , with Y1 ∈ Y1 = {y 1 , . . . , y K } and Y2 ∈ Y2 ⊆ Rdy with absolutely continuous distribution with respect to Lebesgue measure. Let V be a uniformly distributed random variable on [0, 1] independent of Y and U . Call pn (y k ) the frequency of occurrence Pk l of y k in the sample, for k = 1, . . . , K, and call cn (y k ) = l=1 pn (y ) the cumulative frequency. Consider ϑ : Y1 × [0, 1] → R defined by ϑ(y1 , v) = cn (y k ) − vpn (y k ), as illustrated in figure 6. Finally, call Y˜1 = ϑ(Y1 , V ) and Y˜ = (Y˜1 , Y20 )0 . Notice that Y˜1 is a random variable continuously distributed on [0, 1]. y3

0

cn (y 1 )

cn (y 2 )

cn (y 3 )

cn (y K−1 )

1

pn (y 3 )

Figure 6: Observations equal to y 3 are randomized uniformly on (cn (y 2 ), cn (y 3 )].

To test compatibility of data and structure using our randomized data, we need to reformulate the structure in such a way that ˜ θ (Y˜ ) ⇐⇒ U ∈ Γθ (Y ). U ∈Γ

(11)

Calling ϑinv the mapping that “reverses” the randomization, defined as ϑinv (w) = y k , for cn (y k−1 ) < w ≤ cn (y k ) ˜ θ : [0, 1] × Y2 ⇒ U defined by the correspondence Γ ˜ θ ((W, Y20 )0 ) = Γθ (ϑinv (W ), Y20 )0 Γ satisfies (11). The mapping ϑinv simply picks out the observed element of {y 1 , . . . , y K } which could have given rise to the randomization W . Hence the structure is not modified ˜θ. with this definition of Γ 22

Conclusion In this paper we contribute to the understanding of inference in partially identified models with a new proposal for constructing confidence regions for partially identified parameters. We argue that our method has the following advantages over existing approaches: • It clearly distinguishes the treatment of sampling uncertainty and model uncertainty. • It deals with the former with a (computationally reasonable) dilation bootstrap procedure that is performed once and for all on the data set, irrespective of the hypothesized economic structure to be evaluated. • As for model uncertainty, it is treated with a duality principle, which allows to reduce the dimensionality of the optimization problem implicit in the definition of the identified region. As a result, the complexity of this part of the procedure is commensurate with the complexity of the hypothesized economic structure itself. • The procedure involves no choice of deterministic sequence as do the sub-sampling and the adaptive approaches. Since it is based on searching over the parameter space for values for which our statistic is equal to zero, it is still only applicable for low dimensionality parameter spaces. A method for dealing with nuisance parameters still has to be developed for the method to be applicable in a wider context.

Appendix 1: Simulations for Pilot Examples We conducted two simulation experiments to investigate small sample properties of our test of internal consistency on both our pilot examples.

Pilot example 1 continued Given the symmetry apparent in table 1, we can concentrate our analysis of small sample properties of our test on the case where the true value of θ is θ = −1 and P (Z = 1|X = 1) = 0.9. Then, the model is internally consistent if and only if P (Z = 1|X = −1) ≤ η, and we fix the latter to η = 0.5. We investigate size properties of the test for P (Z = 1|X = −1) = 0.5 and power properties with P (Z = 1|X = −1) = 0.6, 0.7, 0.8. Each case for sample sizes n = 500, 1000, 2000.

23

Pr (Z = 1|X = −1) 0.5

0.6

0.7

0.8

500

0.00

0.00

0.17 0.74

n 1000

0.00

0.01

0.53 1.00

2000

0.00

0.05

0.92 1.00

Table 3: Rejection rates for the test of internal consistency. The structure is internally consistent on the left of the double line, and inconsistent on the right of the double line.

In each case, we simulate 100 samples. X is simulated as a binary variable which takes value 1 with probability 0.5, and −1 otherwise, and Z as a binary variable conditioned by the value of X with parameters P (Z = 1|X = 1) and P (Z = 1|X = −1). For each sample, we run the test with 100 dilation bootstrap replications, a level of significance 1 − α = 0.9 and a starting seed value of 7 for the Matlab random number generator. Table 3 shows the simulation results. The conservativeness of the procedure in this particular example is apparent in the actual size of 0 for a nominal size of 0.1 and the very low power against departures of 0.1 from values of Pr(Z = 1|X = −1) which make the structure internally consistent. For larger sample size and larger departures, the power improves rapidly.

Pilot example 2 continued To illustrate the procedure on pilot example 2, we simulated a sample of n = 100 iid U [1/2, 3/2] random variables, and computed the 90%-confidence region for the parameter θ, i.e. the set of θ’s such that thge hypothesis of internal consistency is not rejected at the 0.1 level of significance. The dilation bootstrap was run on the initial simulated sample with 100 bootstrap resamples. Since the equation of the diagonal line is θ1 + θ2 E(Y ) = 1, the confidence band is widest in the neighbourhood of point (0, 1), as plotted on figure 7.

Appendix 2: Dilation Bootstrap Numerical Results The results of lemma 1 are insufficient in that they asymptotic and that they apply to uniformly distributed Y ’s only and for dimension 1, the exact upper bound for the rate of the number of neighbours is provided, but for higher dimensions, they only provide bounding rates without the constant. We ran a simulation experiment to gain insight into the constant term for higher dimensions, and to verify our conjecture that the rates apply to other data-generating processes for the observed variables Y . We considered sample sizes n = 50, 250, and for each, we simulated 100 initial samples of n iid

24

1.25

1

ΘI

0

1

Figure 7: Confidence region for the inequalities pilot example for E(Y ) = 1. The shaded are is the true identified region, whereas the solid line marks the frontier of the confidence region, i.e. the set of first values of θ such that the test of internal consistency is rejected, on a grid of resolution 0.01 × 0.01.

standard multivariate normal random variables with dimensions d = 1, 2, 3, 5. For each sample we computed the minimum number of neighbours necessary for a matching with each of 100 bootstraps samples. The Matlab random number generator was initialized at seed= 1. The quantile curves for the neighbour distributions are plotted on figure 8. The distributions are very highly concentrated around the mean even for sample size n = 50, except maybe for d = 1. As predicted by the theoretical rates, the number of neighbours decreases with dimension, and then becomes stable for d ≥ 3: indeed, it is apparent that the neighbour distributions are insensitive to dimension for d ≥ 3. Table 4 confirms how close the neighbour distributions are to asymptotic values given in lemma 1. To the best of our knowledge, there is no theoretical derivation of the √ constant for d ≥ 2, which in our simulations seems close to 2 for d = 2 and 2 for d ≥ 3. This √ means that our dilation bootstrap procedure constructs clusters of 2n ln ln n nearest neighbours √ for d = 1, 2(ln n)3/2 nearest neighbours for d = 2 and 2 ln n nearest neighbours for d ≥ 3.

25

Figure 8: Neighbour Distributions

Sample size=50

Nb Neighbours

24

19 dimension 1 dimension 2 dimension 3 dimension 5

14

9

4 0.01

0.11

0.21

0.31

0.41

0.51

0.61

0.71

0.81

0.91

Quantile

Sample size=250

56

Nb Neighbours

46 dimension 1 dimension 2 dimension 3 dimension 5

36

26

16

6 0.01

0.11

0.21

0.31

0.41

0.51

0.61

Quantile

26

0.71

0.81

0.91

Dimension Sample size

1 50

2 250

3

50

250

K2 (ln n)3/2

50

5 250

50

250

Theoretical rate

(2n ln ln n)1/2

Simulated number

14.51

32.09

10.10

14.80

8.66

11.29 8.06

10.66

Simulated number Theoretical rate

1.24

1.09

1.30

1.14

2.21

2.04

1.93

K3 ln n

K5 ln n 2.06

Table 4: Comparison of simulated and theoretical rates for the dilation bootstrap nearest neighbour distributions

Appendix 3: Proofs of Theorems in the Main Text Proof of Theorem 1 : Discussion and reader’s guide to the proof The key to the construction of the internal consistency test statistic is the observation that the hypothesis of internal consistency of definition 1 can be seen as an optimization problem: indeed, the following statements are trivially equivalent. (i) There exists a joint probability π ∈ M(P, Vθ ) such that π({(u, y) ∈ Y ×U : u ∈ / Γθ (y)}) = 0, (ii) minπ∈M(P,Vθ ) : π({(u, y) ∈ Y × U : u ∈ / Γθ (y)}) = 0. By duality, statement (ii) above can be shown to imply (iii) supλ

R

gλ,θ (y)dP (y) = 0 where gλ,θ (y) = inf u (1{u ∈ / Γθ (y)} − λ0 m(u)).

One route (which we take in a companion note available on request) is to directly consider the P empirical equivalent of formulation (iii), i.e. supλ nj=1 gλ,θ (Yj )/n and show that with a root-n re-scaling, it converges to the supremum of a P -Brownian bridge over the sub-class comprising R the functions gλ,θ such that gλ,θ (y)dP (y) = 0. Then quantiles may be obtained for the test statistic with either a simulation or a sub-sampling procedure. What we propose instead is to control sampling error prior to considering the optimization problem and taking its dual. Hence the idea to dilate each point in space using the correspondence Jnα : Y ⇒ Y in such a way that the probability of Y ∈ Jnα (Y ∗ ) when Y is distributed according to P and Y ∗ according to Pn is approximately (1 − α) conditionally on the sample (these heuristic statements are made precise using coupling and decoupling arguments). Observing that if

27

Y ∈ Jnα (Y ∗ ), then {U ∈ / Γθ (Jnα (Y ∗ ))} ⊆ {U ∈ / Γθ (Y )}, we see that the event U ∈ / Γθ (Jnα (Y ∗ )) has probability zero under the null that U ∈ Γθ (Y ). Formally, what we show is that under the hypothesis of internal consistency of M(θ, Vθ ), there exists a probability π ∈ M(Pn , Vθ ) such that π({(u, y) ∈ Y × U : u ∈ / Γθ (Jnα (y))}) = 0. We then apply the duality principle to show that the latter is equivalent to Tnα (θ, Vθ ) = 0, which completes the argument since Jnα is constructed in such a way that this happens with probability at least 1 − α asymptotically. Proof: The assumptions of the theorem are the following: 1. The structure M(θ, Vθ ) is internally consistent, hence there exists a joint probability π ∈ M(P, Vθ ) such that π({(u, y) ∈ Y × U : u ∈ / Γθ (y)}) = 0, 2. The dilation Jnα satisfies requirement 6, hence with probability tending to 1−α, there exists a joint probability π ∗ ∈ M(P, Pn ) such that π({(y, v) ∈ Rdy × Rdy : y ∈ / Jnα (v)}) = 0. For the remainder of the proof, all statements are with probability tending to 1 − α. From the assumptions above, there is a coupling π of the pair (Y, U ) such that U ∈ Γθ (Y ) π-almost surely, and there is a coupling π ∗ of the pair (Y ∗ , Y ) such that Y ∈ Jnα (Y ∗ ) π ∗ -almost surely. First step: We need to show that there exists a coupling π ˜ of the pair (U, Y ∗ ) such that U ∈ Γθ (Jnα (Y ∗ )) π ˜ -almost surely. This is achieved with an appeal to the gluing principle (see for instance Villani (2003), lemma 7.6 page 208). We know that the bootstrap variable Y ∗ can be written f (Y, η), where f is a measurable mapping and η is a random variable defined on the Polish space E and independent of Y (in other words, Y ∗ can be obtained from Y using an independent randomization device). Moreover, by theorem 13.1.1 page 487 of Dudley (2002), for each y ∈ Y, there exists a bijective mapping φy : [0, 1] 7→ Y such that φy and φ−1 y are measurable, and m ◦ φy = νy , where m is Lebesgue measure on [0, 1] (i.e. the uniform distribution) and νy is the conditional distribution of U given Y = y. Hence, if we take ζ to be a uniformly distributed random variable on [0, 1] independent of Y , we have for all set B, Pr(φY (ζ) ∈ B) = E [Pr(φy (ζ) ∈ B) | Y = y] = E [νy (B) | Y = y] = ν(B), where ν is the distribution of U . Call π ˜ the distribution of (U, Y ∗ ) = (φY (ζ), f (Y, η)), where (Y, η, ζ) is defined on the product space Y × E × [0, 1] with Y , η and ζ mutually independent. ˜ -almost This coupling is such that the marginal distributions are ν and Pn and U ∈ Γθ (Jnα (Y ∗ )) π surely as required.

28

Second Step: We now need to show that the existence of the coupling π ˜ such that U ∈ Γθ (Jnα ) π ˜ -almost surely, implies that Tnα (θ, Vθ ) = 0. The existence of π ˜ is equivalent to min

π∈M(Pn ,Vθ )

π{(u, v) : u ∈ / Γθ (Jnα (v))} = 0.

The left hand side of the expression above is larger than sup

n X

λ,h j=1

h(Yj ) subject to λ0 m(u) + h(v) ≤ 1{u ∈ / Γθ (Jnα (v))},

where λ ∈ Rdm , and h is a measurable function. Indeed, it suffices to apply π ˜ to the constraint R 1 Pn 0 0 to obtain the claimed inequality: π ˜ (λ m(u) + h(v)) = ν(λ m(u)) + n j=1 h(Yj ) ≤ 1{u ∈ / Γθ (Jnα (v))}d˜ π (u, v) = π ˜ {(u, v) : u ∈ / Γθ (Jnα (v))}. Now, since the quantity to maximize does not depend on u, the program is equivalent to   n X ¢ ¡ 1 hλ (Yj ) with hλ (y) = inf 1{u ∈ sup  / Γθ (Jnα (y))} − λ0 m(u) . u n λ j=1

Hence, the latter expression is smaller than zero. Since zero is achieved when λ = 0, the proof is complete. ¤ Lemma 2. Under assumptions 1 and 2, Vθ is uniformly tight.

Proof of Lemma 2 For M > 1, by assumptions 1, £ ¤ sup ν ({k m (U, θ) k> M }) ≤ sup ν k m(U, θ) k 1{km(U,θ)k>M } → 0 as M → ∞,

ν∈Vθ

ν∈Vθ

hence for ² > 0, there exists M > 0 such that 1 − ² ≤ sup ν ({k m (U, θ) k≤ M }) ν∈Vθ

but by assumption 2, there exists a compact set K such that {k m (U, θ) k≤ M } ⊂ K. ¤ Lemma 3. If Vθ is uniformly tight, then M (P, Vθ ) is uniformly tight.

Proof of Lemma 3 For ² > 0, there exists a compact KY ⊂ Y such that P (KY ) ≥ 1 − ²/2; by tightness of Vθ , there exists also a compact KU ⊂ U such that ν (KU ) ≥ 1−²/2 for all ν ∈ Vθ . For every π ∈ M (P, Vθ ), one has π (KY × KU ) ≥ max (P (KY ) + ν (KU ) − 1, 0) (Fr´echet-Hoeffding lower bound), thus π (KY × KU ) ≥ 1 − ². ¤

29

Proof of Theorem 2 Suppose inf π∈M(P,Vθ ) Eπ [1 {U ∈ / Γθ (Y )}] = 0, we shall show that the infimum is actually attained. Let πn ∈ M (P, Vθ ) a sequence of probability distributions of the joint couple (U, Y ) such that Eπn [1 {U ∈ / Γθ (Y )}] → 0. By Lemma 3, M (P, Vθ ) is uniformly tight, hence by Prohorov’s theorem it is relatively compact. Consequently there exists a subsequence πϕ(n) ∈ M (P, Vθ ) which is weakly convergent to π. One has π ∈ M (P, Vθ ). Indeed, clearly πY = P , and by assumption 2 the sequences of random ¡ ¢ variables m Uϕ(n) , θ are uniformly integrable, therefore by van der Vaart (1998), Theorem 2.20, £ ¡ ¢¤ one has πϕ(n) m Uϕ(n) , θ → π [m (U, θ)], thus π [m (U, θ)] = 0. Therefore, π ∈ M (P, Vθ ). By assumption 3, the set {U ∈ / Γθ (Y )} is open, hence by the Portmanteau lemma (van der Vaart (1998), Lemma 2.2 formulation (v)), lim inf πϕ(n) [{U ∈ / Γθ (Y )}] ≥ π [{U ∈ / Γθ (Y )}] thus π [{U ∈ / Γθ (Y )}] = 0. ¤ Proof of Theorem 3 This result is reproduced from Ekeland, Galichon, and Henry (2007). We need to show that the following two optimization problems (P) and (P ∗ ) have finite solutions, and that they are equal. (P) :

< P, f > subject to Lf ≤ δ − λ0 m

sup (f,λ)∈C 0 ×Rdu

and (P ∗ ) :

< π, δ > subject to L∗ π = P, π ≥ 0, < π, m >= 0.

sup (π,γ)∈M×Rdu

where C 0 is the space of continuous functions of y and u, equipped with the uniform topology, R its dual with respect to the scalar product < Q, f >= f dQ is the space M of signed (Radon) measures on Y ×U equipped with the vague topology (the weak topology with respect to this dual pair), L is the operator defined by L(f )(y, u) = f (y) for all u, and its dual L∗ is the projection of a measure π on Y, and the function δ is defined by δ(y, u) = 1{u ∈ / Γ(y)}. We now see that (P ∗ ) is the dual program of (P): indeed, we have sup

< P, f > subject to Lf ≤ δ − λ0 m

(f,λ)∈C 0 ×Rdu

= =

sup

inf

< P, f > + < π, δ − λ0 m − Lf >

inf

< P, f > + < π, δ > −λ0 < π, m > − < π, Lf >

(f,λ)∈C 0 ×Rdu π≥0, π∈M

sup (f,λ)∈C 0 ×Rdu

π≥0, π∈M

30

inf

π≥0, π∈M

< P, f > + < π, δ > −λ0 < π, m > − < L∗ π, f >

(f,λ)∈C 0 ×Rdu

sup

inf

< π, δ > −λ0 < π, m > + < P − L∗ π, f >,

=

sup

=

(f,λ)∈C 0 ×Rdu π≥0, π∈M

and inf

sup

π≥0, π∈M (f,λ)∈C 0 ×Rdu

=

inf

(π,γ)∈M×Rdu

< π, δ > −λ0 < π, m > + < P − L∗ π, f >

< π, δ > subject to < π, m >= 0, L∗ π = P, π ≥ 0.

We now proceed to prove that the strong duality holds, i.e. that the infimum and supremum can be switched. Under condition (4), by Proposition (2.3) page 52 of Ekeland and Temam (1976), (P) is stable. Hence, by Proposition (2.2) page 51 of Ekeland and Temam (1976), (P) is normal and (P ∗ ) has at least one solution. Finally, since f 7→< P, f > is linear, hence convex and lower semi-continuous, by Proposition (2.1) page 51 of Ekeland and Temam (1976), the two programs are equal and have a finite solution. ¤ Pilot example 1 continued We have λ = (λ1 , λ2 ) ∈ R2 and gλ,θ (x, 0) = min( inf {−λ0 m(ε, x)}; inf {1 − λ0 m(ε, x)}), ε≥−xβ

ε≤−xβ

gλ,θ (x, 1) = min( inf {−λ0 m(ε, x)}; inf {1 − λ0 m(ε, x)}). ε≤−xβ

ε≥−xβ

Hence, λ0 m(ε, x) = [1ε≤0 − η](λ1 (1 + x) + λ2 (1 − x)), thus λ0 m(ε, 1) = 2[1ε≤0 − η]λ1 , λ0 m(ε, −1) = 2[1ε≤0 − η]λ2 . So, for θ = −1, gλ,θ (1, 0) = min(inf {−2[1ε≤0 − η]λ1 }; inf {1 − 2[1ε≤0 − η]λ1 }) ε≥1

ε≤1

= 2ηλ1 + min(0, 1 − 2λ1 ), and similarly, gλ,θ (−1, 0) = 2ηλ2 + min(0, −2λ2 ), gλ,θ (1, 1) = 2ηλ1 + min(0, −2λ1 ), gλ,θ (−1, 1) = 2ηλ2 + min(1, −2λ2 ). We have EP [gλ,θ (Z)] = PX (−1)[PZ|X (0| − 1)gλ,θ (−1, 0) + PZ|X (1| − 1)gλ,θ (−1, 1)] + PX (1)[PZ|X (0|1)gλ,θ (1, 0) + PZ|X (1|1)gλ,θ (1, 1)].

31

Compute [PZ|X (0| − 1)gλ,θ (−1, 0) + PZ|X (1| − 1)gλ,θ (−1, 1)] = 2ηλ2 + PY |X (1| − 1) for λ2 < −1/2, = 2λ2 (η − Pr (1| − 1)) for λ2 ∈ [−1/2, 0], Z|X

= 2(η − 1)λ2 for λ2 > 0. Similarly, [PZ|X (0|1)gλ,θ (1, 0) + PZ|X (1|1)gλ,θ (1, 1)] = 2ηλ1 for λ1 < 0, = 2λ1 (η − Pr (1|1)) for λ1 ∈ [0, 1/2], Z|X

= 2(η − 1)λ1 + Pr (0|1) for λ1 > 1/2. Z|X

The expression attains its maximum for λ2 ∈ [−1/2, 0], the second one for λ1 ∈ [0, 1/2]. Therefore supλ∈R2 EP (hλ,θ=−1 (Z)) = 0 if and only if η ≥ PrZ|X (1| − 1) and η ≤ PrZ|X (1|1). Lemma 4. For all x1 , x2 , x∗1 , x∗2 ∈ R, if the graph of a bijection between (x1 , x2 ) and (x∗1 , x∗2 ) crosses, then max(|x1 − x∗1 |, |x2 − x∗2 |) ≤ max(|x2 − x∗1 |, |x1 − x∗2 |). Proof of Lemma 4 In the trapezium (x1 , x2 , x∗1 , x∗2 ), where the parallel sides are (x1 , x2 ) and (x∗1 , x∗2 ), it is easily seen that when the angle (x∗1 x1 , x1 x2 ) is larger than 90o , as on the left side of figure 9, then |x∗1 − x2 | ≥ |x∗1 − x1 |, and when the angle (x∗1 x1 , x1 x2 ) is smaller than 90o , as on the right side of figure 9, then |x∗1 − x2 | ≥ |x1 − x∗2 |. Hence |x1 − x∗1 | ≤ max(|x2 − x∗1 |, |x1 − x∗2 |). Likewise for |x2 − x∗2 |, which proves the result. ¤

x1

x2

x1 < 90o

> 90o

x∗1

x2

x∗2

x∗1

x∗2

Figure 9: Trapezium configurations.

Proof of Lemma 1 The case d = 1 is derived from (5.3.1) page 163 of Cs¨org˝ o and R´ev´esz (1981). The case d = 2 is a consequence of Leighton and Shor (1989), and the case d ≥ 3 is a consequence of Theorem 1.1 of Shor and Yukich (1991). See also Talagrand (1994). ¤

32

References Andrews, D. (2000): “Inconsistency of the bootstrap when a parameter is on the boundary of the parameter space,” Econometrica, 68, 399–405. Andrews, D., S. Berry, and P. Jia (2003): “Placing bounds on parameters of entry games in the presence of multiple equilibria,” unpublished manuscript. Andrews, D., and P. Guggenberger (2007): “Validity of subsampling and plugin asymptotic inference for parameters defined by moment inequalitites,” unpublished manuscript. Andrews, D., and G. Soares (2007): “Inference for parameters defined by moment inequalitites using generalized moment selection,” unpublished manuscript. Beresteanu, A., and F. Molinari (2006): “Asymptotic properties for a class of partially identified models,” Cemmap Working Papers, CWP10/06. Bierens, H. (1990): “A Consistent Conditional Moment test for Functional Form,” Econometrica, 58, 1443–1458. Blundell, R., M. Browning, and I. Crawford (2003): “Nonparametric engel curves and revealed preference,” Econometrica, 71, 205–240. Brock, B., and S. Durlauf (2002): “A Multinomial choice model of neighborhood effects,” American Economic Review, 92, 298–303. Bugni, F. (2007): “Bootstrap methods for partially identified models,” unpublished manuscript. Chen, X., H. Hong, and E. Tamer (2005): “Measurement error models with auxiliary data,” Review of Economic Studies, 22, 343–366. Chernozhukov, V., H. Hong, and E. Tamer (2007): “Estimation and Confidence Regions for Parameter Sets in Econometric Models,” Econometrica, 75, 1243–1285. ¨ rgo ˝ , M., and P. Re ´ ve ´sz (1981): Strong approximations in probability and statistics. Cso Academic Press. Dudley, R. (2002): Real Analysis and Probability. Cambridge University Press. Ekeland, I., A. Galichon, and M. Henry (2007): “Optimal transportation and the falsifiability of incompletely specified economic models. Part II: moment restrictions on latent variables,” unpublished manuscript. 33

Ekeland, I., and R. Temam (1976): Convex Anaysis and Variational Problems. North Holland Elsevier. Galichon, A., and M. Henry (2006): “Inference in Incomplete Models,” Columbia University Discussion Paper 0506-28. Imbens, G., and C. Manski (2004): “Confidence Intervals for Partially Identified Parameters,” Econometrica, 72, 1845–1859. Jovanovic, B. (1989): “Observable implications of models with multiple equilibria,” Econometrica, 57, 1431–1437. Khan, S., and E. Tamer (2006): “Inference on Randomly Censored Regression Models using Conditional Moment Inequalities,” unpublished manuscript. Leighton, and P. Shor (1989): “Tight bounds for minimax grid matching with applications to the average case analysis of algorithms,” Combinatorica, 9, 161–187. Magnac, T., and E. Maurin (2005): “Partial identification in monotone binary models: discrete regressors and interval data,” unpublished manuscript. Manski, C. (1990): “Nonparametric bounds on treatment effects,” American Economic Review, 80, 319–323. Manski, C. (2004): “Social learning from private experiences: the dynamics of the selection problem,” Review of Economic Studies, 71, 443–458. Molchanov, I. (2005): Theory of Random Sets. London: Springer-Verlag. Pakes, A., J. Porter, K. Ho, and J. Ishii (2004): “Moment Inequalities and Their Application,” unpublished manuscript. Papadimitriou, C., and K. Steiglitz (1982): Combinatorial Optimization: Algorithms and Complexity. Prentice Hall. Rockafellar, R. T., and R. J.-B. Wets (1998): Variational Analysis. Berlin: Springer. Romano, J., and A. Shaikh (2005): “Inference for a Class of Partially Identified Econometric Models,” unpublished manuscript. Rosen, A. (2006): “Confidence sets for partially identified parameters that satisfy a finite number of moment inequalities,” unpublished manuscript. 34

Santos, A. (2007): “Inference in nonparametric instrumental variables with partial identification,” unpublished manuscript. Shaikh, A., and E. Vytlacil (2005): “Threshhold crossing models and bounds on treatment effects: a nonparametric analysis,” NBER Technical Working Paper 0307. Shor, P., and J. Yukich (1991): “Minimax Grid Matching and Empirical measures,” Annals of Probability, 19, 1338–1348. Talagrand, M. (1994): “The transportation cost from the Uniform measure to the empirical measure in dimension greater-than-or-equal-to-3,” Annals of Probability, 22, 919–959. Tamer, E. (2003): “Incomplete Simultaneous Discrete Response Model with Multiple Equilibria,” Review of Economic Studies, 70, 147–165. van der Vaart, A. (1998): Asymptotic Statistics. Cambridge University Press. Villani, C. (2003): Topics in Optimal Transportation. Providence: American Mathematical Society.

35

Dilation Bootstrap: a method for constructing ...

Keywords: Partial identification, incomplete specification test, duality, dilation ...... 15 minutes for n = 1000 on an AMD Opteron (tm) Processor 250 with 4G of ...

381KB Sizes 1 Downloads 243 Views

Recommend Documents

Dilation Invest#1rev.pdf
There was a problem previewing this document. Retrying... Download. Connect more apps... Try one of the apps below to open or edit this item. Dilation ...

Simulation of Markovian models using Bootstrap method
Business. ▻ Computer science and telecommunication ... Queueing Networks (QN) [Little61, Basket et al. ... Stochastic Automata Networks (SAN) [Plateau84].

Simulation of Markovian models using Bootstrap method
equally important role when examining complex systems. There are ... Although numerical solution produces reliable results, it is .... transition matrix storage [9].

A Practical Algorithm for Constructing Oblivious Routing ...
computing, as e.g. multicast routing and data management problems. Furthermore .... on level l ≥ 1 if both endpoints of e are contained in the same level l − 1 ...

A New Method of Constructing Sun-Dials (1767).pdf
ines cut it; and they fhall be the true hour-lines for a horizontal dial; to which fet the hours, as in. fig. 2. Laftly, in fig. 2. draw ^ji? parallel to ACE. in fio". I. and ^jy fliall be the axis or edge of the. ftile cdy that calls the fliadow on

A Simple Randomized Scheme for Constructing Low ...
†Department of Computer Science, Virginia Polytechnic Institute and State University, .... sensor network applications [27]), our scheme constructs low-degree ...

A Duality Transform for Constructing Small Grid ...
May 11, 2014 - We call a CDV Matrix positive if Mij > 0 for all (i, j) ∈ E(G). We call ..... Figure 1: A 2d embedding of G↑ (left), the cone-convex embedding of G (center), and the ..... On the complexity of optimization problems for 3-dimensiona

An algorithmic model for constructing a linkage and ...
1 Department of Statistics, University of Florida, Gainesville, FL 32611, USA. 2 Department of ...... crease of interest in molecular marker technologies and their ...

Bonus play method for a gambling device
Mar 14, 2006 - See application ?le for complete search history. (Us). (56) ... Play and/0r apply ..... As shoWn, there are numerous variations on the theme that.

Constructing Reliable Distributed Communication ... - CiteSeerX
bixTalk, and IBM's MQSeries. The OMG has recently stan- dardized an Event Channel service specification to be used in conjunction with CORBA applications.

constructing connections
CONSTRUCTING CONNECTIONS: MUSEOLOGICAL THEORY AND BLOGGING ... with Web 2.0 include: blogging, wikis, podcasts, tagging, videoblogs, online social .... school age children and increasingly for adults and exhibit making.

Method for processing dross
Nov 20, 1980 - dross than is recovered using prior art cleaning and recovery processes. ..... 7 is an illustration of the cutting edge ofa knife associated with the ...

Method for processing dross
Nov 20, 1980 - able Products from Aluminum Dross", Bur. of Mines. Report of .... the salt bath must be heated at a temperature substan tially above its melting ...

A Novel Method for Travel-Time Measurement for ...
simulation results obtained through use of a simulation program developed by the ... input data is taken from first-arrival travel-time measurements. The .... Data Recovery: ... beginning at 7 msec, at z=0, the free surface, corresponds to a wave.

Method for producing a device for direct thermoelectric energy ...
Sep 5, 2002 - Thus an element With a high atomic mass, i.e. a heavy element, ought to be .... band gap, namely about 0.6 electron volt, is adequate for.

A Novel Method for Travel-Time Measurement for ...
simulation results obtained through use of a simulation program developed by the authors. ... In contemporary modern wireless communications systems.

Method for producing a device for direct thermoelectric energy ...
Sep 5, 2002 - speci?cally, hoW to best use and apply it for the direct conversion of .... carrier concentration of the order of 1018 carriers per cm3 are considered ..... radiation, nuclear element or cell, combustion of fossil fuels,. Waste heat ...

A novel method for measuring semantic similarity for XML schema ...
Enterprises integration has recently gained great attentions, as never before. The paper deals with an essential activity enabling seam- less enterprises integration, that is, a similarity-based schema matching. To this end, we present a supervised a

Method for processing dross
Nov 20, 1980 - the free metal entrained in dross or skimmings obtained from the production of aluminum or aluminum based alloys. In the course of conventional aluminum melting op ..... 7 is an illustration of the cutting edge ofa knife.

A Method for Distributed Optimization for Task Allocation
the graph that underlies the network of information exchange. A case study involving ... firefighting, disaster management, and multi-robot cooperation. [1-3].

Constructing Connections: A Museological Approach to ...
blogs usually include the hosting service name in the web address, such as generic- ... 10. For the purposes of this thesis, conversation is not necessarily oral; the written ..... their best advantage by museums, have the potential to help fulfill t

Collaborative Inquiry in Co-constructing a Better ...
In primary schools, writing is a big problem. Children do not like to write and don‟t write well. ..... Adult and Continuing Education, No. 94, Jossey-Bass. Zhang ...