Submitted to the Annals of Statistics

SUPPLEMENT TO “ESTIMATION IN FUNCTIONAL LINEAR QUANTILE REGRESSION”∗ By Kengo Kato Hiroshima University This supplementary file contains the additional discussion on the connection to nonlinear ill-posed inverse problems, technical proofs omitted in the main body, some useful technical tools and additional simulation results. In this supplementary file, we follow the notation, the numbering and the convention used in the main body. In the technical proofs, we use C > 0 to indicate a generic constant of which the value may change from line to line. APPENDIX A: CONNECTION TO NONLINEAR ILL-POSED INVERSE PROBLEMS This section discusses the connection of our problem of estimating the slope function to nonlinear ill-posed inverse problems. For any fixed u ∈ U, our estimator ˆb(·, u) can be understood as a regularized solution to an empirical version of a nonlinear inverse problem that corresponds to the “normal equation” : (A.1)

A(u, b(·, u)) = 0,

where the map A : U × L2 [0, 1] → L2 [0, 1] is defined by ∫1 A(u, g)(·) = E[{u − 1(Y ≤ 0 g(t)X c (t)dt)}X c (·)] ∫1 = E[{u − FY |X ( 0 g(t)X c (t)dt | X)}X c (·)], u ∈ U, g ∈ L2 [0, 1]. Here FY |X (y | X) is the conditional distribution function of Y given X. For the sake of simplicity, we have ignored the constant term. Observe that for any fixed u ∈ U, the map A(u, ·) : L2 [0, 1] →∑L2 [0, 1] is a nonlinear m ˆ ˆ ˆc operator. In fact, using an approximation Xic ≈ j=1 ξij ϕj =: Xi , our estimator ˆb(·, u) is an approximate solution to an empirical version of (A.1) over the linear subspace spanned by {ϕˆ1 , . . . , ϕˆm }: (A.2) ∗

ˆ ˆb(·, u)) ≈ 0, A(u,

Supported by the Grant-in-Aid for Young Scientists (B) (22730179) from the JSPS.

1 imsart-aos ver. 2013/03/06 file: FunctionQR-supplement-v4.tex date: July 12, 2016

2

K. KATO

where the map Aˆ : U × L2 [0, 1] → L2 [0, 1] is defined by ∫ ˆ g)(·) = En [{u − 1(Yi ≤ 1 g(t)X ˆ ic (t)dt)}X ˆ ic (·)], u ∈ U, g ∈ L2 [0, 1]. A(u, 0 To see (A.2), observe that ˆ ˆb(·, u))(·) = ∑m En [{u − 1(Yi ≤ ∑m ξˆik ˆbk (u))}ξˆij ]ϕˆj (·). A(u, j=1 k=1 The first order condition to (2.4) (in the main body) implies that ∑ ˆ ˆ ˆ En [{u − 1(Yi ≤ m k=1 ξik bk (u))}ξij ] ≈ 0, 1 ≤ j ≤ m, which leads to (A.2) [the discussion here is informal to give an intuition behind our estimator]. Note that solving (2.4) (in the main body) is computationally more appealing than directly searching a solution to (A.2) as the former problem is convex while the latter is not. Meanwhile, as long as the map y 7→ FY |X (y | x) is continuous, for any fixed u ∈ U, the nonlinear inverse problem (A.1) is locally ill-posed at b(·, u) in the sense of Hofmann and Scherzer (1998, Definition 1.1), i.e., there exists a sequence of functions {gn } in a neighborhood of b(·, u) (in L2 [0, 1]) such that A(u, gn ) → A(u, b(·, u)) but gn ↛ b(·, u) in the L2 -norm. To see this, take a sequence of functions {gn } in a neighborhood of b(·, u) such that w w gn → b(·, u) but gn ↛ b(·, u) in the L2 -norm, where → means the weak convergence in L2 [0, 1]. Then, by the weak convergence, we have ∫ 1 ∫ 1 c b(t, u)X c (t)dt. gn (t)X (t)dt → (A.3) 0

0

By the continuity of the map y 7→ FY |X (y | X), (A.3) implies that A(u, gn ) → A(u, b(·, u)) despite gn ↛ b(·, u). This suggests that any sensible estimation procedure based on the normal equation (A.1) has to involve some regularizations. In our case, the regularization is done by restricting the parameter space for b(·, u) to a sequence of finite dimensional subspaces, where the cut-off level m plays a role of regularization parameter. APPENDIX B: PROOF OF THEOREM 3.2 Let Xn+1 be a copy of X independent of the data Dn := {(Y1 , X1 ), . . . , (Yn , Xn )}. Then ˆ Y |X , u) = E[{Q ˆ Y |X (u | Xn+1 ) − QY |X (u | Xn+1 )}2 | Dn ]. E(Q imsart-aos ver. 2013/03/06 file: FunctionQR-supplement-v4.tex date: July 12, 2016

3

FUNCTIONAL QUANTILE REGRESSION c Let Xn+1 = Xn+1 − E[X(t)] =

∑∞

ˆ Y |X (u | Xn+1 ) = a Q ˆ(u) + ∫

c Xn+1 (t) 0

+

m ∑

Observe that

ˆbj (u)ξn+1,j

j=1 1

+ ∫

j=1 ξn+1,j ϕj .

m ∑

ˆbj (u)(ϕˆj − ϕj )(t)dt

j=1 1

¯ˆ ˆb(t, u)dt (E[X(t)] − X(t))

0

and QY |X (u | Xn+1 ) = a(u) +

m ∑ j=1

−1/2

Letting ηn+1,j = κj

∞ ∑

bj (u)ξn+1,j +

bj (u)ξn+1,j .

j=m+1

ξn+1,j , we have

ˆ Y |X (u | Xn+1 ) − QY |X (u | Xn+1 )}2 {Q 2 2   [ ∞ m   ∑  ∑ bj (u)ξn+1,j + (dˆj − dj )(u)ηn+1,j ≤ C (ˆ a − a)2 (u) +     j=m+1 j=1 2  ∫ 1 ∫ 1 ∑ m  c ˆbj (u)(ϕˆj − ϕj )(t) dt + Xn+1 (t)2 dt  0 0  j=1 {∫ 1 }2 ] ¯ ˆb(t, u)dt ˆ . + (E[X(t)] − X(t)) 0

Taking expectation with respect to Xn+1 , we have ˆ Y |X (u | Xn+1 ) − QY |X (u | Xn+1 )}2 | Dn ] E[{Q  2 [ ∫ 1 ∑ ∞ m  ∑ ˆbj (u)(ϕˆj − ϕj )(t) dt ≤ C ∥dˆm (u) − dm (u)∥2ℓ2 + κj b2j (u) +  0  j=1 j=m+1 {∫ 1 }2 ] ¯ˆ + (E[X(t)] − X(t))ˆb(t, u)dt . 0

imsart-aos ver. 2013/03/06 file: FunctionQR-supplement-v4.tex date: July 12, 2016

4

K. KATO

By the proof of Theorem 3.1, we have sup ∥dˆm (u) − dm (u)∥2ℓ2 = OP (m/n) = OP (n−(α+2β−1)/(α+2β) ), and

u∈U ∞ ∑

κj b2j (u) ≤ C

j=m+1

∞ ∑

j −α−2β = O(m−α−2β+1 ) = O(n−(α+2β−1)/(α+2β) ).

j=m+1

Observe that 2  m  ∑ ˆbj (u)(ϕˆj − ϕj )   j=1  2  2 m m  ∑  ∑ ≤2 bj (u)(ϕˆj − ϕj ) +2 (ˆbj − bj )(u)(ϕˆj − ϕj )     j=1 j=1  2  2 m m ∑  ∑  −1/2 ˆ +2 κj (dj − dj )(u)(ϕˆj − ϕj ) =2 bj (u)(ϕˆj − ϕj )     j=1

j=1

≤ 2m

m ∑

b2j (u)(ϕˆj − ϕj )2 + 2∥dˆm (u) − dm (u)∥2ℓ2

j=1

m ∑

2 ˆ κ−1 j ( ϕj − ϕj ) ,

j=1

by which we have 2  ∫ 1 ∑ m  ˆbj (u)(ϕˆj − ϕj )(t) dt  0  j=1

≤ 2m

m ∑

b2j (u)∥ϕˆj − ϕj ∥2 + 2∥dˆm (u) − dm (u)∥2ℓ2

j=1

m ∑

2 ˆ κ−1 j ∥ ϕj − ϕj ∥ .

j=1

By the proof of Theorem 3.1, we see that m

m ∑

b2j (u)∥ϕˆj − ϕj ∥2 ≤ Cm

j=1

m ∑

j −2β ∥ϕˆj − ϕj ∥2 = OP (mn−1 )

j=1

= OP (n−(α+2β−1)/(α+2β) ), ∑ −1 ˆ 2 while by the proof of (5.4) in Appendix C, we have m j=1 κj ∥ϕj − ϕj ∥ = oP (1). Hence we conclude that  2 ∫ 1 ∑ m  ˆbj (u)(ϕˆj − ϕj )(t) dt = OP (n−(α+2β−1)/(α+2β) ). sup  u∈U 0  j=1

imsart-aos ver. 2013/03/06 file: FunctionQR-supplement-v4.tex date: July 12, 2016

5

FUNCTIONAL QUANTILE REGRESSION

Finally, using assumption (A8), we have {∫ 1 }2 ∫ 1 ∫ ¯ˆ ¯ˆ 2 ˆ (E[X(t)] − X(t))b(t, u)dt ≤ (E[X(t)] − X(t)) dt 0

0

1

ˆb2 (t, u)dt

0

= OP (n−1 + ∆γ ) × OP (1) = OP (n−1 ), uniformly in u ∈ U. Taking these together, we conclude that ˆ Y |X (u | Xn+1 ) − QY |X (u | Xn+1 )}2 | Dn ] = OP (n−(α+2β−1)/(α+2β) ). sup E[{Q

u∈U

This completes the proof. APPENDIX C: PROOF OF PROPOSITION 3.1 We here provide a proof for (3.4). Consider the same construction as in Hall and Horowitz (2007). Let ϕ1 (t) ≡ 1 and ϕj+1 (t) = 21/2 cos(jπt) for j ≥ 1. Put ϱj = θj j −β for [n1/(α+2β) ] + 1 ≤ j ≤ 2[n1/(α+2β) ] and ϱj = 0 otherwise where [y] denotes the integer part of y ∈ R and θj is either 0 or ∑ each−α/2 1. Let Z1 , Z2 , · · · ∼ U [−31/2 , 31/2 ] i.i.d. Take X(t) = ∞ j Zj ϕj (t) and j=1 ∑2[n1/(α+2β) ] ϱ(t) = j=[n1/(α+2β) ]+1 ϱj ϕj (t). Note that the former sum is almost surely uniformly convergent in t ∈ [0, 1] and hence X has sample paths almost surely continuous (see for example Marcus, 1975, Theorem 1.1). Consider a sequence of data generating processes ∫ Y =

2[n1/(α+2β) ]

1



ϱ(t)X(t)dt+ϵ = 0

θj j −(α+2β)/2 Zj +ϵ, ϵ ∼ N (0, 1), ϵ ⊥ ⊥ X.

j=[n1/(α+2β) ]+1

Then we have QY |X (u | X) = a(u) +

∫ b(t, u)X(t)dt,

where a(u) = Φ−1 (u), b(t, u) ≡ ϱ(t), fY |X (y|X) = ϕ(y −

∫1 0

ϱ(t)X(t)dt),

by which one sees that assumptions (A4)-(A6) are satisfied. Here ϕ(·) and Φ(·) are the density and the distribution function of the standard normal distribution, respectively. Suppose that α ≤ 3. Then, since for any 0 < γ < α − 1, the function t 7→ cos(t) is γ/2-H¨older continuous (by the periodicity of the cosine function), we have E[(X(s) − X(t))2 ] ≤ C|s − t|γ

∞ ∑

j −α+γ ≤ C ′ |s − t|γ , ∀s, t ∈ [0, 1],

j=1

imsart-aos ver. 2013/03/06 file: FunctionQR-supplement-v4.tex date: July 12, 2016

6

K. KATO

where C and C ′ are some constants. This shows that assumption (A8) is satisfied with 0 < γ < α − 1 when α ≤ 3. For α > 3, K(s, t) is twice continuously differentiable, so that ∫ 1 assumption (A8) is satisfied with 0 < γ ≤ 2. Finally, since E[Y | X] = 0 ϱ(t)X(t)dt, by Hall and Horowitz (2007), for any estimator (t, u) 7→ ¯b(t, u), sup∗



1

sup u∈U

≥ sup∗ ≥ Dn



0 1

E[(¯b(t, u) − b(t, u))2 ]dt

E[(¯b(t, u0 ) − ϱ(t))2 ]dt

0 −(2β−1)/(α+2β)

(u0 is any point in U )

,

] different distributions where sup∗ denotes the supremum over all 2[n of (Y, X) obtained by taking different choices of θ[n1/(α+2β) ]+1 , . . . , θ2[n1/(α+2β) ] , and D > 0 is a constant. The “in-probability” version of the lower bound (3.4) follows from the same reasoning as in the proof of Yuan and Cai (2010, p. 3442) and the Paley-Zygmund inequality (which states that for any nonnegative random variable Z with E[Z 2 ] < ∞, P(Z > λE[Z]) ≥ (1 − λ)2 (E[Z])2 /(E[Z 2 ]) for all λ ∈ (0, 1)). The other assertions follow similarly. This completes the proof. 1/(α+2β)

APPENDIX D: PROOFS OF (5.2), (5.4), (5.6), (5.7) AND (5.9) In this section, we provide proofs of (5.2), (5.4), (5.6), (5.7) and (5.9) omitted in Section 5. Throughout the section, we assume all the conditions of Theorem 3.1. Define the (infeasible) empirical covariance kernel ˆ ∗ (s, t) = En [(Xi (s) − X(s))(X ¯ ¯ K i (t) − X(t))], ∑ ∗ ˆ∗ ˆ∗ ˆ ∗ (s, t) = ∑∞ κ ¯ where X(t) = n−1 ni=1 Xi (t). Let K j=1 ˆ j ϕj (s)ϕj (t) be the ˆ ∗ (s, t) where κ spectral expansion of K ˆ ∗1 ≥ κ ˆ ∗2 ≥ · · · ≥ 0 and {ϕˆ∗j }∞ j=1 is an orthonormal basis of L2 [0, 1]. Without loss of generality, we may assume that ∫ ∫ ∗ ˆ ϕj ϕj ≥ 0, ϕˆj ϕˆ∗j ≥ 0, ∀j ≥ 1, ∫1 ∫ where, to ease the notation, 0 f (t)dt is abbreviated as f for any function f : [0, 1] → R. Define ∫ ¯ ϕˆ∗ , ηˆ∗ = κ−1/2 ξˆ∗ , i = 1, . . . , n; j ≥ 1. ξˆ∗ = (Xi − X) ij

j

ij

j

ij

imsart-aos ver. 2013/03/06 file: FunctionQR-supplement-v4.tex date: July 12, 2016

FUNCTIONAL QUANTILE REGRESSION

7

∫ −1/2 Recall that ηij = κj ξij = κ−1/2 Xic ϕj (where Xic (t) = Xi (t) − E[Xi (t)]) −1/2 ˆ −1/2 ∫ ˆ ¯ˆ ˆ and ηˆij = κ ξij = κ (Xi − X) ϕj for i = 1, . . . , n and j ≥ 1. We will j

j

frequently use the following decomposition: for i = 1, . . . , n and j ≥ 1, ∗ ∗ ηˆij − ηij = ηˆij − ηˆij + ηˆij − ηij , ∫ ∫ −1/2 −1/2 ∗ ˆ ˆ ηˆij − ηˆij = κj (Xi − Xi )ϕj + κj Xi (ϕˆj − ϕˆ∗j ) ∫ ∫ −1/2 −1/2 ¯ˆ ˆ ¯ ¯ ϕˆj − ϕˆ∗j ) − κj (X − X)ϕj − κj X(

ˆ ij1 + ∆ ˆ ij2 − ∆ ˆ j3 − ∆ ˆ j4 , =: ∆ ∫ ∫ ∫ −1/2 −1/2 ∗ ¯ c ϕj − κ−1/2 X ¯ c (ϕˆ∗j − ϕj ) ηˆij − ηij = κj Xic (ϕˆ∗j − ϕj ) − κj X j ˆ ij5 − ∆ ˆ j6 − ∆ ˆ j7 , =: ∆ ∑ ¯ c (t) = n−1 n X c (t). where X i=1 i We prepare some lemmas. For any function : [0, 1]2 → R, define |||R||| = ∫∫ 2 ∫1 R 1/2 2 2 ( R (s, t)dsdt) . Recall that ∥f ∥ = 0 f (t)dt for any f : [0, 1] → R. We have

Lemma D.1.

ˆ i − Xi ∥2 ] = OP (∆γ ), |||K ˆ −K ˆ ∗ |||2 = OP (∆γ ). En [∥X Furthermore, as n → ∞, with probability approaching one, ˆ −K ˆ ∗ |||, 1 ≤ ∀j ≤ m. ∥ϕˆj − ϕˆ∗j ∥ ≤ Cj α+1 |||K Proof. Observe that ˆ i (t) − Xi (t) = X

Li ∑

(Xi (til ) − Xi (t))1(t ∈ [til , ti,l+1 )), t ∈ [0, 1),

l=1

by which we have ˆ i − Xi ∥ 2 = ∥X

Li ∫ ∑ l=1

ti,l+1

(Xi (til ) − Xi (t))2 dt.

til

Taking expectation, we have ˆ i − Xi ∥ 2 ] = E[∥X

Li ∫ ∑

ti,l+1

E[(X(t) − X(til ))2 ]dt

l=1 til Li ∑

≤C

(ti,l+1 − til )γ+1 ≤ C∆γ ,

l=1

imsart-aos ver. 2013/03/06 file: FunctionQR-supplement-v4.tex date: July 12, 2016

8

K. KATO

where we have used assumption (A8). This leads to the first assertion. The second assertion follows from the Schwarz inequality and the first assertion. The third assertion needs some effort. By Bosq (2000, Lemmas 4.2 and 4.3; see also the remark below), we have (D.1)

ˆ ∗ − K|||, sup χ ˆ −K ˆ ∗ |||, sup |ˆ κ∗j − κj | ≤ |||K ˆj ∥ϕˆj − ϕˆ∗j ∥ ≤ 81/2 |||K j≥1

j≥1

ˆ ∗2 . For some where χ ˆj = min{ˆ κ∗j−1 − κ ˆ ∗j , κ ˆ ∗j − κ ˆ ∗j+1 } for j ≥ 2 and χ ˆ1 = κ ˆ ∗1 − κ small constant c > 0, define the event En = { χ ˆj ≥ cj −α−1 , 1 ≤ ∀j ≤ m}. It suffices to show that P(En ) → 1. By the first inequality in (D.1), ˆ ∗ − K||| ≥ C −1 k −α−1 − 2|||K ˆ ∗ − K|||. κ ˆ ∗k − κ ˆ ∗k+1 ≥ κk − κk+1 − 2|||K ˆ ∗ − K||| = OP (n−1/2 ) (which Since k −α−1 ≥ m−α−1 ≍ n−(α+1)/(α+2β) , |||K follows by a simple calculation), and n−1/2 = o(n−(α+1)/(α+2β) ) (which follows by β > α/2 + 1), we have uniformly in 1 ≤ k ≤ m, κ ˆ ∗k − κ ˆ ∗k+1 ≥ C −1 (1 − oP (1))k −α−1 , which leads to that P(En ) → 1 by taking c sufficiently small. Remark D.1. Lemma 4.3 of Bosq (2000) reads as follows: for functions 2 Q, R : [0, 1] in L2 [0, 1]2 of the form ∑∞ ∑∞→ R having the spectral expansions Q(s, t) = j=1 λj ψj (s)ψj (t) and R(s, t) = j=1 νj φj (s)φj (t), where λ1 ≥ ∞ λ2 ≥ · · · ≥ 0, ν1 ≥ ν2 ≥ · · · ≥ 0, and {ψj }∞ j=1 and {φj }j=1 are orthonormal bases for L2 [0, 1], we have: χj ∥φj − ψj ∥ ≤ 81/2 |||R − Q||| for all j ≥ 1 such that χj > 0, where χj = min{λ ∫ j−1 −λj , λj −λj+1 } for j ≥ 2 and χ1 = λ1 −λ2 . Here, we have assumed that φj ψj ≥ 0 for all j ≥ 1. This lemma actually holds with supj≥1 χj ∥φj − ψj ∥ ≤ 81/2 |||R − Q||| since the inequality trivially holds in case of χj = 0. The following useful result was established in Hall and Horowitz (2007). Lemma D.2.

As n → ∞, with probability approaching one,

∥ϕˆ∗j − ϕj ∥2 ≤ 10



−2

(κj − κk )

{∫ ∫

ˆ ∗ − K)(s, t)ϕj (s)ϕk (t)dsdt (K

}2 ,

k:k̸=j

imsart-aos ver. 2013/03/06 file: FunctionQR-supplement-v4.tex date: July 12, 2016

FUNCTIONAL QUANTILE REGRESSION

9

for all 1 ≤ j ≤ m. Furthermore, we have [{∫ ∫ }2 ] ∑ ˆ ∗ − K)(s, t)ϕj (s)ϕk (t)dsdt (κj − κk )−2 E (K = O(j 2 n−1 ), k:k̸=j

uniformly in 1 ≤ j ≤ m. Proof. See Hall and Horowitz (2007, p.83-84). D.1. Proofs of (5.2) and (5.4). We first prove (5.4). By Lemmas D.1 and D.2, we have   m n m ∑  1 ∑∑ ˆ2 α+1 γ ˆ i − Xi ∥ 2 ] κ−1 ∆ ) = oP (1), ∆ij1 ≤ En [∥X j  = OP (m  n i=1 j=1 j=1   n m m ∑  1 ∑∑ ˆ2 ˆj − ϕˆ∗ ∥2 ∆ij2 ≤ En [∥Xi ∥2 ] κ−1 ∥ ϕ j j   n i=1 j=1 j=1 ∑ 3α+2 = OP (1) × OP (∆γ m ) = OP (m3α+3 ∆γ ) = oP (1), j=1 j   m m n ∑   ∑ ∑ 1 −1 ˆ∗ 2 c 2 2 ˆ κj ∥ϕj − ϕj ∥ ∆ij5 ≤ En [∥Xi ∥ ]   n j=1 i=1 j=1 ∑ α+2 = OP (1) × OP (n−1 m ) = OP (mα+3 n−1 ) = oP (1). j=1 j Similarly, we have m ∑

ˆ 2j3 = OP (mα+1 ∆γ ) = oP (1), ∆

j=1 m ∑ j=1 m ∑

ˆ 2j4 = OP (m3α+3 ∆γ ) = oP (1), ∆ ˆ 2j7 = OP (mα+3 n−2 ) = oP (1). ∆

j=1

∑ Using the decomposition Xic (t) = ∞ k=1 ξik ϕk (t), we have ∫ ˆ j6 = κ−1/2 X ¯ c ϕj = κ−1/2 ξ¯j = η¯j , ∆ j j by which we have m ∑

∑m ˆ 2 = OP (E[∑m ∆ ˆ2 ∆ ηj2 ]) = OP (mn−1 ) = oP (1). j6 j=1 j6 ]) = OP ( j=1 E[¯

j=1

imsart-aos ver. 2013/03/06 file: FunctionQR-supplement-v4.tex date: July 12, 2016

10

K. KATO

Finally, by a direct calculation, we have En [∥ηim ∥2 ] = OP (m). Taking these together, we obtain (5.4). We now turn to prove (5.2). Observe that max

1≤i≤n

max

1≤i≤n

max

1≤i≤n

Since



m ∑ j=1 m ∑ j=1 m ∑

ˆ 2ij1 ≤ max ∥X ˆ i − Xi ∥2 × OP (mα+1 ), ∆ 1≤i≤n

ˆ 2 ≤ max ∥Xi ∥2 × OP (m3α+3 ∆γ ), ∆ ij2 1≤i≤n

ˆ 2 ≤ max ∥X c ∥2 × OP (mα+3 n−1 ). ∆ ij5 i 1≤i≤n

j=1

E[X 4 ] ≤ C, we have max ∥Xic ∥2 = OP (n1/2 ),

1≤i≤n

which leads to that max

1≤i≤n

m ∑

ˆ 2 = OP (n1/2 m3α+3 ∆γ ), max ∆ ij2

1≤i≤n

j=1

m ∑

ˆ 2 = OP (mα+3 n−1/2 ). ∆ ij5

j=1

ˆ i − Xi ∥2 ≤ ∑n ∥X ˆ i − Xi ∥2 , we also Using the trivial bound max1≤i≤n ∥X i=1 have m ∑ ˆ 2ij1 = OP (nmα+1 ∆γ ). ∆ max 1≤i≤n

j=1

4 ] = κ−2 E[ξ 4 ] ≤ C for all j ≥ 1 by assumption (A2), we Similarly, since E[ηij ij j have m ∑ 2 max ηij = OP (mn1/2 ). 1≤i≤n

j=0

Taking these together, we have max ∥ˆ ηim ∥2ℓ2 = OP (nmα+1 ∆γ + n1/2 m3α+3 ∆γ + mα+3 n−1/2 + mn1/2 ).

1≤i≤n

Since α > 1, β > α/2 + 1 and m3α+3 ∆γ → 0, there exists a small constant c > 0 (depending on α and β) such that the right side is OP (n−c (n/m)). This implies (5.2), completing the proof.

imsart-aos ver. 2013/03/06 file: FunctionQR-supplement-v4.tex date: July 12, 2016

11

FUNCTIONAL QUANTILE REGRESSION

D.2. Proofs of (5.6) and (5.7). We first prove (5.6). Observe that (hm · ηˆim )2 − (hm · ηim )2 = {hm · (ˆ ηim − ηim )}2 + 2(hm · ηim ){hm · (ˆ ηim − ηim )}, by which we have for all hm ∈ Sm , |En [(hm · ηˆim )2 ] − En [(hm · ηim )2 ]| ≤ En [∥ˆ ηim − ηim ∥2ℓ2 ] + 2(En [(hm · ηim )2 ])1/2 (En [∥ˆ ηim − ηim ∥2ℓ2 ])1/2 . By the proof of (5.4), we have En [∥ˆ ηim − ηim ∥2ℓ2 ] = oP (1). While by Rudelson’s inequality (Theorem E.1 in Appendix E), we have √ [ ] log n m m 2 E sup |En [(h · ηi ) ] − 1| ≤ C E[ max ∥ηim ∥2ℓ2 ], 1≤i≤n m m n h ∈S 4 ] = κ−2 E[ξ 4 ] ≤ C provided that the right side is smaller than 1. Since E[ηij ij j for all j ≥ 1 by assumption (A2), by Lemma E.1 in Appendix E, we have

E[ max ∥ηim ∥2ℓ2 ] = O(mn1/2 ). 1≤i≤n

Therefore, we conclude that sup |En [(hm · ηim )2 ] − 1| = OP (n−1/4 m1/2 (log n)1/2 ) = oP (1),

hm ∈Sm

so that uniformly in hm ∈ Sm , En [(hm · ηˆim )2 ] = En [(hm · ηim )2 ] + oP (1) + OP (1) × oP (1) = 1 + oP (1). This completes the proof of (5.6). We now turn to prove (5.7). Observe that rˆi2 (u) ≤ 2{(ˆ ηim − ηim ) · dm (u)}2 + 2

 ∞  ∑ 

2  dj (u)ηij

j=m+1



.

2 ] = 1 and E[η η ] = 0 for all j ̸= k, we have Since E[ηij ] = 0, E[ηij ij ik 2   ∞ ∞ ∞  ∑  ∑ ∑ E dj (u)ηij  = d2j (u) = κj b2j (u)   j=m+1

j=m+1

≤C

∞ ∑

j=m+1

j −α−2β = O(mn−1 ).

j=m+1

imsart-aos ver. 2013/03/06 file: FunctionQR-supplement-v4.tex date: July 12, 2016

12

K. KATO

Hence, to prove that 2   n  ∑   dj (u)ηij  = OP (m/n), sup En   u∈U j=m+1

it suffices to prove that  2  2   n n     ∑ ∑    −E sup En dj (u)ηij dj (u)ηij  = OP (m/n).     u∈U j=m+1

Defining fu (Xi ) = [

j=m+1

∑∞

j=m+1 dj (u)ηij ,

we wish to show that n ] ∑ E sup {fu (Xi )2 − E[fu (Xi )2 ]} = O(m). u∈U i=1

By the symmetrization inequality, the left side is bounded by n ] [ ∑ 2E sup σi fu (Xi )2 , u∈U i=1

where σ1 , . . . , σn are i.i.d. Rademacher∑random variables independent of −β−α/2 |η | = F (X ); then X1 , . . . , Xn . Observe that |fu (Xi )| ≤ C ∞ ij i j=m+1 j by the contraction principle (see van der Vaart and Wellner, 1996, Proposition A.3.2), the above term is further bounded by n ] [ ∑ 8E max F (Xi ) · sup σi fu (Xi ) 1≤i≤n u∈U i=1 v  u 2  n u √ ∑ u σi fu (Xi )  ≤ 8 E[ max F (Xi )2 ]tE sup 1≤i≤n u∈U i=1

(D.2)

{

≤ O(1) E[ max F (Xi ) ] + 2

1≤i≤n



]} n ∑ E[ max F (Xi )2 ] · E sup σi fu (Xi ) , 1≤i≤n u∈U [

i=1

where the first inequality follows from the Schwarz inequality and the second inequality follows from the Hoffmann-Jorgensen inequality (see van der Vaart and Wellner, 1996, Proposition A.1.6). Observe that max F (Xi ) ≤ C

1≤i≤n

∞ ∑ j=m+1

j −β−α/2 max |ηij |, 1≤i≤n

imsart-aos ver. 2013/03/06 file: FunctionQR-supplement-v4.tex date: July 12, 2016

FUNCTIONAL QUANTILE REGRESSION

13

∑ 2 ] ≤ (E[max 4 1/2 ≤ ( n E[η 4 ])1/2 ≤ Cn1/2 , so and E[max1≤i≤n ηij 1≤i≤n ηij ]) ij i=1 that    ∞ ∞  ∑  ∑  2 j −β−α/2 j −β−α/2 E[ max ηij ] E[ max F (Xi )2 ] ≤ O(1)    1≤i≤n 1≤i≤n j=m+1

= O(n

1/2

m

−2β−α+2

j=m+1

),

which is o(1) because β + α/2 > α + 1 > 2 and m ≍ n1/(2β+α) . Further, because n ∞ n ∑ ∑ ∑ σi fu (Xi ) = dj (u) σi ηij , i=1

j=m+1

i=1

we have [

] ] [ n n ∞ ∑ ∑ ∑ E sup σi fu (Xi ) ≤ C j −β−α/2 E σi ηij u∈U i=1

j=m+1

i=1

∞ ∑

≤ Cn1/2

j −β−α/2 = O(n1/2 m−β−α/2+1 ).

j=m+1

Hence the second term in (D.2) is O(n3/4 m−2β−α+2 ) = O(m) under our assumption. On the other hand, by the proof of (5.4), we have sup En [{(ˆ ηim − ηim ) · dm (u)}2 ] ≤ C

u∈U

m ∑

j −α−2β En [(ˆ ηij − ηij )2 ]

j=1

∑ −α−2β −1 α+2 (n j + ∆γ j 3α+2 )), = OP ( m j=1 j where m ∑

j −2β+2 = O(1)

j=1

and m ∑ j=1

j

−2β+2α+2

  O(1) = O(log n)   O(m−2β+2α+3 )

if −2β + 2α + 2 < −1 if −2β + 2α + 2 = −1 if −2β + 2α + 2 > −1.

imsart-aos ver. 2013/03/06 file: FunctionQR-supplement-v4.tex date: July 12, 2016

14

K. KATO

Since m−2β+2α+3 ≍ n−1 m3α+3 and m3α+3 ≍ n when −2β + 2α + 2 = −1, we have m ∑

j −α−2β (n−1 j α+2 + ∆γ j 3α+2 ) = O(n−1 + ∆γ + n−1 (log n)m3α+3 ∆γ )

j=1

= O(n−1 ). Taking these together, we obtain (5.7). This completes the proof. D.3. Proof of (5.9). Consider the classes of functions { G1 = R × D[0, 1] × Rm+1 ∋ (y, x, η m ) 7→ 1(y ≤ η m · (dm (u) + δ m ))(hm · η m ) } √ : u ∈ U, hm ∈ Sm , δ m = M m/nhm , and { G2 = R × D[0, 1] × Rm+1 ∋ (y, x, η m ) 7→ 1(y ≤ QY |X (u | x))(hm · η m ) } : u ∈ U, hm ∈ Sm . It is relatively standard to see that G1 is a VC subgraph class with VC index bounded by cm for some constant c ≥ 1 (see Belloni et al., 2011, Lemma 18). For G2 , observe first that 1(y ≤ QY |X (u | x)) = 1(FY |X (y|x) ≤ u). Since FY |X (y|x) is a fixed function, it is also shown that G2 is a VC subgraph class with VC index bounded by c′ m for some constant c′ ≥ 1. The conclusion now follows from an application of Theorem 2.6.7 of van der Vaart and Wellner (1996) and a simple covering number calculation. APPENDIX E: USEFUL INEQUALITIES We introduce some useful inequalities. Theorem E.1 (Rudelson’s (1999) inequality). Let Z1 , . . . , Zn be i.i.d. random vectors in Rk with Σ := E[Z1 Z1′ ]. Then, for all k ≥ 2,  √

 n

1 ∑ log k

2 E  Zi Zi′ − Σ  ≤ max{∥Σ∥1/2 E[ max ∥Zi ∥2ℓ2 ], op δ, δ }, δ = C

n 1≤i≤n n i=1

op

where ∥ · ∥op is the operator norm and C is a universal constant.

imsart-aos ver. 2013/03/06 file: FunctionQR-supplement-v4.tex date: July 12, 2016

FUNCTIONAL QUANTILE REGRESSION

15

The expression of Theorem E.1 is slightly different from Rudelson’s original form, but is directly deduced from his proof. Theorem E.1 gives moment bounds on the difference between empirical and population Gram matrices in the operator norm. Recall that for any k × k symmetric matrix A, ∥A∥op = maxv∈Sk−1 |v ′ Av|. To apply Rudelson’s inequality, we have to bound E[max1≤i≤n ∥Zi ∥2ℓ2 ], which is typically implemented by using the following lemma. Lemma E.1. Let X1 , . . . , Xn be arbitrary scalar random variables such that max1≤i≤n E[|Xi |r ] < ∞ for some r ≥ 1. Then, we have E[ max |Xi |] ≤ Cr n1/r , 1≤i≤n

where Cr is a constant depending only on r and max1≤i≤n E[|Xi |r ]. For the proof, see van der Vaart and Wellner (1996, Lemma 2.2.2). In what follows, we introduce “conditional” maximal inequalities. Below we assume the class of functions to be a “pointwise measurable class” to avoid a measurability complication. A class of measurable functions G on a measurable space S is said to be pointwise measurable if there exists a countable class of measurable functions H on S such that for any g ∈ G, there exists a sequence {hm } ⊂ H with hm (x) → g(x) for all x ∈ S. See Chapter 2.3 of van der Vaart and Wellner (1996). This condition is satisfied in our application. Proposition E.1. Let (Ω, A, P) denote the underlying probability space. Let D be a sub σ-field of A. Let {(ui , vi )}ni=1 be a sequence of random variables taking values in some measurable space S such that v1 , . . . , vn are Dmeasurable, the regular conditional distribution of (u1 , . . . , un ) given D exists, and conditional on D, u1 , . . . , un are independent. Let G be a pointwise measurable class of functions on S such that for some D-measurable random ˆ and τˆ, variables B ˆ 1 ≤ ∀i ≤ n, (i) sup |g(ui , vi )| ≤ B, g∈G

(ii) sup En [E[g 2 (ui , vi ) | D]] ≤ τˆ2 , g∈G

ˆ (iii) τˆ ≤ B,

√ almost surely. Suppose that there exist constants A ≥ 3 e and W ≥ 1 such that (E.1)

ˆ G, L2 (Pn )) ≤ (A/ϵ)W , 0 < ∀ϵ ≤ 1, N (Bϵ,

imsart-aos ver. 2013/03/06 file: FunctionQR-supplement-v4.tex date: July 12, 2016

16

K. KATO

where Pn denotes the empirical distribution on S that assigns probability n−1 to each (ui , vi ), i = 1, . . . , n. Let σ1 , . . . , σn be independent Rademacher random variables defined on another probability space. Extend the underlying probability space by the product probability space. Then, we have [ ] E sup |En [σi g(ui , vi )]| | D g∈G

√ ≤ 1(ˆ τ > 0)D 

τˆ2 W n

 ˆ ˆ ˆ AB W B AB  log + log , a.s., τˆ n τˆ

where D is a universal constant. Proposition E.1 is a conditional version of Proposition 2.1 of Gin´e and Guillou (2001). Here, {(ui , vi )}ni=1 are not necessarily independent. However, conditional on D, {ui }ni=1 are independent. Proof of Proposition E.1. The proof is a modification of that of Gin´e and Guillou (2001, Proposition 2.1). For the sake of completeness, we provide the full proof. Without loss of generality we may assume that G contains the 0-function. Suppose that E[supg∈G En [g 2 (ui , vi )] | D] ∧ τˆ > 0. Otherwise the conclusion follows trivially. By Dudley’s inequality (see van der Vaart and Wellner, 1996, Corollary 2.2.8), we have ] [ ∫ θ√ √ n log N (ϵ, G, L2 (Pn ))dϵ, E sup | nEn [σi g(ui , vi )]| | {(ui , vi )}i=1 ≤ D g∈G

0

where θ := (supg∈G En [g 2 (ui , vi )])1/2 and D is a universal constant. Suppose that θ > 0. Using changes of variables, we have ∫ θ/Bˆ √ ∫ θ√ ˆ ˆ G, L2 (Pn ))dϵ log N (ϵ, G, L2 (Pn ))dϵ = B log N (Bϵ, 0

0

√ ∫ ˆ W ≤B 0

(E.2)

ˆ θ/B

√ ∫ ˆ ≤ BA W





ˆ AB/θ

log(A/ϵ)dϵ √ log ϵ dϵ. ϵ2

Integration by parts gives [ √ ]∞ ∫ ∞√ ∫ 1 ∞ 1 log ϵ log ϵ √ dϵ = − + dϵ 2 2 ϵ ϵ 2 c ϵ log ϵ c c √ √ ∫ log c 1 ∞ log ϵ + dϵ, ≤ c 2 c ϵ2 imsart-aos ver. 2013/03/06 file: FunctionQR-supplement-v4.tex date: July 12, 2016

17

FUNCTIONAL QUANTILE REGRESSION

provided that c ≥ e, by which we have √ ∫ ∞√ 2 log c log ϵ dϵ ≤ , if c ≥ e. ϵ2 c c ˆ ≥ A ≥ 3√e > e, we have Since AB/θ √ √ ˆ (E.2) ≤ 2 W θ log(AB/θ), by which we have, using H¨older’s inequality, v [ ] u 2B u ˆ 2 √ A E[(E.2) | D] ≤ 2W tE 1(θ > 0)θ2 log D . θ2 For any fixed c > 0, define f (u) = u log(c/u) if u > 0 and f (0) = 0. Then, f (u) is concave on [0, ∞). Thus, by Jensen’s inequality, the last expression is bounded by √ ˆ2 √ A2 B . 2W E[sup En [g 2 (ui , vi )] | D] × log E[supg∈G En [g 2 (ui , vi )] | D] g∈G Using the decomposition g 2 (ui , vi ) = E[g 2 (ui , vi ) | D] + {g 2 (ui , vi ) − E[g 2 (ui , vi ) | D]}, and the symmetrization inequality conditional on D, we have [ ] E sup En [g 2 (ui , vi )] | D g∈G

[

]

≤ sup En [E[g (ui , vi ) | D]] + 2E sup |En [σi g (ui , vi )]| | D 2

g∈G

2

g∈G

[

]

≤ τˆ + 2E sup |En [σi g (ui , vi )]| | D . 2

2

g∈G

Using now the contraction principle (see van der Vaart and Wellner, 1996, Proposition A.3.2), we have [ ] E sup |En [σi g 2 (ui , vi )]| | {(ui , vi )}ni=1 g∈G

[

]

ˆ sup |En [σi g(ui , vi )]| | ≤ 4BE g∈G

{(ui , vi )}ni=1

,

imsart-aos ver. 2013/03/06 file: FunctionQR-supplement-v4.tex date: July 12, 2016

18

K. KATO

so that

[

]

[

]

ˆ sup |En [σi g(ui , vi )]| | D . E sup En [g (ui , vi )] | D ≤ τˆ + 8BE 2

2

g∈G

g∈G

ˆ 2 . Since for any given c > 0, the map Note that the right side is at most 9B √ u 7→ u log(c/u) is non-decreasing for 0 < u ≤ c/e, and A ≥ 3 e, we have E[sup En [g 2 (ui , vi )] | D] × log g∈G

ˆ log ≤ (ˆ τ 2 + 8BZ) ˆ log ≤ (ˆ τ 2 + 8BZ)

ˆ2 A2 B ˆ τˆ2 + 8BZ ˆ2 A2 B

ˆ log = 2(ˆ τ 2 + 8BZ)

τˆ2 ˆ AB τˆ

,

[

where

ˆ2 A2 B E[supg∈G En [g 2 (ui , vi )] | D]

]

Z := E sup |En [σi g(ui , vi )]| | D . g∈G

Taking these together, we have √ √ nZ ≤ 2D W

√ ˆ ˆ log AB . (ˆ τ 2 + 8BZ) τˆ

Solving this inequality with respect to Z gives the desired bound. Proposition E.2. Consider the same setting as in Proposition E.1. Instead of (i)-(iii) and (E.1), suppose that there is an envelope function G for G such that for some constants A ≥ e and W ≥ 1, N (ϵ∥G∥L2 (Pn ) , G, L2 (Pn )) ≤ (A/ϵ)W , 0 < ∀ϵ ≤ 1. Then, we have for all q ∈ [1, ∞), ( [

])1/q

E sup |En [σi g(ui , vi )]|q | D g∈G

√ ≤ Dn−1/2 (En [E[Gq¯(ui , vi ) | D]])1/¯q W log A a.s.,

where q¯ = q ∨ 2 and D is a constant depending only on q. imsart-aos ver. 2013/03/06 file: FunctionQR-supplement-v4.tex date: July 12, 2016

FUNCTIONAL QUANTILE REGRESSION

19

Proof. Without loss of generality we may assume that G contains the 0-function. By Dudley’s inequality, ( [

√ E sup | nEn [σi g(ui , vi )]|q | {(ui , vi )}ni=1

])1/q

g∈G

∫ ≤D

θ



log N (ϵ, G, L2 (Pn ))dϵ,

0

where θ := (supg∈G En [g 2 (ui , vi )])1/2 ≤ (En [G2 (ui , vi )])1/2 and D is a constant depending only on q. Using changes of variables, we have that the right side is bounded by √ ∫ 1√ 2 1/2 W log(A/ϵ)dϵ. D(En [G (ui , vi )]) 0

If q ≥ 2, then by H¨older’s inequality, E[(En [G2 (ui , vi )])q/2 | D] ≤ E[En [Gq (ui , vi )] | D] = En [E[Gq (ui , vi ) | D]]. On the other hand, if q ∈ [1, 2), E[(En [G2 (ui , vi )])q/2 | D] ≤ (En [E[G2 (ui , vi ) | D]])q/2 . This leads to the desired inequality. APPENDIX F: TABLE AND FIGURES FOR CASE (B) IN SECTION 4

REFERENCES Belloni, A., Chernozhukov, V. and Fern´ andez-Val, I. (2011). Conditional quantile processes based on series or many regressors. arXiv:1105.6154. Bosq, D. (2000). Linear Processes in Function Spaces. Lecture Notes in Statistics. Springer. Gin´e, E. and Guillou, A. (2001). On consistency of kernel density estimators for randomly censored data: rates holding uniformly over adaptive intervals. Ann. Inst. H. Poincar´e Probab. Statist. 37 503-522. Hall, P. and Horowitz, J.L. (2007). Methodology and convergence rates for functional linear regression. Ann. Statist. 35 70-91. Hofmann, B. and Scherzer, O. (1998). Local ill-posedness and source conditions of operator equations in Hilbert space. Inverse Problems 14 1189-1206. Marcus, M.B. (1975). Uniform convergence of random Fourier series. Ark. Mat. 13 107122. Rudelson, M. (1999). Random vectors in the isotropic position. J. Functional Anal. 164 60-72. imsart-aos ver. 2013/03/06 file: FunctionQR-supplement-v4.tex date: July 12, 2016

20

K. KATO

Fig 1. Performance of selection criteria. Case (b). Estimation of (t, u) 7→ b(t, u). n=200,alpha=1.1,normal error

n=500,alpha=1.1,normal error

4

6

8

10

12

14

1.0 0.8

QA−MISE

0.4 2

4

6

8

10

12

14

2

4

6

8

10

12

14

m

n=100,alpha=2,normal error

n=200,alpha=2,normal error

n=500,alpha=2,normal error

4 1

2 4

6

8

10

12

14

0

0

0

2

Fixed IAIC IBIC IGACV

3

QA−MISE

6

QA−MISE

8

Fixed IAIC IBIC IGACV

4

10

15

Fixed IAIC IBIC IGACV

2

10

m

20

m

5

QA−MISE

0.2

0.4 0.2

0.5

2

Fixed IAIC IBIC IGACV

0.6

1.2 0.8

QA−MISE

1.0

Fixed IAIC IBIC IGACV

0.6

1.5 1.0

QA−MISE

2.0

Fixed IAIC IBIC IGACV

1.2

2.5

n=100,alpha=1.1,normal error

2

4

6

m

8

10

12

14

2

4

6

m

8

10

12

14

m

Fig 2. Performance of selection criteria. Case (b). Estimation of QY |X (u | x). n=200,alpha=1.1,normal error

n=500,alpha=1.1,normal error

0.5 0.3

QA−MISE

0.4

Fixed IAIC IBIC IGACV

0.1

0.1

0.1

0.2

0.4 0.3

QA−MISE

Fixed IAIC IBIC IGACV

0.2

0.3

0.4

0.5

Fixed IAIC IBIC IGACV

0.2

QA−MISE

0.5

0.6

n=100,alpha=1.1,normal error

2

4

6

10

12

14

2

4

6

8

10

12

14

2

4

6

8

10

12

14

n=200,alpha=2,normal error

n=500,alpha=2,normal error

0.10

2

4

6

8 m

10

12

14

Fixed IAIC IBIC IGACV

QA−MISE

0.20

QA−MISE

0.25

Fixed IAIC IBIC IGACV

2

4

6

8 m

10

12

14

0.05 0.10 0.15 0.20 0.25

n=100,alpha=2,normal error 0.05 0.10 0.15 0.20 0.25 0.30

m

0.30

m

0.15

QA−MISE

8 m

Fixed IAIC IBIC IGACV

2

4

6

8

10

12

14

m

imsart-aos ver. 2013/03/06 file: FunctionQR-supplement-v4.tex date: July 12, 2016

21

FUNCTIONAL QUANTILE REGRESSION

Fig 3. Performance of selection criteria. Case (b). Estimation of (t, u) 7→ b(t, u). n=200,alpha=1.1,Cauchy error

n=500,alpha=1.1,Cauchy error

6

2.0 1.5

QA−MISE

0.5

2 1 2

4

6

8

10

12

14

2

4

6

8

10

12

14

2

4

6

8

10

12

14

m

m

n=100,alpha=2, Cauchy error

n=200,alpha=2,Cauchy error

n=500,alpha=2,Cauchy error

Fixed IAIC IBIC IGACV

15 QA−MISE

30

QA−MISE

40

Fixed IAIC IBIC IGACV

2

4

6

8

10

12

14

0

0

0

10

5

20

100

Fixed IAIC IBIC IGACV

10

50

20

150

m

50

QA−MISE

Fixed IAIC IBIC IGACV

1.0

4

5

Fixed IAIC IBIC IGACV

3

QA−MISE

10

Fixed IAIC IBIC IGACV

5

QA−MISE

15

2.5

n=100,alpha=1.1,Cauchy error

2

4

6

m

8

10

12

14

2

4

6

m

8

10

12

14

m

Fig 4. Performance of selection criteria. Case (b). Estimation of QY |X (u | x).

2

4

6

8

10

12

14

2

4

6

10

12

14

Fixed IAIC IBIC IGACV

2

4

6

8

10

12

14

m

m

n=100,alpha=2,Cauchy error

n=200,alpha=2,Cauchy error

n=500,alpha=2,Cauchy error

Fixed IAIC IBIC IGACV

0.50 QA−MISE

0.8

QA−MISE

1.0

Fixed IAIC IBIC IGACV

0.4

0.30

0.6

1.5

2.0

Fixed IAIC IBIC IGACV

0.40

2.5

m

1.0

QA−MISE

8

QA−MISE

Fixed IAIC IBIC IGACV

n=500,alpha=1.1,Cauchy error 0.3 0.4 0.5 0.6 0.7 0.8 0.9

QA−MISE

1.5 1.0

QA−MISE

2.0

Fixed IAIC IBIC IGACV

n=200,alpha=1.1,Cauchy error 0.5 0.6 0.7 0.8 0.9 1.0 1.1

n=100,alpha=1.1,Cauchy error

2

4

6

8 m

10

12

14

2

4

6

8 m

10

12

14

2

4

6

8

10

12

14

m

imsart-aos ver. 2013/03/06 file: FunctionQR-supplement-v4.tex date: July 12, 2016

22

K. KATO Table 1 Average numbers of m selected by three criteria in case (b). Standard deviations are given in parentheses. n 100 200 500 100 200 500 100 200 500 100 200 500

α 1.1 1.1 1.1 2 2 2 1.1 1.1 1.1 2 2 2

error dist. normal normal normal normal normal normal Cauchy Cauchy Cauchy Cauchy Cauchy Cauchy

AIC 8.76 (4.33) 8.54 (4.07) 9.09 (3.69) 8.06 (4.59) 7.75 (4.30) 7.69 (3.96) 2.36 (1.53) 2.43 (1.03) 2.86 (0.92) 1.99 (1.42) 2.06 (0.88) 2.30 (0.66)

BIC 3.12 (1.17) 3.47 (1.11) 4.01 (0.99) 2.52 (0.84) 2.72 (0.87) 3.09 (0.67) 1.50 (0.59) 1.76 (0.56) 2.06 (0.48) 1.27 (0.47) 1.44 (0.53) 1.80 (0.45)

GACV 7.25 (3.95) 7.78 (3.86) 8.66 (3.60) 6.37 (4.12) 6.82 (4.03) 7.23 (3.82) 2.25 (1.34) 2.42 (1.01) 2.83 (0.90) 1.86 (1.14) 2.04 (0.83) 2.28 (0.63)

van der Vaart, A.W. and Wellner, J.A. (1996). Weak Convergence and Empirical Processes: With Applications to Statistics. Springer. Yuan, M. and Cai, T. (2010). A reproducing kernel Hilbert space approach to functional linear regression. Ann. Statist. 38 3412-3444. Department of Mathematics Graduate School of Science Hiroshima University 1-3-1 Kagamiyama, Higashi-Hiroshima Hiroshima 739-8526, Japan. E-mail: [email protected]

imsart-aos ver. 2013/03/06 file: FunctionQR-supplement-v4.tex date: July 12, 2016

SUPPLEMENT TO “ESTIMATION IN FUNCTIONAL ...

tion to nonlinear ill-posed inverse problems, technical proofs omitted in the main body, some useful ... where the map A : U × L2[0,1] → L2[0,1] is defined by. A(u, g)(·) = E[{u − 1(Y ≤. ∫ 1. 0 ...... Graduate School of Science. Hiroshima University.

164KB Sizes 0 Downloads 34 Views

Recommend Documents

Supplement to - GitHub
Supplemental Table S6. .... 6 inclusion or exclusion of certain genetic variants in a pharmacogenetic test ..... http://aidsinfo.nih.gov/contentfiles/AdultandAdolescentGL.pdf. .... 2.0 are expected to exhibit higher CYP2D6 enzyme activity versus ...

Supplement to “On the Effect of Bias Estimation on ...
Jan 18, 2017 - All our methods are implemented in software available from the authors' websites and via the ..... For each ξ > 0 and all sufficiently small h sup.

Supplement to “On the Effect of Bias Estimation on ...
Dec 16, 2017 - This reasoning is not an artifact of choosing k even and l = 2, but in other cases ρ → 0 can .... (as the notations never occur in the same place), but in the course of ... ruled out for local polynomial regression, see Section S.II

Supplement to “On the Effect of Bias Estimation on ...
Dec 16, 2017 - K(u)kdu. The three statistics Tus, Tbc, and Trbc share a common structure that is exploited to give a unified theorem statement and proof. For v ∈ {1,2}, ...... 90.5 87.0. 92.4. 0.043. 0.052. Panel B: Summary Statistics for the Estim

supplement to study material - ICSI
Ensure that advertisement giving details relating to oversubscription, basis ... Ensure that no advertisement or distribution material with respect to the issue.

Supplement to: Competition and Nonlinear Pricing in ...
This note is organized as follows: Section 1 provides an alternative method to derive the optimal allocation for Principal 2, following Rochet and Chone (1998) and Basov (2002). Sec- tion 2 provides the proofs left unsolved in the main paper. In sect

supplement to study material - ICSI
(ii) the issuer undertakes to provide market-making for at least two years from ..... buyers if an issuer has not satisfied the basic eligibility criteria and undertakes ...... buyers on proportionate basis as per illustration given in Part C of Sche

Supplement to "Robust Nonparametric Confidence ...
Page 1 ... INTERVALS FOR REGRESSION-DISCONTINUITY DESIGNS”. (Econometrica ... 38. S.2.6. Consistent Bandwidth Selection for Sharp RD Designs .

Supplement to "Efficient Repeated Implementation"
the definition of ψ of ˆg) but induces regime Dj in which, by (A.1), j obtains vj j > πθ(t) θt j . But this is a contradiction. Q.E.D. ... Next define ρ ≡ maxi θ a a [ui(a θ)−ui(a θ)] and ¯δ ≡ ρ ρ+ε . Mechanism ˜g = (M ψ) is def

Mo_Jianhua_Asilomar14_Channel Estimation in Millimeter Wave ...
Mo_Jianhua_Asilomar14_Channel Estimation in Millimeter Wave MIMO Systems with One-Bit Quantization.pdf. Mo_Jianhua_Asilomar14_Channel Estimation ...

Supplement to Dynamic Mixture-Averse Preferences
Mar 2, 2018 - lotteries △(X), where X is any compact metric space. Of particular interest is the special case where X is an interval, e.g., a set of monetary outcomes or the set of continuation values for an ORA representation. We first state a gen

Supplement to “Contributions to the Theory of Optimal Tests”
Tests which depend on the data only through QS, |QST |, and QT are locally unbiased .... Let y2,−1 be the N-dimensional vector whose i-th entry is y2,i−1, and ...

Functional Programming in Scala - GitHub
Page 1 ... MADRID · NOV 21-22 · 2014. The category design pattern · The functor design pattern … ..... Play! ∘ Why Play? ∘ Introduction. Web Dictionary.

ePUB Functional Programming in Java: How functional ...
... performance and parallelization and it will show you how to structure your application so that it will ... a better Java developer. ... building enterprise software.

Supplement
water soluble, rapidly absorbed from the GI tract, has high bioavailability,21,22 and reaches ...... Gage et al197 developed a dosing algorithm based on CYP2C9 ...

Supplement
Additional evidence supports improvements in health-care utilization and psychosocial outcomes. There are few additional ... top 10 in which mortality continues to increase.5– 8 In persons 55 to 74 .... pulmonary rehabilitation programs have been d