STATISTICAL PROPERTIES OF GENERALIZED DISCREPANCIES CHRISTINE CHOIRAT AND RAFFAELLO SERI

Abstract. When testing that a sample of n points in the unit hypercube [0, 1]d comes from a uniform distribution, the Kolmogorov-Smirnov and the Cram´ er-von Mises statistics are simple and well-known procedures. To encompass these measures of uniformity, Hickernell ([Hic96, Hic98a, Hic98b]) introduced the so-called generalized Lp −discrepancies. These discrepancies can be used in numerical integration through Monte Carlo and quasi-Monte Carlo methods, design of experiments, uniformity testing and goodness-of-fit tests. The aim of this paper is to derive the statistical asymptotic properties of these statistics under Monte Carlo sampling. In particular, we show that, under the hypothesis of uniformity of the sample of points, the asymptotic distribution is a complex stochastic integral with respect to a pinned Brownian sheet. On the other hand, if the points are not uniformly distributed, then the asymptotic distribution is Gaussian.

1. Introduction The Koksma-Hlawka inequality (see [Nie92], p. 18) is a well-known bound on the error of Monte Carlo and quasi-Monte Carlo integration: it majorizes the error through the variation of the integrand in the sense of Hardy and Krause times the star discrepancy of the random or quasi-random points ([Nie92], Definition 2.1). The star discrepancy has a remarkable statistical interpretation since, when d

the integrand is defined on the real hypercube [0, 1] , it has the same form as the Kolmogorov-Smirnov statistic to test uniformity. Date: October 22, 2004 (submitted); May 11, 2005 (revised). 2000 Mathematics Subject Classification. Primary 65D30, 60F05, 68U20, 65C05, 11K45. Key words and phrases. Generalized discrepancies, testing uniformity, Monte Carlo, quasi-Monte Carlo, limit distribution. We thank Peter Hellekalek, Søren Johansen, Peter E. Jupp for useful comments on a previous version of this paper and Kendall E. Atkinson and David E. Edmunds for useful references. 1

STATISTICAL PROPERTIES OF GENERALIZED DISCREPANCIES

2

In a series of important papers, Hickernell ([Hic96] and [Hic97], [Hic98a, Hic98b]) has introduced the generalized Lp −discrepancies Dp (Pn ) based on the sample of n points Pn :1 they extend the star discrepancy and allow for measuring the degree of nonuniformity of the sample and the efficiency of the numerical integration procedure. The links of these figures of merit with goodness-of-fit statistics and with optimal design of experiments have been pointed out by Hickernell ([Hic99]) and Fang et al. ([FMW02]). Liang et al. ([LFHL01]) have started an investigation of the asymptotic properties of Dp (Pn ) in the case p = 2: they show that, under the hypothesis of uniformity, D2 (Pn ) converges almost surely to 0 but they do not derive its asymptotic distribution. Then, they obtain two alternative statistics, strictly linked to this one, which can be used to test statistically the efficiency of numerical integration procedures.2 The aim of this paper is to complete the analysis of [LFHL01] and to develop some further properties of the test statistics proposed by Hickernell ([Hic98a, Hic98b]). The interest of these results is more theoretical than practical: indeed, the asymptotic distributions that we obtain are quite difficult to compute (except in the case p = 2, see [CS05b]), but these results yield a precise information about the rate of convergence of these statistics to 0 when the sample Pn is uniformly scattered and the rate of divergence when the sample is not uniformly scattered. As concerns the generalized discrepancy Dp (Pn ), we show that Dp (Pn ) converges almost surely to 0 if and only if the sample Pn is uniformly scattered (as d

defined below) on the unit hypercube [0, 1] . Under this distributional hypothesis, we obtain a rate of convergence towards zero: first of all, we show that Dp (Pn ) is ¡ ¢ √ OP n−1/2 ,3 that is n · Dp (Pn ) converges in distribution to a nontrivial random

1Remark that in D (P ) we explicit the dependence on the indices p and n. p n

2Other authors have recently investigated similar discrepancies (Leeb in [Lee96a, Lee96b, Lee02],

Hoogland and Kleiss in [HK96a, HK96b, HK97], James et al. in [JHK97], Hoogland et al. in [HJK98], van Hameren et al. in [vHKH97]). 3We say that X is O (r ) if, for each ε > 0, there exists M > 0 such that: n n P

n

P where M and ε do not depend on n.

|Xn | >M rn

o

< ε,

∀n,

STATISTICAL PROPERTIES OF GENERALIZED DISCREPANCIES

3

variable whose form is provided in terms of a stochastic integral; moreover, we prove that a Law of the Iterated Logarithm holds for the generalized discrepancy. 2

We show that, in the case p = 2, the asymptotic distribution [D2 (Pn )] can be written as a weighted infinite mixture of χ2 distributions and is a particular instance of what is called Gaussian Chaos. While the asymptotic distribution for p 6= 2 is difficult to obtain, in this special case it can be computed using the algorithm presented in [CS05a]. However, it can be shown that this result holds for a much more general class of discrepancies (the generalized L2 −discrepancies, the classical and the dyadic diaphonies, the weighted spectral test, the serial and the overlapping serial test, etc.). This is the object of a companion paper ([CS05b]). Moreover, we derive the asymptotic distribution of Dp (Pn ) under the hypothesis that the sample Pn does not come from a uniform distribution. In this case, for finite values of p, the asymptotic distribution is Gaussian, but an alternative representation can be given as a stochastic integral. A limited simulation study, in Section 4, confirms these theoretical findings for the case of generalized L2 −discrepancies. The results shown in this paper can be generalized along several lines. A particularly interesting topic would be the extension to sequences of points that are not independent and identically distributed, but of particular relevance to Numerical Analysis, such as scrambled digital nets (see [HHW03] for simulation results). This will be left to future work. Since the results make large use of statistical and probabilistic tools, we need to introduce some notation. In what follows, the term uniformly scattered is used in the sense of Fang and Wang ([FW94], p. 18, see also Remark 1.3) to indicate that the points are uniformly distributed in the sense of Niederreiter ([Nie92], p. 13) (we use the new definition in order to avoid confusion with the related statistical concept). We introduce a notation that is typical of empirical process theory: for a probability space (X , B, P) and a measurable function f : X → Rk , we write R Pf = X f dP to indicate the integral of f with respect to P. If there is any doubt, R we write the variable of integration in capital letter (Pf (X) = X f (x) P (dx)).

STATISTICAL PROPERTIES OF GENERALIZED DISCREPANCIES

4

Unless otherwise stated, we will assume that P is the Lebesgue measure on the unit d

hypercube [0, 1] , also written as λ⊗d .4 In the following, for a n−sample Pn , we 5 define the empirical measures λ⊗d n and the associated integrals as:

P 1 λ⊗d n = n z∈Pn δz , R P 1 λ⊗d f dλ⊗d n f = n = n z∈Pn f (z) , where δz is the Dirac measure concentrated in z. Moreover, we define as |A| the number of points in the set A and as Vol ([0, x)) the volume of the rectangular solid [0, x). For any index set u ⊆ {1, . . . , d}, we denote by |u| its cardinality, by u

[0, 1] the |u| −dimensional unit hypercube and by xu a |u| −dimensional vector containing the elements of x indexed by the elements of u.

2. Asymptotic results for Lp −discrepancies under uniformity The generalized discrepancies Dp (Pn ) have been introduced in Hickernell ([Hic96, Hic98a, Hic98b]) as a generalization of some figures of merit that have arisen in the literature. In the case 1 ≤ p ≤ +L∞ (see equation (3.8b) in [Hic98a]) they are defined by the equation: 

1/p °p X° ° |u| ° Dp (Pn ) =  °β · Dp,u (Pn,u )° p  L

u6=∅

(2.1)  =

 ¯p ¯  1/p ¯ ¯¯ X Y£ ¯ |u|  Y 0 ¤ 1 ¯β · µ0 (xj ) + xj − 1{xj >zj } ¯¯ dxu  , µ (xj ) − ¯ ¯  n [0,1]u ¯ j∈u z∈Pn j∈u

XZ

u6=∅

where u is a subset of the set {1, . . . , d}, Pn,u denotes the projection of the sample u

Pn on the unit cube [0, 1] , β is an arbitrary given positive constant and µ (·) is an

4The ⊗ sign is used to remind that it is the d−dimensional product Lebesgue measure. Strictly ¯

speaking, the uniform distribution should be written as λ⊗d ¯

[0,1]d

, the d−dimensional Lebesgue

measure restricted to [0, 1]d , but with a small abuse of notation we will prefer to write λ⊗d . 5The subscript n reminds that the measure is the empirical counterpart of λ⊗d .

STATISTICAL PROPERTIES OF GENERALIZED DISCREPANCIES

5

arbitrary function satisfying ½ µ∈

df f: ∈ L∞ ([0, 1]) and dx

Z

¾

1

f (x) dx = 0 . 0

Remark that in the case p = +∞, the previous formula becomes: ° ° ° ° D∞ (Pn ) = max °β |u| · D∞,u (Pn,u )° ∞ u L ¯ ¯ ¯Y ¯ X Y£ ¯ ¯ ¤ 1 |u| ¯ 0 0 µ (xj ) − = max ess sup β ¯ µ (xj ) + xj − 1{xj >zj } ¯¯ . u x ∈[0,1]u n ¯j∈u ¯ u z∈Pn j∈u Most of our results can be easily extended to weighted Lp −discrepancies in the spirit Q of [SW98] and [LP03], replacing β |u| with j∈u γj , and in the spirit of [SWW04], substituting β |u| with γd,|u| (refer to these papers for definitions).6 When the sample Pn is uniformly scattered, it is crucial to remark that a recurrent element of the previous formulas can be written as the deviation of the empirical distribution from the uniform one: ¤ 1 X Y£ 0 µ (xj ) + xj − 1{xj >zj } n j∈u z∈Pn j∈u     Y £  Y £  ¤ ¤ = λ⊗d µ0 (xj ) + xj − 1{xj >Zj } − λ⊗d µ0 (xj ) + xj − 1{xj >Zj } n     j∈u j∈u   Y £  ¡ ⊗d ¢ ¤ 0 = λ − λ⊗d µ (x ) + x − 1 . j j {xj >Zj } n   Y

µ0 (xj ) −

j∈u

This helps understanding the forthcoming asymptotic results and justifies this notation that will be used frequently in the following: gu (x, z) =

Y£ ¤ µ0 (xj ) + xj − 1{xj >zj } . j∈u

6The only problem is that, if the weights

Q

γ or γd,|u| are zero for some u, then weighted j∈u j Lp −discrepancies can converge to 0 even if the sample Pn is not uniformly scattered. However, if the weights are bounded away from zero, the asymptotic theory is exactly the same.

STATISTICAL PROPERTIES OF GENERALIZED DISCREPANCIES

6

In the general case of equation (2.1), the choices: µ ¶ 1 1 4 2 µ (x) = − x −x+ , β = 2, M = , 2 6 3 ï ! ¯2 ¯ ¯ ¯ 1 ¯¯ 13 1¯ 1¯ 1 µ (x) = − x − ¯¯ − ¯¯x − ¯¯ + , β = 1, M = , ¯ 2 2 2 6 12 µ (x) =

1 x2 4 − , β = 1, M = , 6 2 3

yield respectively the symmetric, the centered and the star discrepancy (see [LFHL01]). In particular, the star Lp −discrepancies can be written more simply as: 

 °p 1/p X° ° ° |P ∩ [0, x )| n,u u °  Dp∗ (Pn ) =  − Vol ([0, xu ))° ° ° p n L

u6=∅

 =

XZ

u6=∅

[0,1]

1/p ¯ ¯p ¯ |Pn,u ∩ [0, xu )| ¯ ¯ − Vol ([0, xu ))¯¯ dxu  . ¯ u n

The special case ∗ D∞

¯ ¯ ¯ |Pn ∩ [0, x)| ¯ ¯ (Pn ) = sup ¯ − Vol ([0, x))¯¯ n x∈[0,1]d

(see e.g. [Hic98a], p. 316, equation (5.1b), or [Nie92], Definition 2.1), is also called star discrepancy (remark that this definition does not coincide with that of [FW94], p. 33, equation (1.4.2)) and coincides with the Kolmogorov-Smirnov d

statistic for testing uniformity on [0, 1] . On the other hand, D2∗ (Pn ) and D1∗ (Pn ) yield, respectively, the Cram´er-von Mises statistic (see [Hic99]) and the L1 −test of Schmid and Trede ([ST96]; see also [SW86], p. 149). ∗ It is well known that, when n → +∞, the star discrepancy D∞ (Pn ) converges

to 0 if and only if the sample Pn is uniformly scattered. Moreover, under the √ ∗ hypothesis of uniform distribution, nD∞ (Pn ) converges in distribution to a welldefined random variable: this means that the average-case error of a Monte Carlo integration procedure decreases (in a certain average sense) as

√1 . n

A worst-case

∗ error is given by a Law of the Iterated Logarithm (LIL) for the discrepancy D∞ (Pn ),

STATISTICAL PROPERTIES OF GENERALIZED DISCREPANCIES

µq ∗ stating that D∞ (Pn ) = O

7

¶ ln ln n n

P − as ([FW94], p. 19). Similar results holds

also for the Cram´er-von Mises statistics (e.g. the LIL has been proved by [Fin71]). In this Section, we prove that these results hold also for the generalized discrepancies Dp (Pn ). Then, we derive the asymptotic distribution of Dp (Pn ) as a function of a stochastic integral with respect to a pinned Brownian sheet (for the definitions see e.g. [AS87], pp. 1345-1346). Proposition 2.1. Let Pn be given by independent and identically distributed random variables. Then, the following three facts hold: (i): Dp (Pn ) → 0 P − as if and only if Pn is drawn from P (the uniform d

measure on [0, 1] ); on the other hand, Dp (Pn ) 9 0 P∗ − as if and only if Pn is drawn from P∗ 6= P; d

(ii): Under the probability measure P (the uniform measure on [0, 1] ): µ Dp (Pn ) = OP

1 √ n

¶ ; d

(iii): Under the probability measure P (the uniform measure on [0, 1] ), there exists P − as a finite constant K > 0 such that Dp (Pn ) satisfies the limit inequality:

√ nDp (Pn ) ≤ K; lim sup √ n→∞ ln ln n

this can be restated saying that the inequality: r Dp (Pn ) > K

ln ln n , n

holds only for a finite number of indexes n. Remark 2.2. In the special case p = 2, the Law of the Iterated Logarithm could also be proved using the results of [Deh89] or [GKLZ01]. The following Proposition yields an asymptotic distributional result for under the null hypothesis of uniform distribution of the sample.



nDp (Pn )

STATISTICAL PROPERTIES OF GENERALIZED DISCREPANCIES

8 d

Proposition 2.3. Under the probability measure P (the uniform measure on [0, 1] ), the statistic Dp (Pn ) given by (2.1) has the asymptotic distribution:  √

D

nDp (Pn ) −→ 

X

Z β p|u| · [0,1]u

u6=∅

1/p ¯ ¯p ¯ |u| ¯ ¯Z (xu )¯ dxu  ,

© ª where Z|u| (xu ) , u ⊆ {1, . . . , d} is a collection of centered Gaussian processes indexed by the subset u ⊆ {1, . . . , d}. Any Z|u| (xu ) is defined as the stochastic integral:

Z |u|

Z

(xu ) =

[0,1]d

gu (x, y) dB (y)

where B is a d−dimensional pinned Brownian sheet. The processes Z|u| (xu ) are characterized by the covariance functions: h i Y Y Cov Z|u| (xu ) , Z|u| (zu ) = [µ0 (xj ) µ0 (zj ) − xj zj + xj ∨ zj ] − [µ0 (xj ) µ0 (zj )] , j∈u

h i Cov Z|u1 | (xu1 ) , Z|u2 | (zu2 ) =

j∈u

Y

Y

µ0 (xj ) ·

j∈u1 \u2

j∈u2 \u1

µ0 (zj ) ·

  Y 

[µ0 (xj ) µ0 (zj ) − xj zj + xj ∨ zj ] −

j∈u1 ∩u2

Y

µ0 (xj ) µ0 (zj )

j∈u1 ∩u2

for u, u1 , u2 ⊆ {1, . . . , d}.

Remark 2.4. (i) This result can be used as follows to assess uniformity of points. Suppose to observe a set of points Pn believed to come from a uniformly distributed sample. In this case,  √

D

nDp (Pn ) −→ 

X

u6=∅

Z β p|u| · [0,1]u

1/p ¯ ¯p ¯ |u| ¯ ¯Z (xu )¯ dxu 

and

P

©√

  1/p   Z  X ¯ ¯p  ª ¯ |u| ¯  nDp (Pn ) ≤ x ≈ P  β p|u| · Z (x ) dx ≤ x . ¯ u ¯ u   [0,1]u  u6=∅ 

  

,

STATISTICAL PROPERTIES OF GENERALIZED DISCREPANCIES

Therefore, if we want to test that Pn is uniform, we verify whether



9

nDp (Pn ) falls

into the interval [0, γα ], where γα is defined by   1/p   Z   X ¯ ¯p ¯ |u| ¯ p|u|   β · ≤ γα 1−α=P ¯Z (xu )¯ dxu   [0,1]u   u6=∅ ≈P

©√

nDp (Pn ) ≤ γα

ª

and α is 1% or 5%. In this case, if the points are indeed uniformly distributed, we asymptotically make an error only with probability α. (ii) A particularly interesting case arises when p = 2. In this case: D

2

n [Dp (Pn )] −→ =

X

Z [0,1]u

 X

Z

= d

[0,1]

Z

d

[0,1]

[0,1]d

Z β

2|u|

· [0,1]u

u6=∅

β 2|u| ·

u6=∅

Z

Z

X

[0,1]d

gu (x, y) gu (x, z) dB (y) dB (z) dxu  

Z β 2|u| ·



[0,1]u

u6=∅

¯ ¯2 ¯ |u| ¯ ¯Z (xu )¯ dxu

gu (x, y) gu (x, z) dxu Z

dB (y) dB (z)

Z

=

(2.2)



h (y, z) dB (y) dB (z) [0,1]d

[0,1]d

where we have set X

h (y, z) ,

Z β

2|u|

· [0,1]u

u6=∅

gu (x, y) gu (x, z) dxu .

Since Z 0<

h (y, y) dy < ∞, [0,1]d

Z

Z

0<

h (y, z) dydz < ∞, [0,1]d

[0,1]d

the spectrum of the integral operator A defined as: Z h (y, z) m (z) λ⊗d (dz) ,

Am (y) = d

[0,1]

STATISTICAL PROPERTIES OF GENERALIZED DISCREPANCIES

10

d

for y, z ∈ [0, 1] , m ∈ L2 , consists of a sequence of nonnegative eigenvalues (λj ) with P∞ λ1 ≥ λ2 ≥ . . . ≥ 0 and j=1 λj < ∞, and corresponding eigenvectors (φj ) that can P∞ be taken orthonormal. Therefore, h (y, z) can be written as j=1 λj · φj (y) φj (z) in L2 and (2.2) becomes: X u6=∅

Z β

2|u|

· [0,1]u

Z ∞ ¯2 ¯ X ¯ ¯ |u| λj · ¯Z (xu )¯ dxu =

[0,1]d

j=1

=

∞ X

[0,1]d

φj (y) φj (z) dB (y) dB (z) !2

ÃZ λj ·

[0,1]d

j=1

Now, the quantities defined by Xj =

Z

³R [0,1]d

φj (y) dB (y)

.

´ φj (y) dB (y) , for j = 1, . . . , ∞, are

independent standard Gaussian random variables. Therefore, we can write: 2

D

n [Dp (Pn )] −→

∞ X

λj · Xj2

j=1

where (λj ) and (Xj ) are defined as above. The random variable

P∞ j=1

λj · Xj2 is

a linear combination of chi-squared random variables and is called a second order Gaussian chaos. This result has also been obtained in [CS05b] in a completely different way (using the properties of degenerate V −statistics). The eigenvalues P (λj ) and the distribution of the random variable λj Xj2 can be approximated through the algorithms exposed in [CS05a]. 3. Asymptotic results for Lp −discrepancies under nonuniformity A centered and scaled version of Dp (Pn ) converges to a well-defined random variable under the alternative too, that is when the sample Pn does not come from d

the uniform distribution on [0, 1] . While for p = ∞ the asymptotic distribution is very complicated, in the case p < ∞ the limit distribution reduces to a normal random variable with a complex variance (see [Ang83] for the case of the Cram´ervon Mises statistic, i.e. p = 2 and [Rag73] for the case of the Kolmogorov-Smirnov statistic, i.e. p = ∞). We give an explicit formula for the variance and we provide an alternative representation as a function of Brownian sheets.

STATISTICAL PROPERTIES OF GENERALIZED DISCREPANCIES

11

Proposition 3.1. Under the probability measure P∗ 6= λ⊗d (i.e. P∗ is not the d

uniform measure on [0, 1] ), we have: P∗ − as

Dp (Pn ) −→ Dp (P∞ ) , n→∞

where

 Dp (P∞ ) = 

X

1/p

Z p

β |u|p · [0,1]u

u6=∅

|(P − P∗ ) gu (x, Z)| dxu 

.

Proposition 3.2. If p < ∞, under the probability measure P∗ 6= λ⊗d (i.e. P∗ is d

not the uniform measure on [0, 1] ) with cdf F ∗ , we have: √

N

D

n [Dp (Pn ) − Dp (P∞ )] −→

p−1

p [Dp (P∞ )]

where N is a centered Gaussian random variable with variance given by: (3.1)



σ 2 = P∗ p

X

p−1

β |u|p · [0,1]d

u6=∅

= p2

2

Z

X X

[(P − P∗ ) gu (x, Z)] Z

· [gu (x, y) − P∗ gu (x, Y)] dxu 

Z p−1

β (|u|+|v|)p · d

d

[0,1]

u6=∅ v6=∅

[(P − P∗ ) gu (x, Z)]

[0,1]

· {P∗ [gu (x, y) gv (x, y)] − P∗ gu (x, Y) P∗ gv (x, Y)} dxu dxv . N can alternatively be expressed as the following stochastic integral: N =p·

X

Z p−1

β |u|p · [0,1]

u6=∅

d

where

[(P − P∗ ) gu (x, Z)]

|u|

· ZF ∗ (xu ) dxu

Z |u| ZF ∗

(xu ) =

[0,1]d

gu (x, y) dBF ∗ (y) ,

and BF ∗ is the centered Gaussian process characterized by variance: Cov [BF ∗ (x) , BF ∗ (z)] = F ∗ (x ∧ z) − F ∗ (x) · F ∗ (z) . Remark 3.3. (i) The asymptotic distribution when p = ∞ is nonstandard and could be obtained as in [Rag73] and [SW86] (p. 177).

p−1

· [(P − P∗ ) gv (x, Z0 )]

STATISTICAL PROPERTIES OF GENERALIZED DISCREPANCIES

12

(ii) We go back to the framework of Remark 2.4 (i). Suppose, now, that the points √ are not uniformly distributed. Then, the probability that nDp (Pn ) falls into the interval [0, γα ] is approximately given by: P

©√

=P

ª ©√ ª nDp (Pn ) ∈ [0, γα ] = P nDp (Pn ) ≤ γα

©√

n [Dp (Pn ) − Dp (P∞ )] ≤ γα − ¶ µ √ γα − nDp (P∞ ) ≈Φ σ

where σ 2 is the asymptotic variance of 0, as n → ∞, we get:

µ Φ

γα −







nDp (P∞ )

ª

n [Dp (Pn ) − Dp (P∞ )]. Since Dp (P∞ ) >

nDp (P∞ ) σ

¶ → 0,

and the test rejects the null hypothesis with probability converging to 1. (iii) Equation (3.1) can be used to derive the variance of the Cram´er-von Mises statistic under the alternative, as [Ang83]. Since the Cram´er-von Mises statistic is the star discrepancy with p = 2 and d = 1, we have β = 1, M =

4 3

and

g1 (x, z) = −1{x>z} and we recover the formulas (stated for a general choice of P) on page 2480 of [Ang83]. 4. A simulation study In the following we will show some of the previous results using the generalized L2 −discrepancies as an example. We just consider the L2 −case since this is the framework of [LFHL01]. We will let n ∈ {25, 50, 100, 200, 400} and d ∈ {1, 2, 5}.7 2

4.1. Finite sample distribution of n·[D2 (Pn )] . The following simulation study 2

shows some characteristics of n · [D2 (Pn )] for several sample sizes. For every graphic, we have drawn 10, 000 times a sample Pn of size n of uniform independent d

2

random variables on [0, 1] . We have calculated n · [D2 (Pn )] for each of the 10, 000 samples and for the three statistics proposed by [Hic98a], that is the centered, the star and the symmetric one. Then we have represented the density (as a histogram 7The simulations have been performed using Ox Professional 3.0 (see [Doo01]).

STATISTICAL PROPERTIES OF GENERALIZED DISCREPANCIES

13

and a kernel estimator) and the Q − Q plot with respect to a Gaussian random variable with the same mean and the same variance. For d = 1, we have represented just the centered discrepancy since the others are equal up to a scalar multiplication, the constant being 1 for the star and 4 for the symmetric discrepancy. 2

For n varying and d fixed, the distribution of n · [D2 (Pn )] is remarkably stable : apart from some fluctuations in the upper tail of the distribution, the overall form is almost the same, especially when the behavior is compared with what happens under the alternative hypothesis of nonuniformity. Moreover, another feature of the finite sample distributions is the fact that, as long as d increases, the distribution of 2

n · [D2 (Pn )] appears to be less skewed: this is compatible with the fact that, when both n and d go to infinity at suitable rates, the asymptotic distribution of the scaled discrepancy is a Gaussian distribution. These two stylized facts are reflected by the results of [CS05b], for the case of L2 −discrepancies and related quantities.

2

4.2. [D2 (Pn )] under the alternative. The following simulated example shows 2

that [D2 (Pn )] approaches a normal distribution when the points are not uniformly d

distributed on [0, 1] . For each graphic, we have drawn 10, 000 samples Pn of size n of Beta(2, 2) independent random variables on [0, 1]. Therefore, we expect to have: √

à n

2

[D2 (Pn )] − .004762 .01810

! D

→ N (0, 1) .

The graphics are similar to the previous ones, but we have subtracted the true mean from the values of the discrepancies and, divided by the true standard devi√ ation and multiplied by n. We have represented only the centered discrepancy since the others are equal up to a scalar multiplication. The convergence towards a Gaussian random variable is evident, but slower than the convergence towards a second order Gaussian Chaos under the null (this is compatible with the convergence rates expressed by the Berry-Ess´een bounds of [CS05b]).

STATISTICAL PROPERTIES OF GENERALIZED DISCREPANCIES

14

5. Proofs Proof of Proposition 2.1. We start with a general majorization result for Dp (Pn ) that will be used several times in the following of the proof. Using Theorem 5.2 in [Hic98a], it is possible to show that any Lp −discrepancy Dp (Pn ) can be majorized by a constant K times the corresponding Lp −star discrepancy Dp∗ (Pn ) (see [Hic98a], Section 5.1, for a definition). Indeed, taking β1 = β, µ1 = µ, β2 = 1, µ2 =

1 6



x2 2 ,

we have: µ ¶ 1 x2 Dp (Pn ) ≤ K q, 1, β, − , µ · Dp∗ (Pn ) 6 2

where q −1 + p−1 = 1. Moreover from the fact that µ ∈ X∞ (see [Hic98a], equation ³ ´ 2 (2.4)), µ0 is bounded and K = K q, 1, β, 16 − x2 , µ is finite. Writing Dp∗ (Pn ) as in [Hic98a] (page 316, equation 5.1a), we have: °p X° ° |Pn,u ∩ [0, xu )| ° £ ∗ ¤p ° Dp (Pn ) = − Vol ([0, xu ))° ° ° p n

(5.1)

L

u6=∅

u

where Pn,u denotes the projection of the sample Pn on the unit cube [0, 1] . There° ° ° ° |P ∩[0,x )| fore, from Lyapunov’s inequality, ° n,u n u − Vol ([0, xu ))° can be majorized Lp

by the L

(5.2)



distance. This leads to the majorization formula:

Dp (Pn ) ≤ K · Dp∗ (Pn ) ≤ K ·

 X 

p

∗ [D∞ (Pn,u )]

u6=∅

 p1  

.

Using this result, the first part of (i) is trivial, since Pn,u is a uniform sample on u

∗ [0, 1] , for any u, and the Glivenko-Cantelli Theorem implies that D∞ (Pn,u ) → 0

almost surely. As concerns the second part of (i), using Theorem 5.2 in [Hic98a] it is possible to show that any Lp −discrepancy Dp (Pn ) can be minorized by a constant times the corresponding Lp −star discrepancy Dp∗ (Pn ) (see [Hic98a], Section 5.1, for a definition). Indeed, taking β1 = 1, µ1 =

1 6



x2 2 , β2

= β, µ2 = µ, we have:

µ ¶ 1 x2 Dp∗ (Pn ) ≤ K q, β, 1, µ, − · Dp (Pn ) 6 2

STATISTICAL PROPERTIES OF GENERALIZED DISCREPANCIES

³ where q −1 + p−1 = 1. K q, β, 1, µ, 16 −

x2 2

15

´ is strictly positive since it is greater

than max {β, β s }. Moreover, the generalized Lp −star discrepancy can be minorized ° ° ° ° as Dp∗ (Pn ) ≥ ° |Pn ∩[0,x)| − Vol ([0, x))° p and the latter converges to a strictly n L

positive constant as long as n → ∞. To prove (ii), we just need to show that for any δ there exists a λ such that: sup P

©√

n∈N

ª nDp (Pn ) ≥ λ ≤ δ.

This derives from the application of (5.2): ½

λ P Dp (Pn ) ≥ √ n

¾ ≤P

 X 

µ p

∗ [D∞ (Pn,u )] ≥

u6=∅

 ¶p 

λ √

½ p ∗ P [D∞ (Pn,u )] ≥

X

1 d (2 − 1) u6=∅ ½ ¾ X m ∗ ≤ P D∞ (Pn,u ) ≥ √ , n

≤ (5.3)



K n µ

λ √

¶p ¾

K n

u6=∅

where we have set m = λ/

h¡ ¢1 i 2d − 1 p K .

Here, an inequality of Kiefer and

Wolfowitz (see [KW58]) (indeed a multivariate extension of the Dvoretzky-KieferWolfowitz or DKW inequality) implies the existence of positive constants c and c0 such that: ¾ ½ © ª m ∗ ≤ c0 · exp −cm2 P D∞ (Pn,u ) ≥ √ n

(5.4)

for all n > 0, m ≥ 0 and u. Therefore, from (5.3) and (5.4): ½

λ P Dp (Pn ) ≥ √ n

¾

( ¡ d ¢ ≤ c · 2 − 1 · exp −

)

cλ2

0

2

(2d − 1) p K 2

and the RHS is independent of n. As concerns (iii), remark that



∗ n · D∞ (Pn,u ) is exactly the empirical process

u

of the sample Pn,u on [0, 1] and then, by the LIL (see e.g. [Kie61], [FW94], p. 19) we have

√ lim sup n→∞

∗ n · D∞ (Pn,u ) 1 √ ≤√ . 2 ln ln n

STATISTICAL PROPERTIES OF GENERALIZED DISCREPANCIES

16

This yields the result. Proof of Proposition 2.3. First of all, we recall the definitions of the empirical ⊗|u|

measures λ⊗d n and λn

for u ⊆ {1, . . . , d} and the associated integrals:

(5.5)

P 1 λ⊗d n = n z∈Pn δz , R P 1 λ⊗d f dλ⊗d n f = n = n z∈Pn f (z) ,

P ⊗|u| λn = n1 z∈Pn δzu , R P ⊗|u| ⊗|u| λn f = f dλn = n1 z∈Pn f (zu ) , u

where δz is the Dirac measure in z and zu is the projection of z on [0, 1] . Remarking that Z

Y

0

µ (xj ) = λ

⊗d

gu (x, Z) =

j∈u

[0,1]d

gu (x, z) dz,

(2.1) can be written as:  Dp (Pn ) = 

(5.6)

X

Z β p|u| · [0,1]d

u6=∅

Since, by Proposition 2.1−(ii)  √

nDp (Pn ) = 

X



nDp (Pn ) = OP (1), (5.6) becomes:

Z β p|u| ·

u6=∅

1/p ¯¡ ⊗d ¯ ¢ p ¯ λ − λ⊗d gu (x, Z)¯ dx . n

[0,1]d

1/p ¯ ¯√ ¡ ⊗d ¢ p ¯ n λ − λ⊗d gu (x, Z)¯ dx , n

and we are led to consider the asymptotic behavior of the empirical process (see [vdV98], p. 266): Z|u| n (xu ) =

¢ √ ¡ ⊗d n λ − λ⊗d gu (x, Z) . n d

We need to show that, uniformly in x ∈ [0, 1] and for any u ⊆ {1, . . . , d}, |u|

Zn (xu ) converges in the space of bounded functions towards a stochastic integral with respect to a |u| −dimensional pinned Brownian sheet. In order to do so, we show that the class of functions: n o d F = gu (x, ·) , for any x ∈ [0, 1] , u ⊆ {1, . . . , d} is λ⊗d −Donsker (see [vdVW96], p. 81). The elements of F are linear combinations Q Q of a finite number of elements of the form j∈u1 (µ0 (xj ) + xj ) · j∈u2 1{xj >·j } for

STATISTICAL PROPERTIES OF GENERALIZED DISCREPANCIES

17

u1 ∩ u2 = ∅ and u1 ∪ u2 = u ⊆ {1, . . . , d}. From Example 2.1.3 in [vdVW96], nQ o u2 1 , x ∈ [0, 1] , u ⊆ {1, . . . , d} is a λ⊗d −Donsker class; the propu 2 {x >· } 2 j j j∈u2 o nQ u1 0 , u ⊆ {1, . . . , d} erty is preserved by multiplication with the class (µ (x ) + x ) , x ∈ [0, 1] j j u 1 1 j∈u1 (remark that µ0 (·) ∈ L∞ ([0, 1]) and is therefore bounded by a constant) and by linear combination. |u|

Therefore, Zn (xu ) converges uniformly over F to the stated limit, and simple manipulations show that Z|u| (xu ) can be expressed as: Z |u|

Z

(xu ) =

[0,1]d

gu (x, y) dB (y) ,

where B denotes a pinned Brownian sheet (the multidimensional analogue of a Brownian bridge). The Gaussian process Z|u| (xu ) is characterized by the same |u|

mean and covariance of Zn (xu ). Now, for finite p, the function k(fu )k : f 7→

hP u6=∅

β p|u|

R

p

[0,1]

d

|fu | dx

i1/p

([Hic98a], p. 300) is continuous and therefore the result derives from the Continuous Mapping Theorem (see [Kal97], Theorem 3.27). When p = ∞, on the other hand, we have to apply the Argmax Continuous Mapping Theorem ([vdV98], Corollary 5.58). The limiting process has continuous sample paths and maxima of Gaussian processes are unique by Lemma 2.6 in [KP90]. As concerns the uniform ´ ³ |u| tightness of the maximum of Zn (xu ) , it means that n

¯√ ¡ ¯ ¢ sup ¯ n λ⊗d − λ⊗d gu (x, Y)¯ = OP (1) , n x

that is for any ε there exists an M such that: ½ ¾ ¯√ ¡ ¯ ¢ ¯ sup P sup ¯ n λ⊗d − λ⊗d (x, Y) > M ≤ ε. g u n

(5.7)

n∈N

x

First we decompose gu (x, y) as:      Y £ X  Y Y£ ¤ ¤ xj − 1{xj >yj } µ0 (xj ) · µ0 (xj ) + xj − 1{xj >yj } = gu (x, y) =    0 0 0 j∈u

u ⊆u

j∈u\u

j∈u

STATISTICAL PROPERTIES OF GENERALIZED DISCREPANCIES

18

and we write: ¯√ ¡ ¯ ¢ gu (x, Y)¯ sup ¯ n λ⊗d − λ⊗d n x ¯ ¯    ¯X  Y Y £ ¯¯  √ ¡ ¯ ¤ ¢ ⊗d ⊗d 0 ¯ = sup ¯ xj − 1{xj >yj } ¯¯ µ (xj ) · n λ − λn ¯   0  x ¯ 0 0 j∈u u ⊆u j∈u\u ¯  ¯ ¯ ¯ ¯ ¯ Y ¯ Y £ ¯¯ X ¯ ¯ ¯ ¢ ¤ ¡ √ ¯ x − 1 ≤ sup ¯¯ µ0 (xj )¯¯ · sup ¯¯ n λ⊗d − λ⊗d {xj >yj } n  0 j ¯¯ x 0 ¯ xu0 ¯ j∈u u0 ⊆ u u\u ¯j∈u\u0 ° °  ° ° Y £ X ° ° ¤ ¡ ¢ √ 0 u\u | ° ⊗d ⊗d | ≤ m · ° n λ − λn xj − 1{xj >yj } ° °   ° ° ∞ j∈u0 u0 ⊆ u L

where we have used the fact that supxj |µ0 (xj )| ≤ m. Therefore (5.7) becomes: ½

¯ ¯√ ¡ ¢ sup P sup ¯ n λ⊗d − λ⊗d gu (x, Y)¯ > M n

¾

x

n∈N

° °   ° ° Y £  ° ° ¤ ¡ ¢ √ ⊗d ⊗d ° x − 1 m|u\u | · ° n λ − λ ≤ sup P > M j {x >y } n j j ° °  0  n∈N  0 ° ∞ ° j∈u u ⊆u L ° °   ° Y £  X ° °√ ¡ ⊗d ° ¤ ¢ M ⊗d ° P ° · − 1 . ≤ sup n λ − λ > {·j >yj } n ° °  0 j 2|u| m|u\u0 |  n∈N 0 ° °  X

0

j∈u

u ⊆u

L∞

Using inequality (5.4), we get: ½ ¾ ¯√ ¡ ¯ ¢ ¯ sup P sup ¯ n λ⊗d − λ⊗d g (x, Y) > M u n x

n∈N

X

≤ sup n∈N

u0 ⊆ u

X

°  ° ° Y £ ° °¡ ⊗d ° ¢ ¤ ⊗d ° P ° λ − λ x − 1 j {xj >·j } n ° °  0 ° ° j∈u

½

2

M ≤ c · exp −c · 2|u| 2|u\u0 | 2 m u0 ⊆ u 0

Setting M =

2|u| max{1,m|u| } √ c

r ·

³ ln

(

¾ 0

|u|

≤c ·2

2|u| c0 ε

 

>

M √ 2|u| m|u\u0 | n 

L∞

M2 © ª · exp −c · 2|u| 2 max 1, m2|u|

) .

´ we get the uniform tightness required.

Proof of Proposition 3.1. The proof is trivial and uses repeated application of the as version of the Slutsky Theorem (see [Dav94], p. 286, Theorem 18.8 (i)),

STATISTICAL PROPERTIES OF GENERALIZED DISCREPANCIES

19

starting from the fact that, under the alternative: Z 1 X gu (x, z) = P∗n gu (x, Z) −→ gu (x, z) dP∗ (z) = P∗ gu (x, Z) , n→∞ [0,1]d n

P∗ −as.

z∈Pn

Proof of Proposition 3.2. Our strategy of proof is to link the asymptotic p

distribution of Dp (Pn ) to that of [Dp (Pn )] using the standard delta method, then p

to express the asymptotic distribution of [Dp (Pn )] in terms of an empirical process using the functional delta method (see [vdV98], Chapter 20) and then to derive the asymptotic distribution of this empirical process. We divide the proof in six steps.

(1) We rewrite (2.1) as: ¯ ¯p 1/p ¯ ¯ Y ¯ ¯ ¯ Dp (Pn ) =  β |u|p · µ0 (xj ) − Z (xu ; P∗n )¯¯ dxu  ¯ u [0,1] ¯ ¯ j∈u u6=∅ 

(5.8)

X

Z

where 1 X gu (x, z) n z∈Pn Z = gu (x, z) dP∗n (z) .

Z (xu ; P∗n ) =

[0,1]d

From Proposition 3.1, as long as n → ∞, Dp (Pn ) converges to Dp (P∞ ) R with Z (xu ; P∗ ) = [0,1]d gu (x, z) dP∗ (z): ¯ ¯p 1/p ¯ ¯ Y ¯ ¯ ¯ β |u|p · Dp (P∞ ) =  µ0 (xj ) − Z (xu ; P∗ )¯¯ dxu  . ¯ u [0,1] ¯j∈u ¯ u6=∅ 

(5.9)

X

Z

Remark that Z (xu ; P∗n ) is a random perturbation of Z (xu ; P∗ ) and √

n [Z (xu ; P∗n ) − Z (xu ; P∗ )] =

is an empirical process.



n (P∗n − P∗ ) gu (x, Z)

STATISTICAL PROPERTIES OF GENERALIZED DISCREPANCIES

20

(2) First of all, applying the standard delta method to the function φ (x) = xp , we get: √

p

p

n ([Dp (Pn )] − [Dp (P∞ )] ) =

(5.10) √

n (Dp (Pn ) − Dp (P∞ )) =



p−1

np [Dp (P∞ )]

· (Dp (Pn ) − Dp (P∞ )) + oP (1)

√ p p n ([Dp (Pn )] − [Dp (P∞ )] ) p−1

p [Dp (P∞ )]

+ oP (1) . p

p

Therefore, we are led to study the behavior of ([Dp (Pn )] − [Dp (P∞ )] ). √ p p Using (5.8) and (5.9), n ([Dp (Pn )] − [Dp (P∞ )] ) becomes: √

p

p

n ([Dp (Pn )] − [Dp (P∞ )] )

(5.11) =

X

β |u|p ·



¯ ¯p ¯ ¯p  ¯ ¯ ¯  ¯¯ Y Y ¯ ¯ ¯ 0 ∗ ¯ 0 ∗ ¯ ¯ ¯ n µ (xj ) − Z (xu ; Pn )¯ − ¯ µ (xj ) − Z (xu ; P )¯ dxu . ¯ u  [0,1] ¯ ¯ ¯ ¯  Z

j∈u

u6=∅

j∈u

(3) Then, we use the functional delta method (see van der Vaart, 1998, Chapter 20) to approximate the previous formula. Define the statistical functional φu as:

¯ ¯p ¯ ¯ ¯Y 0 ¯ ¯ φu (G) = µ (xj ) − G (xu )¯¯ dxu . ¯ u [0,1] ¯j∈u ¯ Z

The difference φu (Z (xu ; P∗ )) − φu (Z (xu ; P∗n )) appears in (5.11) and it is interesting to approximate it through a simpler one (the objective is to obtain a functional that is linear in (P∗n − P∗ ), that is an empirical process). The right way to simplify this formula is to show that this functional is Hadamard differentiable at G = Z (xu ; P∗ ), and the same holds for any u 6= ∅.8 Therefore, we have to show that, for every gt → g, when t ↓ 0,

8We say that map φ : D → E defined on a subset D of a normed space D containing θ, is φ φ

Hadamard differentiable at θ if there exists a continuous, linear map φ0θ : D → E such that

° ° ° φ (θ + tht ) − φ (θ) − φ0 (h)° → 0 ° ° θ t E

as t ↓ 0 and for every ht → h (such that θ + tht is contained in the domain of φ for all small t > 0). Loosely speaking, this means that the difference 1t · [φ (θ + tht ) − φ (θ)] can be approximated by the linear function φ0θ (h), whose asymptotic behavior is often much simpler to study.

STATISTICAL PROPERTIES OF GENERALIZED DISCREPANCIES

21

there exists a function φ0u (g) such that: ° ° ° φu (G + t · gt ) − φu (G) ° 0 ° − φu (g)° ° ° → 0. t

(5.12)

In this formula, the reader can identify t = √1n , G (xu ) = Z (xu ; P∗ ), √ (G + t · gt ) (xu ) = Z (xu ; P∗n ), gt (xu ) = n [Z (xu ; P∗n ) − Z (xu ; P∗ )]. Moreover, k·k is the uniform norm k·kL∞ and 

Z φ0u (g) = −p ·

 u

[0,1]

Y

p−1 µ0 (xj ) − G (xu )

g (xu ) dxu .

j∈u

Now, φu (Z (xu ; P∗ )) − φu (Z (xu ; P∗n )) becomes: ¯ ¯p ¯ ¯p  ¯ ¯ ¯  ¯¯ Y Y ¯ ¯ ¯ 0 ∗ ¯ 0 ∗ ¯ ¯ ¯ µ (x ) − Z (x ; P ) − µ (x ) − Z (x ; P ) j u j u n ¯ ¯ ¯ ¯  dxu [0,1]u ¯ ¯ ¯ ¯

Z

j∈u

j∈u

¯ ¯p ¯ ¯p ¯ ¯ ¯ ¯ Z Z ¯Y 0 ¯ ¯Y 0 ¯ ¯ ¯ ¯ = µ (xj ) − (G + t · gt ) (xu )¯ dxu − µ (xj ) − G (xu )¯¯ dxu ¯ ¯ u u [0,1] ¯j∈u [0,1] ¯j∈u ¯ ¯ ¯ ¯p ¯p ¯ ¯ ·¯ ¸ Z ¯ ¯Y 0 ¯ ¯ (t·gt )(xu ) ¯ ¯ ¯ ¯ ¢ ¡ Q − 1 dxu = µ (x ) − G (x ) · 1 − j u ¯ ¯ ¯ µ0 (xj )−G(xu ) ¯ [0,1]u ¯j∈u ¯ j∈u (5.13) Z =

¯ ¯p ¯ ¯ ·µ ¶p ¸ ¯Y 0 ¯ (t·g )(x ) t u ¯ ¯ ¢ − 1 dxu µ (xj ) − G (xu )¯ · 1 − ¡Q ¯ µ0 (xj )−G(xu ) [0,1]u ¯j∈u ¯ j∈u

STATISTICAL PROPERTIES OF GENERALIZED DISCREPANCIES

where we take t small enough to have (t · gt ) (xu ) ≤

³Q j∈u

22

´ µ0 (xj ) − G (xu ) .

Now, using (5.13), (5.12) can be majorized as: ¯p ¯Q ¯p o n¯Q °R ¯ ¯ ¯ ¯ ° 0 0 µ (x ) − (G + t · g ) (x ) − µ (x ) − G (x ) dxu ¯ ¯ ¯ u j t u j u ¯ ° [0,1] j∈u j∈u ° ° t ° (5.14) Z

 

+p · [0,1]u

Y

p−1 µ0 (xj ) − G (xu )

j∈u

° ° ° ° g (xu ) dxu ° ° °

¯ ° ¯p °Z ¯Y ¯ ·µ ¶p ¸ ° ¯ ¯ (t·gt )(xu ) 0 −1 ¯ ¯ ¡ ¢ Q µ · t 1 − − 1 dxu =° (x ) − G (x ) j u ¯ ° ¯ µ0 (xj )−G(xu ) ° [0,1]u ¯j∈u ¯ j∈u °  p−1 ° Z ° Y ° 0  g (xu ) dxu ° µ (xj ) − G (xu ) +p · u ° [0,1] j∈u ° ° ¯ ¯p °Z ¯ ¯ ¶ µ ° ¯Y 0 ¯ (t·gt )(xu ) −1 ° ¯ ¯ ¡ ¢ =° + Et (u, x) dxu µ (xj ) − G (xu )¯ · t −p Q ¯ µ0 (xj )−G(xu ) ° [0,1]u ¯j∈u ¯ j∈u °  p−1 ° Z ° Y ° 0  +p · µ (xj ) − G (xu ) g (xu ) dxu ° ° [0,1]u j∈u ° °  p−1 ° Z ° Y °  = °p · µ0 (xj ) − G (xu ) · (g (xu ) − gt (xu )) dxu u ° [0,1] j∈u ° ¯ ¯p ° ¯ ¯ ° Z Y ¯ ¯ ° 0 ¯ ¯ ° +t−1 · µ (x ) − G (x ) · E (u, x) dx j u ¯ t u° ¯ [0,1]u ¯j∈u ¯ ° ° ° ¯ ¯ °Z ° ¯ ¯p−1 ° ° Y ¯ ¯ ° ° ¯ ≤p·° µ0 (xj ) − G (xu )¯¯ · |g (xu ) − gt (xu )| dxu ° ¯ ° [0,1]u ¯ ° ¯ j∈u ° ° ° ¯ ¯p ° ° ¯ ¯ ° Z ° −1 ¯Y 0 ¯ ° ° ¯ ¯ + °t · µ (xj ) − G (xu )¯ · Et (u, x) dxu ° ¯ ° u [0,1] ¯j∈u ° ¯ ° where Et is defined as µ Et (u, x) , 1 − ¡Q

(t·gt )(xu ) j∈u

¶p ¢

µ0 (xj )−G(xu )

− 1 + p ¡Q

(t·gt )(xu ) j∈u

¢.

µ0 (xj )−G(xu )

STATISTICAL PROPERTIES OF GENERALIZED DISCREPANCIES

23

Now, (5.14) converges to 0, since: ° ° ¯ ¯p−1 °Z ° ¯Y ¯ ° ° ¯ ¯ ° ° 0 ¯ ¯ µ (x ) − G (x ) · |g (x ) − g (x )| dx ° ° j u u t u u ¯ ° [0,1]u ¯¯ ° ¯ j∈u ° ° ° ° ¯ ¯p−1 ° °Z ¯Y ¯ ° ° ¯ ¯ ° ° 0 ¯ ¯ µ dx (x ) − G (x ) ≤° ° · kg (xu ) − gt (xu )k → 0 u j u ¯ ° ° [0,1]u ¯¯ ¯ j∈u ° ° and ° ¯ ¯p ° ° ¯ ° ¯Y Z ° ¯ ¯ ° −1 0 ¯ ¯ °t · µ (xj ) − G (xu )¯ · Et dxu ° ° ° ¯ [0,1]u ¯j∈u ° ¯ ° ° ¯ ¯p ° °Z ¯Y ¯ ° ° ¯ ¯ ° 0 ¯ ¯ dxu ° · t−1 kEt (u, x)k ≤° µ (x ) − G (x ) j u ° ¯ ¯ ° ° [0,1]u ¯j∈u ¯ ° → 0. This means that (5.11) becomes: √

p

p

n ([Dp (Pn )] − [Dp (P∞ )] )

=

X

β |u|p ·



u6=∅

=

X

β |u|p ·

u6=∅

= −p ·

X



n [φu (Z (xu ; P∗n )) − φu (Z (xu ; P∗ ))] · µ ¶ ¸ √ n [Z (xu ; P∗n ) − Z (xu ; P∗ )] √ n φu Z (xu ; P∗ ) + − φu (Z (xu ; P∗ )) n 

Z



β |u|p ·

u6=∅

[0,1]u

Y

p−1 µ0 (xj ) − Z (xu ; P∗ )

j∈u

(5.15) ·



n [Z (xu ; P∗n ) − Z (xu ; P∗ )] dxu + oP (1) .

(4) Let F ∗ be the cdf of P∗ . Then, we show that: (5.16)



D

|u|

n (Z (xu ; P∗n ) − Z (xu ; P∗ )) −→ ZF ∗ (xu ) ,

STATISTICAL PROPERTIES OF GENERALIZED DISCREPANCIES

24

for any u ⊆ {1, . . . , d}, where Z |u|

ZF ∗ (xu ) =

[0,1]d

gu (x, y) dBF ∗ (y) ,

and BF ∗ is the centered Gaussian process characterized by variance: Cov [BF ∗ (x) , BF ∗ (z)] = F ∗ (x ∧ z) − F ∗ (x) · F ∗ (z) . The method is the same used in the proof of Proposition 2.3 with the same class of functions F and, indeed, Z|u| can be obtained setting P∗ = λ⊗d . (5) Combining (5.10), (5.15) and (5.16), we get the final distribution. |u|

(6) The variance of N can be computed substituting the expression of ZF ∗ (xu ) into N and exchanging the order of integration: N =p·

X

p−1

β |u|p · [0,1]d

u6=∅



Z

p

= [0,1]d

"Z

Z

X

[(P − P∗ ) gu (x, Z)]

· [0,1]d

p−1

u6=∅

[0,1]d

gu (x, y) dBF ∗ (y) dxu 

Z β |u|p ·

#

[(P − P∗ ) gu (x, Z)]

· gu (x, y) dxu  dBF ∗ (y) .

This is a stochastic integral with respect to a Gaussian process and V (N ) can be computed as in [vdV98], p. 269.

6. Graphs References [Ang83]

J.E. Angus, On the asymptotic distribution of Cram´ er-von Mises one-sample test statistics under an alternative, Comm. Statist. A—Theory Methods 12 (1983), no. 21, 2477–2482. MR 85c:62040

[AS87]

R.J. Adler and G. Samorodnitsky, Tail behaviour for the suprema of Gaussian processes with applications to empirical processes, Ann. Probab. 15 (1987), no. 4, 1339– 1351. MR 88j:60073

[CS05a]

C. Choirat and R. Seri, The asymptotic distribution of quadratic discrepancies, Proceedings of MC2QMC (D. Talay and H. Niederreiter, eds.), Springer Verlag, 2005, Forthcoming.

[CS05b]

, Statistical properties of quadratic discrepancies, Working paper (2005).

STATISTICAL PROPERTIES OF GENERALIZED DISCREPANCIES

25

2 5

1

n = 25 0

0 0.00

0.25

0.50

0.75

1.00

1.25

0.0

0.5

1.0

1.5

2.0

0.0

0.5

1.0

1.5

2.0

0.0

0.5

1.0

1.5

2.0

0.0

0.5

1.0

1.5

2.0

0.0

0.5

1.0

1.5

2.0

2 5

1

n = 50 0

0 0.00

0.25

0.50

0.75

1.00

1.25 2

5

1

n = 100 0

0 0.00

0.25

0.50

0.75

1.00

1.25 2

5

1

n = 200 0

0 0.00

0.25

0.50

0.75

1.00

1.25 2

5

1

n = 400 0

0 0.00

0.25

0.50

0.75

1.00

1.25

Density

Q−Q Plot

2

Figure 6.1. Centered n · [D2 (Pn )] for d = 1

2.5

2

n = 25

0.0 0

0.0

0.5

1.0

1.5

2

n = 50

0.0

0.5

1.0

1.5

2.0

2.5

0.0

0.5

1.0

1.5

2.0

2.5

0.0

0.5

1.0

1.5

2.0

2.5

0.0

0.5

1.0

1.5

2.0

2.5

0.0

0.5

1.0

1.5

2.0

2.5

2.5 0.0

0

n = 100

0.0

0.5

1.0

1.5 2.5

2

0.0 0

0.0

0.5

1.0

1.5 2.5

2

n = 200

0.0 0

n = 400

0.0

0.5

1.0

1.5 2.5

2

0.0 0

0.0

0.5

1.0

1.5

Density

Q−Q Plot

2

Figure 6.2. Centered n · [D2 (Pn )] for d = 2

[Dav94]

J. Davidson, Stochastic limit theory, Advanced Texts in Econometrics, The Clarendon Press Oxford University Press, New York, 1994. MR 97k:60002

STATISTICAL PROPERTIES OF GENERALIZED DISCREPANCIES

2.5

2

n = 25

26

0.0 0

0.0

0.5

1.0

1.5

2.0

2.5

0.5

1.0

1.5

2.0

2.5

3.0

0.0

0.5

1.0

1.5

2.0

2.5

3.0

0.0

0.5

1.0

1.5

2.0

2.5

3.0

0.0

0.5

1.0

1.5

2.0

2.5

3.0

0.0

0.5

1.0

1.5

2.0

2.5

3.0

2.5

2

n = 50

0.0

0.0 0

n = 100

0.0

0.5

1.0

1.5

2.0

2.5 2.5

2

0.0 0

0.0

0.5

1.0

1.5

2.0

2.5 2.5

2

n = 200

0.0 0

n = 400

0.0

0.5

1.0

1.5

2.0

2.5 2.5

2

0.0 0

0.0

0.5

1.0

1.5

2.0

2.5

Density

Q−Q Plot

2

Figure 6.3. Star n · [D2 (Pn )] for d = 2

0.5

10

n = 25

0 0.0

0

2

4

6

8

0.5

0.0

2.5

5.0

7.5

10.0

2.5

5.0

7.5

10.0

2.5

5.0

7.5

10.0

2.5

5.0

7.5

10.0

5.0

7.5

10.0

10

n = 50

0 0.0

0

2

4

6

8

0.5

0.0 10

n = 100

0 0.0

0

2

4

6

8

0.5

0.0 10

n = 200

0 0.0

0

2

4

6

8

0.5

0.0 10

n = 400

0 0.0

0

2

4

6

8

0.0

2.5

Density

Q−Q Plot

2

Figure 6.4. Symmetric n · [D2 (Pn )] for d = 2

[Deh89]

H. Dehling, Complete convergence of triangular arrays and the law of the iterated logarithm for U -statistics, Statist. Probab. Lett. 7 (1989), no. 4, 319–321. MR 90h:60028

STATISTICAL PROPERTIES OF GENERALIZED DISCREPANCIES

5

1

n = 25 0

0 0

1

2

3

4

1

2

3

4

4

5

1

2

3

4

5

1

2

3

4

5

1

2

3

4

5

1

2

3

4

5

0 0

1

2

3

4 5

1

n = 200

0 0

1

2

3

4 5

1

n = 400 0

3

5

1

0

2

0 0

n = 100 0

1 5

1

n = 50 0

27

0 0

1

2

3

4

Density

Q−Q Plot

2

Figure 6.5. Centered n · [D2 (Pn )] for d = 5

0.50 10 0.25

n = 25

0 0.0

2.5

5.0

7.5

10.0

12.5

0.0

2.5

5.0

7.5

10.0

12.5

15.0

17.5

2.5

5.0

7.5

10.0

12.5

15.0

17.5

2.5

5.0

7.5

10.0

12.5

15.0

17.5

2.5

5.0

7.5

10.0

12.5

15.0

17.5

2.5

5.0

7.5

10.0

12.5

15.0

17.5

0.50 10 0.25

n = 50

0 0.0

2.5

5.0

7.5

10.0

12.5

0.0

0.50 10

n = 100

0.25 0 0.0

2.5

5.0

7.5

10.0

12.5

0.0

0.50 10

n = 200

0.25 0 0.0

2.5

5.0

7.5

10.0

12.5

0.0

0.50 10

n = 400

0.25 0 0.0

2.5

5.0

7.5

10.0

12.5

0.0

Density

Q−Q Plot

2

Figure 6.6. Star n · [D2 (Pn )] for d = 5

[Doo01]

J.A. Doornik, Ox: An object-oriented matrix language, 4th ed., Timberlake Consultants Press, London, 2001.

STATISTICAL PROPERTIES OF GENERALIZED DISCREPANCIES

28

0.10 50 0.05

n = 25

0

10

20

30

40

50

60

0 10

20

30

40

50

60

20

30

40

50

60

20

30

40

50

60

20

30

40

50

60

30

40

50

60

0.10 50 0.05

n = 50

0

10

20

30

40

50

60

0 10

0.10 50

n = 100

0.05 0

10

20

30

40

50

60

0 10

0.10 50

n = 200

0.05 0

10

20

30

40

50

60

0 10

0.10 50

n = 400

0.05 0

10

20

30

40

50

60

0 10

20

Density

Q−Q Plot

2

Figure 6.7. Symmetric n · [D2 (Pn )] for d = 5

0.50

10

0.25

n = 25

0 −2

0

2

4

6

−2.5

0.50 0.25

n = 50

0

2

4

6

−2.5

0.50 0.25

7.5

10.0

12.5

0.0

2.5

5.0

7.5

10.0

12.5

0.0

2.5

5.0

7.5

10.0

12.5

0 0

2

4

6

−2.5

0.50

10

0.25

0 −2

0

2

4

6

−2

0.50

n = 400

5.0

10

−2

n = 200

2.5

0 −2

n = 100

0.0

10

0

2

4

6

10

0.25

0 −2

0

2

4

6

−2.5

0.0

2.5

5.0

7.5

10.0

12.5

Q−Q Plot

Density

2

Figure 6.8. Convergence of [D2 (Pn )] towards a normal random variable

[Fin71]

H. Finkelstein, The law of the iterated logarithm for empirical distributions, Ann. Math. Statist. 42 (1971), 607–615. MR 44 #4803

STATISTICAL PROPERTIES OF GENERALIZED DISCREPANCIES

29

[FMW02] K.-T. Fang, C.-X. Ma, and P. Winker, Centered L2 -discrepancy of random sampling and Latin hypercube design, and construction of uniform designs, Math. Comp. 71 (2002), no. 237, 275–296 (electronic). MR 2002h:65024 [FW94]

K.-T. Fang and Y. Wang, Number-theoretic methods in statistics, Monographs on Statistics and Applied Probability, vol. 51, Chapman & Hall, London, 1994. MR 95g:65189

[GKLZ01] E. Gin´ e, S. Kwapie´ n, R. LataÃla, and J. Zinn, The LIL for canonical U -statistics of order 2, Ann. Probab. 29 (2001), no. 1, 520–557. MR 2002k:60083 [HHW03] H.S. Hong, F.J. Hickernell, and G. Wei, The distribution of the discrepancy of scrambled digital (t, m, s)-nets, Math. Comput. Simulation 62 (2003), no. 3-6, 335–345. MR 2004e:11080 [Hic96]

F.J. Hickernell, Quadrature error bounds with applications to lattice rules, SIAM J. Numer. Anal. 33 (1996), no. 5, 1995–2016. MR 97m:65050

[Hic97]

, Erratum: “Quadrature error bounds with applications to lattice rules” [SIAM J. Numer. Anal. 33 (1996), no. 5, 1995–2016;], SIAM J. Numer. Anal. 34 (1997), no. 2, 853–866. MR 2000h:65042

[Hic98a]

, A generalized discrepancy and quadrature error bound, Math. Comp. 67 (1998), no. 221, 299–322. MR 98c:65032

[Hic98b]

, Lattice rules: how well do they measure up?, Random and quasi-random point sets, Lecture Notes in Statist., vol. 138, Springer, New York, 1998, pp. 109–166. MR 2000b:65007

[Hic99]

, Goodness-of-fit statistics, discrepancies and robust designs, Statist. Probab. Lett. 44 (1999), no. 1, 73–78.

[HJK98]

J. Hoogland, F. James, and R. Kleiss, Quasi-Monte Carlo, discrepancies and error estimates, Monte Carlo and quasi-Monte Carlo methods 1996 (Salzburg), Lecture Notes in Statist., vol. 127, Springer, New York, 1998, pp. 266–276. MR 99d:65070

[HK96a]

J. Hoogland and R. Kleiss, Discrepancy-based error estimates for quasi-monte carlo. I: General formalism, Comput. Phys. Comm. 98 (1996), no. 1–2, 111–127.

[HK96b]

, Discrepancy-based error estimates for quasi-monte carlo. II: Results in one dimension, Comput. Phys. Comm. 98 (1996), no. 1–2, 128–136.

[HK97]

, Discrepancy-based error estimates for quasi-monte carlo. III: Error distribution and central limits, Comput. Phys. Comm. 101 (1997), no. 1–2, 21–30.

[JHK97]

F. James, J. Hoogland, and R. Kleiss, Multidimensional sampling for simulation and integration: Measures, discrepancies and quasi-random numbers, Comput. Phys. Comm. 99 (1997), no. 2–3, 180–220.

STATISTICAL PROPERTIES OF GENERALIZED DISCREPANCIES

[Kal97]

30

O. Kallenberg, Foundations of modern probability, Probability and its Applications (New York), Springer-Verlag, New York, 1997. MR 99e:60001

[Kie61]

J. Kiefer, On large deviations of the empiric D. F. of vector chance variables and a law of the iterated logarithm, Pacific J. Math. 11 (1961), 649–660. MR 24 #A1732

[KP90]

J. Kim and D. Pollard, Cube root asymptotics, Ann. Statist. 18 (1990), no. 1, 191–219. MR 91f:62059

[KW58]

J. Kiefer and J. Wolfowitz, On the deviations of the empiric distribution function of vector chance variables, Trans. Amer. Math. Soc. 87 (1958), 173–186. MR 20 #5519

[Lee96a]

´ H. Leeb, The asymptotic distribution of diaphony in one dimension, GERAD - Ecole ´ des Hautes Etudes Commerciales, Montr´ eal (1996).

[Lee96b]

, A weak law for diaphony, RIST++, Research Institute for Software Technology, University of Salzburg (1996).

[Lee02]

, Asymptotic properties of the spectral test, diaphony, and related quantities, Math. Comp. 71 (2002), no. 237, 297–309 (electronic). MR 2002j:11086

[LFHL01] J.-J. Liang, K.-T. Fang, F.J. Hickernell, and R. Li, Testing multivariate uniformity and its applications, Math. Comp. 70 (2001), no. 233, 337–355. MR 2001f:62032 [LP03]

G. Leobacher and F. Pillichshammer, Bounds for the weighted Lp discrepancy and tractability of integration, J. Complexity 19 (2003), no. 4, 529–547. MR 2004f:65030

[Nie92]

H. Niederreiter, Random number generation and quasi-Monte Carlo methods, CBMSNSF Regional Conference Series in Applied Mathematics, vol. 63, Society for Industrial and Applied Mathematics (SIAM), Philadelphia, PA, 1992. MR 93h:65008

[Rag73]

M. Raghavachari, Limiting distributions of Kolmogorov-Smirnov type statistics under the alternative, Ann. Statist. 1 (1973), 67–73. MR 49 #11696

[ST96]

F. Schmid and M. Trede, An L1 -variant of the Cram´ er-von Mises test, Statist. Probab. Lett. 26 (1996), no. 1, 91–96. MR 97a:62109

[SW86]

G.R. Shorack and J.A. Wellner, Empirical processes with applications to statistics, Wiley Series in Probability and Mathematical Statistics: Probability and Mathematical Statistics, John Wiley & Sons Inc., New York, 1986. MR 88e:60002

[SW98]

I.H. Sloan and H. Wo´ zniakowski, When are quasi-Monte Carlo algorithms efficient for high-dimensional integrals?, J. Complexity 14 (1998), no. 1, 1–33. MR 99d:65384

[SWW04] I.H. Sloan, X. Wang, and H. Wo´ zniakowski, Finite-order weights imply tractability of multivariate integration, J. Complexity 20 (2004), no. 1, 46–74. MR 2004j:65034 [vdV98]

A.W. van der Vaart, Asymptotic statistics, Cambridge Series in Statistical and Probabilistic Mathematics, Cambridge University Press, Cambridge, 1998. MR 2000c:62003

STATISTICAL PROPERTIES OF GENERALIZED DISCREPANCIES

31

[vdVW96] A.W. van der Vaart and J.A. Wellner, Weak convergence and empirical processes, Springer Series in Statistics, Springer-Verlag, New York, 1996. MR 97g:60035 [vHKH97] A van Hameren, R. Kleiss, and J. Hoogland, Gaussian limits for discrepancies. I. Asymptotic results, Comput. Phys. Comm. 107 (1997), no. 1-3, 1–20. MR 99k:65009

` degli Studi dell’Insubria, Via Ravasi 2, 21100 Dipartimento di Economia, Universita Varese, Italy E-mail address: [email protected] ` degli Studi dell’Insubria, Via Ravasi 2, 21100 Dipartimento di Economia, Universita Varese, Italy E-mail address: [email protected]

Statistical Properties of Generalized Discrepancies

Here, an inequality of Kiefer and. Wolfowitz (see [KW58]) (indeed a multivariate extension of the Dvoretzky-Kiefer-. Wolfowitz or DKW inequality) implies the existence of positive constants c and c such that: (5.4). P. {. D∗. ∞ (Pn,u) ≥ m. √ n. } ≤ c · exp. {. −cm2. } for all n > 0, m ≥ 0 and u. Therefore, from (5.3) and (5.4):. P.

498KB Sizes 2 Downloads 202 Views

Recommend Documents

statistical properties of generalized discrepancies ...
vestigation of the asymptotic properties of D2 (Pn) and have derived two alternative statistics, strictly linked to this one, that can be used to test statistically the efficiency of numerical integration procedures. Recently, other authors have inve

Statistical Properties of Generalized Discrepancies and ...
Liang et al. ([11]) have started an investigation of the asymp- totic properties of D2 (Pn) and have derived two alternative statistics, strictly linked to this one, that can be used to test statisti- cally the efficiency of numerical integration pro

Generalized discrepancies on the sphere
2 Department of Quantitative Methods, School of Economics and Business Management, Universidad de Navarra, Pamplona,. Spain. Generalized discrepancies ...

Qualitative properties of generalized principal ...
[3] H. Berestycki, L. Nirenberg, S.R.S. Varadhan, The principal eigenvalue and maxi- mum principle for second-order elliptic operators in general domains, ...

On the Network Properties of Generalized Hypercubes
Due to advances in technology, it is now feasible to build large-scale parallel ... low degree, regularity, small diameter, large bisection width and high fault .... For any pair (u, v) and (u , v ) there is an auto- morphism σ of .... even me or m1

A statistical examination of the evolution and properties ...
Sep 22, 2009 - FreeBSD ports. ○ 12,010 ports. ○ 6,556 written in C. ○ 1,109,697 files. ○ 447,612 in C. AnsiC. C++. Shell. Perl. Java. PHP. Python. Lisp.

Statistical Properties of the Warped Discrete Cosine ...
of the TIMIT database (Dialect Region: North Midland) to form the train and test ... SNR (dB). Vowel Class Separability in Babble noise. WDCTC. MFCC. Figure 1: ...

Statistical Verification of Probabilistic Properties with ...
probabilistic model checking, PRISM [17], relies on iterative methods to verify properties with unbounded until. Each iteration involves a matrix–vector multi- plication, which in the worst case is O(n2), but often O(n) (for sparse models), where n

Statistical Properties and Memory of Excursions in ...
The correlation method applied to the return intervals shows a weak correlation for both groups with changes as the threshold q increases. The DFA analysis.

Statistical Properties and Memory of Excursions in ...
statistics of return intervals between long excursions above ... The DFA analysis confirms the presence ... behavior (shuffled records) is observed in original data.

Statistical Properties and Memory of Excursions in ...
cial indexes [13], x-ray solar flares [14], climate records. [15], or even long heartbeat ..... invariance in the nonstationarity of human heat rate. Phys. Rev Lett 2001 ...

A Generalized Links and Text Properties Based Forum ...
Digital Forensics Lab. Cryptography ... information on the software used to create them. Cai et al. .... 1) Signature Generation and Signature Based Clustering:.

Generalized image models and their application as statistical models ...
Jul 20, 2004 - exploit the statistical model to aid in the analysis of new images and .... classically employed for the prediction of the internal state xПtч of a ...

Exploring Systematic Discrepancies in DFT ... - ACS Publications
May 5, 2017 - agreement with prior literature findings after accounting for different .... of 35Cl |CQ| and ηQ for glycine hydrochloride, calculated with CRYSTAL software using the neutron structure ...... CrystEngComm 2016, 18, 6213−6232.

Exploring Statistical Properties for Semantic Annotation ...
kth most common word. Proof by ... tures,” J. Machine Learning Research, vol. 3, 2003. [6] David M. Blei, ... Statistical Learning in Computer Vision, 2004.

Properties of Water
electron presence. Electron density model of H2O. 1. How many hydrogen atoms are in a molecule of water? 2. How many oxygen atoms are in a molecule of ...

properties
Type. Property Sites. Address. Zip. Code. Location. East or West. Site. Acres. Main Cross Streets. Status. Price. Bldg. (GSF). Year. Built. 1 Building. Brady School.

Unsupervised Learning of Generalized Gamma ...
model (GΓMM) to implement an effective statistical analysis of .... models in fitting SAR image data histograms for most cases. [5]. ..... the greatest for large u.

2003_C_c_bmc_7-Mechanical Properties of Concrete_Reinforced ...
There was a problem previewing this document. Retrying... Download. Connect more apps... Try one of the apps below to open or edit this item. 2003_C_c_bmc_7-Mechanical Properties of Concrete_Reinforced with AR-Glass Fibers.pdf. 2003_C_c_bmc_7-Mechani

properties of materials
ENERGY-solar energy-solar cells,solar cooker-flat plate collector for power ... circuits-active,Reactive and Apparent power,Power factor,Resonance in R-L-C.

2001_C_a_pacrim_fly_ash_Mechanical Properties of Concrete with ...
2001_C_a_pacrim_fly_ash_Mechanical Properties of Concrete with Flyash.pdf. 2001_C_a_pacrim_fly_ash_Mechanical Properties of Concrete with Flyash.pdf.