Power in High-Dimensional Testing Problems Anders Bredahl Kock and David Preinerstorfer∗ University of Oxford, Aarhus University and CREATES ECARES, Universit´e libre de Bruxelles September 2017

Abstract Fan et al. (2015) recently introduced a method for increasing asymptotic power of tests in high-dimensional testing problems. If applicable to a given test, their power enhancement principle leads to an improved test that has the same asymptotic size, uniformly non-inferior asymptotic power, and is consistent against a strictly broader range of alternatives than the initially given test. We study under which conditions this method can be applied and show the following: In asymptotic regimes where the dimensionality of the parameter space is fixed as sample size increases, there often exist tests that can not be further improved by the power enhancement principle. When the dimensionality can increase with sample size, however, there typically is a range of “slowly” diverging rates for which every test with asymptotic size smaller than one can be improved with the power enhancement principle. We also address under which conditions the latter statement even extends to all rates at which the dimensionality increases with sample size.

1

Introduction

The effect of dimensionality on power properties of tests has witnessed a lot of research in recent years. One common goal is to construct tests with good asymptotic size and power properties for testing problems where the length of the parameter vector involved in the hypothesis to be tested increases with sample size. In the context of high-dimensional cross-sectional testing problems Fan et al. (2015) introduced a power enhancement principle, which essentially works as follows: given an initial test, one tries to find another test that has asymptotic size zero and is consistent against sequences of alternatives the initial test is not consistent against. If such an auxiliary test, a power enhancement component of the initial test, can be found, one can construct a test that has better asymptotic properties than the initial test. In particular, one can obtain a test that (i) has the same asymptotic size as the initial test, (ii) has uniformly non-inferior ∗ Address

for correspondence: [email protected] and [email protected].

1

asymptotic power and (iii) is consistent against all sequences of alternatives the auxiliary test is consistent against. As a consequence of (iii) the improved test is consistent against sequences of alternatives the initial test is not consistent against. Fan et al. (2015) illustrate the power enhancement principle by showing how an initial test based on a weighted Euclidean norm of an estimator can be made consistent against sparse alternatives, which it could previously not detect, by incorporating a power enhancement component based on the supremum norm of the estimator. The existence of a suitable power enhancement component in the specific situation they consider, however, does not answer the following general questions: • Under which conditions does a test admit a power enhancement component? • And, similarly, do there exist tests for which no power enhancement components exist? In this paper we address these questions in a general setup. In the sequel we call tests that do (not) admit a power enhancement component asymptotically (un)enhanceable. We first consider the classical asymptotic setting, where the dimension of the parameter vector being tested remains fixed as sample size tends to infinity. Under fairly weak assumptions on the model we prove (cf. Theorem 4.1) that in this framework tests exist that are asymptotically unenhanceable. That is, in such settings there exist tests that can not be further improved by the power enhancement principle. Furthermore, such tests exist with any asymptotic size α ∈ (0, 1]. The situation changes drastically when the dimension increases (unboundedly) with the sample size. Here we show (cf. Theorem 5.1) that if the models under consideration satisfy a mild “fixed-dimensional” (i.e., “marginal”) local asymptotic normality (LAN) assumption, then there always exist growth rates of the dimension of the parameter vector such that every test with asymptotic size less than one is asymptotically enhanceable. Furthermore, these growth rates can be chosen to be arbitrarily slow, but can not be chosen to be arbitrarily rapid in general. This is somewhat surprising, as one may have conjectured that the behavior from the fixed-dimensional case breaks down only when the dimension of the parameter vector increases sufficiently rapidly. We show, however, that when the dimension increases very quickly, the testing problem may become so difficult that no test has asymptotic power higher than its asymptotic size against any deviation from the null hypothesis (cf. Example 2). Then no power enhancement components exist and every test is asymptotically unenhanceable. Finally, we show in Theorem 5.4 that under a fairly natural additional assumption ruling out a behavior as in Example 2, even for any growth rate of the dimension of the parameter vector every test of asymptotic size less than one is asymptotically enhanceable. As a consequence, in many high-dimensional testing problems the power enhancement principle is applicable to any test with asymptotic size smaller than one. In particular, every test with asymptotic size smaller than one has “blind spots” of inconsistency that can be removed by applying the power enhancement principle. However, our results also imply that it is impossible to remove all of them (without sacrificing the size constraint). Therefore, in practice, one needs to think carefully about which removable “blind spots” of inconsistency of a given tests one 2

wants to eliminate. We would like to stress that even if a test is asymptotically enhanceable, the test might still be “optimal” within a restricted class of tests (e.g., satisfying certain invariance properties), or the test might still have “optimal detection properties” against certain subsets of the alternative. Hence, our findings are not in contradiction with such results. Instead, they provide an alternative perspective on power properties in high-dimensional testing problems.

1.1

Related literature

The setting we consider in our main results (Theorems 5.1 and 5.4) requires neither independently nor identically distributed data, and covers many practical situations of interest. On the other hand, for concrete high-dimensional testing problems, and under suitable assumptions on how fast the dimension of the parameter to be tested is allowed to increase with sample size, many articles have considered the construction of tests with good size and power properties. Testing problems in one- or two-sample multivariate location models are analyzed in Dempster (1958), Bai and Saranadasa (1996), Srivastava and Du (2008), Srivastava et al. (2013), Cai et al. (2014), and Chakraborty and Chaudhuri (2017); in this context the articles Pinelis (2010, 2014), where the asymptotic efficiency of tests based on different p-norms relative to the Euclidean-norm is studied, need to be mentioned. In regression models power properties of F-tests when the dimension increases with sample size have been investigated in Wang and Cui (2013), Zhong and Chen (2011), and Steinberger (2016). In the context of testing hypotheses on large covariance matrices properties of tests were studied in Ledoit and Wolf (2002), Srivastava (2005), Bai et al. (2009), and Onatski et al. (2013, 2014). For properties of tests for high-dimensional testing problems arising in spatial statistics we refer to Cai et al. (2013), Ley et al. (2015), and Cutting et al. (2017). An article that obtains results somewhat similar to ours is Janssen (2000), where local power properties of goodness-of-fit tests are studied. For such testing problems it is shown, among other things, that any test can have high local asymptotic power relative to its asymptotic size only against alternatives lying in a finite-dimensional subspace of the parameter space. Although related, our results are qualitatively different, because asymptotic enhanceability is an intrinsically non-local concept (cf. Remark 3.1), and because we do not consider testing problems with infinitely many parameters for any sample size. Instead we consider situations where the number of parameters can increase with sample size at different rates. Results on power properties of tests in situations where the sample size is fixed while the number of parameters diverges to infinity have been obtained in Lockhart (2016), who shows that in such scenarios the asymptotic power of invariant tests (w.r.t. various subgroups of the orthogonal group) against contiguous alternatives coincides with their asymptotic size.

3

2

Framework

The general framework in this article is a double array  Ωn,d , An,d , {Pn,d,θ : θ ∈ Θd }

for n ∈ N and d ∈ N,

(2.1)

where for every n ∈ N and d ∈ N the tuple (Ωn,d , An,d ) is a measurable space, i.e., the sample space, and {Pn,d,θ : θ ∈ Θd } is a set of probability measures on that space, i.e., the set of possible distributions of the data observed. For every d ∈ N the parameter space Θd is assumed to be a subset of Rd and to contain a neighborhood of the origin. The two indices n ∈ N and d ∈ N should be interpreted as “sample size” and as the “dimension of the parameter space”, respectively. Expectation w.r.t. Pn,d,θ is denoted by En,d,θ . We consider the situation where one wants to test (possibly after a suitable re-parameterization) whether or not the unknown parameter vector θ equals zero. Such problems have been studied extensively in the classical asymptotic framework where d is fixed and n → ∞, i.e., properties of sequences of tests for the testing problem H0 : θ = 0 ∈ Θd

against

H1 : θ ∈ Θd \ {0}

are studied in the sequence of experiments  Ωn,d , An,d , {Pn,d,θ : θ ∈ Θd }

for n ∈ N.

In contrast to such an analysis, the framework we are interested in is the more general situation in which d = d(n) is a non-decreasing sequence. More precisely, we study properties of sequences of tests for the sequence of testing problems H0 : θ = 0 ∈ Θd(n)

against

H1 : θ ∈ Θd(n) \ {0}

(2.2)

in the corresponding sequence of experiments (Ωn,d(n) , An,d(n) , {Pn,d(n),θ : θ ∈ Θd(n) })

for n ∈ N,

(2.3)

for all possible rates d(n) at which the dimension of the parameter space can increase with n. The following running example illustrates our framework (for several more examples we refer to the end of Section 5.1). Here, for i ∈ N, λi denotes Lebesgue measure on the Borel sets of Ri . Example 1 (Linear regression model). Consider the linear regression model yi = x0i,d θ + ui , i = 1, ..., n where θ ∈ Θd = Rd . One must distinguish between the cases where the covariates xi,d are fixed or random: Fixed covariates: Here the sample space Ωn,d equals Rn , An,d is the corresponding Borel σ-field, and xi,d = (x1i , . . . , xdi )0 for X = (xkl )∞ k,l=1 a given double array of real numbers. Assuming that 4

the error terms ui are i.i.d. with u1 ∼ F having λ1 -density f , it follows that yi has λ1 -density gyi (y) = f (y − x0i,d θ). Hence, Pn,d,θ is the distribution with λn -density g(y1 ,...,yn ) (z1 , ..., zn ) = Qn i=1 gyi (zi ). Random covariates: Here the sample space Ωn,d equals

×ni=1(R × Rd) and An,d is the corre-

sponding Borel σ-field. Letting the error terms be as in the case of fixed covariates, and the xi,d now be i.i.d. and independent of ui with distribution Kd on the Borel sets of Rd , we have that Pn,d,θ is the n-fold product of the measure with density f (y − x0 θ) w.r.t. (λ1 ⊗ Kd )(y, x).

3

Asymptotic enhanceability

Recently Fan et al. (2015) introduced a method – the power enhancement principle – that can be used to improve asymptotic power properties of tests in high-dimensional testing problems. After re-formulating their main idea in terms of tests instead of test statistics, a corresponding power enhancement principle can be formulated in our general context: Let d(n) be a nondecreasing sequence of natural numbers and let ϕn : Ωn,d(n) → [0, 1] be measurable, i.e., ϕn is a sequence of tests for (2.2) in (2.3). Suppose that it is possible to find another sequence of tests νn : Ωn,d(n) → [0, 1] with asymptotic size 0, i.e., lim sup En,d(n),0 (νn ) = 0,

(3.1)

n→∞

and so that νn is consistent against a sequence θn ∈ Θd(n) which the initial test ϕn is not consistent against, i.e., 1 = lim En,d(n),θn (νn ) > lim inf En,d(n),θn (ϕn ). n→∞

n→∞

(3.2)

In this case ϕn and νn can be combined into the test ψn = min(ϕn + νn , 1),

(3.3)

which has the following properties (as is easy to verify): 1. ψn has the same asymptotic size as ϕn . 2. ψn ≥ ϕn , implying that ψn has nowhere smaller power than ϕn . 3. ψn is consistent against the sequence of alternatives θn (which ϕn is not consistent against). This method of obtaining a sequence of tests ψn with improved asymptotic properties from a given sequence ϕn is applicable whenever νn with the above properties can be determined. A test ϕn for which there exists such a corresponding test νn , i.e., an enhancement component, will subsequently be called asymptotically enhanceable. For simplicity this is summarized in the following definition. 5

Definition 3.1. Given a non-decreasing sequence d(n) in N, a sequence of tests ϕn : Ωn,d(n) → [0, 1] is called asymptotically enhanceable, if there exists a sequence of tests νn : Ωn,d(n) → [0, 1] and a sequence θn ∈ Θd(n) so that (3.1) and (3.2) hold. The sequence νn will then be called an enhancement component of ϕn . Before we move on to formulate our main question, we make two observations: Remark 3.1. Any sequence θn as in Definition 3.1 must be such that Pn,d(n),θn and Pn,d(n),0 are not contiguous. Hence, asymptotic enhanceability as introduced in Definition 3.1 is a “non-local” property in the sense that whether or not a sequence of tests can be asymptotically enhanced, depends only on its power properties against sequences of alternatives that are not contiguous to Pn,d(n),0 . Remark 3.2. In the context of Definition 3.1 one could argue that instead of (3.1) one should require the stronger property Pn,d(n),0 (νn = 0) → 1 as n → ∞ (guaranteeing “better” size properties of the enhanced test min(ϕn +νn , 1) in finite samples). In particular, the constructions in Fan et al. (2015) (formulated in terms of test statistics) are based on a corresponding property. But note that if νn is a sequence of tests as in Definition 3.1, the sequence νn∗ = 1{νn ≥ 1/2} is a sequence of tests as in Definition 3.1 that furthermore satisfies Pn,d(n),0 (νn∗ = 0) → 1 as n → ∞. Hence, requiring existence of tests so that Pn,d(n),0 (νn = 0) → 1 holds instead of (3.1) would lead to an equivalent definition.

4

Main question

The power enhancement principle tells us how one can improve a sequence of tests ϕn provided an enhancement component νn is available. The obvious question now of course is: when does such an enhancement component actually exist? Similarly, in a situation where there are many possible enhancement components νn available (each improving power against a different sequence θn ), one can repeatedly apply the power enhancement principle. Then, the question arises: when should one stop enhancing? It is quite tempting to argue that one should keep enhancing until a test is obtained that is not enhanceable anymore. But this suggestion is certainly practical only if there exists a test that can not be asymptotically enhanced. Hence, a question that immediately arises is: Does there exist a sequence of tests with asymptotic size smaller than one that can not be asymptotically enhanced? Note that if the size requirement is dropped in the above question, then the answer is trivially yes, since one can then choose ϕn ≡ 1, a test that is obviously not asymptotically enhanceable. But this is of no practical use. One would expect the answer to the above question to depend on the dimensionality d(n). We first consider the fixed-dimensional case d(n) ≡ d ∈ N, in which it turns out that there often 6

exist sequences of tests that are not asymptotically enhanceable. We shall present a result that supports this claim in the i.i.d. case under an L2 -differentiability (cf. Definition 1.103 of Liese and Miescke (2008)) and a separability condition on the model: Assumption 1. For every n ∈ N it holds that n

Ωn,d =

× i=1

Ω,

An,d =

n O

A,

and Pn,d,θ =

i=1

n O

Pd,θ for every θ ∈ Θd ,

i=1

where Ω = Ω1,d , A = A1,d and Pd,θ = P1,d,θ . The family {Pd,θ : θ ∈ Θd } is L2 -differentiable at 0 with nonsingular information matrix. Furthermore, for every ε > 0 so that Θd contains a θ with kθk2 ≥ ε there exists a sequence of tests ψn,d (ε) : Ωn,d → [0, 1] so that as n → ∞ En,d,0 (ψn,d (ε)) → 0

and

inf

θ∈Θd :kθk≥ε

En,d,θ (ψn,d (ε)) → 1.

Assumption 1 often holds and sufficient conditions are discussed in chapter 10.2 of van der Vaart (2000) where a Bernstein-von Mises theorem is established under the same set of model assumptions. The proof idea of the subsequent result is taken from the proof of Lemma 10.3 in the same reference. Theorem 4.1. Let d(n) ≡ d for some d ∈ N and assume that Assumption 1 holds. Then, for every α ∈ (0, 1] there exists a sequence of tests with asymptotic size α that is not asymptotically enhanceable. The proof of Theorem 4.1 is given in Section 7.1. The theorem affirmatively answers the question raised above under weak assumptions in case the non-decreasing sequence d(n) is constant (eventually). Hence we have, at least in some generality, answered the question raised above for the fixed-dimensional case, and the question can thus be sharpened to: Does there exist a sequence of tests with asymptotic size smaller than one that can not be asymptotically enhanced if d(n) diverges with n? This now constitutes our main question. A natural first reaction could be to conjecture that, although the typical fixed-dimensional behavior described in the above theorem might not hold for some sequences d(n) that grow “too fast”, it still remains valid if d(n) diverges “slowly enough”. Surprisingly, we shall see in the following section that this naive conjecture is generally false.

5

Asymptotic enhanceability in high dimensions

In this section we present our main results concerning the question raised in Section 4. The setting described in Section 2 is very general and we need to impose some further structural properties on the double array of experiments (2.1) to answer the question. Our main assumption imposes 7

only a marginal local asymptotic normality (LAN) condition on the double array (cf. Definition 6.63 in Liese and Miescke (2008) concerning local asymptotic normality) and is as follows: Assumption 2 (Marginal LAN). There exists a sequence hn > 0 so that for every fixed d ∈ N d Hn,d := {h ∈ Rd : h−1 n h ∈ Θd } ↑ R ,

(5.1)

and so that the sequence of experiments   : En,d = Ωn,d , An,d , {Pn,d,h−1 h ∈ H } n,d n h

for n ∈ N

(5.2)

is locally asymptotically normal with positive definite information matrix Id . Note that Assumption 2 only imposes LAN to hold for fixed d as n → ∞. Put differently, LAN is only imposed in classical “fixed-dimensional” experiments in which it has been verified in many setups as illustrated in the examples further below. Frequently hn can be chosen as √ n (in principle we could extend our results to situations where hn is a sequence of invertible matrices that also depends on d, but for the sake of simplicity we omit this generalization). Note further that LAN is only assumed to hold at the origin.

5.1

Examples

Before we answer the main question of Section 4, we briefly discuss under which additional assumptions our running example satisfies Assumption 2. Furthermore, we provide several references to other experiments that are LAN for fixed d, merely to illustrate the generality of our results. Example 1 continued. Fixed covariates: Assume that f is absolutely continuous with derivative f 0 such that 0 < R If = (f 0 /f )2 dF < ∞. Suppose further that the double array X has the following properties: denoting Xn,d = (x1,d , ..., xn,d )0 , for every fixed d and as n → ∞ we have

1 0 n Xn,d Xn,d



Qd where Qd has full rank (implying that eventually rankXn,d = d holds), and max1≤i≤d 0 0 (Xn,d (Xn,d Xn,d )−1 Xn,d )i,i → 0. It then follows from Theorems 2.3.9 and 2.4.2 in Rieder (1994)

that for every fixed d the corresponding sequence of experiments En,d in (5.2) is LAN with √ hn = n and Id = If Qd being positive definite. Random covariates: Let the error terms satisfy the same assumptions as in the case of fixed R covariates. If, furthermore, for every d the matrix Kd = xx0 dKd (x) ∈ Rd×d has full rank d, it follows from Theorems 2.3.7 and 2.4.6 in Rieder (1994) that the corresponding experiment En,d √ in (5.2) is LAN for every fixed d with hn = n and Id = If Kd being positive definite. Further examples: Local asymptotic normality for fixed d is often satisfied: For example, L2 -differentiable models with i.i.d. data are covered via Theorem 7.2 in van der Vaart (2000). Many examples of models being L2 -differentiability and subsequently LAN for fixed d, including 8

exponential families, can be found in Chapter 12.2 of Lehmann and Romano (2006), while generalized linear models are covered in Pupashenko et al. (2015). Various time series models have been studied in, e.g., Davies (1973), Swensen (1985), Kreiss (1987), Garel and Hallin (1995) and Hallin et al. (1999). For more details and further references on LAN in time series models see also the monographs Dzhaparidze (1986) and Taniguchi and Kakizawa (2000).

5.2

Asymptotic enhanceability for “slowly” diverging d(n)

We first show that for arrays satisfying Assumption 2 there always exists a range of unbounded sequences d(n) (dimensions of the parameter space) in which every test with asymptotic size less than one is asymptotically enhanceable. In Theorem 5.4 further below we then show that this statement even extends to any unbounded sequence d(n) under additional structural assumptions on the experiments. Theorem 5.1. Suppose the double array of experiments (2.1) satisfies Assumption 2. Then, there exists a non-decreasing unbounded sequence p(n) in N, so that for any non-decreasing unbounded sequence d(n) in N satisfying d(n) ≤ p(n) every sequence of tests with asymptotic size smaller than one is asymptotically enhanceable. The proof is given in Sections 7.2 and 7.3. In words Theorem 5.1 shows the following: if d(n) is any sequence as in the statement of the theorem and ϕn is a given sequence of tests of asymptotic size smaller than one in the corresponding sequence of experiments, then there exists a sequence of enhancement components νn and corresponding tests ψn (cf. (3.3) and the discussion preceding Definition 3.1) so that: 1. ψn has the same asymptotic size as ϕn ; 2. ψn ≥ ϕn holds, guaranteeing that ψn has uniformly non-inferior asymptotic power compared to ϕn ; 3. ψn is consistent against a sequence of alternatives which ϕn is not consistent against. Remark 5.2. Actually, inspection of the part of the proof isolated in Section 7.2 reveals that a stronger statement than 3. above can be achieved: there even exists a sequence of alternatives against which ψn is consistent, but against which ϕn has asymptotic power equal to its asymptotic size. We would like to discuss two implications of Theorem 5.1: • Concerning the constructive value of Theorem 5.1: The theorem shows that at least in certain regimes (but cf. also Section 5.3) any test of asymptotic size smaller than one can benefit from an application of the power enhancement principle. In particular, every such test has removable “blind spots” of inconsistency. Therefore, if some of these are of major practical relevance, it can be worthwhile to try to remove these via an application of the 9

power enhancement principle. The particular “blind spots” we exploit in the proof are exhibited in the argument isolated in Section 7.2, see in particular Equation (7.4). • Theorem 5.4 also comes with a distinct warning: since the theorem applies equally well to ψn and any further enhanced test, no test will be entirely without removable “blind spots”. These “blind spots” are (implicitly or explicitly) determined by the choice of a test. This underscores the importance of carefully selecting the “right” test for a specific problem at hand. Finally, it is also worth noting that while Theorem 5.1 guarantees the existence of a test νn , and thus a corresponding test ψn as above, it does not indicate how such a sequence of tests can be obtained from ϕn (although the part of the proof in Section 7.2 gives some insights into how certain enhancement components can be obtained for a given test ϕn ). Theorem 5.1 shows that every test with asymptotic size less than one is asymptotically enhanceable as long as the dimension of the parameter space diverges sufficiently slowly. This is somewhat surprising, as one might have expected the result of Theorem 4.1 of typical existence of asymptotically unenhanceable tests in the case of d(n) ≡ d to carry over to the case of slowly diverging d(n). Given Theorem 5.1, intuition would further suggest that every test must also be asymptotically enhanceable when the dimension of the parameter space increases very quickly, as this only makes the testing problem “more difficult” thus broadening the scope for increasing the power of a test. As a consequence, one would be led to believe that under the same set of assumptions, the statement in the theorem can be extended to all diverging sequences d(n). However, this intuition is again flawed: asymptotically unenhanceable tests can exist under the assumptions of Theorem 5.1 when the dimension of the parameter space increases sufficiently fast. Intuitively, the reason for this is that for d(n) increasing sufficiently quickly, the testing problem can (without further assumptions than marginal LAN) become so hard that any test will have asymptotic power equal to its asymptotic size against any sequence of alternatives. Then, every test is asymptotically unenhanceable as no enhancement components exist. The following example illustrates this. We denote by Nm (µ, Σ) the m-variate Gaussian distribution with mean µ and covariance matrix Σ. Example 2. Let Ωn,d =

×ni=1 Rd and let An,d be the Borel sets of ×ni=1 Rd. 3

Set Pn,d,θ equal

d

to the n-fold product of Nd (θ, d Id ), and let Θd = (−1, 1) . Assumption 2 is obviously satisfied √ (with hn = n and Id = d−3 Id ). We now show that for d(n) = n no test is asymptotically enhanceable. To this end, it suffices to show that any sequence of tests νn : Ωn,d(n) → [0, 1] so that limn→∞ En,d(n),0 (νn ) = 0 must also satisfy lim En,d(n),θn (νn ) = 0

n→∞

for any sequence θn ∈ Θd(n) .

By sufficiency of the vector of sample means, we may assume that νn is a measurable function thereof, which (since d(n) = n) is distributed as Nn (θ, n2 In ). It hence suffices to verify 10

that the total variation distance between Nn (θn , n2 In ) and Nn (0, n2 In ), or equivalently between Nn (n−1 θn , In ) and Nn (0, In ), converges to 0 as n → ∞. But since each coordinate of θn is bounded in absolute value by 1, and thus kn−1 θn k2 ≤ n−1/2 → 0, this follows from, e.g., Example 2.3 in DasGupta (2008). Summarizing, Theorem 4.1, Theorem 5.1 and Example 2 show that without further assumptions there are three asymptotic regimes one must distinguish concerning the asymptotic enhanceability of tests. 1. Fixed d: when the dimension of the parameter space is fixed asymptotically unehanceable tests of size less than one often exist; 2. Slowly diverging d(n): every test of asymptotic size less than one is asymptotically enhanceable. 3. Quickly diverging d(n): here asymptotically unenhanceable tests may exist. Interestingly, the two outer regimes of fixed d and d(n) diverging quickly are most similar: here unenhanceable tests can exist and even typically do so when d is fixed. However, in the intermediate regime of slowly diverging d(n) Theorem 5.1 establishes a markedly different behavior, as every test of asymptotic size less than one is enhanceable here.

5.3

Asymptotic enhanceability for any non-decreasing unbounded d(n)

Theorem 5.1 showed that when the dimension of the parameter space diverges sufficiently slowly, then every test of asymptotic size less than one is asymptotically enhanceable. Without additional assumptions, this ceases to be the case in general when the dimension increases sufficiently quickly as illustrated by Example 2. We shall next give sufficient conditions under which the statement in Theorem 5.1 extends to all unboundedly increasing regimes d(n). Informally, this can be achieved by ensuring that for all natural numbers d1 < d2 and n the testing problem  concerning a zero restriction on the parameter vector in Ωn,d2 , An,d2 , {Pn,d2 ,θ : θ ∈ Θd2 } nests (in a suitable sense) as a sub-problem a testing problem that is equivalent to the testing problem  concerning a zero restriction on the parameter vector in Ωn,d1 , An,d1 , {Pn,d1 ,θ : θ ∈ Θd1 } . One assumption that achieves this is as follows: Assumption 3. For all natural numbers d1 < d2 , we have that Θdd12 := {θ ∈ Θd2 : θi = 0 if d1 < i ≤ d2 }

equals

Θd1 × {0}d2 −d1 ,

and for every n ∈ N: 1. For every test ϕ : Ωn,d2 → [0, 1] there exists a test ϕ0 : Ωn,d1 → [0, 1] so that En,d2 ,θ (ϕ) = En,d1 ,(θ1 ,...,θd1 )0 (ϕ0 ) for every θ ∈ Θdd12 . 11

(5.3)

2. For every test ϕ0 : Ωn,d1 → [0, 1] there exists a test ϕ : Ωn,d2 → [0, 1] so that En,d1 ,θ (ϕ0 ) = En,d2 ,(θ0 ,0)0 (ϕ) for every θ ∈ Θd1 . For more discussion of the notion of equivalence of testing problems underlying Assumption 3 we refer to Chapter 4 in Strasser (1985) (note that the discussion there is for dominated experiments which we do not require). The following observation is sometimes useful (e.g., for regression models with fixed regressors) in verifying the preceding assumption (and thus also the weaker Assumption 4 given below) for special cases. Remark 5.3. If the sample space does not depend on the dimensionality of the parameter space, i.e., Ωn,d = Ωn and An,d = An holds for every n ∈ N and every d ∈ N, Assumption 3 is satisfied if for all natural numbers d1 < d2 it holds that (5.3), and that θ ∈ Θdd12 implies Pn,d2 ,θ = Pn,d1 ,(θ1 ,...,θd1 )0 . For in this case one can simply use ϕ0 ≡ ϕ in Part 1, and ϕ ≡ ϕ0 in Part 2. In our running example, Assumption 3 holds in the fixed covariates case, and also in the random covariates case under an additional assumption on the family Kd : Example 1 continued. Since Θd = Rd condition (5.3) obviously holds. Fixed covariates: Since Ωn,d and An,d do not depend on d it follows immediately from the observation in Remark 5.3 that Assumption 3 is satisfied. Random covariates: In this case further conditions on Kd for d ∈ N are necessary. Recall that Kd is a probability measure on the Borel sets of Rd . Given two natural numbers d1 < d2 associate with Kd2 its “marginal distribution” Kd1 ,d2 (A) = Kd2 (A × Rd2 −d1 )

for every Borel set A ⊆ Rd1 .

If for any two natural numbers d1 < d2 it holds that Kd1 = Kd1 ,d2 , then Assumption 3 is seen to be satisfied by a sufficiency argument. See Section 7.4 for details. Note that Assumption 3 imposes restrictions to hold for every n ∈ N. Since asymptotic enhanceability concerns large-sample properties of tests, it is not surprising that a (weaker) asymptotic version of Assumption 3 suffices for our purpose. The weaker version we work with is as follows: Assumption 4. For all natural numbers d1 < d2 we have (5.3), and for any two non-decreasing unbounded sequences r(n) and d(n) in N so that r(n) < d(n) the following holds: 1. For every sequence of tests ϕn : Ωn,d(n) → [0, 1], there exists a sequence of tests ϕ0n : Ωn,r(n) → [0, 1] so that sup En,d(n),θ (ϕn ) − En,r(n),(θ1 ,...,θr(n) )0 (ϕ0n ) → 0 as n → ∞. r(n)

θ∈Θd(n)

12

(5.4)

2. For every sequence of tests ϕ0n : Ωn,r(n) → [0, 1], there exists a sequence of tests ϕn : Ωn,d(n) → [0, 1] so that sup En,r(n),θ (ϕ0n ) − En,d(n),(θ0 ,0)0 (ϕn ) → 0 as n → ∞.

θ∈Θr(n)

The following theorem states that if in addition to Assumption 2 also Assumption 4 is satisfied, then for any non-decreasing unbounded sequence d(n) of natural numbers every test of asymptotic size less than one is enhanceable. Theorem 5.4. Suppose the double array of experiments (2.1) satisfies Assumptions 2 and 4. Then, for every non-decreasing and unbounded sequence d(n) in N every sequence of tests with asymptotic size smaller than one is asymptotically enhanceable. The proof of this theorem is given in Section 7.5. While Theorem 5.1 established the existence of a range of sufficiently slowly non-decreasing unbounded d(n) along which every test is asymptotically enhanceable, Theorem 5.4 strengthens this property to hold for any non-decreasing unbounded d(n). This stronger conclusion comes from adding Assumption 4, which now allows one to transfer properties of experiments with slowly increasing d(n) (established through Theorem 5.1) to statements about (subexperiments of) experiments with quickly increasing sequences d(n).

6

Conclusion

In the present paper we have studied the asymptotic enhanceability of tests, which concerns the applicability of the power enhancement principle. After showing that in fixed-dimensional regimes there often exist tests that are not asymptotically enhanceable, we have shown that under only a marginal LAN assumption every test with asymptotic size smaller than one is asymptotically enhanceable when the dimension of the parameter space diverges sufficiently slowly with the sample size. Under further, quite natural, assumptions this enhanceability statement extends to all regimes in which the dimensionality of the testing problem diverges with sample size. As a practical consequence, in such situations, as every test possesses removable “blind spots” of inconsistency, the statistician must prioritize! In particular one has to think very carefully about which of the “blind spots” of a test are most important to remove by the power enhancement principle of Fan et al. (2015), as not all of them can be addressed.

7

Appendix

Throughout, given a random variable (or vector) x defined on a probability space (F, F, Q) the image measure induced by x is denoted by Q ◦ x. Furthermore, “⇒” denotes weak convergence.

13

7.1

Proof of Theorem 4.1

The statement trivially holds for α = 1. Let α ∈ (0, 1). Suppose we could construct a sequence of tests ϕ∗n : Ωn,d → [0, 1] with the property that for some ε > 0 so that B(ε) = {θ ∈ Rd : kθk2 < ε} $ Θd (recall that Θd is throughout assumed to contain an open neighborhood of the origin) the following holds: En,d,0 (ϕ∗n ) → α, and for any sequence θn ∈ B(ε) so that n1/2 kθn k2 → ∞ it holds that En,d,θn (ϕ∗n ) → 1. Given such a sequence of tests, we could define tests ϕn = min(ϕ∗n + ψn,d (ε), 1) (cf. Assumption 1), and note that ϕn has asymptotic size α, and has the property that En,d,θn (ϕn ) → 1 for any sequence θn ∈ Θd so that n1/2 kθn k2 → ∞. But tests with the latter property are certainly not asymptotically enhanceable, because tests νn : Ωn,d → [0, 1] can satisfy En,d,0 (νn ) → 0 and En,d,θn (νn ) → 1 only if θn ∈ Θd satisfies n1/2 kθn k2 → ∞. To see this recall that convergence of n1/2 kθn k2 along a subsequence n0 together with the maintained i.i.d. and L2 -differentiability assumption implies mutual contiguity of the sequences Pn0 ,d,0 and Pn0 ,d,θn0 (this can be verified easily using, e.g., results in Section 1.5 of Liese and Miescke (2008) and Theorem 6.26 in the same reference). It hence remains to construct such a sequence ϕ∗n . To this end, denote by L : Ω → Rd (measurable) an L2 derivative of {Pd,θ : θ ∈ Θd } at 0. In the following we denote expectation w.r.t. Pd,θ by Ed,θ . By assumption the information matrix Ed,0 (LL0 ) = Id is positive definite. Let C > 0 and define LC = L1{kLk2 ≤ C}. Since Ed,0 (LC L0 ) and M (C) = Ed,0 ((LC − Ed,0 (LC ))(LC − Ed,0 (LC ))0 ) converge to Id as C → ∞ (by the Dominated Convergence Theorem and Ed,0 (L) = 0, for the latter see Proposition 1.110 in Liese and Miescke (2008)), there exists a C ∗ so that Ed,0 (LC ∗ L0 ) and M := M (C ∗ ) are non-singular. Now, by the L2 -differentiability assumption (using again Proposition 1.110 in Liese and Miescke (2008)), there exists an ε > 0 and a c > 0 so that B(ε) $ Θd , and so that kEd,θ (LC ∗ ) − Ed,0 (LC ∗ )k2 ≥ ckθk2

holds for every

θ ∈ B(ε).

(7.1)

×ni=1 Ω the functions Zn(θ) := n−1/2 nni=1(LC (ωi,n) − Ed,θ (LC )) for θ ∈ Θd, where ωi,n denotes the i-th coordinate projection on ×i=1 Ω, and set Zn (0) = Zn . It is easy to verify P

Define on





that Pn,d,θn ◦ Zn (θn ) is tight for any sequence θn ∈ Θd , and that by the central limit theorem Pn,d,0 ◦ Zn ⇒ Nd (0, M ). Finally, let ϕ∗n : Ωn,d → [0, 1] be the indicator function of the set {kZn k2 ≥ Qα }, where Qα denotes the 1 − α quantile of the distribution of the Euclidean norm of a Nd (0, M )-distributed random vector. By construction En,d,0 (ϕ∗n ) → α. It remains to verify En,d,θn (ϕ∗n ) → 1 for any sequence θn ∈ B(ε) so that n1/2 kθn k2 → ∞. Let θn be such a sequence. By the triangle inequality kZn k2 ≥ n1/2 kEd,θn (LC ∗ ) − Ed,0 (LC ∗ )k2 − kZn (θn )k2 . Hence, 1 − En,d,θn (ϕ∗n ) is not greater (cf. (7.1)) than Pn,d,θn (cn1/2 kθn k2 − Qα ≤ kZn (θn )k2 ) → 0, the convergence following from Pn,d,θn ◦ Zn (θn ) being tight, and cn1/2 kθn k2 → ∞.

14

7.2

Proof of Theorem 5.1

The proof is based on the following proposition, which is proven in Section 7.3. The proof of the second statement in the proposition is constructive. Proposition 7.1. Suppose the double array (2.1) satisfies Assumption 2, and for every d ∈ N 0 let v1,d , . . . , vd,d be an orthogonal basis of eigenvectors of Id so that vi,d Id vi,d = 1 for i = 1, . . . , d.

Then, there exists a non-decreasing unbounded sequence p(n) > 0 and an M ∈ N, so that for every non-decreasing unbounded sequence of natural numbers d(n) ≤ p(n): 1. For every n ≥ M and i = 1, . . . , d(n) it holds that p θi,n := h−1 n max( log(d(n))/2, 1) vi,d(n) ∈ Θd(n) ,

(7.2)

and every sequence of tests ϕn : Ωn,d(n) → [0, 1] satisfies d(n) −1

En,d(n),0 (ϕn ) − d(n)

X

En,d(n),θi,n (ϕn ) → 0

as n → ∞.

i=1

2. For every sequence 1 ≤ i(n) ≤ d(n) there exists a sequence of tests νn : Ωn,d(n) → [0, 1] so that En,d(n),0 (νn ) → 0 and En,d(n),θi(n),n (νn ) → 1 as n → ∞. To prove Theorem 5.1, choose for each d ∈ N an arbitrary orthogonal basis as in Proposition 7.1 to obtain a corresponding sequence p(n), and let d(n) ≤ p(n) be non-decreasing and unbounded. Let the sequence of tests ϕn : Ωn,d(n) → [0, 1] be of asymptotic size α < 1, i.e., lim supn→∞ En,d(n),0 (ϕn ) = α < 1. According to Definition 3.1 we need to show that lim inf n→∞ En,d(n),θn (ϕn ) < 1 for a sequence θn ∈ Θd(n) for which a sequence of tests νn : Ωn,d(n) → [0, 1] exists so that lim En,d(n),0 (νn ) = 0

n→∞

and

lim En,d(n),θn (νn ) = 1.

n→∞

(7.3)

But Part 1 of Proposition 7.1 implies existence of a sequence 1 ≤ i(n) ≤ d(n) so that lim sup En,d(n),θi(n),n (ϕn ) ≤ α < 1,

(7.4)

n→∞

and Part 2 of Proposition 7.1 verifies existence of a sequence of tests νn as in Equation (7.3) for θn = θi(n),n .

7.3

Proof of Proposition 7.1

The proof is divided into three steps. First we construct a sequence p(n). Then, we verify that the first and second part of Proposition 7.1, respectively, is satisfied for this sequence.

15

7.3.1

Step 1: Construction of the sequence p(n)

For every d ∈ N denote a central sequence and the (positive definite and symmetric) information matrix corresponding to (5.2) by Zn,d : Ωn,d → Rd and by Id , respectively (cf. Definition 6.63 in Liese and Miescke (2008)). By Theorem 6.76 in Liese and Miescke (2008), the following holds for every fixed d ∈ N: there exists a sequence c(n, d) > 0 satisfying c(n, d) → ∞ as n → ∞, so that the family of probability measures {Qn,d,h : h ∈ Hn,d } on (Ωn,d , An,d ) defined via   dQn,d,h ∗ = exp h0 Zn,d − Kn,d (h) , dPn,d,0 Kn,d (h) = log(

R Ωn,d

(7.5)

∗ ∗ exp(h0 Zn,d )dPn,d,0 ) and Zn,d = Zn,d 1{kZn,d k2 ≤ c(n, d)}, satisfies

lim |Kn,d (h) − .5h0 Id h| = 0

n→∞

for every h ∈ Rd ,

(7.6)

and lim d1 (Pn,d,h−1 , Qn,d,h ) = 0 n h

n→∞

for every h ∈ Rd .

(7.7)

Here d1 denotes the total variation distance, cf. Strasser (1985) Definition 2.1. Furthermore (e.g., Theorem 6.72 in Liese and Miescke (2008)), for every fixed d ∈ N and as n → ∞ Pn,d,h−1 ◦ Zn,d ⇒ Nd (Id h, Id ) n h

for every h ∈ Rd .

(7.8)

Next, define the sequence  1/2 ai = max( .5 log(i) , 1)

for i ∈ N,

˜d = which (i) is positive, (ii) diverges to ∞, and satisfies (iii) i−1 exp(a2i ) → 0. Now, let H  d ˜ 0, ad v1,d , . . . , ad vd,d and Hd = a−2 d Hd \ {0}. By Hn,d ↑ R (as n → ∞) and by Equations (7.6), (7.7), (7.8) (and the continuous mapping theorem together with e0 Id e = a−2 d for every e ∈ Hd ), for every d ∈ N there exists an N (d) ∈ N so that n ≥ N (d) implies (firstly) ˜d + H ˜ d ⊆ Hn,d , H where, for A ⊆ Rd , the set A + A denotes {a + b : a ∈ A, b ∈ A}, and (secondly) max

˜ d +H ˜d) h∈(H

+

max

|Kn,d (h) − .5h0 Id h| + max d1 (Pn,d,h−1 , Qn,d,h ) n h

˜ d ×Hd (h,e)∈H

˜d h∈H



 −1 dw Pn,d,h−1 ◦ (e0 Zn,d ), N1 (e0 Id h, a−2 . d ) ≤d n h

Here dw (., .) denotes a metric on the set of probability measures on the Borel sets of R that generates the topology of weak convergence, cf. Dudley (2002) pp. 393 for specific examples. Note also that we can (and do) choose N (1) < N (2) < . . .. Obviously, there exists a non-

16

decreasing unbounded sequence p(n) in N that satisfies N (p(n)) ≤ n for every n ≥ N (1) =: M . Hence, the two previous displays still hold for n ≥ M when d is replaced by p(n). Moreover, the two previous displays also hold for n ≥ M when d is replaced by any sequence of non-decreasing natural numbers d(n) ≤ p(n). The latter implying that for any such sequence d(n) that is also unbounded we have ˜ d(n) + H ˜ d(n) ⊆ Hn,d(n) H

for n ≥ M

(7.9)

and that (as n → ∞) max

˜ d(n) +H ˜ d(n) ) h∈(H

|Kn,d(n) (h) − .5h0 Id(n) h| → 0

(7.10)

max d1 (Pn,d(n),h−1 , Qn,d(n),h ) → 0, n h

(7.11)

˜ d(n) h∈H

and max

˜ d(n) ×Hd(n) (h,e)∈H

  −2 0 0 dw Pn,d(n),h−1 ◦ (e Z ), N (e I h, a ) → 0. 1 n,d(n) d(n) h d(n) n

(7.12)

We shall now verify that the sequence p(n) and the natural number M defined above have the required properties. Let d(n) ≤ p(n) be an unbounded non-decreasing sequence of natural numbers. 7.3.2

Step 2: Verification of Part 1

˜ d(n) ⊆ Hn,d(n) for n ≥ M (cf. also (5.1)). Equation (7.2) follows from (7.9) which implies H Now, let ϕn : Ωn,d(n) → [0, 1] be a sequence of tests. For h ∈ Hn,d(n) abbreviate Pn,d(n),h−1 = n h Q Pn,h and Qn,d(n),h = Qn,h , and denote expectation w.r.t. Pn,h and Qn,h by EP n,h and En,h , P 1 respectively. Furthermore, define for n ≥ M the measures Pn = d(n) ˜ d(n) \{0} Pn,h , and h∈H P 1 similarly Qn = d(n) h∈H˜ d(n) \{0} Qn,h . Since for n ≥ M

X

En,d(n),0 (ϕn ) − d(n)−1

EP n,h (ϕn ) ≤ d1 (Pn,0 , Pn )

˜ n \{0} h∈H

(cf. Strasser (1985) Lemma 2.3), it suffices to verify d1 (Pn,0 , Pn ) → 0. From (7.11) we see that it suffices to show that d1 (Qn,0 , Qn ) → 0. Since Qn  Qn,0 = Pn,0 by (7.5), d21 (Qn,0 , Qn ) equals (e.g., Strasser (1985) Lemma 2.4)  2 dQn   1 EQ − 1 ≤ EQ n,0 2 n,0 dQn,0 

It remains to verify that lim sup EP n,0 n→∞



dQn dPn,0

2

!2 dQn −1 = EP n,0 dQn,0

dQn dPn,0

!2 − 1.

≤ 1: Let ad(n) = a(n), kn,i = Kn,d(n) (a(n)vi,d(n) ),

∗ 0 ∗ kn,i,j = Kn,d(n) (a(n)vi,d(n) + a(n)vj,d(n) ), and let zn,i = vi,d(n) Zn,d(n) . Let n ≥ M . From (7.5)

17

we see that d(n)

X dQn ∗ = d(n)−1 exp(a(n)zn,i − kn,i ) dPn,0 i=1 and    ∗ ∗ EP = exp kn,i,j − kn,i − kn,j . n,0 exp a(n)zn,i − kn,i exp a(n)zn,j − kn,j 2  dQn is not greater than the sum of Thus, EP n,0 dPn,0 d(n)−1 exp a2 (n)



 max exp kn,i,i − 2kn,i − a2 (n)

and

1≤i≤d(n)

max 1≤i
 exp kn,i,j − kn,i − kn,j .

But the first sequcence converges to 0, and the second to 1. This follows from i−1 exp(a2i ) → 0, and since the sequences

max |kn,i − .5a2 (n)|,

1≤i≤d(n)

max |kn,i,i − 2a2 (n)|, and

1≤i≤d(n)

max 1≤i
|kn,i,j −

a2 (n)| all converge to 0 by Equation (7.10). 7.3.3

Step 3: Verification of Part 2

0 Given a sequence 1 ≤ i(n) ≤ d(n) define tn = a(n)−1 vi(n),d(n) Zn,d(n) and let νn = 1{tn ≥ 1/2}.

By definition  EP n,0 (νn ) = Pn,0 ◦ tn [.5, ∞) .

(7.13)

˜ d(n) and a(n)−1 vi(n),d(n) ∈ Hd(n) , it follows from (7.12) that Since 0 ∈ H dw (Pn,0 ◦ tn , N1 (0, a(n)−2 )) → 0. But a(n) → ∞ hence implies (via the triangle inequality, together with dw -continuity of (µ, σ 2 ) 7→ N1 (µ, σ 2 ) on R × [0, ∞), N1 (µ, 0) being interpreted as δµ , i.e., point mass at µ) that Pn,0 ◦ tn ⇒ δ0 . From the Portmanteau Theorem it hence follows that the sequence in (7.13) converges to  ˜ d(n) , δ0 [.5, ∞) = 0. Concerning asymptotic power let vn = a(n)vi(n),d(n) . Note that vn ∈ H a(n)−1 vi(n),d(n) ∈ Hd(n) and Equation (7.12) implies dw (Pn,vn ◦ tn , N1 (1, a(n)−2 )) → 0, hence  Pn,vn ◦ tn ⇒ δ1 , and thus EP n,vn (νn ) = Pn,vn ◦ tn [.5, ∞) → 1.

7.4

Verification of Assumption 4 for the random covariates case in our running example

For convenience, denote a generic element of Ωn,d =

×ni=1(R × Rd) by zd = (y, x(1), . . . , x(d))

for y, x(1) , . . . , x(d) ∈ Rn . Let d1 < d2 and n be natural numbers. We start with Part 2: given ϕ0 : Ωn,d1 → [0, 1] define ϕ : Ωn,d2 → [0, 1] as ϕ(zd2 ) = ϕ0 (zd1 ). Then, for every θ ∈ Θd1 , the expectation En,d2 ,(θ0 ,0)0 (ϕ) obviously coincides with En,d1 ,θ (ϕ0 ) if Kd1 = Kd1 ,d2 . Part 1 can be

18

verified by a sufficiency argument: Consider the experiment (Ωn,d2 , An,d2 , {Pn,d2 ,θ : θ ∈ Θdd12 }),

(7.14)

define the map T : Ωn,d2 → Ωn,d1 as T (zd2 ) = zd1 , and note that T is sufficient for (7.14) (e.g., Theorem 20.9 in Strasser (1985)). Note further that Pn,d2 ,θ ◦ T = Pn,d1 ,θ holds for every θ ∈ Θdd12 under our additional assumption that Kd1 = Kd1 ,d2 . Part 1 now follows from Corollaries 22.4 and 22.6 in Strasser (1985).

7.5

Proof of Theorem 5.4

Let d(n) be a non-decreasing and unbounded sequence in N, and let ϕn : Ωn,d(n) → [0, 1] be of asymptotic size α < 1. We apply Theorem 5.1 to obtain a sequence p(n) as in that theorem. Let r(n) ≡ min(p(n), d(n) − 1), a non-decreasing unbounded sequence that eventually satisfies r(n) ∈ N and r(n) < d(n). By Part 1 of Assumption 4 there exists a sequence of tests ϕ0n : Ωn,r(n) → [0, 1] so that (5.4) holds. In particular ϕ0n has asymptotic size α. Therefore, by Theorem 5.1 (applied with “d(n) ≡ r(n)”), ϕ0n is asymptotically enhanceable, i.e., there exist tests νn0 : Ωn,r(n) → [0, 1] and a sequence θn ∈ Θr(n) so that En,r(n),0 (νn0 ) → 0 and 1 = lim En,r(n),θn (νn0 ) > lim inf En,r(n),θn (ϕ0n ) = lim inf En,r(n),(θn0 ,0)0 (ϕn ), n→∞

n→∞

n→∞

the second equality following from (5.4). By Part 2 of Assumption 4 tests νn : Ωn,d(n) → [0, 1] exist so that En,d(n),0 (νn ) → 0 and En,d(n),(θn0 ,0)0 (νn ) → 1. Hence ϕn is asymptotically enhanceable.

Acknowledgements Financial support by the Danish National Research Foundation (Grant DNRF 78, CREATES) is gratefully acknowledged. We would like to thank Michael Jansson and Davy Paindaveine for helpful comments and discussions.

References Bai, Z., Jiang, D., Yao, J.-F. and Zheng, S. (2009). Corrections to lrt on large-dimensional covariance matrix by rmt. Annals of Statistics, 37 3822–3840. Bai, Z. and Saranadasa, H. (1996). Effect of high dimension: by an example of a two sample problem. Statistica Sinica 311–329. Cai, T., Fan, J. and Jiang, T. (2013). Distributions of angles in random packing on spheres. The Journal of Machine Learning Research, 14 1837–1864.

19

Cai, T., Liu, W. and Xia, Y. (2014). Two-sample test of high dimensional means under dependence. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 76 349–372. Chakraborty, A. and Chaudhuri, P. (2017). Tests for high-dimensional data based on means, spatial signs and spatial ranks. Annals of Statistics, 45 771–799. Cutting, C., Paindaveine, D. and Verdebout, T. (2017). Testing uniformity on highdimensional spheres against monotone rotationally symmetric alternatives. Annals of Statistics, 45 1024–1058. DasGupta, A. (2008). Asymptotic theory of statistics and probability. Springer. Davies, R. B. (1973). Asymptotic inference in stationary gaussian time-series. Advances in Applied Probability, 5 469497. Dempster, A. P. (1958). A high dimensional two sample significance test. Annals of Mathematical Statistics, 29 995–1010. Dudley, R. M. (2002). Real analysis and probability. Cambridge University Press. Dzhaparidze, K. (1986). Parameter estimation and hypothesis testing in spectral analysis of stationary time series. Springer. Fan, J., Liao, Y. and Yao, J. (2015). Power enhancement in high-dimensional cross-sectional tests. Econometrica, 83 1497–1541. Garel, B. and Hallin, M. (1995). Local asymptotic normality of multivariate arma processes with a linear trend. Annals of the Institute of Statistical Mathematics, 47 551–579. Hallin, M., Taniguchi, M., Serroukh, A. and Choy, K. (1999). Local asymptotic normality for regression models with long-memory disturbance. Annals of Statistics, 27 2054–2080. Janssen, A. (2000). Global power functions of goodness of fit tests. Annals of Statistics, 28 239–253. Kreiss, J.-P. (1987). On adaptive estimation in stationary arma processes. Annals of Statistics 112–133. Ledoit, O. and Wolf, M. (2002). Some hypothesis tests for the covariance matrix when the dimension is large compared to the sample size. Annals of Statistics, 30 1081–1102. Lehmann, E. L. and Romano, J. P. (2006). Testing statistical hypotheses. Springer. Ley, C., Paindaveine, D. and Verdebout, T. (2015). High-dimensional tests for spherical location and spiked covariance. Journal of Multivariate Analysis, 139 79 – 91.

20

Liese, F. and Miescke, K. J. (2008). Statistical Decision Theory. Springer. Lockhart, R. A. (2016). Inefficient best invariant tests. arXiv preprint arXiv:1608.05994. Onatski, A., Moreira, M. J. and Hallin, M. (2013). Asymptotic power of sphericity tests for high-dimensional data. Annals of Statistics, 41 1204–1231. Onatski, A., Moreira, M. J. and Hallin, M. (2014). Signal detection in high dimension: The multispiked case. Annals of Statistics, 42 225–254. Pinelis, I. (2010). Asymptotic efficiency of p-mean tests for means in high dimensions. arXiv preprint arXiv:1006.0505. Pinelis, I. (2014). Schur2-concavity properties of gaussian measures, with applications to hypotheses testing. Journal of Multivariate Analysis, 124 384 – 397. Pupashenko, D., Ruckdeschel, P. and Kohl, M. (2015). L2 differentiability of generalized linear models. Statistics & Probability Letters, 97 155–164. Rieder, H. (1994). Robust asymptotic statistics. Springer. Srivastava, M. S. (2005). Some tests concerning the covariance matrix in high dimensional data. Journal of the Japan Statistical Society, 35 251–272. Srivastava, M. S. and Du, M. (2008). A test for the mean vector with fewer observations than the dimension. Journal of Multivariate Analysis, 99 386 – 402. Srivastava, M. S., Katayama, S. and Kano, Y. (2013). A two sample test in high dimensional data. Journal of Multivariate Analysis, 114 349 – 358. Steinberger, L. (2016). The relative effects of dimensionality and multiplicity of hypotheses on the F-test in linear regression. Electronic Journal of Statistics, 10 2584–2640. Strasser, H. (1985). Mathematical theory of statistics, vol. 7. Walter de Gruyter. Swensen, A. R. (1985). The asymptotic distribution of the likelihood ratio for autoregressive time series with a regression trend. Journal of Multivariate Analysis, 16 54–70. Taniguchi, M. and Kakizawa, Y. (2000). Asymptotic theory of statistical inference for time series. Springer. van der Vaart, A. W. (2000). Asymptotic statistics, vol. 3. Cambridge University Press. Wang, S. and Cui, H. (2013). Generalized F test for high dimensional linear regression coefficients. Journal of Multivariate Analysis, 117 134–149. Zhong, P.-S. and Chen, S. X. (2011). Tests for high-dimensional regression coefficients with factorial designs. Journal of the American Statistical Association, 106 260–274. 21

Power in High-Dimensional Testing Problems - ULB

In the sequel we call tests ... As a consequence, in many high-dimensional testing problems the power .... for all possible rates d(n) at which the dimension of the parameter space can increase ..... that the total variation distance between Nn(θn,n2In) and Nn(0,n2In), ...... for regression models with long-memory disturbance.

397KB Sizes 2 Downloads 321 Views

Recommend Documents

Power in High-Dimensional Testing Problems - ULB
In regression models power properties of F-tests when the ...... But a(n) → ∞ hence implies (via the triangle inequality, together with dw-continuity of (µ, σ2) ↦→.

Dual Problems in Property Testing - MIT CSAIL Theory of Computation
Thm 3: The following dual problems are equivalent to the original problems: 1. Testing whether a string is far from a code. *. 2. Testing whether a function is far from monotone. **. 3. Testing whether a distribution is far from uniform. ***. > Teste

Dual Problems in Property Testing - MIT CSAIL Theory of Computation
Testing whether a function is far from monotone. **. 3. Testing whether a distribution is far from uniform. ***. > Testers via equivalence to the original problem ...

Fine-scale genetic structure and gene dispersal in Centaurea ... - ULB
+32 2 650 9169; fax: +32 2 650 9170; e-mail: ..... the model (for the best fitting a,b parameters) using a ..... The pollen dispersal kernel best fitting the data had a.

Fine-scale genetic structure and gene dispersal in Centaurea ... - ULB
Our model is Centaurea corymbosa Pourret (Asteraceae), ... within a 3 km2 area along the French Mediterranean ..... defined previously (using the best fitting a, b parameters ..... Programme Diversitas, Fragmented Populations network,.

Testing Problems using Finite Automata with Timing ...
2.1 I/O Automata. The majority of the testing techniques discussed in the literature are based on a variant FSM: the famous Mealy machine [7]. This fact is based.

Making Rain on Arid Regions The GESHEM Rain System - ULB
The project consists in using a small mesoscale solar absorber surface at the ground to collect solar energy and produce a strong upward heat flux in order to ...

pdf-149\electrical-power-equipment-maintenance-and-testing ...
Try one of the apps below to open or edit this item. pdf-149\electrical-power-equipment-maintenance-and-testing-power-engineering-willis-by-paul-gill.pdf.

2. Treaty Making Power Problems: A Comparative Study and ...
ASEAN Sectorial Mutual Recognition Arrangements (MRAs), ASEAN- Australia-New Zealand FTA, and ASEAN-India FTA. 6 Similarly, the. ASEAN Charter and ...

Read New PDF Power Practice Problems for the ...
Computer PE Exam Book Online. Book Synopsis. Power Practice Problems for the Electrical and Computer. PE Exam contains over ... support material.Topics.

Lazy Learners at work: the Lazy Learning Toolbox ... - ULB CoDE
In Section 2 we will introduce the Lazy Learning method in formal ... O nce obtained the local first-degree polynomial approximation, a prediction ... O n the basis of this information a final prediction .... collaboration of Iridia with the H oneywe

Making Rain on Arid Regions The GESHEM Rain System - ULB
laboratory experiments provide conditions for an artificial HI prototype to produce rainfalls : 1. Proximity of sea shore: interaction of the convective heat plume of ...

conventional dc electrochemical techniques in corrosion testing
Electrochemical methods in corrosion testing have been used ever since the electrochemical ... electrochemical techniques have been employed to both speed data ..... Conversion of polarization resistance to corrosion current requires ...

Introduction to Testing in Java
Big bang theory season 8 complete 1080p. ... Sincethe 19th century it has widely been accepted that which forevermoreshall bethestate has ... friendsand colleagues. o Open-ended questionscan generatelargeamounts of ... Thejapanese pdf.

Symbol Implementation in Software Testing
“Software testing is an exploration accompanied to provide sponsors with information about the superiority of the product or provision under test. ... culture of software development, a challenging association may be separate from the expansion tea

Lessons Learned in Software Testing
[PDF BOOK] Lessons Learned in Software Testing: A Context Driven .... technology itself is often made a scapegoat, as few people in the organization Welcome!

Testing Normality in SPSS.pdf
... the Dependent List and click Plots... Step 3: Uncheck the boxes except for Normality plots with tests. Click Continue. Step 4: Read the result! The Q-Q plot sure ...