Early stopping for non-parametric regression: An optimal data-dependent stopping rule

Garvesh Raskutti∗ Department of Statistics University of California, Berkeley Berkeley, CA 94720 [email protected]

Martin Wainwright Department of Statistics and Department of EECS University of California, Berkeley Berkeley, CA 94720 [email protected]

Bin Yu Department of Statistics and Department of EECS University of California, Berkeley Berkeley, CA 94720 [email protected]

Abstract The goal of non-parametric regression is to estimate an unknown function f ∗ based on n i.i.d. observations of the form yi = f ∗ (xi ) + wi , where {wi }ni=1 are additiveP noise variables. Simply choosing a function to minimize the least-squares n 1 2 loss 2n i=1 (yi − f (xi )) will lead to “overfitting”, so that various estimators are based on different types of regularization. The early stopping strategy is to run an iterative algorithm such as gradient descent for a fixed but finite number of iterations. Early stopping is known to yield estimates with better prediction accuracy than those obtained by running the algorithm for an infinite number of iterations. Although bounds on this prediction error are known for certain function classes and step size choices, the bias-variance tradeoffs for arbitrary reproducing kernel Hilbert spaces (RKHSs) and arbitrary choices of step-sizes have not been wellunderstood to date. In this paper, we derive upper bounds on both the L2 (Pn ) and L2 (P) error for arbitrary RKHSs, and provide an explicit and easily computable data-dependent stopping rule. In particular, it depends only on the sum of step-sizes and the eigenvalues of the empirical kernel matrix for the RKHS. For Sobolev spaces and finite-rank kernel classes, we show that our stopping rule yields estimates that achieve the statistically optimal rates in a minimax sense.

1 Introduction The phenomenon of overfitting is ubiquitous throughout statistics, and is particularly problematic in non-parametric problems. For estimating regression functions and other infinite-dimensional quantities, some form of regularization is essential: it prevents overfitting, and thereby improves the prediction accuracy for future (unseen) samples. The most classical form of regularization is based on adding a penalty to the objective function, such as the least-squares loss, that measures fit to the data. An alternative and algorithmic approach to regularization is based on early stopping of an iterative algorithm, such as gradient descent, applied to the loss function. Such approaches are often referred to as boosting algorithms in the statistics literature. In practice, early stopping of gradient descent and other iterative algorithms has been found to improve prediction performance for many ∗ Use footnote for providing further information about author (webpage, alternative address)—not for acknowledging funding agencies.

1

problems– for instance, see the papers [1, 3, 4, 6, 8, 13, 14] and references therein. Developing theoretical bounds for early stopping of iterative methods is beneficial for two reasons. First, iterative algorithms present quite a natural regularization path indexed by iteration number t. Second, early stopping has the potential to yield improvements in both statistical performance (reduced prediction error) and computational complexity (reduced number of iterations). In this paper, we study these issues in the context of a standard non-parametric regression model, in which we make observations of the form yi = f ∗ (xi ) + wi , for i = 1, 2, . . . , n. (1) Here {wi }ni=1 is an i.i.d. sequence of standard normal N (0, 1), and {xi }ni=1 is a sequence of design points in X ⊂ R, sampled i.i.d. according to some unknown distribution P. The function f ∗ is fixed but unknown, and assumed to belong to some reproducing kernel Hilbert space H. Our main contribution is a precise analysis of a simple estimator that runs a form of gradient descent on the least-squares objective n X b )= 1 R(f (yi − f (xi ))2 , (2) 2n i=1

and to exploit this analysis to derive a simple data-dependent stopping rule. For various kernel classes, we show that the function estimate obtained by this stopping rule has a prediction error that, with high probability, achieves the statistically optimal rates for non-parametric regression. In more detail, our main result (Theorem 1) provides probabilistic upper bounds on both the empirical L2 (Pn ) error and population L2 (P) error for a certain form of gradient descent. Based on these bounds, we establish a data-dependent stopping rule Tb that is easy to compute. In rough terms, this stopping rule is based on the first time that a running sum of the step sizes in gradient descent increase above a critical threshold determined by the eigenvalues of the empirical kernel matrix for the underlying RKHS. For the case of finite-rank kernel classes and Sobolev spaces, we prove that the function estimate fbTb produced by our stopping rule has a statistical error that is within constant factors of the minimax optimal rates. Consequently, apart from constant factors, the bounds from our analysis are unimprovable. our bound can Our proof is based on a combination of analytic techniques from past work [3] with methods from empirical process theory (e.g., [11]); this combination allows us to derive sharp probabilistic upper bounds.

2 Background and problem formulation In this section, we provide some background on reproducing kernel Hilbert spaces (RKHSs), the problem of non-parametric regression, and the iterative updates for gradient descent. 2.1 Reproducing kernel Hilbert spaces Given a subset X ⊂ R and a probability measure P on X , we consider a Hilbert space H ⊂ L2 (P), meaning a family of functions g : X → R, with kgkL2(P) < ∞, and an associated inner product h·, ·iH under which H is complete. The space H is a reproducing kernel Hilbert space (RKHS) if there exists a symmetric function K : X × X → R+ such that: (a) for each x ∈ X , the function K(·, x) belongs to the Hilbert space H, and (b) we have the reproducing relation f (x) = hf, K(·, x)iH for all f ∈ H. Any such kernel function must be positive semidefinite; under suitable regularity conditions, Mercer’s theorem [9] guarantees that the kernel has an eigenexpansion of the form ∞ X K(x, x′ ) = λk φk (x)φk (x′ ), (3) k=1

where λ1 ≥ λ2 ≥ λ3 ≥ . . . ≥ 0 are a non-negative sequence of eigenvalues, and {φk }∞ k=1 are the associated eigenfunctions, taken to be orthonormal in L2 (P). The decay rate of the eigenvalues will play a crucial role in our analysis. Throughout the paper, assume that all functions are uniformly bounded. That is kf k∞ : = sup |f (x)| ≤ kf kH ≤ 1. x

2

(4)

This boundedness condition (4) is satisfied for Sobolev spaces, as well as Hilbert spaces with eigenfunctions based on trigonometric functions. 2.2 Our gradient descent method on least-squares objective For i = 1, 2, . . . , n, let (xi , yi ) be a collection of n i.i.d. samples, and consider the empirical leastsquares risk: n X 2 b ) := 1 R(f yi − f (xi ) . 2n i=1

(5)

b ∈ Rn×n with entries Furthermore, we define the empirical kernel matrix K

1 K(xi , xj ). n Using the representer theorem [7], any f , the gradient step we consider is: b ij = [K]

b − ft (X)). ft+1 (X) = ft (X) + αt K(Y

(6)

Hence the gradient descent iteration for f (X) depends only on the step-size αt and the empirical b We assume that f0 (X) = 0. kernel matrix K.

3 Main Result

We are now ready to state our main result, and to derive some of its consequences for specific cases of reproducing kernel Hilbert spaces. Our main result provides a stopping rule, or more precisely, a procedure by which to select an iteration number T at which the gradient descent procedure should Pt−1 be halted. The stopping rule depends on the running sum of the step-sizes ηt : = k=0 αk , as well as the eigenvalues of the empirical kernel matrix. Given the function fT obtained after T rounds, we provide upper bounds on the L2 (Pn ) error kft − f ∗ k2n : =

n 2 1X ft (xi ) − f ∗ (xi ) , n i=1

as well as the error kft − f ∗ k22 = E[(ft (X) − f ∗ (X))2 ]. 3.1 Data-dependent stopping rule The empirical kernel complexity is defined in terms of the function  X n 1/2  bn (δ, H) : = √1 Q , min λbi , δ 2 n i=1 bj } of the empirical matrix. Note that this matrix (and hence which depends on the eigenvalues {λ these eigenvalues) are easily computed from the data. Given a set of positive stepsizes {αk }, we consider a stopping rule of the form:   t 1 1b X −1/2 ≥ Q (( α ) , H) , (7) Tb : = max n k t>0 ηt 4 k=0

Pt

where ηt : = k=0 αk is the running sum of the step sizes. Note that the stopping rule (7) is computable and data-dependent since it depends only on the eigenvalues of the empirical kernel b matrix K.

Throughout our analysis, we focus on valid sequences of stepsizes, meaning that they are strictly b  I , and Pt αk → ∞ positive and non-increasing, with initial stepsize α1 chosen such that α1 K k=0 as t → ∞. 3

3.2 Bounds on squared L2 (P) error In order to relate our bounds to the optimal statistical rates, we define a second function that measures the complexity of the kernel operator underlying the RKHS. In particular, recalling that the kernel operator has a non-negative sequence {λk } of eigenvalues, we define the population kernel complexity 1/2 1 X Qn (δ, H) : = √ min{λk , δ 2 } , n k≥1

Using this complexity measure, the critical rate δn2 is defined via the equation δn2 = 256Qn(δn , H). (8) As will be clarified, for various kernel classes, this quantity corresponds to the statistically optimal rate. Theorem 1. Under the observation model (1), suppose that we apply the gradient descent iteration (6) with a valid sequence of stepsizes, starting at f0 = 0. Then there are universal constants ci , i = 0, 1, 2 such that all of the following events hold with probability at least 1 − c1 exp(−c2 nδn2 ): At the stopping time Tb, we have kfbb − f ∗ k2 ≤ c0 δ 2 , and (9a) T

n

n

kfbTb − f ∗ k22 ≤ 4 c0 δn2 .

(9b)

Theorem 1 is a general result that applies to any reproducing kernel Hilbert space. Let us now illustrate some of its consequences for special choices of RKHSs that are of interest in practice. We begin with the class of finite-rank kernels, which includes (among other examples) linear functions, as well as polynomial functions of any fixed degree. Corollary 1. Consider any kernel class H with finite rank m. Then using the function estimate obtained by the stopping rule (7), m kfbTb − f ∗ k22 ≤ c0 n with probability greater than 1 − c1 exp(−c2 m). 2 It is worth noting that for a rank m-kernel, the rate m n is minimax optimal in terms of squared L (P) error (e.g., see the paper [10] for details).Next we present a result for the RKHS’s with infinitely manyc eigenvalues, but whose eigenvalues decay at a rate λk ≃ (1/k)2ν for some parameter ν > 1/2. Among other examples, this type of scaling covers the case of Sobolev spaces, say consisting of functions with ν derivatives (e.g., [2, 5]).

Corollary 2. Consider the kernel class H with eigenvalue decay λk ≃ (1/k)2ν for some ν > 1/2. Then the function estimate fbTb obtained by the stopping rule (7) satisfies the bound c0 kfbTb − f ∗ k22 ≤ 2ν 2ν+1 n 1 with probability greater than 1 − c1 exp(−c2 n 2ν+1 ). 3.3 Comments on results For Sobolev kernels with smoothness ν, past work by Yao et al. [13] established upper bounds on the 2ν−1 2ν squared L2 (P) error that scale n− 2ν+2 ) for Sobolev spaces. Note that this is slower than the n− 2ν+1 rate that follows from our analysis. The improvement is likely due to greater care in controlling the estimation and approximation error terms, using techniques from empirical process theory (see e.g. [11]). In both of the previous results, our stopping rule yielded minimax-optimal rates—more specifically, 2ν − 2ν+1 m for Sobolev spaces (see e.g. [12]). Corollary 2 provides n for m-rank kernel classes and n the parallel result in the random design case to the fixed design case in Theorem 3 of B¨uhlmann and Yu [3]. Although these are two interesting classes of kernels, it would be interesting to show that the stopping rule (7) yields minimax optimal rates for general RKHSs with arbitrary eigenvalue decay. 4

References [1] P. Bartlett and M. Traskin. Adaboost is consistent. Journal of Machine Learning Research, 8:2347–2368, 2007. [2] M. S. Birman and M. Z. Solomjak. Piecewise-polynomial approximations of functions of the classes Wpα . Math. USSR-Sbornik, 2(3):295–317, 1967. [3] P. Buhlmann and B.Yu. Booting with l2 loss: Regression and classification. Journal of American Statistical Association, 98:324–340, 2003. [4] Y. Freund and R. Schapire. A decision-theoretic generalization of on-line learning and an application to boosting. Journal of Computer and System Sciences, 55(1):119–139, 1997. [5] C. Gu. Smoothing spline ANOVA models. Springer Series in Statistics. Springer, New York, NY, 2002. [6] Wenxin Jiang. Process consistency for adaboost. Annals of Statistics, 32:13–29, 2004. [7] G. Kimeldorf and G. Wahba. Some results on Tchebycheffian spline functions. Jour. Math. Anal. Appl., 33:82–95, 1971. [8] L. Mason, J. Baxter, P. Bartlett, and M. Frean. Boosting algorithms as gradient descent. In Neural Information Processing Systems (NIPS), December 1999. [9] J. Mercer. Functions of positive and negative type and their connection with the theory of integral equations. Philosophical Transactions of the Royal Society A, 209:415–446, 1909. [10] G. Raskutti, M. J. Wainwright, and B. Yu. Minimax-optimal rates for sparse additive models over kernel classes via convex programming. Technical Report arXiv:0910.2042, UC Berkeley, Department of Statistics, 2010. [11] S. van de Geer. Empirical Processes in M-Estimation. Cambridge University Press, 2000. [12] Y. Yang and A. Barron. Information-theoretic determination of minimax rates of convergence. Annals of Statistics, 27(5):1564–1599, 1999. [13] Y. Yao, L. Rosasco, and A. Caponnetto. On early stopping in gradient descent learning. Constructive Approximation, 26:289–315, 2007. [14] T. Zhang and B. Yu. Boosting with early stopping: Convergence and consistency. Annal of Statistics, 33:1538–1579, 2005.

5

Early stopping for non-parametric regression: An ...

University of California, Berkeley. Berkeley, CA 94720 ... For Sobolev spaces and finite-rank kernel classes, we show that our stopping rule yields estimates that ...

65KB Sizes 1 Downloads 162 Views

Recommend Documents

Nonparametric Regression with Infinite Order Flat ... - TL McMurry, LLC
Apr 24, 2003 - All programming was done in Mathematica. Before results are presented, a few words should be said about the choice of the parameters c and h. The simulations were performed using the kernel defined by equation (3) with c = 0.05 and b =

Asymptotic Theory of Nonparametric Regression with Spatial Data
Gaussian Maximum Likelihood Estimation. As far as this paper is concerned, models defined on a regular lattice are discussed. In general, spatial ...

rdrobust: An R Package for Robust Nonparametric ... - The R Journal
(2008), IK, CCT, Card et al. (2014), and references therein. .... Direct plug-in (DPI) approaches to bandwidth selection are based on a mean. The R Journal Vol.

Nonparametric Estimation of an Instrumental ...
Oct 6, 2009 - ϕ(Z) is not the conditional expectation function E(Y |Z). ... Integral equation of the first kind and recovering its solution ϕ is an ill-posed inverse.

Nonparametric Estimation of an Instrumental ...
in the second step we compute the regularized bayesian estimator of ϕ. We develop asymptotic analysis in a frequentist sense and posterior consistency is ...

TWO EFFICIENT STOPPING CRITERIA FOR ...
Email: [email protected]. ABSTRACT ... teria need not to store any extra data, and can reduce the .... sion of the last iteration, and the S3 criterion needs to store.

An Effective Tree-Based Algorithm for Ordinal Regression
Abstract—Recently ordinal regression has attracted much interest in machine learning. The goal of ordinal regression is to assign each instance a rank, which should be as close as possible to its true rank. We propose an effective tree-based algori

Nonparametric Hierarchical Bayesian Model for ...
results of alternative data-driven methods in capturing the category structure in the ..... free energy function F[q] = E[log q(h)] − E[log p(y, h)]. Here, and in the ...

Robust Nonparametric Confidence Intervals for ...
results are readily available in R and STATA using our companion software ..... mance in finite samples by accounting for the added variability introduced by.

Nonparametric Hierarchical Bayesian Model for ...
employed in fMRI data analysis, particularly in modeling ... To distinguish these functionally-defined clusters ... The next layer of this hierarchical model defines.

(EBMAL) for Regression
†Human Research and Engineering Directorate, U.S. Army Research Laboratory, Aberdeen Proving Ground, MD USA ... ¶Brain Research Center, National Chiao-Tung University, Hsinchu, Taiwan ... Administration (NHTSA) [36], 2.5% of fatal motor vehicle ..

A Robust Stopping Criterion for Agglomerative ...
ters are homogeneous in terms of speaker identity before every merging, by ... a pair of heterogeneous clusters where there are a small num- ber of feature samples ... computes difference between H0 and HA in terms of average log-likelihood ...

Stopping Unsafe Practices
Q. Failure to operate within the incident command system, e.g., accountability, freelancing. R. Not observing safe practices unique to wildland fire incidents. S. Being complacent on automatic alarm or gas leak incidents. T. Failure to properly maint

DOWNLOAD An Interdisciplinary Approach to Early Childhood ...
Education and Care: Perspectives from Australia (Routledge Research in. Early Childhood Education) pdf by Susanne Garvis Full KINDLE ePub. Online.

Dušan IVKOVIĆ, early offense.pdf
opponent's court with the quality rival who knows to use all the advantages. offered by his own court. Although it's hard to resist, I've never looked forward to ...

What Model for Entry in First&Price Auctions? A Nonparametric ...
Nov 22, 2007 - project size, we find no support for the Samuelson model, some support for the Levin and Smith ..... Define the cutoff 's a function of N as 's+N,.

A nonparametric hierarchical Bayesian model for group ...
categories (animals, bodies, cars, faces, scenes, shoes, tools, trees, and vases) in the .... vide an ordering of the profiles for their visualization. In tensorial.

Boosting Methodology for Regression Problems
Decision-support Systems, with an Application in. Gastroenterology” (with discussion), Journal of the Royal. Statistical Society (Series A), 147, 35-77. Zheng, Z. and G.I. Webb [1998]. “Lazy Bayesian Rules,”. Technical Report TR C98/17, School

Learning and Smooth Stopping
Oct 12, 2009 - University of Exeter Business School, Streatham Court, Rennes ...... the distance between the distributions is large, and so learning is relatively ...

Bagging for Gaussian Process Regression
Sep 1, 2008 - rate predictions using Gaussian process regression models. ... propose to weight the models by the inverse of their predictive variance, and ...

Sampling Algorithms and Coresets for lp Regression
Email: [email protected]. ‡Computer Science, University of Pennsylvania, Philadelphia,. PA 19107. Work done while the author was visiting Yahoo! Research. Email: [email protected] ficient sampling algorithms for the classical ℓp regres- sion p