scispace - formally typeset
Open AccessJournal ArticleDOI

Optimal Global Rates of Convergence for Nonparametric Regression

Charles J. Stone
- 01 Dec 1982 - 
- Vol. 10, Iss: 4, pp 1040-1053
TLDR
In this article, it was shown that the optimal rate of convergence for an estimator of an unknown regression function (i.e., a regression function of order 2p + d) with respect to a training sample of size n = (p - m)/(2p + 2p+d) is O(n−1/n−r) under appropriate regularity conditions, where n−1 is the optimal convergence rate if q < q < \infty.
Abstract
Consider a $p$-times differentiable unknown regression function $\theta$ of a $d$-dimensional measurement variable Let $T(\theta)$ denote a derivative of $\theta$ of order $m$ and set $r = (p - m)/(2p + d)$ Let $\hat{T}_n$ denote an estimator of $T(\theta)$ based on a training sample of size $n$, and let $\| \hat{T}_n - T(\theta)\|_q$ be the usual $L^q$ norm of the restriction of $\hat{T}_n - T(\theta)$ to a fixed compact set Under appropriate regularity conditions, it is shown that the optimal rate of convergence for $\| \hat{T}_n - T(\theta)\|_q$ is $n^{-r}$ if $0 < q < \infty$; while $(n^{-1} \log n)^r$ is the optimal rate if $q = \infty$

read more

Citations
More filters
Journal ArticleDOI

De-noising by soft-thresholding

TL;DR: The authors prove two results about this type of estimator that are unprecedented in several ways: with high probability f/spl circ/*/sub n/ is at least as smooth as f, in any of a wide variety of smoothness measures.
Journal ArticleDOI

Locally Weighted Regression: An Approach to Regression Analysis by Local Fitting

TL;DR: Locally weighted regression as discussed by the authors is a way of estimating a regression surface through a multivariate smoothing procedure, fitting a function of the independent variables locally and in a moving fashion analogous to how a moving average is computed for a time series.
Journal ArticleDOI

Adapting to Unknown Smoothness via Wavelet Shrinkage

TL;DR: In this article, the authors proposed a smoothness adaptive thresholding procedure, called SureShrink, which is adaptive to the Stein unbiased estimate of risk (sure) for threshold estimates and is near minimax simultaneously over a whole interval of the Besov scale; the size of this interval depends on the choice of mother wavelet.
Journal ArticleDOI

Matching As An Econometric Evaluation Estimator

TL;DR: In this article, a rigorous distribution theory for kernel-based matching is presented, and the method of matching is extended to more general conditions than the ones assumed in the statistical literature on the topic.
Book

A Probabilistic Theory of Pattern Recognition

TL;DR: The Bayes Error and Vapnik-Chervonenkis theory are applied as guide for empirical classifier selection on the basis of explicit specification and explicit enforcement of the maximum likelihood principle.