scispace - formally typeset
Search or ask a question
Journal ArticleDOI

A Measure of Asymptotic Efficiency for Tests of a Hypothesis Based on the sum of Observations

01 Dec 1952-Annals of Mathematical Statistics (Institute of Mathematical Statistics)-Vol. 23, Iss: 4, pp 493-507
TL;DR: In this paper, it was shown that the likelihood ratio test for fixed sample size can be reduced to this form, and that for large samples, a sample of size $n$ with the first test will give about the same probabilities of error as a sample with the second test.
Abstract: In many cases an optimum or computationally convenient test of a simple hypothesis $H_0$ against a simple alternative $H_1$ may be given in the following form. Reject $H_0$ if $S_n = \sum^n_{j=1} X_j \leqq k,$ where $X_1, X_2, \cdots, X_n$ are $n$ independent observations of a chance variable $X$ whose distribution depends on the true hypothesis and where $k$ is some appropriate number. In particular the likelihood ratio test for fixed sample size can be reduced to this form. It is shown that with each test of the above form there is associated an index $\rho$. If $\rho_1$ and $\rho_2$ are the indices corresponding to two alternative tests $e = \log \rho_1/\log \rho_2$ measures the relative efficiency of these tests in the following sense. For large samples, a sample of size $n$ with the first test will give about the same probabilities of error as a sample of size $en$ with the second test. To obtain the above result, use is made of the fact that $P(S_n \leqq na)$ behaves roughly like $m^n$ where $m$ is the minimum value assumed by the moment generating function of $X - a$. It is shown that if $H_0$ and $H_1$ specify probability distributions of $X$ which are very close to each other, one may approximate $\rho$ by assuming that $X$ is normally distributed.
Citations
More filters
Journal ArticleDOI
TL;DR: In this paper, a general central limit theorem for probabilities of large deviations for sequences of random variables satisfying certain natural analytic conditions has been proved, which has wide applications to combinatorial structures and to the distribution of additive arithmetical functions.
Abstract: We prove a general central limit theorem for probabilities of large deviations for sequences of random variables satisfying certain natural analytic conditions. This theorem has wide applications to combinatorial structures and to the distribution of additive arithmetical functions. The method of proof is an extension of Kubilius’ version of Cram er’s classical method based on analytic moment generating functions. We thus generalize Cram er’s and Kubilius’ theorems on large deviations.

73 citations


Cites methods from "A Measure of Asymptotic Efficiency ..."

  • ...Actually, we shall follow Kubilius’ method [26, Ch. IX] which is more suitable for our purposes....

    [...]

Proceedings ArticleDOI
Vitaly Feldman1
17 May 2008
TL;DR: It is shown that evolvability is equivalent to learnability by a restricted form of statistical queries, and it is proved that for any fixed distribution D over the instance space, every class of functions learnable by SQs over D is evolvable over D.
Abstract: Valiant has recently introduced a framework for analyzing the capabilities and the limitations of the evolutionary process of random change guided by selection. In his framework the process of acquiring a complex functionality is viewed as a substantially restricted form of PAC learning of an unknown function from a certain set of functions. Valiant showed that classes of functions evolvable in his model are also learnable in the statistical query (SQ) model of Kearns and asked whether the converse is true. We show that evolvability is equivalent to learnability by a restricted form of statistical queries. Based on this equivalence we prove that for any fixed distribution D over the instance space, every class of functions learnable by SQs over D is evolvable over D. Previously, only the evolvability of monotone conjunctions of Boolean variables over the uniform distribution was known. On the other hand, we prove that the answer to Valiant's question is negative when distribution-independent evolvability is considered. To demonstrate this, we develop a technique for proving lower bounds on evolvability and use it to show that decision lists and linear threshold functions are not evolvable in a distribution-independent way. This is in contrast to distribution-independent learnability of decision lists and linear threshold functions in the statistical query model.

73 citations

Journal ArticleDOI
TL;DR: An equation-based adaptive search mechanism that uses an estimate of the popularity of a resource in order to choose the parameters of random walk such that a targeted performance level is achieved by the search.

73 citations


Cites methods from "A Measure of Asymptotic Efficiency ..."

  • ...Similarly an analogy may be established between uniform sampling and random walk by considering Chernoff bound [6,11] on a sequence of Bernoulli trials....

    [...]

Book ChapterDOI
01 Jan 2010
TL;DR: This book focuses on Alfred Renyi’s seminal work on information theory to derive a set of estimators to apply entropy and divergence as cost functions in adaptation and learning.
Abstract: It is evident from Chapter 1 that Shannon’s entropy occupies a central role in information-theoretic studies. Yet, the concept of information is so rich that perhaps there is no single definition that will be able to quantify information properly. Moreover, from an engineering perspective, one must estimate entropy from data which is a nontrivial matter. In this book we concentrate on Alfred Renyi’s seminal work on information theory to derive a set of estimators to apply entropy and divergence as cost functions in adaptation and learning. Therefore, we are mainly interested in computationally simple, nonparametric estimators that are continuous and differentiable in terms of the samples to yield well-behaved gradient algorithms that can optimize adaptive system parameters. There are many factors that affect the determination of the optimum of the performance surface, such as gradient noise, learning rates, and misadjustment, therefore in these types of applications the entropy estimator’s bias and variance are not as critical as, for instance, in coding or rate distortion theories. Moreover in adaptation one is only interested in the extremum (maximum or minimum) of the cost, with creates independence from its actual values, because only relative assessments are necessary. Following our nonparametric goals, what matters most in learning is to develop cost functions or divergence measures that can be derived directly from data without further assumptions to capture as much structure as possible within the data’s probability density function (PDF).

73 citations

Proceedings ArticleDOI
09 Jan 2001
TL;DR: In this article, the existence of small-sized equitable graph colorings is used to prove sharp Chernoff-Hoeffding type concentration results for sums of random variables with dependence.
Abstract: Chernoff-Hoeffding bounds are sharp tail probability bounds for sums of bounded independent random variables. Often we cannot avoid dependence among random variables involved in the sum. In some cases the theory of martingales has been used to obtain sharp bounds when there is a limited amount of dependence. This paper will present a simple but powerful new technique that uses the existence of small sized equitable graph colorings to prove sharp Chernoff-Hoeffding type concentration results for sums of random variables with dependence. This technique also allows us to focus on the dependency structure of the random variables and in cases where the dependency structure is a tree or an outerplanar graph, it allows us to derive bounds almost as sharp as those obtainable had the random variables been mutually independent. This technique connects seemingly unrelated topics: extremal graph theory and concentration inequalities. The technique also motivates several open questions in equitable graph colorings, positive answers for which will lead to surprisingly strong Chernoff-Hoeffding type bounds.

73 citations

References
More filters