scispace - formally typeset
Open AccessJournal ArticleDOI

A Measure of Asymptotic Efficiency for Tests of a Hypothesis Based on the sum of Observations

Herman Chernoff
- 01 Dec 1952 - 
- Vol. 23, Iss: 4, pp 493-507
Reads0
Chats0
TLDR
In this paper, it was shown that the likelihood ratio test for fixed sample size can be reduced to this form, and that for large samples, a sample of size $n$ with the first test will give about the same probabilities of error as a sample with the second test.
Abstract
In many cases an optimum or computationally convenient test of a simple hypothesis $H_0$ against a simple alternative $H_1$ may be given in the following form. Reject $H_0$ if $S_n = \sum^n_{j=1} X_j \leqq k,$ where $X_1, X_2, \cdots, X_n$ are $n$ independent observations of a chance variable $X$ whose distribution depends on the true hypothesis and where $k$ is some appropriate number. In particular the likelihood ratio test for fixed sample size can be reduced to this form. It is shown that with each test of the above form there is associated an index $\rho$. If $\rho_1$ and $\rho_2$ are the indices corresponding to two alternative tests $e = \log \rho_1/\log \rho_2$ measures the relative efficiency of these tests in the following sense. For large samples, a sample of size $n$ with the first test will give about the same probabilities of error as a sample of size $en$ with the second test. To obtain the above result, use is made of the fact that $P(S_n \leqq na)$ behaves roughly like $m^n$ where $m$ is the minimum value assumed by the moment generating function of $X - a$. It is shown that if $H_0$ and $H_1$ specify probability distributions of $X$ which are very close to each other, one may approximate $\rho$ by assuming that $X$ is normally distributed.

read more

Citations
More filters
Journal ArticleDOI

On a two-truths phenomenon in spectral graph clustering.

TL;DR: This work provides a clear and concise demonstration of a “two-truths” phenomenon for spectral graph clustering in which the first step—spectral embedding—is either Laplacian spectral embedding, wherein one decomposes the normalized LaplACian of the adjacency matrix, or adjacenciescripts given by a decomposition of theAdjacency Matrix itself.
Journal ArticleDOI

On Minimal Modulo 2 Sums of Products for Switching Functions

TL;DR: Every symmetric function of 2m+1 variables has a modulo 2 sum of products realization with at most 3m terms; but there are functions of n variables which require at least 2n/n log 2 3 terms for sufficiently large n.
Journal ArticleDOI

On the sample complexity of noise-tolerant learning

TL;DR: This paper shows a general lower bound of Ω(log( 1 δ ) e(1 − 2η) 2 ) on the number of examples required for PAC learning in the presence of classification noise and demonstrates the optimality of the generalLower bound by providing a noise-tolerant learning algorithm for the class of symmetric Boolean functions which uses a sample size within a constant factor of this bound.
Book ChapterDOI

Graph Sparsification in the Semi-streaming Model

TL;DR: This paper provides a one pass $\tilde{O}(n/\epsilon^{2})$ space algorithm that produces a sparsification that approximates each cut to a (1 + *** ) factor, and shows that $\Omega(n \log \frac1\ep silon)$ space is necessary for a onePass streaming algorithm to approximate the min-cut, improving upon the *** (n ) lower bound that arises from lower bounds for testing connectivity.
Journal ArticleDOI

A Randomized Parallel Sorting Algorithm with an Experimental Study

TL;DR: A novel variation on sample sort which uses only two rounds of regular all-to-all personalized communication in a scheme that yields very good load balancing with virtually no overhead, and its performance is invariant over the set of input distributions unlike previous efficient algorithms.
References
More filters