scispace - formally typeset
Search or ask a question
Journal ArticleDOI

A Measure of Asymptotic Efficiency for Tests of a Hypothesis Based on the sum of Observations

01 Dec 1952-Annals of Mathematical Statistics (Institute of Mathematical Statistics)-Vol. 23, Iss: 4, pp 493-507
TL;DR: In this paper, it was shown that the likelihood ratio test for fixed sample size can be reduced to this form, and that for large samples, a sample of size $n$ with the first test will give about the same probabilities of error as a sample with the second test.
Abstract: In many cases an optimum or computationally convenient test of a simple hypothesis $H_0$ against a simple alternative $H_1$ may be given in the following form. Reject $H_0$ if $S_n = \sum^n_{j=1} X_j \leqq k,$ where $X_1, X_2, \cdots, X_n$ are $n$ independent observations of a chance variable $X$ whose distribution depends on the true hypothesis and where $k$ is some appropriate number. In particular the likelihood ratio test for fixed sample size can be reduced to this form. It is shown that with each test of the above form there is associated an index $\rho$. If $\rho_1$ and $\rho_2$ are the indices corresponding to two alternative tests $e = \log \rho_1/\log \rho_2$ measures the relative efficiency of these tests in the following sense. For large samples, a sample of size $n$ with the first test will give about the same probabilities of error as a sample of size $en$ with the second test. To obtain the above result, use is made of the fact that $P(S_n \leqq na)$ behaves roughly like $m^n$ where $m$ is the minimum value assumed by the moment generating function of $X - a$. It is shown that if $H_0$ and $H_1$ specify probability distributions of $X$ which are very close to each other, one may approximate $\rho$ by assuming that $X$ is normally distributed.
Citations
More filters
Journal ArticleDOI
TL;DR: The results generalize the classical result for G(n, p), when p=clogn/n and show that, under reasonably weak assumptions, the connectivity threshold of the model can be determined.
Abstract: We find conditions for the connectivity of inhomogeneous random graphs with intermediate density. Our results generalize the classical result for Gn, p, when p=clogn/n. We draw n independent points Xi from a general distribution on a separable metric space, and let their indices form the vertex set of a graph. An edge i, j is added with probability min1,i¾?Xi,Xjlogn/n, where i¾?i¾?0 is a fixed kernel. We show that, under reasonably weak assumptions, the connectivity threshold of the model can be determined. © 2013 Wiley Periodicals, Inc. Random Struct. Alg., 45, 408-420, 2014

37 citations


Additional excerpts

  • ...We use the binomial Chernoff bound [5, 11, 12]: If ξ ∼ binomial(n, p) and t > 0 then min (P(ξ ≤ tnp),P(ξ ≥ tnp)) ≤ e−f , where we write f (x) = x log x − x + 1....

    [...]

Journal ArticleDOI
TL;DR: This paper provides an analysis of a natural d-round tournament over n = 2d players and demonstrates that the tournament possesses a surprisingly strong ranking property.
Abstract: This paper provides an analysis of a natural d-round tournament over n = 2d players and demonstrates that the tournament possesses a surprisingly strong ranking property. The ranking property of this tournament is used to design efficient sorting algorithms for several models of parallel computation: a comparator network of depth $c\\cdot\lg n$, $c\approx 7.44$, that sorts the vast majority of the n! possible input permutations; an $O(\lg n)$-depth hypercubic comparator network that sorts the vast majority of permutations; a hypercubic sorting network with nearly logarithmic depth; an $O(\lg n)$-time randomized sorting algorithm for any hypercubic machine (other such algorithms have been previously discovered, but this algorithm has a significantly smaller failure probability than any previously known algorithm); and a randomized algorithm for sorting n O(m)-bit records on an $(n\lg n)$-node omega machine in $O(m+\lg n)$ bit steps.

37 citations

Journal ArticleDOI
TL;DR: In this article, it was shown that prior communication, although unable to create nonlocality, leads to wirings not only beyond LOSRs but also not contained in a much broader class of (nonlocality-generating) global Wirings.
Abstract: Bell nonlocality can be formulated in terms of a resource theory with local-hidden variable models as resourceless objects. Two such theories are known, one built upon local operations assisted by shared randomness (LOSRs) and the other one allowing, in addition, for prior-to-input classical communication. We show that prior communication, although unable to create nonlocality, leads to wirings not only beyond LOSRs but also not contained in a much broader class of (nonlocality-generating) global wirings. Technically, this is shown by proving that it can improve the statistical distinguishability between Bell correlations optimized over all fixed measurement choices. This has implications in nonlocality quantification, and leads us to a natural universal definition of Bell nonlocality measures. To end up with, we also consider the statistical strength of nonlocality proofs. We point out some issues of its standard definition in the resource-theoretic operational framework, and suggest simple fixes for them. Our findings reveal nontrivial features of the geometry of the set of wirings and may have implications in the operational distinguishability of nonlocal behaviors.

37 citations

Proceedings Article
06 Jun 2016
TL;DR: In this article, the authors studied the structure and learnability of sums of independent integer random variables (SIIRVs) of order n 2 Z+ and showed that the optimal sample complexity of this learning problem is (( k = 2 ) p log(1= )).
Abstract: We study the structure and learnability of sums of independent integer random variables (SIIRVs). For k 2 Z+, a k-SIIRV of order n2 Z+ is the probability distribution of the sum of n mutually independent random variables each supported onf0; 1;:::;k 1g. We denote bySn;k the set of allk-SIIRVs of ordern. How many samples are required to learn an arbitrary distribution inSn;k? In this paper, we tightly characterize the sample and computational complexity of this problem. More precisely, we design a computationally efficient algorithm that uses e O(k= 2 ) samples, and learns an arbitrary k-SIIRV within error ; in total variation distance. Moreover, we show that the optimal sample complexity of this learning problem is (( k= 2 ) p log(1= )); i.e., we prove an upper bound and a matching information-theoretic lower bound. Our algorithm proceeds by learning the Fourier transform of the target k-SIIRV in its effective support. Its correctness relies on the approximate sparsity of the Fourier transform of k-SIIRVs ‐ a structural property that we establish, roughly stating that the Fourier transform ofk-SIIRVs has small magnitude outside a small set. Along the way we prove several new structural results about k-SIIRVs. As one of our main structural contributions, we give an efficient algorithm to construct a sparse proper -cover for Sn;k; in total variation distance. We also obtain a novel geometric characterization of the space of k-SIIRVs. Our characterization allows us to prove a tight lower bound on the size of -covers for Sn;k ‐ establishing that our cover upper bound is optimal ‐ and is the key ingredient in our tight sample complexity lower bound. Our approach of exploiting the sparsity of the Fourier transform in distribution learning is general, and has recently found additional applications. In a subsequent work Diakonikolas et al. (2015c), we use a generalization of this idea to obtain the first computationally efficient learning algorithm for Poisson multinomial distributions. In Diakonikolas et al. (2015b), we build on our Fourier-based approach to obtain the fastest known proper learning algorithm for Poisson binomial distributions (2-SIIRVs).

37 citations

Journal ArticleDOI
TL;DR: In this paper, the authors compared the performance of the Kolmogorov-Smirnov test, Berk-Jones test, score test and their integrated versions in the context of testing the goodness-of-fit of a heavy tailed distribution function.

37 citations

References
More filters