scispace - formally typeset
Search or ask a question
Journal ArticleDOI

A Measure of Asymptotic Efficiency for Tests of a Hypothesis Based on the sum of Observations

01 Dec 1952-Annals of Mathematical Statistics (Institute of Mathematical Statistics)-Vol. 23, Iss: 4, pp 493-507
TL;DR: In this paper, it was shown that the likelihood ratio test for fixed sample size can be reduced to this form, and that for large samples, a sample of size $n$ with the first test will give about the same probabilities of error as a sample with the second test.
Abstract: In many cases an optimum or computationally convenient test of a simple hypothesis $H_0$ against a simple alternative $H_1$ may be given in the following form. Reject $H_0$ if $S_n = \sum^n_{j=1} X_j \leqq k,$ where $X_1, X_2, \cdots, X_n$ are $n$ independent observations of a chance variable $X$ whose distribution depends on the true hypothesis and where $k$ is some appropriate number. In particular the likelihood ratio test for fixed sample size can be reduced to this form. It is shown that with each test of the above form there is associated an index $\rho$. If $\rho_1$ and $\rho_2$ are the indices corresponding to two alternative tests $e = \log \rho_1/\log \rho_2$ measures the relative efficiency of these tests in the following sense. For large samples, a sample of size $n$ with the first test will give about the same probabilities of error as a sample of size $en$ with the second test. To obtain the above result, use is made of the fact that $P(S_n \leqq na)$ behaves roughly like $m^n$ where $m$ is the minimum value assumed by the moment generating function of $X - a$. It is shown that if $H_0$ and $H_1$ specify probability distributions of $X$ which are very close to each other, one may approximate $\rho$ by assuming that $X$ is normally distributed.
Citations
More filters
Journal ArticleDOI
TL;DR: In this article, the authors proposed a scheme for obtaining relief by means of randomized routing of packets on simple extensions of the well-known omega networks for the renewal context in which packets are generated at the sources continually and asynchronously.
Abstract: Parallel communication algorithms and networks are central to large-scale parallel computing and, also, data communications. This paper identifies adverse source-destination traffic patterns and proposes a scheme for obtaining relief by means of randomized routing of packets on simple extensions of the well-known omega networks. Valiant and Aleliunas have demonstrated randomized algorithms, for a certain context which we call nonrenewal, that complete the communication task in time O(log N) with overwhelming probability, where N is the number of sources and destinations. Our scheme has advantages because it uses switches of fixed degree, requires no scheduling, and, for the nonrenewal context, is as good in proven performance. The main advantage of our scheme comes when we consider the renewal context in which packets are generated at the sources continually and asynchronously. Our algorithm extends naturally from the nonrenewal context. In the analysis in the renewal context we, first, explicitly identify the maximum traffic intensities in the internal links of the extended omega networks over all source-destination traffic specifications that satisfy loose bounds. Second, the benefits of randomization on the stability of the network are identified. Third, exact results, for certain restricted models for sources and transmission, and approximate analytic results, for quite general models, are derived for the mean delays. These results show that, in the stable regime, the maximum mean time from source to destination is asymptotically proportional to log N. Numerical results are presented.

47 citations

Proceedings ArticleDOI
01 Oct 1992
TL;DR: The results extend the Chernoff-Hoeffding bounds to certain types of random variables which are not stochastic ally independent, and believe that these results are of independent interest, and merit further study.
Abstract: Certain types of routing, scheduling and resource allocation problems in a distributed setting can be modeled as edge coloring problems. We present fast and simple randomized algorithms for edge coloring a graph, in the synchronous distributed point–to–point model of comput ation. Our algorithms compute an edge–coloring of a graph G with n nodes and maximum degree A with at most (1.6 + E)A + logz+d n colors with high probability (arbitrarily close to 1), for any fixed c, 6>0. To analyze the performance of our algorithms, we introduce new techniques for proving upper bounds on the tail probabilities of certain random variables. ChernoffHoeflding bounds are fundamental tools that are used very frequently in estimating tail probabilities. However, they assume stochastic independence among certain random variables, which may not always hold. Our results extend the Chernoff-Hoeffding bounds to certain types of random variables which are not stochastic ally independent. We believe that these results are of independent interest, and merit further study.

46 citations

Book ChapterDOI
07 Jul 2012
TL;DR: This work presents a simple algorithm that uses the notion of cross-entropy to find an optimal importance sampling distribution and uses a naturally defined low dimensional vector of parameters to specify this distribution and thus avoids the intractable explicit representation of a transition matrix.
Abstract: Statistical model checking avoids the exponential growth of states associated with probabilistic model checking by estimating probabilities from multiple executions of a system and by giving results within confidence bounds Rare properties are often important but pose a particular challenge for simulation-based approaches, hence a key objective for statistical model checking (SMC) is to reduce the number and length of simulations necessary to produce a result with a given level of confidence Importance sampling can achieves this, however to maintain the advantages of SMC it is necessary to find good importance sampling distributions without considering the entire state space Here we present a simple algorithm that uses the notion of cross-entropy to find an optimal importance sampling distribution In contrast to previous work, our algorithm uses a naturally defined low dimensional vector of parameters to specify this distribution and thus avoids the intractable explicit representation of a transition matrix We show that our parametrisation leads to a unique optimum and can produce many orders of magnitude improvement in simulation efficiency We demonstrate the efficacy of our methodology by applying it to models from reliability engineering and biochemistry

46 citations


Cites background from "A Measure of Asymptotic Efficiency ..."

  • ..., the Chernoff and Hoeffding bounds [4,12]) that relate absolute error, confidence and the required number of simulations to achieve them, independent of the probability of the property....

    [...]

Journal ArticleDOI
01 Jan 2020
TL;DR: This paper introduces systematic and user‐friendly schemes for developing non‐asymptotic lower bounds of tail probabilities and establishes matching upper and lower bounds for extreme value expectation of the sum of independent sub‐Gaussian and sub‐exponential random variables.
Abstract: The non-asymptotic tail bounds of random variables play crucial roles in probability, statistics, and machine learning. Despite much success in developing upper bounds on tail probability in literature, the lower bounds on tail probabilities are relatively fewer. In this paper, we introduce systematic and user-friendly schemes for developing non-asymptotic lower bounds of tail probabilities. In addition, we develop sharp lower tail bounds for the sum of independent sub-Gaussian and sub-exponential random variables, which match the classic Hoeffding-type and Bernstein-type concentration inequalities, respectively. We also provide non-asymptotic matching upper and lower tail bounds for a suite of distributions, including gamma, beta, (regular, weighted, and noncentral) chi-square, binomial, Poisson, Irwin-Hall, etc. We apply the result to establish the matching upper and lower bounds for extreme value expectation of the sum of independent sub-Gaussian and sub-exponential random variables. A statistical application of signal identification from sparse heterogeneous mixtures is finally considered.

46 citations


Cites background from "A Measure of Asymptotic Efficiency ..."

  • ...random variables ([26] and [12], Theorem 1; also see [39], Proposition 14....

    [...]

  • ...…John Wiley & Sons, Ltd. 1 of 11 https://doi.org/10.1002/sta4.314 characterized the asymptotic tail probability for the sum of i.i.d. random variables (Lyons & Pemantle, 1992, and Chernoff, 1952, Theorem 1; also see Van der Vaart, 2000, Proposition 14.23): suppose Z1, … , Zk are i.i.d. copies of Z....

    [...]

  • ...The Chernoff–Cramèr bound (Chernoff (1952), Theorem 1; also see, e.g., Boucheron et al. (2013), and Wainwright (2019)), with a generic statement given below, has been a basic tool to develop upper bounds of tail probabilities....

    [...]

Book ChapterDOI
25 Feb 1993
TL;DR: The properties of the complexity measure C b (·) is studied, which is an encoding e: D → {0,1}b, so that for any y, f(x,y) can be determined (quickly) by probing e(x).
Abstract: A static data structure problem consists of a set of data D, a set of queries Q and a function f with domain D × Q. Given a space bound b, a (good) solution to the problem is an encoding e: D → {0,1}b, so that for any y, f(x,y) can be determined (quickly) by probing e(x). The worst case number of probes needed is C b (f), the bit probe complexity of f. We study the properties of the complexity measure C b (·).

46 citations

References
More filters