scispace - formally typeset
Search or ask a question
Journal ArticleDOI

On the composition of elementary errors

01 Jan 1928-Scandinavian Actuarial Journal (Taylor & Francis Group)-Vol. 1928, Iss: 1, pp 13-74
TL;DR: In this paper, the authors define a variable V(t) the probability function of a quantity z, which may assume certain real values with certain probabilistic properties, and call V t the probability of z having exactly the value t.
Abstract: 1. By a variable in the sense of the Theory of Probability we mean a quantity z, which may assume certain real values with certain probahilities. We shall call V(t) the probability function of z if, for every real t, V(t) is equal to the probabiliby that z has a value < t, increased by half the probability that z has exactly the value t.
Citations
More filters
Journal ArticleDOI
S. S. Shapiro1, M. B. Wilk1
TL;DR: In this article, a new statistical procedure for testing a complete sample for normality is introduced, which is obtained by dividing the square of an appropriate linear combination of the sample order statistics by the usual symmetric estimate of variance.
Abstract: The main intent of this paper is to introduce a new statistical procedure for testing a complete sample for normality. The test statistic is obtained by dividing the square of an appropriate linear combination of the sample order statistics by the usual symmetric estimate of variance. This ratio is both scale and origin invariant and hence the statistic is appropriate for a test of the composite hypothesis of normality. Testing for distributional assumptions in general and for normality in particular has been a major area of continuing statistical research-both theoretically and practically. A possible cause of such sustained interest is that many statistical procedures have been derived based on particular distributional assumptions-especially that of normality. Although in many cases the techniques are more robust than the assumptions underlying them, still a knowledge that the underlying assumption is incorrect may temper the use and application of the methods. Moreover, the study of a body of data with the stimulus of a distributional test may encourage consideration of, for example, normalizing transformations and the use of alternate methods such as distribution-free techniques, as well as detection of gross peculiarities such as outliers or errors. The test procedure developed in this paper is defined and some of its analytical properties described in ? 2. Operational information and tables useful in employing the test are detailed in ? 3 (which may be read independently of the rest of the paper). Some examples are given in ? 4. Section 5 consists of an extract from an empirical sampling study of the comparison of the effectiveness of various alternative tests. Discussion and concluding remarks are given in ?6. 2. THE W TEST FOR NORMALITY (COMPLETE SAMPLES) 2 1. Motivation and early work This study was initiated, in part, in an attempt to summarize formally certain indications of probability plots. In particular, could one condense departures from statistical linearity of probability plots into one or a few 'degrees of freedom' in the manner of the application of analysis of variance in regression analysis? In a probability plot, one can consider the regression of the ordered observations on the expected values of the order statistics from a standardized version of the hypothesized distribution-the plot tending to be linear if the hypothesis is true. Hence a possible method of testing the distributional assumptionis by means of an analysis of variance type procedure. Using generalized least squares (the ordered variates are correlated) linear and higher-order

16,906 citations

Journal ArticleDOI
TL;DR: In this article, a general method for calculating the limiting distributions of these criteria is developed by reducing them to corresponding problems in stochastic processes, which in turn lead to more or less classical eigenvalue and boundary value problems for special classes of differential equations.
Abstract: The statistical problem treated is that of testing the hypothesis that $n$ independent, identically distributed random variables have a specified continuous distribution function $F(x)$. If $F_n(x)$ is the empirical cumulative distribution function and $\psi(t)$ is some nonnegative weight function $(0 \leqq t \leqq 1)$, we consider $n^{\frac{1}{2}} \sup_{-\infty

3,082 citations

Book
04 Dec 1998
TL;DR: The most useful parts of large-sample theory are accessible to scientists outside statistics and certainly to master's-level statistics students who ignore most of measure theory as discussed by the authors, which constitutes a coherent body of concepts and results that are central to both theoretical and applied statistics.
Abstract: This introductory book on the most useful parts of large-sample theory is designed to be accessible to scientists outside statistics and certainly to master’s-level statistics students who ignore most of measure theory. According to the author, “the subject of this book, first-order large- sample theory, constitutes a coherent body of concepts and results that are central to both theoretical and applied statistics.” All of the other existing books published on the subject over the last 20 years, from Ibragimov and Has’minskii in 1979 to the most recent by Van der Waart in 1998 have a common prerequisite in mathematical sophistication (measure theory in particular) that do not make the concepts available to a wide audience.

1,182 citations


Cites background from "On the composition of elementary er..."

  • ...This modification of the problem is due to Cramér (1946), who gave the following conditions for the existence of a consistent sequence θ̂n of local maxima of the likelihood function (satisfied, for instance, by the smoothed version of Example 7....

    [...]

  • ...15) For a proof of this result, see, for example, Cramér (1946) or Lehmann and Casella (1998)....

    [...]

  • ...For k = 2, one finds (Cramér (1946))...

    [...]

Journal ArticleDOI
TL;DR: In this paper, an empirical sampling study of the sensitivities of nine statistical procedures for evaluating the normality of a complete sample was carried out, including W (Shapiro and Wilk, 1965), (standard third moment), b 2 (standard fourth moment), KS (Kolmogorov-Smirnov), CM (Cramer-Von Mises), WCM (weighted CM), D (modified KS), CS (chi-squared) and u (studentized range).
Abstract: Results are given of an empirical sampling study of the sensitivities of nine statistical procedures for evaluating the normality of a complete sample. The nine statistics are W (Shapiro and Wilk, 1965), (standard third moment), b 2 (standard fourth moment), KS (Kolmogorov-Smirnov), CM (Cramer-Von Mises), WCM (weighted CM), D (modified KS), CS (chi-squared) and u (Studentized range). Forty-five alternative distributions in twelve families and five sample sizes were studied. Results are included on the comparison of the statistical procedures in relation to groupings of the alternative distributions, on means and variances of the statistics under the various alternatives, on dependence of sensitivities on sample size, on approach to normality as measured by the W statistic within some classes of distribution, and on the effect of misspecification of parameters on the performance of the simple hypothesis test statistics. The general findings include: (i) The W statistic provides a generally superio...

1,093 citations

Journal ArticleDOI
TL;DR: In this paper, it was shown that for a statistic such as the mean of a sample of size $n, or the ratio of two such means, a satisfactory approximation to its probability density, when it exists, can be obtained nearly always by the method of steepest descents.
Abstract: It is often required to approximate to the distribution of some statistic whose exact distribution cannot be conveniently obtained. When the first few moments are known, a common procedure is to fit a law of the Pearson or Edgeworth type having the same moments as far as they are given. Both these methods are often satisfactory in practice, but have the drawback that errors in the "tail" regions of the distribution are sometimes comparable with the frequencies themselves. The Edgeworth approximation in particular notoriously can assume negative values in such regions. The characteristic function of the statistic may be known, and the difficulty is then the analytical one of inverting a Fourier transform explicitly. In this paper we show that for a statistic such as the mean of a sample of size $n$, or the ratio of two such means, a satisfactory approximation to its probability density, when it exists, can be obtained nearly always by the method of steepest descents. This gives an asymptotic expansion in powers of $n^{-1}$ whose dominant term, called the saddlepoint approximation, has a number of desirable features. The error incurred by its use is $O(n^{-1})$ as against the more usual $O(n^{-1/2})$ associated with the normal approximation. Moreover it is shown that in an important class of cases the relative error of the approximation is uniformly $O(n^{-1})$ over the whole admissible range of the variable. The method of steepest descents was first used systematically by Debye for Bessel functions of large order (Watson [17]) and was introduced by Darwin and Fowler (Fowler [9]) into statistical mechanics, where it has remained an indispensable tool. Apart from the work of Jeffreys [12] and occasional isolated applications by other writers (e.g. Cox [2]), the technique has been largely ignored by writers on statistical theory. In the present paper, distributions having probability densities are discussed first, the saddlepoint approximation and its associated asymptotic expansion being obtained for the probability density of the mean $\bar{x}$ of a sample of $n$. It is shown how the steepest descents technique is related to an alternative method used by Khinchin [14] and, in a slightly different context, by Cramer [5]. General conditions are established under which the relative error of the saddlepoint approximation is $O(n^{-1})$ uniformly for all admissible $\bar{x}$, with a corresponding result for the asymptotic expansion. The case of discrete variables is briefly discussed, and finally the method is used for approximating to the distribution of ratios.

1,004 citations