scispace - formally typeset
Search or ask a question
Author

Galen R. Shorack

Bio: Galen R. Shorack is an academic researcher. The author has an hindex of 2, co-authored 2 publications receiving 2737 citations.

Papers
More filters
Book
01 Apr 1986
TL;DR: In this paper, a broad cross-section of the literature available on one-dimensional empirical processes is summarized, with emphasis on real random variable processes as well as a wide-ranging selection of applications in statistics.
Abstract: Here is the first book to summarize a broad cross-section of the large volume of literature available on one-dimensional empirical processes. Presented is a thorough treatment of the theory of empirical processes, with emphasis on real random variable processes as well as a wide-ranging selection of applications in statistics. Featuring many tables and illustrations accompanying the proofs of major results, coverage includes foundations - special spaces and special processes, convergence and distribution of empirical processes, alternatives and processes of residuals, integral tests of fit and estimated empirical processes and martingale methods.

2,774 citations

Journal ArticleDOI
TL;DR: In this paper, the empirical df of a sample from the uniform (0, 1) df $I was defined, and conditions on λ n > 1 were given that determine whether λn = 0 or 1.
Abstract: Let $\Gamma_n$ denote the empirical df of a sample from the uniform (0, 1) df $I$. Let $\xi_{nk}$ denote the $k$th smallest observation. Let $\lambda_n > 1$. Let $A_n$ denote the event that $\Gamma_n$ intersects the line $\lambda_n I$ on [0, 1] and let $B_n$ denote the event that $\Gamma_n$ intersects the line $I/\lambda_n$ on $\lbrack\xi_{n1}, 1\rbrack$. Conditions on $\lambda_n$ are given that determine whether $P(A_n \mathrm{i.o.})$ and $P(B_n \mathrm{i.o.})$ equal 0 or 1. Results for $A_n$ (for $B_n$) are related to upper class sequences for $1/(n\xi_{n1})$ (for $n\xi_{n2})$. Upper class sequences for $n\xi_{nk}$, with $k > 1$, are characterized. In the case of nonidentically distributed random variables, we present the result analogous to $P(A_n \mathrm{i.o.}) = 0$.

29 citations


Cited by
More filters
Journal ArticleDOI
TL;DR: In this article, Modelling Extremal Events for Insurance and Finance is discussed. But the authors focus on the modeling of extreme events for insurance and finance, and do not consider the effects of cyber-attacks.
Abstract: (2002). Modelling Extremal Events for Insurance and Finance. Journal of the American Statistical Association: Vol. 97, No. 457, pp. 360-360.

2,729 citations

Book
16 Apr 2013
TL;DR: How to Construct Nonparametric Regression Estimates * Lower Bounds * Partitioning Estimates * Kernel Estimates * k-NN Estimates * Splitting the Sample * Cross Validation * Uniform Laws of Large Numbers
Abstract: Why is Nonparametric Regression Important? * How to Construct Nonparametric Regression Estimates * Lower Bounds * Partitioning Estimates * Kernel Estimates * k-NN Estimates * Splitting the Sample * Cross Validation * Uniform Laws of Large Numbers * Least Squares Estimates I: Consistency * Least Squares Estimates II: Rate of Convergence * Least Squares Estimates III: Complexity Regularization * Consistency of Data-Dependent Partitioning Estimates * Univariate Least Squares Spline Estimates * Multivariate Least Squares Spline Estimates * Neural Networks Estimates * Radial Basis Function Networks * Orthogonal Series Estimates * Advanced Techniques from Empirical Process Theory * Penalized Least Squares Estimates I: Consistency * Penalized Least Squares Estimates II: Rate of Convergence * Dimension Reduction Techniques * Strong Consistency of Local Averaging Estimates * Semi-Recursive Estimates * Recursive Estimates * Censored Observations * Dependent Observations

1,931 citations

Proceedings ArticleDOI
26 Aug 2001
TL;DR: An efficient algorithm for mining decision trees from continuously-changing data streams, based on the ultra-fast VFDT decision tree learner is proposed, called CVFDT, which stays current while making the most of old data by growing an alternative subtree whenever an old one becomes questionable, and replacing the old with the new when the new becomes more accurate.
Abstract: Most statistical and machine-learning algorithms assume that the data is a random sample drawn from a stationary distribution. Unfortunately, most of the large databases available for mining today violate this assumption. They were gathered over months or years, and the underlying processes generating them changed during this time, sometimes radically. Although a number of algorithms have been proposed for learning time-changing concepts, they generally do not scale well to very large databases. In this paper we propose an efficient algorithm for mining decision trees from continuously-changing data streams, based on the ultra-fast VFDT decision tree learner. This algorithm, called CVFDT, stays current while making the most of old data by growing an alternative subtree whenever an old one becomes questionable, and replacing the old with the new when the new becomes more accurate. CVFDT learns a model which is similar in accuracy to the one that would be learned by reapplying VFDT to a moving window of examples every time a new example arrives, but with O(1) complexity per example, as opposed to O(w), where w is the size of the window. Experiments on a set of large time-changing data streams demonstrate the utility of this approach.

1,790 citations

Journal ArticleDOI
TL;DR: In this paper, the authors proposed a simple modification of a conventional method of moments estimator for a discrete response model, replacing response probabilities that require numerical integration with estimators obtained by Monte Carlo simulation.
Abstract: This paper proposes a simple modification of a conventional method of moments estimator for a discrete response model, replacing response probabilities that require numerical integration with estimators obtained by Monte Carlo simulation. This method of simulated moments (MSM) does not require precise estimates of these probabilities for consistency and asymptotic normality, relying instead on the law of large numbers operating across observations to control simulation error, and hence can use simulations of practical size. The method is useful for models such as high-dimensional multinomial probit (MNP), where computation has restricted applications.

1,621 citations

Proceedings ArticleDOI
05 Dec 1989
TL;DR: An exact characterization of the ability of the rate monotonic scheduling algorithm to meet the deadlines of a periodic task set and a stochastic analysis which gives the probability distribution of the breakdown utilization of randomly generated task sets are represented.
Abstract: An exact characterization of the ability of the rate monotonic scheduling algorithm to meet the deadlines of a periodic task set is represented. In addition, a stochastic analysis which gives the probability distribution of the breakdown utilization of randomly generated task sets is presented. It is shown that as the task set size increases, the task computation times become of little importance, and the breakdown utilization converges to a constant determined by the task periods. For uniformly distributed tasks, a breakdown utilization of 88% is a reasonable characterization. A case is shown in which the average-case breakdown utilization reaches the worst-case lower bound of C.L. Liu and J.W. Layland (1973). >

1,582 citations