Book ChapterDOI
On Some Bayesian Solutions of the Neyman-Scott Problem
Malay Ghosh
- pp 267-276
Reads0
Chats0
TLDR
In this article, a class of hierarchical Bayes (HB) estimators of σ 2 is introduced, and a subclass, each member of which dominates the best multiple of S (S being the error sum of squares) under the relative squared error loss L(a, σ2) = (a σ -2 -1)2.Abstract:
One of the two celebrated examples of Neyman and Scott (1948) is that in a fixed effects one-way analysis of variance model with normal homoscedastic errors, the maximum likelihood estimator of the error variance, say σ 2 is inconsistent as the cell-size remains fixed, but the number of cells grows to infinity. The UMVUE, or the best multiple estimator of the error sum of squares does not suffer from this drawback. However, the best multiple estimator is inadmissible, as it is dominated by a Stein-type estimator. The Stein-estimator, on the other hand, being non-smooth, is itself inadmissible. The present paper introduces a class of hierarchical Bayes (HB) estimators of σ 2, and identifies a subclass, each member of which dominates the best multiple of S (S being the error sum of squares) under the relative squared error loss L(a, σ 2) = (a σ -2 -1)2. Included in our class is the Brewster-Zidek (1974) estimator of σ 2. Also, we have provided analytic expressions for the risk improvement of the HB estimators over the best multiple estimator. Numerical values of the percentage risk improvement are given in some special cases. These calculations indicate that the risk-improvement over the best multiple estimator can often be quite substantial.read more
Citations
More filters
Journal ArticleDOI
The Selection of Prior Distributions by Formal Rules
Robert E. Kass,Larry Wasserman +1 more
TL;DR: In this paper, a review of techniques for constructing non-informative priors is presented and some of the practical and philosophical issues that arise when they are used are discussed.
Journal ArticleDOI
An overview of robust Bayesian analysis
James O. Berger,Elías Moreno,Luis R. Pericchi,M. Jesús Bayarri,José M. Bernardo,Juan Antonio Cano,Julián de la Horra,Jacinto Martín,David Rios-Insua,Bruno Betrò,Anirban DasGupta,Paul Gustafson,Larry Wasserman,Joseph B. Kadane,Cid Srinivasan,Michael Lavine,Anthony O'Hagan,Wolfgang Polasek,Christian P. Robert,Constantinos Goutis,Fabrizio Ruggeri,G. Salinetti,Siva Sivaganesan +22 more
TL;DR: An overview of the subject of robust Bayesian analysis is provided, one that is accessible to statisticians outside the field, and recent developments in the area are reviewed.
Book ChapterDOI
Maximum Likelihood Estimates
TL;DR: In this paper, a two-tailed permutation test (unpaired) for testing significance of observing (S1,S2) is proposed, where k denotes the power set of the integers 1,..., k. Input: (·, ·): function of two samples S 1: first sample S 2: second sample 1: stat (S 1,S 2) 2: pool S 1 + S 2 3: n,m |S 1|, |S 2| 4: dist.
Journal ArticleDOI
Bayesian reference prior analysis on the ratio of variances for the balanced one-way random effect model
TL;DR: Tibshirani, R., Biometrika 76 (1989) 604-608 as discussed by the authors used grouped ordering reference prior approach to analyze one-way random effect model.
Journal ArticleDOI
Minimax estimators of a normal variance
TL;DR: In this paper, a new class of minimax estimators for the estimation of unknown variance of a multivariate normal distribution is presented. And it is shown that a sequence of estimators in this class converges to the Stein's truncated estimator.
References
More filters
Journal ArticleDOI
Consistent Estimates Based on Partially Consistent Observations
Jerzy Neyman,Elizabeth L. Scott +1 more
Journal ArticleDOI
Parametric Empirical Bayes Inference: Theory and Applications
TL;DR: In this paper, a review of the state of the art in multiparameter shrinkage estimators with emphasis on the empirical Bayes viewpoint, particularly in the case of parametric prior distributions, is presented.
Journal ArticleDOI
Consistency of the maximum likelihood estimator in the presence of infinitely many incidental parameters
J. Kiefer,Jacob Wolfowitz +1 more
TL;DR: In this paper, the authors considered the problem of consisteently estimating 0 (as n --* X ), where the chance variables were assumed to be scalars, and the parameters 0 and ai may be vectors.
Journal ArticleDOI
Asymptotic Properties of Conditional Maximum-Likelihood Estimators
TL;DR: In this article, a maximum-likelihood estimator based on the conditional distribution given minimal sufficient statistics for the incidental parameters is proposed, and it is proved that conditional maximum likelihood estimates in the regular case are consistent and asymptotically normally distributed.
Related Papers (5)
A note on covariance estimation in the unbiased estimator of risk framework
Bala Rajaratnam,Dario Vincenzi +1 more