scispace - formally typeset
Search or ask a question
Author

Chris Chatfield

Bio: Chris Chatfield is an academic researcher from University of Bath. The author has contributed to research in topics: Time series & Series (mathematics). The author has an hindex of 47, co-authored 108 publications receiving 14482 citations. Previous affiliations of Chris Chatfield include Drexel University & International Institute of Minnesota.


Papers
More filters
Book
01 Jan 2017
TL;DR: In this paper, simple descriptive techniques for time series estimation in the time domain forecasting stationary processes in the frequency domain spectral analysis bivariate processes linear systems state-space models and the Kalman filter non-linear models multivariate time series modelling some other topics.
Abstract: Simple descriptive techniques probability models for time series estimation in the time domain forecasting stationary processes in the frequency domain spectral analysis bivariate processes linear systems state-space models and the Kalman filter non-linear models multivariate time series modelling some other topics.

3,694 citations

Book
01 Jan 1980
TL;DR: In this article, the multivariate normal distribution is used for principal component analysis and multivariate analysis of covariance and related topics, as well as multi-dimensional scaling and cluster analysis.
Abstract: Part One. Multivariate distributions. Preliminary data analysis. Part Two: Finding new underlying variables. Principal component analysis. Factor analysis. Part Three: Procedures based on the multivariate normal distribution. The multivariate normal distribution. Procedures based on normal distribution theory. The multivariate analysis of variance. The multivariate analysis of covariance and related topics. Part Four: Multi-dimensional scaling and cluster analysis. Multi-dimensional scaling. Cluster analysis.

1,856 citations

Journal ArticleDOI
TL;DR: The effects of model uncertainty, such as too narrow prediction intervals, and the non-trivial biases in parameter estimates which can follow data-based modelling are reviewed.
Abstract: This paper takes a broad, pragmatic view of statistical inference to include all aspects of model formulation. The estimation of model parameters traditionally assumes that a model has a prespecified known form and takes no account of possible uncertainty regarding the model structure. This implicitly assumes the existence of a 'true' model, which many would regard as a fiction. In practice model uncertainty is a fact of life and likely to be more serious than other sources of uncertainty which have received far more attention from statisticians. This is true whether the model is specified on subject-matter grounds or, as is increasingly the case, when a model is formulated, fitted and checked on the same data set in an iterative, interactive way. Modern computing power allows a large number of models to be considered and data-dependent specification searches have become the norm in many areas of statistics. The term data mining may be used in this context when the analyst goes to great lengths to obtain a good fit. This paper reviews the effects of model uncertainty, such as too narrow prediction intervals, and the non-trivial biases in parameter estimates which can follow data-based modelling. Ways of assessing and overcoming the effects of model uncertainty are discussed, including the use of simulation and resampling methods, a Bayesian model averaging approach and collecting additional data wherever possible. Perhaps the main aim of the paper is to ensure that statisticians are aware of the problems and start addressing the issues even if there is no simple, general theoretical fix.

995 citations

Book
01 Jan 2001
TL;DR: Chris Chatfield looks at the chequered history of forecasting and asks how to do it better using time-series data, and what statistical techniques and models might help us.
Abstract: As we go through life, everyone makes forecasts all the time, often without realising it. Sadly these forecasts are often (very) inaccurate. Chris Chatfield looks at the chequered history of forecasting and asks how we might do it better using time-series data, and what statistical techniques and models might help us.

795 citations

Journal ArticleDOI

397 citations


Cited by
More filters
Journal ArticleDOI
TL;DR: In this paper, the authors provide guidance for substantive researchers on the use of structural equation modeling in practice for theory testing and development, and present a comprehensive, two-step modeling approach that employs a series of nested models and sequential chi-square difference tests.
Abstract: In this article, we provide guidance for substantive researchers on the use of structural equation modeling in practice for theory testing and development. We present a comprehensive, two-step modeling approach that employs a series of nested models and sequential chi-square difference tests. We discuss the comparative advantages of this approach over a one-step approach. Considerations in specification, assessment of fit, and respecification of measurement models using confirmatory factor analysis are reviewed. As background to the two-step approach, the distinction between exploratory and confirmatory analysis, the distinction between complementary approaches for theory testing versus predictive application, and some developments in estimation methods also are discussed.

34,720 citations

Book
08 Sep 2000
TL;DR: This book presents dozens of algorithms and implementation examples, all in pseudo-code and suitable for use in real-world, large-scale data mining projects, and provides a comprehensive, practical look at the concepts and techniques you need to get the most out of real business data.
Abstract: The increasing volume of data in modern business and science calls for more complex and sophisticated tools. Although advances in data mining technology have made extensive data collection much easier, it's still always evolving and there is a constant need for new techniques and tools that can help us transform this data into useful information and knowledge. Since the previous edition's publication, great advances have been made in the field of data mining. Not only does the third of edition of Data Mining: Concepts and Techniques continue the tradition of equipping you with an understanding and application of the theory and practice of discovering patterns hidden in large data sets, it also focuses on new, important topics in the field: data warehouses and data cube technology, mining stream, mining social networks, and mining spatial, multimedia and other complex data. Each chapter is a stand-alone guide to a critical topic, presenting proven algorithms and sound implementations ready to be used directly or with strategic modification against live data. This is the resource you need if you want to apply today's most powerful data mining techniques to meet real business challenges. * Presents dozens of algorithms and implementation examples, all in pseudo-code and suitable for use in real-world, large-scale data mining projects. * Addresses advanced topics such as mining object-relational databases, spatial databases, multimedia databases, time-series databases, text databases, the World Wide Web, and applications in several fields. *Provides a comprehensive, practical look at the concepts and techniques you need to get the most out of real business data

23,600 citations

Journal ArticleDOI
TL;DR: A new coefficient is proposed to summarize the relative reduction in the noncentrality parameters of two nested models and two estimators of the coefficient yield new normed (CFI) and nonnormed (FI) fit indexes.
Abstract: Normed and nonnormed fit indexes are frequently used as adjuncts to chi-square statistics for evaluating the fit of a structural model A drawback of existing indexes is that they estimate no known population parameters A new coefficient is proposed to summarize the relative reduction in the noncentrality parameters of two nested models Two estimators of the coefficient yield new normed (CFI) and nonnormed (FI) fit indexes CFI avoids the underestimation of fit often noted in small samples for Bentler and Bonett's (1980) normed fit index (NFI) FI is a linear function of Bentler and Bonett's non-normed fit index (NNFI) that avoids the extreme underestimation and overestimation often found in NNFI Asymptotically, CFI, FI, NFI, and a new index developed by Bollen are equivalent measures of comparative fit, whereas NNFI measures relative fit by comparing noncentrality per degree of freedom All of the indexes are generalized to permit use of Wald and Lagrange multiplier statistics An example illustrates the behavior of these indexes under conditions of correct specification and misspecification The new fit indexes perform very well at all sample sizes

21,588 citations

Journal ArticleDOI
TL;DR: In this article, a step-by-step guide to wavelet analysis is given, with examples taken from time series of the El Nino-Southern Oscillation (ENSO).
Abstract: A practical step-by-step guide to wavelet analysis is given, with examples taken from time series of the El Nino–Southern Oscillation (ENSO). The guide includes a comparison to the windowed Fourier transform, the choice of an appropriate wavelet basis function, edge effects due to finite-length time series, and the relationship between wavelet scale and Fourier frequency. New statistical significance tests for wavelet power spectra are developed by deriving theoretical wavelet spectra for white and red noise processes and using these to establish significance levels and confidence intervals. It is shown that smoothing in time or scale can be used to increase the confidence of the wavelet spectrum. Empirical formulas are given for the effect of smoothing on significance levels and confidence intervals. Extensions to wavelet analysis such as filtering, the power Hovmoller, cross-wavelet spectra, and coherence are described. The statistical significance tests are used to give a quantitative measure of change...

12,803 citations

Book
23 Nov 2005
TL;DR: The treatment is comprehensive and self-contained, targeted at researchers and students in machine learning and applied statistics, and deals with the supervised learning problem for both regression and classification.
Abstract: A comprehensive and self-contained introduction to Gaussian processes, which provide a principled, practical, probabilistic approach to learning in kernel machines. Gaussian processes (GPs) provide a principled, practical, probabilistic approach to learning in kernel machines. GPs have received increased attention in the machine-learning community over the past decade, and this book provides a long-needed systematic and unified treatment of theoretical and practical aspects of GPs in machine learning. The treatment is comprehensive and self-contained, targeted at researchers and students in machine learning and applied statistics. The book deals with the supervised-learning problem for both regression and classification, and includes detailed algorithms. A wide variety of covariance (kernel) functions are presented and their properties discussed. Model selection is discussed both from a Bayesian and a classical perspective. Many connections to other well-known techniques from machine learning and statistics are discussed, including support-vector machines, neural networks, splines, regularization networks, relevance vector machines and others. Theoretical issues including learning curves and the PAC-Bayesian framework are treated, and several approximation methods for learning with large datasets are discussed. The book contains illustrative examples and exercises, and code and datasets are available on the Web. Appendixes provide mathematical background and a discussion of Gaussian Markov processes.

11,357 citations