scispace - formally typeset
Open AccessJournal ArticleDOI

General-to-specific Modeling: An Overview and Selected Bibliography

TLDR
The theory of reduction is reviewed, the approach of general-to-specific modeling is summarized, and the econometrics of model selection are discussed, noting that general- to- specific modeling is the practical embodiment of reduction.
Abstract
This paper discusses the econometric methodology of general-to-specific modeling, in which the modeler simplifies an initially general model that adequately characterizes the empirical evidence within his or her theoretical framework. Central aspects of this approach include the theory of reduction, dynamic specification, model selection procedures, model selection criteria, model comparison, encompassing, computer automation, and empirical implementation. This paper thus reviews the theory of reduction, summarizes the approach of general-to-specific modeling, and discusses the econometrics of model selection, noting that general-to-specific modeling is the practical embodiment of reduction. This paper then summarizes fifty-seven articles key to the development of general-to-specific modeling.

read more

Content maybe subject to copyright    Report

Board of Go vernors of the Federal Reserv e System
In ter na tional Finance Discussion Papers
Number 838
Augu st 2005
General-to-specicModeling:
An Overview and Selected Bibliography
Julia Campos, Neil R. Ericsson, and David F. Hendry
NOTE: International Finance Discussion Papers are preliminary materials circulated
to stimulate discussion and critical comment. Referen ces to Internation al F inance
Discussion Papers (other than an ackn owledgm ent that the writer has had access to
unpu blished material) should be cleared with the author or authors. Recent IFDPs
are a vailable at www .federa lreserve.gov/pubs/ifdp / on the Web. This paper can
be dow nloaded without charge from the Social Science Research Network electronic
library at www.ssrn.com.

General-to-specicModeling:
An Overview and Selected Bibliography
JuliaCampos,NeilR.Ericsson,andDavidF.Hendry
Abstract : This paper discusses the econ om etric methodology of general-to-specic
modeling, in wh ich the modeler simplies an initially general model that adequa tely
c haracterizes the empirical evidence within his or her theoretical framework. Central
aspects of this app roach inclu de the theory of reduction, dynamic specication, model
selection procedures, model selection criteria, model comparison, encompassing, com-
puter automa tion, and empirical implementation. This paper thus reviews the theory
of reduction, summa rizes the appro ach of general-to-specic modeling, and discusses
the econom etr ics of model selection, noting that general-to-specicmodelingisthe
practical em bodiment of reduction. This paper then summ ar izes ft y-s even articles
k e y to the development of general-to-specicmodeling.
Keywor ds: coin tegration, conditional m odels, data min ing, diagnostic testing, dy-
namic specication, econometric methodology, encom passin g, equilibrium correction
models, error correction models, exogeneity, general-to-specic modeling, model com-
parison, model design, model evaluation , model selection, non-nested hypotheses,
PcG ets, PcGive, reduction, specic-to-general modeling.
JEL classications:C1,C5.
Forthcoming as the editors’ introduction to a two-v olume set of readings entitled General-to-
SpecicModelling, Julia Campos, Neil R. Ericsson, and David F. Hendry (eds.), Edward Elgar
Publishing, Cheltenham, 2005. The rst author is a professor of econometrics in the Departamento
de Economía e Historia Económica, Facultad de Economía y Empresa, Universidad de Salamanca,
Salamanca 37008 España (Spain). The second author is a sta economist in the Division of In-
ternational Finance, Board of Governors of the Federal Reserve System, Washington, D.C. 20551
U.S.A. The third author is an ESRC Professorial Researc h Fellow and the head of the Economics
Department at the University of Oxford, Oxford, England. They may be reached on the Internet at
jcampos@usal.es, ericsson@frb.gov, and david.hendry@economics.ox.ac.uk respectively. The views
in this paper are solely the responsibility of the authors and should not be interpreted as reecting
the views of the Board of Governors of the Federal Reserve System or of any other person associated
with the Federal Reserve System. The authors are indebted to Jonathan Halket, Jaime Marquez,
and Kristian Rogers for helpful comments; and to Nicola Mills for seeing the production of this
work through to completion. Scientic Workplace (MacKich an Soft ware, Inc., Poulsbo, Washing-
ton, U.S.A.) eased the preparation of this paper in L
A
T
E
X. This discussion paper is available from the
authors and at www.federalreserve.gov/pubs/ifdp/2005/838/default.htm on the WorldWide Web.

1 Motiva t ion and Overvie w
This paper focuses on a central method for selecting useful empirical models, called
general-to-specicmodeling. In this method, the m odeler simplies an initially gene ral
model that adequately characterizes the empirical eviden ce within his or her theo-
retical framew ork. While the methodological, statist ical, and emp irical foun d ations
for general-to-specic modeling ha ve been laid do w n over the last several decades, a
burst of activity has occurred in the last half-dozen ye ars, stimulated in fair part b y
Hoover and Perez’s (1999a) dev elopm ent and analysis of a comp uter algorithm for
general-to-specic modeling. The papers discussed herein detail ho w the subject has
advanced to its presen t stage of success and convey the promise of these dev elopm ents
for future empirical researc h . The remainde r of this o verview mo tivates the in te rest in
general-to-specic modeling and summarizes the structure of the subsequent sections
(Sections 2—5).
Econ omists have long sought to develop quantitative models of economic beha v ior
b y blending economic theory with data evidence. The task has pro ved an arduous
onebecauseofthenatureoftheeconomymodeled,theeconomictheory,andthe
data evidence. The economy is a complicated, dynamic, nonlinear, sim u ltaneous,
high-dimensional, and ev olving en tit y; social systems alter o ver time; law s change;
and tec hnological innovations occur. Thus, the target is not only a moving one; it be-
ha v es in a distinctly nonstationary manner, both evolving over time and being subject
to sudden and unanticipated shifts. Econom ic theories are highly abstract and sim-
plied ; and they also chan ge ov er time, with conicting rival explanation s sometimes
coexisting. T he data evidence is tarnished: economic magnitud es are inaccurately
measured and subject to substantiv e revisions, and man y important variables are not
even observable. The data themselv es are often time series wher e samples are short,
highly aggregated, heterogeneous, time-dependen t, and in ter-dependen t. Econom et-
ric modeling of econom ic time series has nev e rtheless strived to disco ver sustainab le
and interpreta ble relationships between observed economic variab les. T his paper
focuses on general-to-specic modeling, in wh ich the modeler simplies an initially
general model that adequately c haracterizes the empirical evidence within his or her
theoretical framework. T h is method has proved useful in practice for selecting em-
pirical economic models.
The diculties of empirical modeling are well reected in the slow ness of empirical
progress, providing plen ty of ammunitio n for critics. Howev er , part of the problem
may be internal to the discipline, deriving from inappropriate modeling methods.
The “con ventional” approac h insists on a complete theoretical model of the phe-
nomena of interest prior to data analysis, leaving the empirical evidence as little
more than quantitative clothing. Unfortunately, the com plexity and nonstationar-
it y of economies ma kes it impro b able than an yone–ho wever brilliant–co uld deduce
apriorithe multitude of quantitative equations c haracterizin g the behavior of mil-
lions of disparate and competing agents. Without a radical change in the discipline’s
meth odology, empirical progress seems doomed to remain slo w.
The situation is not as bleak as just described, for two reasons. First, the accu-
mulation of kno w ledge is progressive, implying that one does not need to know all
1

the answers at the start. Otherwise, no science could have advanced. Although the
best empirical model at any given time ma y be supplanted later, it can provide a
springboard for further discovery. Data-b ase d m odel selection need not raise serious
concerns: this im plication is established below and is demonstrated b y the actual
behavior of model selection algorithms.
Second, inconsistencies between the implications of any conjectured model and
the observ ed data are often easy to detect. The ease of model rejection worries some
economists, yet it is also a powerful advantage by helping sort out which models are
empirically adequate and which are not. Constructive progress may still be dicu lt
because “w e don’t know what w e don’t kno w, and so w e cannot know how best to
ndoutwhatwedontknow. Thedichotomybetweenmodeldestructionandmodel
construction is an old one in the p hilosop hy of science. Wh ile critical evaluation
of empirical evidence is a destructive use of econometrics, it can also establish a
legitimate basis for empirical models.
To undertake empirical modeling, one m ust begin b y assuming a probabilit y struc-
ture for the data, which is tantamou nt to conjecturing the data generating process.
Because the economic mechanism is itself unkn own, the relevant probability struc-
ture is also unknown, so one must proceed iterativ ely: conjecture the data generation
process (DG P ), develop the associated probabilit y theory, use that theory for model-
ing empirical evidence, and revise the starting point when the results do not match
consistently. The developm ent of economet ric theory highlig hts this iterativ e pro-
gression: from stationarit y assum ptions, through integrated-cointegrated systems, to
general nonstationa ry mixing processes, as empirica l evidence revealed the inadequacy
of each earlier step. Further developmen ts will undoubtedly occur, leading to a still
more useful foundation for empirical modeling. See Hendr y (1995a) for an extensive
treatment of progressive researc h strategies.
Ha ving postulated a reasonable probability basis for the DGP, including the pro-
cedures used for data measurem ent and its collection, the next issue concerns what
classes of model migh t be useful. The theory of reduction (discussed in Section 2)
explains ho w empirical models arise and what their status is, noting that they are
not facsimiles of the DGP. Specically, empirical models describe the beha v ior of a
relatively small set of variables–often from one to seve ral h und red–and neve r the
many millions of distinct variab les present in most economies.
A k ey concept here is that of the local DG P, which is the probability mec h anism
in the space of those variables under analysis. The theory of reduction shows how
the local DGP arises as a simplication of a vastly more general DGP involving
millions of variables. The usefulness of a given local DGP depends on it capturin g
sustainable links, whic h in turn depends par tly on the theor etica l fram e work and
partly on data accuracy. An econometric model cannot do better than capture the
salien t characteristics of its corresponding local DGP . The extent to wh ich the model
does capture those characteristics depends both on its specication at least embedding
the local DG P and on the goodness of its selection.
Ther e are thu s two distinct conceptual steps in modeling, albeit ones closely re-
lated in practice. First, specify a useful information set for a “well-behaved” local
DGP. Second, select a “good” empirical model of that local DGP.
2

A viable methodology for empirical modeling is an in tegral component of ac hieving
the second step. Despite the controversy surrounding every aspect of econometric
methodology, the “LSE” (or London Sch ool of Economics) approac h has emerged as
a leading me thodology for empirical modeling; see Hend ry (1993) for an o verview.
One of the LSE appro ach’s main tenets is general-to-specic modeling, sometimes
abbreviated as Gets. In general-to-specic modeling, empirical analysis starts with a
general statistical model that captures the essential c ha racteristics of the underlyin g
dataset, i.e., that general model is congruen t. Then , that general model is reduced in
comp lex ity by elimin atin g statistically insignicant variables, checking the va lidity of
the reductions at ev ery stage to ensure congruence of the n ally selected model.
The papers discussed below articulate many reasons for adopting a genera l-to-
specic approac h. First amongst these reasons is that general-to-specic modeling
implem ents the theory of reduction in an empirical context. Section 2 summar izes
the theory of reduction, and Section 3 discusses general-to-specicmodelingasthe
empirica l analogue of reduction .
General-to -specic modeling also has excellent c ha racteristics for model selection,
as documented in Monte Carlo studies of autom atic general-to -specic modeling al-
gorithms. Hoo ver and Perez (1999a) we re the rsttoevaluatetheperformanceof
general-to-specic modeling as a general approach to econometric model building. To
analyze the general-to-specic approach systematica lly, Hoo ver and Perez mechanized
the dec is ions in gene ral-to - s pecic modeling by coding them in a compu ter algorithm.
In doing so, Hoover and Perez also mad e important advances in practical modeling.
To appreciate Hoover and Perez’s con trib utions to general-to-specic m odeling,
consider the m ost basic steps that such an algorithm follo ws.
1. Ascertain that the general statistical model is congruent.
2. Eliminate a variable (or variab les) that satises the selection (i.e., simp lication)
criteria.
3. Chec k that the simplied model remains congruent.
4. Contin u e steps 2 and 3 until none of the rem a inin g variables can be eliminated.
P agan (1987) and other critics have argued that the outcome of general-to-specic
modeling ma y depend on the simplication path chosen– tha t is, on the order in
which var iables are eliminated an d on the data transformations adop ted and so the
selected m odel might vary with the in vestigator. Many redu ction paths certainly
could be considered from aninitialgeneralmodel.
Hoover and Perez (1999a) turned this potential draw b ack in to a virtue b y explor-
ing man y feasible paths and seeing which models result. When searches do lead to
dierent model selections, encom passing tests can be used to discrim ina te between
these models, with only the surviving (po ssibly non-nested) specication s retained. If
multiple models are found that are both congruen t and encompassing, a new general
model can be formed from their union, and the simplicationprocessisthenre-
applied. If that union model re-occurs, a nal selection am on g the competing models
can be made by using (say) information criteria. Oth erw ise, a unique, congruent,
encom passin g reduction has been loca ted.
3

Citations
More filters
Journal ArticleDOI

Promoting novelty, rigor, and style in energy social science: towards codes of practice for appropriate methods and research design

TL;DR: It is the hope that this Review will inspire more interesting, robust, multi-method, comparative, interdisciplinary and impactful research that will accelerate the contribution that energy social science can make to both theory and practice.
Journal ArticleDOI

Forecasting and operational research : a review

TL;DR: It is argued that the unique contribution that OR can continue to make to forecasting is through developing models that link the effectiveness of new forecasting methods to the organizational context in which the models will be applied.
Journal ArticleDOI

Mortality Density Forecasts: An Analysis of Six Stochastic Mortality Models

TL;DR: In this paper, the authors investigate the uncertainty of forecasts of future mortality generated by a number of previously proposed stochastic mortality models, with the conclusion that model risk can be significant.
ReportDOI

What Segments Equity Markets

TL;DR: In this paper, a new, valuation-based measure of world equity market segmentation is proposed, where the authors identify a country's political risk profile and its stock market development as two additional local segmentation factors as well as the U.S. corporate credit spread as a global segmentation factor.
Journal ArticleDOI

Mortality density forecasts: An analysis of six stochastic mortality models

TL;DR: In this paper, the suitability of six stochastic mortality models for forecasting future mortality and estimating the density of mortality rates at different ages is discussed, with a focus on the plausibility of their forecasts: biological reasonableness, predicted levels of uncertainty, and the robustness of the forecasts relative to the sample period used to fit the model.
References
More filters

Estimating the dimension of a model

TL;DR: In this paper, the problem of selecting one of a number of models of different dimensions is treated by finding its Bayes solution, and evaluating the leading terms of its asymptotic expansion.
Journal ArticleDOI

Distribution of the Estimators for Autoregressive Time Series with a Unit Root

TL;DR: In this article, the limit distributions of the estimator of p and of the regression t test are derived under the assumption that p = ± 1, where p is a fixed constant and t is a sequence of independent normal random variables.
Journal ArticleDOI

Specification Tests in Econometrics

Jerry A. Hausman
- 01 Nov 1978 - 
TL;DR: In this article, the null hypothesis of no misspecification was used to show that an asymptotically efficient estimator must have zero covariance with its difference from a consistent but asymptonically inefficient estimator, and specification tests for a number of model specifications in econometrics.
Journal ArticleDOI

Statistical analysis of cointegration vectors

TL;DR: In this paper, the authors consider a nonstationary vector autoregressive process which is integrated of order 1, and generated by i.i.d. Gaussian errors, and derive the maximum likelihood estimator of the space of cointegration vectors and the likelihood ratio test of the hypothesis that it has a given number of dimensions.
Related Papers (5)
Frequently Asked Questions (12)
Q1. What are the contributions in this paper?

This paper discusses the econometric methodology of general-to-specific modeling, in which the modeler simplifies an initially general model that adequately characterizes the empirical evidence within his or her theoretical framework. This paper thus reviews the theory of reduction, summarizes the approach of general-to-specific modeling, and discusses the econometrics of model selection, noting that general-to-specific modeling is the practical embodiment of reduction. This paper then summarizes fifty-seven articles key to the development of general-to-specific modeling. 

This paper details how the subject has advanced to its present stage of success and should convey the promise of these developments for future empirical research. 

A variety of approaches are still extant, such as “classical” and “Bayesian”, and with varying degrees of emphasis on least squares, instrumental variables, maximum likelihood, method of moments, and so on. 

By establishing the link between DGPs and empirical models, the theory of reduction justifies both evaluative and constructive aspects of econometric modeling. 

While the construction of econometric models requires solving more complicated problems than the order of a polynomial, Anderson’s results have three direct implications for general-to-specific modeling. 

Computers have long played an essential role in the operational development of econometrics, particularly for estimation, inference, and simulation. 

Non-exogenous variables are called predetermined if consistent and asymptotically unbiased estimates of the parameters of the subsystem of non-exogenous non-predetermined variables can be obtained separately from those of the subsystem of predetermined variables. 

Phillips shows that errors correlated with explanatory variables can be avoided by including in the estimatedequation all factors affecting directly or indirectly the explained variable as well as exogenous factors. 

To illustrate the use of a pre-search filter, suppose that the general unrestricted model has 40 variables, of which 30 are irrelevant and 10 are relevant, and where the latter have population t-values equal to two in absolute value. 

The extent to which the model does capture those characteristics depends both on its specification at least embedding the local DGP and on the goodness of its selection. 

The interpretation of test statistics in the final model may be at issue because of the potential consequences of empirical model selection. 

Hendry and Mizon’s procedure consists of formulating a congruent VAR that embeds all the contending dynamic structural econometric models (SEMs) being evaluated, and testing which of those models is a valid reduction of that VAR.