scispace - formally typeset
Search or ask a question
Journal ArticleDOI

Estimation of ergodic agent-based models by simulated minimum distance

01 Feb 2015-Journal of Economic Dynamics and Control (Masanao Aoki and Stephen J Turnovsky)-Vol. 51, pp 148-165
TL;DR: In this article, the authors show how to consistently estimate ergodic models by simulated minimum distance techniques, both in a long-run equilibrium and during an adjustment phase, under a variety of conditions.
About: This article is published in Journal of Economic Dynamics and Control.The article was published on 2015-02-01 and is currently open access. It has received 122 citations till now. The article focuses on the topics: Ergodic theory.

Summary (6 min read)

1 Introduction

  • Agent-based (AB) models are sometimes considered as a candidate to replace or at least complement dynamic stochastic general equilibrium (DSGE) as the standard tool for macroeconomic analysis.
  • When present, this is often limited to some ad-hoc calibration of the relevant parameters; this resembles the state of the art in DSGE modeling a few years ago, which has now moved forward toward more formal estimation.
  • The authors then turn to some basic questions concerning estimation of AB models:.
  • Therefore, the aggregate properties of an AB model remain hidden in the complexity of the relations among the different elements and the different layers (micro, macro and possibly meso) of the system.
  • In section 6 the authors give two examples of estimation of ergodic AB models: in the first, estimation is 3In this paper they focus on estimation of ergodic models.

2 Little AB models grow big

  • AB models have long been considered as theoretical exercises aimed at investigating the macro effects arising from the interaction of many individuals –each following possibly simple rules of behavior– or the individual routines/strategies underlying some observed macro phenomenon (Richiardi, 2012).
  • Here, empirical validation is crucial, as the lack of empirical relevance is the ultimate critique that has been moved to DSGE models.
  • While in the last decade the literature on DSGE models has evolved from simple calibration to proper estimation, the AB macro models are still lagging behind, although some sophisticated examples of calibration are starting to close the gap (Bianchi et al., 2007, 2008; Fabretti, 2012; Cirillo and Gallegati, 2012).6.
  • As such, estimation is concerned with the properties of the estimators and the quantification of the uncertainty around the estimates.
  • Winker and Gilli (2001) and Gilli and Winker (2003) estimate respectively two and three parameters of an AB model of the foreign exchange market, by employing the method of simulated moments (MSM).

3 Estimation of DSGE models

  • In DSGE models aggregation is generally not a problem, thanks to a very low level of heterogeneity.
  • A common strategy to solve the models involves the linearization of first order conditions and constraints by means of a Taylor approximation around the steady state.9.
  • In presence of nonstationarity, the model is rescaled and detrended, in order to make it stationary around a balanced growth path; the trend is often assumed to follow an exogenous stochastic process.
  • Turning to estimation methods, old vintage DSGE models (as most current AB macro models) where mainly calibrated: the values of the parameters were chosen according to some external knowledge, theoretical belief or empirical criterion.
  • The tide has however turned in favor of a more formal estimation approach.

3.1 Maximum likelihood

  • This is the standard method for estimating DSGE models, so it deserves a little more space.
  • This raises serious identification issues (see subsection 5.2.2 below).
  • If however, the state space representation is not linear and the shocks are not normal, filtering becomes more complicated and must be performed by means of simulation (for instance, with the use of the so-called particle filters –see below): the conditional distribution of the states given the past observations and the parameters is replaced by a simulated distribution.
  • In turn, the flatness of the likelihood 12Strictly speaking, stochastic singularity is a feature of linearized DSGE models, but it may also have implications for the estimation of nonlinear models depending on the extent to which they differ from their linearized counterparts.
  • 17 Alternatively, one could think of using simulated empirical maximum likelihood to approximate the likelihood function, as for instance in Kristensen and Shin (2012).

3.2 Simulated minimum distance

  • The main alternatives to ML are GMM, MSM and II (or extended method of simulated moments, EMSM).
  • SMD summarizes the real system and the artificial system by means of a set of functions on the observed and artificial data and minimizes the difference between the “short summaries” of the two systems.
  • Often, a vector autoregression (VAR) is used as meta-model.
  • The latter is a weaker restriction because it is possible to find independent moments that incorporate information about more variables than those that are linearly independent.” (Ruge-Murcia, 2007, p. 2600).
  • He concludes that moment-based estimation methods (GMM and MSM) compare very favorably to the more widely used method of ML.

4 AB models as recursive systems

  • This is an essential feature as “[t]he elimination of simultaneous equations allows us to get results from a 19Indirect inference has a very interesting application which allows to estimate a model without (directly using) the data.
  • While eq. (2) has an explicit analytical formulation in DSGE models, it remains only implicitly defined by the micro transition equations (1) in AB models.
  • 21Here and in the following the authors use “behavioral rules” and similar terms in a loose sense that encompasses the actual intentional behaviors of individuals as well as other factors such as technology etc. 22For a discussion of the Markov chain representation of AB models, see Izquierdo et al. (2009).
  • In general it may happen that, for a given MC, some projections are Markov and others are not.
  • Let Z0 = {X0, s} be the set of initial conditions of the simulation, where X0 is the initial state of the system and s stands for the seed(s) used by the random number generator(s) in the simulation, which determine(s) the evolution of ξt and κt.

4.1 Equilibrium

  • One important difference between DSGE and AB models lies in the definition of equilibrium.
  • On the contrary, if Yt is stationary but not ergodic, different (absorbing and/or transient) equilibria are obtained, for the same values of the parameters, depending on the initial conditions.
  • Its properties are then suited for estimation, also known as equilibrium.
  • This new regularity breaks down when GDP reaches its steady state: it is therefore a transient statistical equilibrium.

5 Estimation of AB models

  • In the following the authors assume that the model is correctly specified: this is a fundamental hypothesis, which implies that the model perfectly describes the real world, that the structural parameters exist and have a finite value and that all the parameters in the real system are represented in the model.
  • This can be seen directly in eq. (3): while it is always theoretically possible to solve for Y t by substituting in eqs.
  • At an empirical level, ergodicity of the true DGP cannot be assessed in the real data, as Nature only plays once, and is therefore assumed (or derived from the assumption of correct specification).
  • To summarize, there are two issues in estimation of AB models, and they both descend from the fact that the properties of the model are not analytically known: the objective function used for estimation must be numerically evaluated, and ergodicity and stationarity of the aggregated series used for estimation must be tested.
  • The first problem is common to other modeling strategies (as in DSGE), and calls for the use of simulation-based econometric techniques (Stern, 1997, 2000).

5.1 Consistency in minimum distance and simulated minimum distance esti-

  • Mators Define a set of estimators ĥn and a vector of functions h(θ), where the estimators are functions of the data, the functions h(θ) are the mapping between the model and the estimators and θ are the structural parameters.
  • The minimum distance estimator belongs to the class of extremum estimators, which also includes maximum likelihood, nonlinear least square and generalized method of moments.
  • The theorem states that if Q̂n(θ) converges to Q0(θ) for all θ ∈ Θ and Q0(θ) is maximized only at the true parameters, then the limit of the maximum θ̂ is the maximum of the limit θ0.
  • If it is possible (in terms of computational time) to simulate the theoretical moments for each possible combination of the parameters, the technique gives the value of the objective function for all θ ∈ Θ, so that a θ̂ is chosen such that the objective function is maximized.
  • The authors now turn to investigate under which conditions an AB model can be consistently estimated by SMD.

5.2 Consistency conditions and AB models

  • Consistency conditions for AB models refer to the properties of the statistics used for estimation.
  • The attention therefore turns from the objective function to the constituents of the objective function.
  • Preliminarily, note that in a computer program as an AB model the objective function is discrete by construction, since only a countable number of values of the parameters can be tested.
  • 32Fixing the random numbers across iterations is required to achieve uniform convergence (McFadden, 1989; Pakes and Pollard, 1989).

5.2.1 Convergence

  • The first essential condition for convergence of the objective function is the convergence of its elements to their true value.
  • Each statistics used in the objective function should therefore be tested for stationarity in the simulated data.
  • The second is that the observed moments converge in probability to their expected value and in turn the objective function Q̂n(θ) converges in probability to Q0(θ).
  • Note that the authors are talking here about convergence in time.
  • This is a crucial point which distinguishes estimation from many calibration exercises, where cross-sectional statistics are used to characterize a model.37.

5.2.2 Identification

  • The objective function does not display a unique minimum), under-identification (when the objective function is independent of certain structural parameters), partial identification (when some parameters enter the objective function only as a linear combination of other parameters), weak identification (when the objective function has a unique minimum, but its curvature is very flat at the minimum).
  • As they argue, “[t]hese problems turn out to be inherent to DSGE models –they are typically produced because the mapping from the structural parameters to the coefficients of the solution is ill-behaved– and standard choices of data summaries and of objective functions may make identification deficiencies worse” (p. 432).
  • A common practice when dealing with identification issues is to reduce the number of parameters to be estimated (for instance by dropping observationally equivalent parameters from the specification) or calibrate some of the parameters and estimate the others, conditional on the calibrated values.
  • Monotonicity implies that the only point in the parameter space in which all h(θ) = h0 is θ0.

5.3 Small-sample estimation bias

  • If the moments are a nonlinear function of the parameters, the authors get a bias in the estimates.
  • This is easy to show in the case of perfect identification when there is only one parameter and one moment.
  • The observed moment can be written as h(θ0, ζn) = h(θ0) + ζn (10) The observed moment is the “true moment” plus an error ζn which depends on the sample (and on sample size).
  • In the same way it is possible to show that if the moment function is concave, the estimated parameter is upward biased if the moment function is increasing and downward biased if the moment function is decreasing.
  • For simplicity the authors omit here to consider this additional source of variability.

5.4 Summary: estimation of ergodic models

  • As the authors have seen, estimating an ergodic AB model by means of simulated minimum distance is conceptually straightforward, the crucial step being the choice of appropriate longitudinal statistics (longitudinal “moments”) to match simulated and real data.
  • In order for these longitudinal moments to be meaningful quantities, the time series considered (both real and simulated) must be stationary: (weak) stationarity implies, in loose terms, that the theoretical mean and variance are constant in time.
  • Stationarity can be a permanent property of the time series, that is, once it becomes stationary, the time series remains stationary.
  • This the authors call an absorbing equilibrium of the model, with respect to the time series considered.
  • The observed moments are then constructed by averaging over the available real data.

6 Examples

  • While the second is estimated in a transient equilibrium.the authors.
  • In both cases, because the moments used for estimation are non-linear, the authors get a small-sample bias (of predictable direction).
  • All estimation strategies are explored by means of Monte Carlo experiments: pseudo-true data are created by running the model with some chosen value of the parameters; then, estimates are obtained by matching the simulated moments to the moments computed on the pseudo-true data.
  • This procedure is repeated many times in order to get a distribution of the estimates.

6.1 Estimation in an absorbing equilibrium

  • The model used here is an AB stock market model proposed in Cliff and Bruten (1997) to reproduce the experimental results obtained by Smith (1962), showing how a small number of inexperienced traders converge rapidly to a competitive equilibrium under the double auction mechanism (Smith, 1962, p.157).
  • In each period traders look at the book and define a target price τi(t): traders increase their target price if the last trade occurred at a high price, and lower it otherwise –see Cliff and Bruten (1997) for details.
  • The behavior around the equilibrium price, on the other hand, depends on β.
  • The variance is convex in the parameter: this introduces a downward bias in the estimates.
  • Figure 3 shows the shape of the objective function, for one particular Monte Carlo experiment (pseudo-true series) using the variance (left panel) and the standard deviation (right panel).

6.2 Estimation in a transient equilibrium

  • In that paper the authors estimate an AB version of the well known Bass (1969) model, which provides a mathematical explanation of the different stages of product adoption described in Rogers (1962) (innovators, early adopters, early majority, large majority, and laggards), and formalizes the crucial distinction between innovators and imitators.
  • The authors therefore focus on the adjustment process, looking for regularities which define the transient equilibria of the system.
  • First, the authors get an estimate p̂ for the external influence parameter, conditional on a specific value of the population parameter m, by matching the observed and simulated adoption rate at t = 0.
  • Because the moment τ is not linear in m (figure 5), the authors get an upward small sample bias.

7 Conclusions

  • In this paper the authors have identified simulated minimum distance as a natural approach to estimation of AB models.
  • The theoretical quantities or statistics used for characterizing the model conditional on the values of the parameters (the moments, for instance), for which no analytical expression is available in AB models, are replaced by their simulated counterparts.
  • This requires that these statistics are appropriately chosen so that their estimates in the simulated data converge to the theoretical values.
  • One can think of many open issues and avenues for research.the authors.
  • Finally, one could think of DSGE and AB models as a mechanism for generating priors, rather than a model of the data.

Did you find this useful? Give us your feedback

Figures (5)
Citations
More filters
Book ChapterDOI
TL;DR: A review of the existing validation techniques for agent-based models in economics can be found in this paper, where the authors sketch a simple theoretical framework that conceptualizes existing validation approaches, which examine along three different dimensions: (i) comparison between artificial and real-world data; (ii) calibration and estimation of model parameters; and (iii) parameter space exploration.
Abstract: Since the survey by Windrum et al. (Journal of Artificial Societies and Social Simulation 10:8, 2007), research on empirical validation of agent-based models in economics has made substantial advances, thanks to a constant flow of high-quality contributions. This Chapter attempts to take stock of such recent literature to offer an updated critical review of the existing validation techniques. We sketch a simple theoretical framework that conceptualizes existing validation approaches, which we examine along three different dimensions: (i) comparison between artificial and real-world data; (ii) calibration and estimation of model parameters; and (iii) parameter space exploration. Finally, we discuss open issues in the field of ABM validation and estimation. In particular, we argue that more research efforts should be devoted toward advancing hypothesis testing in ABM, with specific emphasis on model stationarity and ergodicity.

194 citations

Journal ArticleDOI
TL;DR: In this article, the authors present a critical discussion of the theoretical, empirical and political-economy pitfalls of the DSGE-based approach to policy analysis and suggest that a more fruitful research avenue to pursue is to explore alternative theoretical paradigms, which can escape the strong theoretical requirements of neoclassical models (e.g., equilibrium, rationality, representative agent, etc.).
Abstract: The Great Recession seems to be a natural experiment for macroeconomics showing the inadequacy of the predominant theoretical framework - the New Neoclassical Synthesis - grounded on the DSGE model. In this paper, we present a critical discussion of the theoretical, empirical and political-economy pitfalls of the DSGE-based approach to policy analysis. We suggest that a more fruitful research avenue to pursue is to explore alternative theoretical paradigms, which can escape the strong theoretical requirements of neoclassical models (e.g., equilibrium, rationality, representative agent, etc.). We briefly introduce one of the most successful alternative research projects - known in the literature as agent-based computational economics (ACE) - and we present the way it has been applied to policy analysis issues. We then provide a survey of agent-based models addressing macroeconomic policy issues. Finally, we conclude by discussing the methodological status of ACE, as well as the (many) problems it raises.

173 citations

Posted Content
01 Jan 2015
TL;DR: Issues surrounding variance stability, sensitivity analysis, spatio-temporal analysis, visualization, and effective communication of all these to non-technical audiences, such as various stakeholders are examined.
Abstract: The proliferation of agent-based models (ABMs) in recent decades has motivated model practitioners to improve the transparency, replicability, and trust in results derived from ABMs. The complexity of ABMs has risen in stride with advances in computing power and resources, resulting in larger models with complex interactions and learning and whose outputs are often high-dimensional and require sophisticated analytical approaches. Similarly, the increasing use of data and dynamics in ABMs has further enhanced the complexity of their outputs. In this article, we offer an overview of the state-of-the-art approaches in analyzing and reporting ABM outputs highlighting challenges and outstanding issues. In particular, we examine issues surrounding variance stability (in connection with determination of appropriate number of runs and hypothesis testing), sensitivity analysis, spatio-temporal analysis, visualization, and effective communication of all these to non-technical audiences, such as various stakeholders.

157 citations

Journal ArticleDOI
TL;DR: The critical choices that must be made in developing an ABM, namely the modelling of decision processes and social networks, are identified and advice on how these challenges might be overcome is offered.
Abstract: We review agent-based models (ABM) of human migration with respect to their decision-making rules. The most prominent behavioural theories used as decision rules are the random utility theory, as implemented in the discrete choice model, and the theory of planned behaviour. We identify the critical choices that must be made in developing an ABM, namely the modelling of decision processes and social networks. We also discuss two challenges that hamper the widespread use of ABM in the study of migration and, more broadly, demography and the social sciences: (a) the choice and the operationalisation of a behavioural theory (decision-making and social interaction) and (b) the selection of empirical evidence to validate the model. We offer advice on how these challenges might be overcome.

150 citations


Cites background or methods from "Estimation of ergodic agent-based m..."

  • ...The parameters most in line with the data can be estimated consistently using simulated minimum distance (Grazzini and Richiardi 2015)....

    [...]

  • ...Those are the values that are the most likely or that minimise a function of the distance between observations and simulations (see, e.g. Kennedy and O’Hagan 2001; Grazzini and Richiardi 2015)....

    [...]

Journal ArticleDOI
TL;DR: In this article, a surrogate meta-model is proposed for parameter space exploration and calibration of agent-based models (ABM) combining supervised machine learning and intelligent sampling to reduce computation time.

148 citations

References
More filters
Book
01 Jan 1962
TL;DR: A history of diffusion research can be found in this paper, where the authors present a glossary of developments in the field of Diffusion research and discuss the consequences of these developments.
Abstract: Contents Preface CHAPTER 1. ELEMENTS OF DIFFUSION CHAPTER 2. A HISTORY OF DIFFUSION RESEARCH CHAPTER 3. CONTRIBUTIONS AND CRITICISMS OF DIFFUSION RESEARCH CHAPTER 4. THE GENERATION OF INNOVATIONS CHAPTER 5. THE INNOVATION-DECISION PROCESS CHAPTER 6. ATTRIBUTES OF INNOVATIONS AND THEIR RATE OF ADOPTION CHAPTER 7. INNOVATIVENESS AND ADOPTER CATEGORIES CHAPTER 8. DIFFUSION NETWORKS CHAPTER 9. THE CHANGE AGENT CHAPTER 10. INNOVATION IN ORGANIZATIONS CHAPTER 11. CONSEQUENCES OF INNOVATIONS Glossary Bibliography Name Index Subject Index

38,750 citations

Journal ArticleDOI
TL;DR: Upon returning to the U.S., author Singhal’s Google search revealed the following: in January 2001, the impeachment trial against President Estrada was halted by senators who supported him and the government fell without a shot being fired.

23,419 citations


Additional excerpts

  • ...…AB version of the well known Bass (1969) model, which provides a mathematical explanation of the different stages of product adoption described in Rogers (1962) (innovators, early adopters, early majority, large majority, and laggards), and formalizes the crucial distinction between innovators…...

    [...]

Journal ArticleDOI
TL;DR: In this article, the authors argue that the style in which their builders construct claims for a connection between these models and reality is inappropriate, to the point at which claims for identification in these models cannot be taken seriously.
Abstract: Existing strategies for econometric analysis related to macroeconomics are subject to a number of serious objections, some recently formulated, some old. These objections are summarized in this paper, and it is argued that taken together they make it unlikely that macroeconomic models are in fact over identified, as the existing statistical theory usually assumes. The implications of this conclusion are explored, and an example of econometric work in a non-standard style, taking account of the objections to the standard style, is presented. THE STUDY OF THE BUSINESS cycle, fluctuations in aggregate measures of economic activity and prices over periods from one to ten years or so, constitutes or motivates a large part of what we call macroeconomics. Most economists would agree that there are many macroeconomic variables whose cyclical fluctuations are of interest, and would agree further that fluctuations in these series are interrelated. It would seem to follow almost tautologically that statistical models involving large numbers of macroeconomic variables ought to be the arena within which macroeconomic theories confront reality and thereby each other. Instead, though large-scale statistical macroeconomic models exist and are by some criteria successful, a deep vein of skepticism about the value of these models runs through that part of the economics profession not actively engaged in constructing or using them. It is still rare for empirical research in macroeconomics to be planned and executed within the framework of one of the large models. In this lecture I intend to discuss some aspects of this situation, attempting both to offer some explanations and to suggest some means for improvement. I will argue that the style in which their builders construct claims for a connection between these models and reality-the style in which "identification" is achieved for these models-is inappropriate, to the point at which claims for identification in these models cannot be taken seriously. This is a venerable assertion; and there are some good old reasons for believing it;2 but there are also some reasons which have been more recently put forth. After developing the conclusion that the identification claimed for existing large-scale models is incredible, I will discuss what ought to be done in consequence. The line of argument is: large-scale models do perform useful forecasting and policy-analysis functions despite their incredible identification; the restrictions imposed in the usual style of identification are neither essential to constructing a model which can perform these functions nor innocuous; an alternative style of identification is available and practical. Finally we will look at some empirical work based on an alternative style of macroeconometrics. A six-variable dynamic system is estimated without using 1 Research for this paper was supported by NSF Grant Soc-76-02482. Lars Hansen executed the computations. The paper has benefited from comments by many people, especially Thomas J. Sargent

11,195 citations


"Estimation of ergodic agent-based m..." refers background in this paper

  • ...However, Sims (1980) warns against judging macro models only in terms of fit with the data, and suggests that large scale models may fit the data well but may perform poorly out-of-sample due to non-credible identification restrictions....

    [...]

Journal ArticleDOI
TL;DR: In this article, the authors developed a model of staggered prices along the lines of Phelps (1978) and Taylor (1979, 1980), but utilizing an analytically more tractable price-setting technology.

8,580 citations


"Estimation of ergodic agent-based m..." refers methods in this paper

  • ...Similarly, heterogeneity in pricing decisions by firms is modeled according to a variant of the Calvo (1983) or Taylor (1980) mechanism (see Dixon and Bihan, 2012), which allows a solution without explicitly tracking the distribution of prices across firms. (9)Linearization however is not neutral: it eliminates asymmetries, threshold effects and many other interesting phenomena. Other solution methods, which however also involve some degree of approximation, are projection algorithms and value function iteration (Fernández-Villaverde, 2010). (10)An intermediate approach is to calibrate some parameters, and then estimate the others conditional on the values of the calibrated set. This raises serious identification issues (see subsection 5.2.2 below). (11)See Canova (2007), Tovar (2008) and Fernández-Villaverde (2010) for an introduction....

    [...]

  • ...Similarly, heterogeneity in pricing decisions by firms is modeled according to a variant of the Calvo (1983) or Taylor (1980) mechanism (see Dixon and Bihan, 2012), which allows a solution without explicitly tracking the distribution of prices across firms....

    [...]

Journal ArticleDOI
01 Apr 1993
TL;DR: An algorithm, the bootstrap filter, is proposed for implementing recursive Bayesian filters, represented as a set of random samples, which are updated and propagated by the algorithm.
Abstract: An algorithm, the bootstrap filter, is proposed for implementing recursive Bayesian filters. The required density of the state vector is represented as a set of random samples, which are updated and propagated by the algorithm. The method is not restricted by assumptions of linear- ity or Gaussian noise: it may be applied to any state transition or measurement model. A simula- tion example of the bearings only tracking problem is presented. This simulation includes schemes for improving the efficiency of the basic algorithm. For this example, the performance of the bootstrap filter is greatly superior to the standard extended Kalman filter.

8,018 citations


"Estimation of ergodic agent-based m..." refers background in this paper

  • ...Since their introduction in the early 1990s (Gordon et al., 1993), particle filters –also known as sequential Monte Carlo methods– have become a popular solution of optimal estimation problems in nonlinear nonGaussian scenarios....

    [...]

Frequently Asked Questions (18)
Q1. What are the contributions mentioned in the paper "Estimation of ergodic agent-based models by simulated minimum distance" ?

In this paper the authors show how to circumvent these difficulties and under which conditions ergodic models can be consistently estimated by simulated minimum distance techniques, both in a long-run equilibrium and during an adjustment phase. 

Of course, the analysis and recommendations contained in this paper must be regarded as a mere introduction to the problem of estimating AB models ( in their companion paper ( Grazzini and Richiardi, 2014 ) the authors extend the analysis to estimation of non-ergodic models ). Third, the possibility of applying Bayesian methods should be investigated, not only in conjunction with simulated maximum likelihood but also with the other estimation procedures. However, including priors allows for a more general estimation procedure, and leaves open the possibility of using little informative priors. In this respect, the authors can think of many open issues and avenues for research. 

In particular, since the simulated moments (or other summary measures) are used as an estimate of the theoretical moments, it is crucial to know whether the corresponding estimators are consistent. 

In particular, DSGE models not only feature a large number of parameters, but share with AB models an important aspect of complex systems: they include many nonlinear feedback effects. 

The minimum distance estimator belongs to the class of extremum estimators, which also includes maximum likelihood, nonlinear least square and generalized method of moments. 

The tests are used to assess whether the statistics of interest are constant in time and across runs: the stationarity test uses samples from a given simulation run, while the ergodicity test uses samples from different runs. 

35If the real data cannot be tested for stationarity due to small sample size, the assumption of correct specification must again be invoked. 

A vector of aggregate variables Y t is defined as a (vectorial) function over the state of the system, that is as a projection from X to Y :Y t = G(Xt,κt). 

The first one is stochastic singularity, which arises when a small number of structural shocks is used to generate predictions about a large number of observable variables: the model then predicts a deterministic linear combination of observable variables which causes the likelihood to be 0 with probability 1.12 Solutions to this problem involve reducing the number of observable variables on which inference is made (or using a projection from the set of observables to a smaller set of composed indicators), increasing the number of shocks, or adding measurement errors. 

Because the theoretical moments cannot be analytically derived, the authors proceed by simulating them: stationarity (and ergodicity) ensure that the empirical means computed on the artificial data converge to the theoretical ones. 

This is because of the curse of dimensionality, the fact that the convergence of any estimator to the true value of a smooth function defined on a space of high dimension (the parameter space) is very slow (De Marchi, 2005; Weeks, 1995). 

if the state space is nonlinear an extended Kalman filter (XKF) can be implemented on a linearized system around the steady state; this provides sub-optimal inference, but still relies on the Gaussian assumption. 

The consistency conditions for extremum estimators, including the minimum distance estimatordefined in equation (7), are given in the following theorem (Newey and McFadden, 1994, p.2121):Theorem If there is a function Q0(θ) such that (i) Q0(θ) is uniquely maximized at θ0; (ii) Θ is compact; (iii) Q0(θ) is continuous; (iv) Q̂n(θ) converges uniformly in probability to Q0(θ), then θ̂n p−→ θ0. 

For some of them direct estimation techniques can be used, as the models are simple enough to derive a closed form solution for the distribution of relevant statistics. 

As stated by Canova and Sala (2009, p. 448), when models are under- or weaklyidentified “reasonable estimates are obtained not because the model and the data are informative but because auxiliary restrictions make the likelihood of the data (or a portion of it) informative. 

This is because “ML estimation is limited by the number of linearly independent variables while moment-based estimation is limited by the number of linearly independent moments. 

the only way to analyze the mapping of (X0,θ) into Y t is by means of Monte Carlo analysis, by simulating the model for different initial states and values of the parameters, and repeating each simulation experiment many times to obtain a distribution of Y t. 

An intermediate approach is to calibrate some parameters, and then estimate the others conditional on the values of the calibrated set.