scispace - formally typeset
Search or ask a question
Journal ArticleDOI

Well-Posedness And Accuracy Of The Ensemble Kalman Filter In Discrete And Continuous Time

TL;DR: In this paper, a systematic analysis of the ensemble Kalman filter (EnKF) is presented, in particular to do so in the small ensemble size limit, where the authors view the method as a state estimator, and not as an algorithm which approximates the true filtering distribution.
Abstract: The ensemble Kalman filter (EnKF) is a method for combining a dynamical model with data in a sequential fashion Despite its widespread use, there has been little analysis of its theoretical properties Many of the algorithmic innovations associated with the filter, which are required to make a useable algorithm in practice, are derived in an ad hoc fashion The aim of this paper is to initiate the development of a systematic analysis of the EnKF, in particular to do so in the small ensemble size limit The perspective is to view the method as a state estimator, and not as an algorithm which approximates the true filtering distribution The perturbed observation version of the algorithm is studied, without and with variance inflation Without variance inflation well-posedness of the filter is established; with variance inflation accuracy of the filter, with resepct to the true signal underlying the data, is established The algorithm is considered in discrete time, and also for a continuous time limit arising when observations are frequent and subject to large noise The underlying dynamical model, and assumptions about it, is sufficiently general to include the Lorenz '63 and '96 models, together with the incompressible Navier-Stokes equation on a two-dimensional torus The analysis is limited to the case of complete observation of the signal with additive white noise Numerical results are presented for the Navier-Stokes equation on a two-dimensional torus for both complete and partial observations of the signal with additive white noise

Summary (4 min read)

1 Introduction

  • The algorithm is used in oceanography, oil reservoir simulation and weather prediction [BVLE98, EVL00, Kal03, ORL08], for example.
  • Its behaviour is not well understood.
  • Section 4 contains theoretical analyses of the perturbed observation EnKF, without and with variance inflation.

2.1 Filtering Distribution

  • The authors assume that the observed dynamics are governed by an evolution equation du dt = F (u) (2.1) which generates a one-parameter semigroup.
  • The authors also assume that K ⊂ H is another Hilbert space, which acts as the observation space.
  • The authors assume that noisy observations are made in K every h time units and write Ψ = Ψh.
  • That is, given the distribution uj |Yj as well as the observation yj+1, find the distribution of uj+.
  • The authors refer to the sequence P(uj |Yj) as the filtering distribution.

2.2 Assumptions

  • To write down the EnKF as the authors do in section 3, and indeed to derive the continuum limit of the EnKF, as they do in section 5, they need make no further assumptions about the underlying dynamics and observation operator other than those made above.
  • In infinite dimensions the existence of a global attractor in V follows from the techniques in [Tem97] for the Navier-Stokes equation by application of more subtle inequalities relating to the bilinear operator B – see section 2.2 in Chapter III of [Tem97].
  • Other equations arising in dissipative fluid mechanics can be treated similarly.
  • Whilst the preceding assumptions on the underlying dynamics apply to a range of interesting models arising in applications, the following assumptions on the observation model are rather restrictive; however the authors have been unable to extend the analysis without making them.
  • The following consequence of Assumption 2.3 will be useful to us.

3.1 The Algorithm

  • The analysis step is achieved by performing a randomised version of the Kalman update formula, and using the empirical covariance of the prediction ensemble to compute the Kalman gain.
  • There are many variants on the basic EnKF idea and the authors will study the perturbed observation form of the method.
  • The sequence of minimisers vj+1 can be written down explicitly by simply solving the quadratic minimization problem.
  • This straightforward exercise yields the following result.

3.3 Connection to Randomized Maximum Likelihood

  • The analysis step of EnKF can be understood in terms of the Randomised Maximum Likelihood (RML) method widely used in oil reservoir history matching applications [ORL08].
  • The authors will now briefly describe this method.
  • There are several reasons that this update step only produces approximate samples from the filtering distribution.
  • The decision to use the empirical distribution instead of say the push-forward of the covariance uj |Yj gives a huge advantage to the EnKF in terms of computational efficiency.

3.5 Variance Inflation

  • The minimization step of the EnKF computes an update which is a compromise between the model predictions and the data.
  • This compromise is weighted by the empirical covariance on the model and the fixed noise covariance on the data.
  • The model typically allows for unstable divergence of trajectories, whilst the data tends to stabilize.
  • Variance inflation is a technique of adding stability to the algorithm by increasing the size of the model covariance in order to weight the data more heavily.
  • Furthermore, by adding a positive definite operator, one eliminates the null-space of Ĉj+1 (which will always be present if the number of ensemble members is smaller than the dimension of H) effectively preventing the ensemble from becoming degenerate.

4 Discrete-Time Estimates

  • The authors will derive long-time estimates for the discrete-time EnKF, under the Assumptions 2.3 and 2.5 on the dynamics and observation models respectively.
  • The authors study the algorithm without and then with variance inflation.
  • The technique is to consider evolution of the error between the filter and the true signal underlying the data.
  • (4.1) Throughout this section the authors use E to denote expectation with respect to the independent i.i.d. noise sequences {ξj} and {ξ(k)j } and initial conditions u0 and v (k) 0 .

4.1 Well-Posedness Without Variance Inflation

  • Note also that Ĉj+1 has rank K and let Pj+1 denote projection into the finite dimensional subspace orthogonal to the kernel of Ĉj+1.
  • Here the authors have used the fact that Pj+1 projects onto a space of dimension at most K.
  • The preceding result shows that the EnKF is well-posed and does not blow-up faster than exponentially.
  • The authors now show that, with the addition of variance inflation, a stronger result can be proved, implying accuracy of the EnKF.

4.3 Accuracy With Variance Inflation

  • The authors will focus on the variance inflation technique with A = α2I .
  • Again assuming H = I and Γ = γ2I , the EnKF ensemble is governed by the following update equations.
  • The authors will now show that with variance inflation, one obtains much stronger long-time estimates than without it.
  • The proof is almost identical to the proof of Theorem 4.2.
  • Choosing α large enough to ensure θ < 1 will result in filter boundedness; furthermore, if the observational noise standard deviation γ is small then choosing α large enough results in filter accuracy.

5 Derivation Of The Continuous Time Limit

  • In this section the authors formally derive the continuous time scaling limits of the EnKF. (5.1b) The final step is to find an SDE for which the above represents a reasonable numerical scheme.
  • Of course, this depends crucially on the choice of scaling parameter s.
  • In fact, it is not hard to show that the one non-trivial limiting SDEs corresponds to the choice s =.
  • The stabilizing term, which draws the ensemble member back towards the truth, and the noise, both act only orthogonal to the null-space of the empirical covariance of the set of particles.

6 Continuous-Time Estimates

  • Under Assumptions 2.3 and 2.5.the authors.
  • As can be seen in [DZ92, Theorem 4.17], these conditions are sufficient to utilise Itô’s formula.
  • The authors note that, in the case of (6.1), it is not unreasonable to assume the existence of strong solutions.
  • In finite dimensions this is in fact a consequence of the mean square estimate provided by the theorem; since the authors have been unable to prove this in the rather general infinite dimensional setting, however, they make it an assumption.
  • The above argument does not work, but nevertheless it is still informative to see why it doesn’t work.

7 Numerical Results

  • In this section the authors confirm the validity of the theorems derived in the previous sections for variants of the EnKF when applied to the dynamical system (2.2).
  • Furthermore, the authors extend their numerical explorations beyond the strict range of validity of the theory and, in particular, consider the case of partial observations.
  • The authors conduct all of their numerical experiments in the case of the incompressible Navier-Stokes equation on a two dimensional torus.
  • The filter is always inaccurate when used without inflation.
  • The authors thus turn to study the effect of inflation and note that their results indicate the filter can then always be made accurate, even in the case of partial observations, provided that sufficiently many low Fourier modes are observed.

7.1 Setup

  • J∇ with J the canonical skew-symmetric matrix.
  • The true initial condition u0 is randomly drawn from N(0, ν2A−2).
  • The authors use the notation m(t) to denote the mean of the ensemble.
  • The method used to approximate the forward model is a modification of a fourth-order Runge-Kutta method, ETD4RK [CM02], in which the Stokes semi-group is computed exactly by working in the incompressible Fourier basis {ψm(x)}m∈Z2\{0}, and Duhamel’s principle (variation of constants formula) is used to incorporate the nonlinear term.

7.2.1 Full observations

  • Here the authors consider observations made at all numerically resolved, and hence observable, wavenumbers in the system; hence K = 322, not including padding in the spectral domain which avoids aliasing.
  • Observations of the full-field are made every J = 20 time-steps.
  • In Figure 2 there is no variance inflation and, whilst typical ensemble members remain bounded on the timescales shown, the error between the ensemble mean and the truth is O(1); indeed comparison with Figure 1 shows that the error in the mean is in fact worse than that of an ensemble evolving without access to data.
  • Using variance inflation removes this problem and filter accuracy is obtained: see Figure 3.

7.2.2 Partial observations

  • The observations are again made every J = 20 time-steps, but the authors will now consider observing only projections inside and outside a ring of radius |kλ| = 5 in Fourier space.
  • Figure 4 shows that when observing all Fourier modes inside a ring of radius |kλ| = 5 the filter is accurate over long time-scales.
  • In contrast, Figure 5 shows that observing all Fourier modes outside a ring of radius |kλ| = 5 does not provide enough information to induce accurate filtering.

7.3 Continuous Time

  • And its relation to the underlying truth governed by (2.2), by means of numerical experiments.the authors.
  • The authors thereby illustrate and extend the results of section 6.
  • The Navier-Stokes equation 2.2 itself is solved by the method described in section 7.1.
  • The stochastic process is also diagonalized in the Fourier basis (7.1) and then time-approximated by the EulerMaruyama scheme [KP92].

7.3.1 Full observations

  • Here the authors consider observations made at all numerically resolved, and hence observable, wavenumbers in the system; hence K = 322, not including padding in the spectral domain which avoids aliasing.
  • Figure 6 shows that, without inflation, the ensemble remains bounded, but the mean is inaccurate, on the time-scales of interest.
  • In contrast Figure 7 demonstrates that inflation leads to accurate reconstruction of the truth via the ensemble mean.

8 Conclusions

  • The authors have developed a method for the analysis of the EnKF.
  • Instead of viewing it as an algorithm designed to accurately approximate the true filtering distribution, which it cannot do, in general, outside Gaussian scenarios and in the large ensemble limit, the authors study it as an algorithm for signal estimation in the finite (possibly small) ensemble limit.
  • These positive results about the EnKF are encouraging and serve to underpin its perceived effectiveness in applications.
  • On the other hand it is important to highlight that their analysis applies only to fully observed dynamics and interesting open questions remain concerning the partially observed case.
  • These numerical results demonstrate two interesting potential extensions of their theory: (i) to strengthen well-posedness to obtain boundedness of trajectories, at least in mean square; (ii) to extend well-posedness and accuracy results to certain partial observation scenarios.

Did you find this useful? Give us your feedback

Figures (9)

Content maybe subject to copyright    Report

Well-Posedness And Accuracy Of The Ensemble Kalman Filter In
Discrete And Continuous Time
October 14, 2013
D.T.B. Kelly, K.J.H. Law, A.M. Stuart
The University of Warwick, Email: David.Kelly@Warwick.ac.uk
Abstract
The ensemble Kalman filter (EnKF) is a method for combining a dynamical model with data in a sequential
fashion. Despite its widespread use, there has been little analysis of its theoretical properties. Many of the
algorithmic innovations associated with the filter, which are required to make a useable algorithm in practice, are
derived in an ad hoc fashion. The aim of this paper is to initiate the development of a systematic analysis of the
EnKF, in particular to do so in the small ensemble size limit. The perspective is to view the method as a state
estimator, and not as an algorithm which approximates the true filtering distribution. The perturbed observation
version of the algorithm is studied, without and with variance inflation. Without variance inflation well-posedness
of the filter is established; with variance inflation accuracy of the filter, with resepct to the true signal underlying the
data, is established. The algorithm is considered in discrete time, and also for a continuous time limit arising when
observations are frequent and subject to large noise. The underlying dynamical model, and assumptions about it,
is sufficiently general to include the Lorenz ’63 and ’96 models, together with the incompressible Navier-Stokes
equation on a two-dimensional torus. The analysis is limited to the case of complete observation of the signal with
additive white noise. Numerical results are presented for the Navier-Stokes equation on a two-dimensional torus
for both complete and partial observations of the signal with additive white noise.
1 Introduction
In recent years the ensemble Kalman filter (EnKF) [Eve06] has become a widely used methodology for combin-
ing dynamical models with data. The algorithm is used in oceanography, oil reservoir simulation and weather
prediction [BVLE98, EVL00, Kal03, ORL08], for example. The basic idea of the method is to propagate an
ensemble of particles to describe the distribution of the signal given data, employing empirical second order
statistics to update the distribution in a Kalman-like fashion when new data is acquired. Despite the widespread
use of the method, its behaviour is not well understood. In contrast with the ordinary Kalman filter, which ap-
plies to linear Gaussian problems, it is difficult to find a mathematical justification for EnKF. The most notable
progress in this direction can be found in [LGMT
+
10, MCB11], where it is proved that, for linear dynamics,
the EnKF approximates the usual Kalman filter in the large ensemble limit. This analysis is however far from
being useful for practitioners who typically run the method with small ensemble size on nonlinear problems.
Furthermore there is an accumulation of numerical evidence showing that the EnKF, and related schemes such
as the extended Kalman filter, can “diverge” with the meaning of “diverge” ranging from simply loosing the
true signal through to blow-up [IKJ02, MH08, GM13]. The aim of our work is to try and build mathematical
foundations for the analysis of the EnKF, in particular with regards to well-posedness (lack of blow-up) and
accuracy (tracking the signal over arbitrarily long time-intervals). To make progress on such questions it is
necessary to impose structure on the underlying dynamics and we choose to work with dissipative quadratic
systems with energy-conserving nonlinearity, a class of problems which has wide applicability [MW06] and
which has proved to be useful in the development of filters [MH12].
In section 3 we derive the perturbed observation form of the EnKF and demonstrate how it links to the
randomized maximum likelihood method (RML) which is widely used in oil reservoir simulation [ORL08].
1
arXiv:1310.3167v1 [math.PR] 11 Oct 2013

We also introduce the idea of variance inflation, widely used in many practical implementations of the EnKF
[And07]. Section 4 contains theoretical analyses of the perturbed observation EnKF, without and with variance
inflation. Without variance inflation we are able only to prove bounds which grow exponentially in the discrete
time increment underlying the algorithm (Theorem 4.2); with variance inflation we are able to prove filter
accuracy and show that, in mean square with respect to the noise entering the algorithm, the filter is uniformly
close to the true signal for all large times, provided enough inflation is employed (Theorem 4.4). These results,
and in particular the one concerning variance inflation, are similar to the results developed in [BLL
+
12] for the
3DVAR filter applied to the Navier-Stokes equation and for the 3DVAR filter applied to the Lorenz ’63 model
in [LSS14], as well as the similar analysis developed in [MLPvL13] for the 3DVAR filter applied to globally
Lipschitz nonlinear dynamical systems. In section 5 we describe a continuous time limit in which data arrives
very frequently, but is subject to large noise. If these two effects are balanced appropriately a stochastic (partial)
differential equation limit is found and it is instructive to study this limiting continuous time process. This
idea was introduced in [BLSZ] for the 3DVAR filter and is here employed for the EnKF filter. The primary
motivation for the continuous time limit is to obtain insight into the mechanisms underlying the EnKF; some of
these mechanisms are more transparent in continuous time. In section 6 we analyze the well-posedness of the
continuous time EnKF (Theorem 6.2). Section 7 contains numerical experiments which illustrate and extend
the theory, and section 8 contains some brief concluding remarks.
Throughout the sequel we use the following notation. Let H be a separable Hilbert space with norm |·| and
inner product , ·i. For a linear operator C on H, we will write C 0 (resp. C > 0) when C is self-adjoint
and positive semi-definite (resp. positive definite). Given C > 0, we will denote
|·|
C
def
=
C
1/2
(·)
.
2 Set-Up
2.1 Filtering Distribution
We assume that the observed dynamics are governed by an evolution equation
du
dt
= F (u) (2.1)
which generates a one-parameter semigroup Ψ
t
: H H. We also assume that K H is another Hilbert
space, which acts as the observation space. We assume that noisy observations are made in K every h time
units and write Ψ = Ψ
h
. We define u
j
= u(jh) for j N and, assuming that u
0
is uncertain and modelled as
Gaussian distributed, we obtain
u
j+1
= Ψ(u
j
) , with u
0
N (m
0
, C
0
)
for some initial mean m
0
and covariance C
0
. We are given the observations
y
j+1
= Hu
j+1
+ Γ
1/2
ξ
j+1
, with ξ
j
N (0, I) i.i.d. ,
where H L(H, K) is the observation operator and Γ L(H, H) with Γ 0 is the covariance operator of
the observational noise; the i.i.d. noise sequence {ξ
j
} is assumed independent of u
0
. The aim of filtering is to
approximate the distribution of u
j
given Y
j
= {y
`
}
j
`=1
using a sequential update algorithm. That is, given the
distribution u
j
|Y
j
as well as the observation y
j+1
, find the distribution of u
j+1
|Y
j+1
. We refer to the sequence
P(u
j
|Y
j
) as the filtering distribution.
2.2 Assumptions
To write down the EnKF as we do in section 3, and indeed to derive the continuum limit of the EnKF, as we do in
section 5, we need make no further assumptions about the underlying dynamics and observation operator other
than those made above. However, in order to analyze the properties of the EnKF, as we do in sections 4 and 6,
we will need to make structural assumptions and we detail these here. It is worth noting that the assumptions we
2

make on the underlying dynamics are met by several natural models used to test data assimilation algorithms.
In particular, the 2D Navier-Stokes equations on a torus, as well as both Lorenz ’63 and ’96 models, satisfy
Assumptions 2.3 [MW06, MH12, Tem97].
Assumption 2.3. (Dynamics Model) Suppose there is some Banach space V, equipped with norm k·k, that can
be continuously embedded into H. We assume that (2.1) has the form
du
dt
+ Au + B(u, u) = f , (2.2)
where A : H H is an unbounded linear operator satisfying
hAu, ui λ kuk
2
, (2.3)
for some λ > 0, B is a symmetric bilinear operator B : V × V H and f : R
+
H. We furthermore assume
that B satisfies the identity
hB(u, u), ui = 0 , (2.4)
for all u H and also
hB(u, v), vi c kuk kvk |v| , (2.5)
for all u, v, w H, where c > 0 depends only on the bilinear form. We assume that the equation (2.2) has
a unique weak solution for all u(0) H, and generates a one-parameter semigroup Ψ
t
: V V which may
be extended to act on H. Finally we assume that there exists a global attractor Λ V for the dynamics, and
constant R > 0 such that for any initial condition u
0
Λ, we have that sup
t0
ku(t)k R.
Remark 2.4. In the finite dimensional case the final assumption on the existence of a global attractor does not
need to be made as it is a consequence of the preceding assumptions made. To see this note that
1
2
d|u|
2
dt
+ λkuk
2
hf, ui. (2.6)
The continuous embedding of V, together with the Cauchy-Schwarz inequality, implies the existence of a strictly
positive constant such that
1
2
d|u|
2
dt
+ |u|
2
1
2δ
|f|
2
+
δ
2
|u|
2
(2.7)
for all δ > 0. Choosing δ = gives the existence of an absorbing set and hence a global attractor by Theorem
1.1 in Chapter I of [Tem97]. In infinite dimensions the existence of a global attractor in V follows from the
techniques in [Tem97] for the Navier-Stokes equation by application of more subtle inequalities relating to the
bilinear operator B see section 2.2 in Chapter III of [Tem97]. Other equations arising in dissipative fluid
mechanics can be treated similarly.
Whilst the preceding assumptions on the underlying dynamics apply to a range of interesting models arising
in applications, the following assumptions on the observation model are rather restrictive; however we have been
unable to extend the analysis without making them. We will demonstrate, by means of numerical experiments,
that our results extend beyond the observation scenario employed in the theory
Assumption 2.5. (Observation Model) The system is completely observed so that K = H and H = I.
Furthermore the i.i.d. noise sequence {ξ
j
} is white so that ξ
1
N (0, Γ) with Γ = γ
2
I.
The following consequence of Assumption 2.3 will be useful to us.
Lemma 2.6. Let Assumptions 2.3 hold. Then there is β R such that, for any v
0
Λ, h > 0 and w
0
H,
|Ψ
h
(v
0
) Ψ
h
(w
0
)| e
βh
|v
0
w
0
| .
3

Proof. Let v, w denote the solutions of (2.2) with initial conditions v
0
, w
0
respectively; define e = v w. Then
de
dt
+ Ae + 2B(v, e) B(e, e) = 0 , (2.8)
with e(0) = v
0
w
0
. Taking the inner-product with e, using (2.3), (2.4) and (2.5), and choosing δ = λ/(2cR),
gives
1
2
d
dt
|e|
2
+ λkek
2
2ckvkkek|e|
2cRkek|e|
cR(δkek
2
+ δ
1
|e|
2
)
=
λ
2
kek
2
+
2
λ
(cR)
2
|e|
2
.
Thus
d
dt
|e|
2
4
λ
(cR)
2
|e|
2
and the desired result follows from an application of the Gronwall inequality.
3 The Ensemble Kalman Filter
3.1 The Algorithm
The idea of the EnKF is to represent the filtering distribution through an ensemble of particles, to propagate
this ensemble under the model to approximate the mapping P(u
j
|Y
j
) to P(u
j+1
|Y
j
) (refered to as prediction
in the applied literature), and to update the ensemble distribution to include the data point Y
j+1
by using a
Gaussian approximation based on the second order statistics of the ensemble (refered to as analysis in the
applied literature).
The prediction step is achieved by simply flowing forward the ensemble under the model dynamics, that is
bv
(k)
j+1
= Ψ(v
(k)
j
) , for k = 1 . . . K.
The analysis step is achieved by performing a randomised version of the Kalman update formula, and using the
empirical covariance of the prediction ensemble to compute the Kalman gain. There are many variants on the
basic EnKF idea and we will study the perturbed observation form of the method.
The algorithm proceeds as follows.
1. Set j = 0 and draw an independent set of samples {v
(k)
0
}
K
k=1
from N(m
0
, C
0
).
2. (Prediction) Let bv
(k)
j+1
= Ψ(v
(k)
j
) and define
b
C
j+1
as the empirical covariance of {bv
(k)
j+1
}
K
k=1
. That is,
b
C
j+1
=
1
K
K
X
k=1
(bv
(k)
j+1
¯v
j+1
) (bv
(k)
j+1
¯v
j+1
) ,
where ¯v
j+1
=
1
K
P
K
k=1
bv
j+1
denotes the ensemble mean.
3. (Observation) Make an observation y
j+1
= Hu
j+1
+ Γ
1/2
ξ
j+1
. Then, for each k = 1 . . . K, generate an
artificial observation
y
(k)
j+1
= y
j+1
+ Γ
1/2
ξ
(k)
j+1
,
where ξ
(k)
j+1
are N(0, I) distributed and pairwise independent.
4

4. (Analysis) Let v
(k)
j+1
be the minimiser of the functional
J(v) =
1
2
|y
(k)
j+1
v|
2
Γ
+
1
2
|bv
(k)
j+1
v|
b
C
j+1
.
5. Set j 7→ j + 1 and return to step 2.
The name “perturbed observation EnKF” follows from the construction of the artificial observations y
(k)
j+1
which
are found by perturbing the given observation with additional noise. The sequence of minimisers v
j+1
can be
written down explicitly by simply solving the quadratic minimization problem. This straightforward exercise
yields the following result.
Proposition 3.2. The sequence {v
(k)
j
}
j0
is defined by the equation
(I +
b
C
j+1
H
T
Γ
1
H)v
(k)
j+1
= bv
(k)
j+1
+
b
C
j+1
H
T
Γ
1
y
(k)
j+1
,
for each k = 1, . . . , K.
Hence, collecting the ingredients from the preceding, the defining equations of the EnKF are given by
(I +
b
C
j+1
H
T
Γ
1
H)v
(k)
j+1
= Ψ(v
(k)
j+1
) +
b
C
j+1
H
T
Γ
1
y
(k)
j+1
(3.1a)
y
(k)
j+1
= y
j+1
+ Γ
1/2
ξ
(k)
j+1
(3.1b)
¯v
j+1
=
1
K
K
X
k=1
Ψ(v
(k)
j+1
) (3.1c)
b
C
j+1
=
1
K
K
X
k=1
Ψ(v
(k)
j+1
) ¯v
j+1
Ψ(v
(k)
j+1
) ¯v
j+1
. (3.1d)
There are other representations of the EnKF that are more algorithmically convenient, but the formulae (3.1)
are better suited to our analysis.
3.3 Connection to Randomized Maximum Likelihood
The analysis step of EnKF can be understood in terms of the Randomised Maximum Likelihood (RML) method
widely used in oil reservoir history matching applications [ORL08]. We will now briefly describe this method.
Suppose that we have a random variable u and that u N( bm,
b
C). Moreover, let G be some linear operator and
suppose we observe
y = Gu + ξ where ξ N(0, Γ) .
One can use Bayes’ theorem to write down the conditional density P(u|y). In practice however, it is often
sufficient (or sometimes even better) to simply have a collection of samples {u
(k)
}
K
k=1
from the conditional
distribution, rather than the density itself. RML is a method of taking samples from the prior N( bm,
b
C) and
turning them into samples from the posterior. This is achieved as follows, given bu
(k)
N( bm,
b
C) (samples
from the prior), define u
(k)
for each k = 1 . . . K by u
(k)
= argmin
u
J
(k)
(u) where
J
(k)
(u) =
1
2
|y Gu + Γ
1/2
ξ
(k)
|
2
Γ
+
1
2
|u bu
(k)
|
2
b
C
,
where ξ
(k)
N(0, I) and independent of ξ. The u
(k)
are then draws from the posterior distribution of u|y
which is a Gaussian with mean m and covariance C. Since one can explicitly write down (m, C), it may be
checked that the u
(k)
defined as above are independent random variables of the form u
(k)
= m + C
1/2
ζ
(k)
,
where ζ
(k)
N (0, I) i.i.d. and are hence draws from the desired posterior, as we know show.
5

Citations
More filters
Book
10 Sep 2015
TL;DR: In this paper, the authors provide a systematic treatment of the mathematical underpinnings of work in data assimilation, covering both theoretical and computational approaches, and develop a unified mathematical framework in which a Bayesian formulation of the problem provides the bedrock for the derivation, development and analysis of algorithms; the many examples used in the text, together with the algorithms which are introduced and discussed, are all illustrated by MATLAB software detailed in the book and made freely available online.
Abstract: This book provides a systematic treatment of the mathematical underpinnings of work in data assimilation, covering both theoretical and computational approaches. Specifically the authors develop a unified mathematical framework in which a Bayesian formulation of the problem provides the bedrock for the derivation, development and analysis of algorithms; the many examples used in the text, together with the algorithms which are introduced and discussed, are all illustrated by the MATLAB software detailed in the book and made freely available online. The book is organized into nine chapters: the first contains a brief introduction to the mathematical tools around which the material is organized; the next four are concerned with discrete time dynamical systems and discrete time data; the last four are concerned with continuous time dynamical systems and continuous time data and are organized analogously to the corresponding discrete time chapters. This book is aimed at mathematical researchers interested in a systematic development of this interdisciplinary field, and at researchers from the geosciences, and a variety of other scientific fields, who use tools from data assimilation to combine data with time-dependent models. The numerous examples and illustrations make understanding of the theoretical underpinnings of data assimilation accessible. Furthermore, the examples, exercises and MATLAB software, make the book suitable for students in applied mathematics, either through a lecture course, or through self-study.

246 citations

Journal ArticleDOI
TL;DR: The goal of this paper is to analyze the EnKF when applied to inverse problems with fixed ensemble size, and to demonstrate that the conclusions of the analysis extend beyond the linear inverse problem setting.
Abstract: The ensemble Kalman filter (EnKF) is a widely used methodology for state estimation in partial, noisily observed dynamical systems, and for parameter estimation in inverse problems. Despite its widespread use in the geophysical sciences, and its gradual adoption in many other areas of application, analysis of the method is in its infancy. Furthermore, much of the existing analysis deals with the large ensemble limit, far from the regime in which the method is typically used. The goal of this paper is to analyze the method when applied to inverse problems with fixed ensemble size. A continuous-time limit is derived and the long-time behavior of the resulting dynamical system is studied. Most of the rigorous analysis is confined to the linear forward problem, where we demonstrate that the continuous time limit of the EnKF corresponds to a set of gradient flows for the data misfit in each ensemble member, coupled through a common pre-conditioner which is the empirical covariance matrix of the ensemble. Numerical results demonstrate that the conclusions of the analysis extend beyond the linear inverse problem setting. Numerical experiments are also given which demonstrate the benefits of various extensions of the basic methodology.

206 citations

Journal ArticleDOI
TL;DR: This review article will address the two principal components of the cardiovascular system: arterial circulation and heart function, and systematically describe all aspects of the problem, ranging from data imaging acquisition to the development of reduced-order models that are of paramount importance when solving problems with high complexity, which would otherwise be out of reach.
Abstract: Mathematical and numerical modelling of the cardiovascular system is a research topic that has attracted remarkable interest from the mathematical community because of its intrinsic mathematical difficulty and the increasing impact of cardiovascular diseases worldwide. In this review article we will address the two principal components of the cardiovascular system: arterial circulation and heart function. We will systematically describe all aspects of the problem, ranging from data imaging acquisition, stating the basic physical principles, analysing the associated mathematical models that comprise PDE and ODE systems, proposing sound and efficient numerical methods for their approximation, and simulating both benchmark problems and clinically inspired problems. Mathematical modelling itself imposes tremendous challenges, due to the amazing complexity of the cardiocirculatory system, the multiscale nature of the physiological processes involved, and the need to devise computational methods that are stable, reliable and efficient. Critical issues involve filtering the data, identifying the parameters of mathematical models, devising optimal treatments and accounting for uncertainties. For this reason, we will devote the last part of the paper to control and inverse problems, including parameter estimation, uncertainty quantification and the development of reduced-order models that are of paramount importance when solving problems with high complexity, which would otherwise be out of reach.

176 citations

Journal ArticleDOI
TL;DR: This work combines modern data assimilation methods with Wikipedia access logs and CDC influenza-like illness (ILI) reports to create a weekly forecast for seasonal influenza, and adjusts the initialization and parametrization of a disease model to determine systematic model bias.
Abstract: Infectious diseases are one of the leading causes of morbidity and mortality around the world; thus, forecasting their impact is crucial for planning an effective response strategy. According to the Centers for Disease Control and Prevention (CDC), seasonal influenza affects 5% to 20% of the U.S. population and causes major economic impacts resulting from hospitalization and absenteeism. Understanding influenza dynamics and forecasting its impact is fundamental for developing prevention and mitigation strategies. We combine modern data assimilation methods with Wikipedia access logs and CDC influenza-like illness (ILI) reports to create a weekly forecast for seasonal influenza. The methods are applied to the 2013-2014 influenza season but are sufficiently general to forecast any disease outbreak, given incidence or case count data. We adjust the initialization and parametrization of a disease model and show that this allows us to determine systematic model bias. In addition, we provide a way to determine where the model diverges from observation and evaluate forecast accuracy. Wikipedia article access logs are shown to be highly correlated with historical ILI records and allow for accurate prediction of ILI data several weeks before it becomes available. The results show that prior to the peak of the flu season, our forecasting method produced 50% and 95% credible intervals for the 2013-2014 ILI observations that contained the actual observations for most weeks in the forecast. However, since our model does not account for re-infection or multiple strains of influenza, the tail of the epidemic is not predicted well after the peak of flu season has passed.

167 citations

Journal ArticleDOI
TL;DR: In this paper, the authors established a simple, systematic and rigorous framework for the nonlinear analysis of EnKF and ESRF with arbitrary ensemble size, focusing on the dynamical properties of boundedness and geometric ergodicity.
Abstract: The ensemble Kalman filter (EnKF) and ensemble square root filter (ESRF) are data assimilation methods used to combine high dimensional, nonlinear dynamical models with observed data. Despite their widespread usage in climate science and oil reservoir simulation, very little is known about the long-time behavior of these methods and why they are effective when applied with modest ensemble sizes in large dimensional turbulent dynamical systems. By following the basic principles of energy dissipation and controllability of filters, this paper establishes a simple, systematic and rigorous framework for the nonlinear analysis of EnKF and ESRF with arbitrary ensemble size, focusing on the dynamical properties of boundedness and geometric ergodicity. The time uniform boundedness guarantees that the filter estimate will not diverge to machine infinity in finite time, which is a potential threat for EnKF and ESQF known as the catastrophic filter divergence. Geometric ergodicity ensures in addition that the filter has a unique invariant measure and that initialization errors will dissipate exponentially in time. We establish these results by introducing a natural notion of observable energy dissipation. The time uniform bound is achieved through a simple Lyapunov function argument, this result applies to systems with complete observations and strong kinetic energy dissipation, but also to concrete examples with incomplete observations. With the Lyapunov function argument established, the geometric ergodicity is obtained by verifying the controllability of the filter processes; in particular, such analysis for ESQF relies on a careful multivariate perturbation analysis of the covariance eigen-structure.

70 citations

References
More filters
Book
01 Jun 1992
TL;DR: In this article, a time-discrete approximation of deterministic Differential Equations is proposed for the stochastic calculus, based on Strong Taylor Expansions and Strong Taylor Approximations.
Abstract: 1 Probability and Statistics- 2 Probability and Stochastic Processes- 3 Ito Stochastic Calculus- 4 Stochastic Differential Equations- 5 Stochastic Taylor Expansions- 6 Modelling with Stochastic Differential Equations- 7 Applications of Stochastic Differential Equations- 8 Time Discrete Approximation of Deterministic Differential Equations- 9 Introduction to Stochastic Time Discrete Approximation- 10 Strong Taylor Approximations- 11 Explicit Strong Approximations- 12 Implicit Strong Approximations- 13 Selected Applications of Strong Approximations- 14 Weak Taylor Approximations- 15 Explicit and Implicit Weak Approximations- 16 Variance Reduction Methods- 17 Selected Applications of Weak Approximations- Solutions of Exercises- Bibliographical Notes

6,284 citations

Book
10 Sep 1993
TL;DR: In this article, the authors give bounds on the number of degrees of freedom and the dimension of attractors of some physical systems, including inertial manifolds and slow manifolds.
Abstract: Contents: General results and concepts on invariant sets and attractors.- Elements of functional analysis.- Attractors of the dissipative evolution equation of the first order in time: reaction-diffusion equations.- Fluid mechanics and pattern formation equations.- Attractors of dissipative wave equations.- Lyapunov exponents and dimensions of attractors.- Explicit bounds on the number of degrees of freedom and the dimension of attractors of some physical systems.- Non-well-posed problems, unstable manifolds. lyapunov functions, and lower bounds on dimensions.- The cone and squeezing properties.- Inertial manifolds.- New chapters: Inertial manifolds and slow manifolds the nonselfadjoint case.

5,038 citations

Book ChapterDOI
01 Jan 1985
TL;DR: In this paper, the authors return to the possible solutions X t (ω) of the stochastic differential equation where W t is 1-dimensional "white noise" and where X t satisfies the integral equation in differential form.
Abstract: We now return to the possible solutions X t (ω) of the stochastic differential equation (5.1) where W t is 1-dimensional “white noise”. As discussed in Chapter III the Ito interpretation of (5.1) is that X t satisfies the stochastic integral equation or in differential form (5.2) .

4,144 citations

Book
01 Jan 2006
TL;DR: In this paper, the authors define a statistical analysis scheme for estimating an oil reservoir simulator and an ocean prediction system based on the En-KF model, and propose a sampling strategy for the EnKF and square root analysis schemes.
Abstract: Statistical definitions.- Analysis scheme.- Sequential data assimilation.- Variational inverse problems.- Nonlinear variational inverse problems.- Probabilistic formulation.- Generalized Inverse.- Ensemble methods.- Statistical optimization.- Sampling strategies for the EnKF.- Model errors.- Square Root Analysis schemes.- Rank issues.- Spurious correlations, localization, and inflation.- An ocean prediction system.- Estimation in an oil reservoir simulator.

2,206 citations

Journal ArticleDOI
TL;DR: In this article, it is shown that the observations must be treated as random variables at the analysis steps, which results in a completely consistent approach if the covariance of the ensemble of model states is interpreted as the prediction error covariance, and there are no further requirements on the ensemble Kalman filter method.
Abstract: This paper discusses an important issue related to the implementation and interpretation of the analysis scheme in the ensemble Kalman filter. It is shown that the observations must be treated as random variables at the analysis steps. That is, one should add random perturbations with the correct statistics to the observations and generate an ensemble of observations that then is used in updating the ensemble of model states. Traditionally, this has not been done in previous applications of the ensemble Kalman filter and, as will be shown, this has resulted in an updated ensemble with a variance that is too low. This simple modification of the analysis scheme results in a completely consistent approach if the covariance of the ensemble of model states is interpreted as the prediction error covariance, and there are no further requirements on the ensemble Kalman filter method, except for the use of an ensemble of sufficient size. Thus, there is a unique correspondence between the error statistics from the ensemble Kalman filter and the standard Kalman filter approach.

1,801 citations

Frequently Asked Questions (13)
Q1. What have the authors contributed in "Well-posedness and accuracy of the ensemble kalman filter in discrete and continuous time" ?

The aim of this paper is to initiate the development of a systematic analysis of the EnKF, in particular to do so in the small ensemble size limit. The perturbed observation version of the algorithm is studied, without and with variance inflation. 

These numerical results demonstrate two interesting potential extensions of their theory: ( i ) to strengthen well-posedness to obtain boundedness of trajectories, at least in mean square ; ( ii ) to extend well-posedness and accuracy results to certain partial observation scenarios. Furthermore the authors highlight the fact that their results have assumed exact solution of the underlying differential equation model ; understanding how filtering interacts with numerical approximations, and potentially induces numerical instabilities, is a subject which requires further investigation ; this issue is highlighted in [ GM13 ]. 

The method used to approximate the forward model is a modification of a fourth-order Runge-Kutta method, ETD4RK [CM02], in which the Stokes semi-group is computed exactly by working in the incompressible Fourier basis {ψm(x)}m∈Z2\\{0}, and Duhamel’s principle (variation of constants formula) is used to incorporate the nonlinear term. 

The idea of the EnKF is to represent the filtering distribution through an ensemble of particles, to propagate this ensemble under the model to approximate the mapping P(uj |Yj) to P(uj+1|Yj) (refered to as prediction in the applied literature), and to update the ensemble distribution to include the data point Yj+1 by using a Gaussian approximation based on the second order statistics of the ensemble (refered to as analysis in the applied literature). 

The perturbed observations noise contribution will act to prevent the particles from synchronizing which, in their absence, could happen. 

1|Yj is certainly not Gaussian in general, unless the dynamics are linear, hence the RML method becomes an approximation of samples. 

However the fact that the noise is effectively finite dimensional, due to the presence of the finite rank covariance operator C, does mean that existence of strong solutions may well be established on a case-by-case basis in some infinite dimensional settings. 

The prediction step is achieved by simply flowing forward the ensemble under the model dynamics, that isv̂ (k) j+1 = Ψ(v (k) j ) , for k = 1 . . .K.The analysis step is achieved by performing a randomised version of the Kalman update formula, and using the empirical covariance of the prediction ensemble to compute the Kalman gain. 

The aim of their work is to try and build mathematical foundations for the analysis of the EnKF, in particular with regards to well-posedness (lack of blow-up) and accuracy (tracking the signal over arbitrarily long time-intervals). 

The authors thus turn to study the effect of inflation and note that their results indicate the filter can then always be made accurate, even in the case of partial observations, provided that sufficiently many low Fourier modes are observed. 

In infinite dimensions the existence of a global attractor in V follows from the techniques in [Tem97] for the Navier-Stokes equation by application of more subtle inequalities relating to the bilinear operator B – see section 2.2 in Chapter III of [Tem97]. 

v(k) j+1 = Ψ(v (k) j ) + (α2 γ2 The author+ 1 γ2 Ĉj+1)y (k) j+1 . (4.4)The authors will now show that with variance inflation, one obtains much stronger long-time estimates than without it. 

from the second equation the authors must have 1 − s/2 ≥ 1/2, for otherwise the stochastic terms would diverge when summed up, in accordance with the central limit theorem.