Aug 3, 2018 - Errors in Ensemble Data Assimilation with Non-local Model .... the regression coefficients (entries of the Kalman gain matrices). Vall`es and ...
Methods to Mitigate Loss of Variance Due to Sampling Errors in Ensemble Data Assimilation with Non-local Model Parameters Johann M. Lacerdaa , Alexandre A. Emerickb , Adolfo P. Piresc a
Petrobras and UENF b Petrobras c UENF
Abstract Ensemble data assimilation methods are among the most successful techniques for assisted history matching. However, these methods suffer from sampling errors caused by the limited number of ensemble members employed in practical applications. A typical consequence of sampling errors is an excessive loss of ensemble variance. In practice, the posterior ensemble tends to underestimate the uncertainty range in the parameter values and production predictions. Distance-based localization is the standard method for mitigating sampling errors in ensemble data assimilation. A properly designed localization matrix works remarkably well to improve the estimate of gridblock properties such as porosity and permeability. However, field history-matching problems also include non-local model parameters, i.e., parameters with no spatial location. Examples of non-local parameters include relative permeability curves, fluid contacts, global property multipliers, among others. In these cases, we cannot use distance-based localization. This paper presents an investigation on several methods proposed in the literature that can be applied to non-local model parameters to mitigate erroneous loss of ensemble variance. We use a small synthetic history-matching problem to evaluate the performance of the methods. For this problem, we were able to compute a reference solution with a very large ensemble. We compare the methods mainly in terms of the preservation of the ensemble variance without compromising the ability of matching the observed data. We also analyse the robustness and difficulty to select proper configurations for each method. Keywords: Ensemble data assimilation, history matching, ensemble smoother, Preprint submitted to Journal of Petroleum Science and Engineering
August 3, 2018
sampling errors, localization, covariance inflation
1
1. Introduction
2
Ensemble-based methods proved to be useful techniques to solve history-matching
3
problems. In spite of their success in several field applications reported in the lit-
4
erature (Skjervheim et al., 2007; Bianco et al., 2007; Evensen et al., 2007; Haugen
5
et al., 2008; Cominelli et al., 2009; Zhang and Oliver, 2011; Emerick and Reynolds,
6
2011a, 2013b; Chen and Oliver, 2014; Emerick, 2016; Luo et al., 2018b), some chal-
7
lenges remain. Perhaps the main concerns are related to sampling errors. Because
8
of the computational cost of reservoir simulations, the data assimilation in reser-
9
voir engineering is done with small-sized ensembles, typically on the order of 100
10
members. Unfortunately, the use of small ensembles causes inaccurate estimation
11
of covariance matrices introducing spurious correlations between predicted data and
12
model parameters. The consequence of spurious correlation is typically manifested
13
as significant underestimation of posterior variances (Aanonsen et al., 2009). Even
14
an absolute collapse of ensemble can arise in some extreme cases. It is interesting to
15
note that sampling errors can be exacerbated by nonlinearity in the model. Raanes
16
et al. (2018) illustrate this fact using a simple univariate problem with a linear and a
17
nonlinear model, both cases with the same posterior (Gaussian) distribution. They
18
showed that for a linear model the sampling errors are quickly attenuated, and the
19
ensemble statistics converge to the exact ones. However, for the nonlinear model
20
the sampling errors are chronic.
21
Localization (Houtekamer and Mitchell, 2001) is a widely used method to mit-
22
igate problems due to spurious correlations and has the additional advantage of
23
increasing the degrees of freedom to assimilate data. Most of localization methods
24
rely on distances, which means that they can be applied when both, model variables
25
and observations, have associated spatial locations. Although it is not necessar-
26
ily obvious to specify an appropriate size for the localization region, the number
27
of publications with successful applications is quite extensive; see, e.g., (Chen and
28
Oliver, 2010b, 2014; Emerick and Reynolds, 2011a,b; Emerick, 2016). There are two
29
main localization procedures in the literature: Schur-product localization and do2
30
main localization. Sakov and Bertino (2011) compared Schur product and domain
31
localization and concluded that both methods produce similar results for typical
32
geophysical data assimilation problems. However, Schur-product localization may
33
be more accurate for “strong” (accurate measurements) data assimilation problems.
34
Nevertheless, the performance of both methods is highly dependent of the choice of
35
the localization region.
36
However, important parameters in history matching such as permeability and
37
porosity multipliers, relative permeability curves, analytical aquifer parameters, rock
38
and fluid compressibilities, water-oil and gas-oil depth contacts may not have asso-
39
ciated spatial relation with typical observations such as oil, water and gas rates,
40
pressure measurements and 4D seismic. For these cases, a non-distance dependent
41
localization scheme is necessary. Here, we refer to these parameters as “non-local”
42
parameters1 .
43
There exists localization schemes designed for general covariance structures with-
44
out making the assumption that covariance depends on the distance between model
45
parameters and data. However, these methods have been rarely applied to history-
46
matching problems. Perhaps, the first work on a non-distance localization scheme
47
is the hierarchical filter from Anderson (2007a). In this method, an ensemble of Ne
48
members is divided into Nh sub-ensembles of size Ne /Nh , which are used to correct
49
the regression coefficients (entries of the Kalman gain matrices). Vall`es and Næv-
50
dal (2009) tested the hierarchical filter in the Brugge case (Peters et al., 2010) and
51
concluded that it improved the diversity in the updated models compared to the
52
standard ensemble Kalman filter (EnKF). Zhang and Oliver (2010) proposed that
53
instead of Nh ensembles of Ne members, we can resample a single ensemble with
54
replacement to generate Nb bootstrapped ensembles. They tested the method in
55
a synthetic history matching problem and concluded that it improved the results 1
The terminology localization was introduced after (Houtekamer and Mitchell, 2001) based on
the idea of limiting the spatial location of the updates, removing long-distance correlations. Hence, for non-local model parameters the terminology localization seems less appropriate. Nevertheless, here we still use the term localization for historical reasons and because it is well diffused in the literature.
3
56
obtained by EnKF with a small ensemble (Ne = 30) for the estimation of grid-
57
block permeabilities. Furrer and Bengtsson (2007) present a simple expression that
58
can be used as a non-distance localization scheme. They minimized term-by-term
59
the norm of the difference between the true forecast covariance matrix and the lo-
60
calized estimate ignoring the positive-definiteness constraint. The final expression
61
depends on the true covariance. The authors suggest to replace the true covariance
62
by the ensemble estimate to construct a non-distance localization function, which is
63
computationally inexpensive and straightforward to implement. Bishop and Hodyss
64
(2007) proposed to obtain the localization matrix by raising each element of a sam-
65
ple correlation matrix of model variables to a certain power. To the best of our
66
knowledge, there are no history-matching applications of non-distance localizations
67
schemes from Furrer and Bengtsson (2007) and Bishop and Hodyss (2007) reported
68
in the literature. More recently, Luo et al. (2018) proposed a localization scheme
69
that does not rely on the spatial locations and update model parameters using only
70
the observations that have relatively high correlations with them. They tested the
71
method to update gridblock porosity and permeability in the presence of seismic
72
data showing that the proposed method avoided ensemble collapse.
73
Another procedure that is quite common in the literature to compensate the un-
74
derestimation of posterior variances in ensemble data assimilation is covariance infla-
75
tion (Anderson and Anderson, 1999). Covariance inflation is often used in oceanogra-
76
phy and numerical weather prediction applications (Anderson, 2007b; Hamill et al.,
77
2001; Li et al., 2009), but has rarely been applied in reservoir history-matching prob-
78
lems. Perhaps one of the few works using covariance inflation for history matching
79
was presented by Chen and Oliver (2010a). They concluded that covariance infla-
80
tion was useful to compensate for the insufficiency of the ensemble variability for the
81
Brugge case. In covariance inflation, the forecast ensemble is replaced by another
82
one with the same mean but with an inflated variance. The inflation factor is a
83
number slightly greater than one, but its choice is problem dependent. Typically a
84
higher inflation factor is required for smaller ensembles.
85
Ensemble methods such as EnKF and ensemble smoother use a “perturbed ob-
86
servation scheme,” which means that a random vector is added to the measurements 4
87
before the analysis. However, it has been shown that the perturbed observations is
88
an additional source sampling errors (Anderson, 2001; Whitaker and Hamill, 2002;
89
Tippett et al., 2003). This fact motivated the development of deterministic or
90
square-root schemes which avoid the need of perturbed observations. In spite of
91
being standard methods in areas such as oceanography and numerical weather pre-
92
diction, deterministic schemes have been rarely used in reservoir history matching.
93
A review on the main square-root schemes can be found in (Tippett et al., 2003;
94
Sakov and Oke, 2008a). Among these methods, the deterministic EnKF (DEnKF)
95
(Sakov and Oke, 2008b) has the practical advantage of having an implementation
96
very close to the standard EnKF. In fact, both methods compute the same Kalman
97
gain, allowing Schur product localization. Besides avoiding the need of perturbed
98
observations, DEnKF has a “built-in” covariance inflation, which makes the method
99
less prone to variance underestimation than EnKF.
100
Despite of the aforementioned works on non-distance localization schemes and
101
covariance inflation, our experience is that these methods are less effective than
102
distance-based localization in eliminating sampling errors in gridblock properties
103
such as porosity and permeability. Emerick (2012, Chap. 4) compared some of these
104
methods and concluded that only distance-based localization was able to prevent a
105
severe variance reduction in two synthetic history matching problems. Most of the
106
previous works, including the comparisons presented in (Emerick, 2012), consider
107
only the problem of localization of gridblock parameters. However, as previously
108
mentioned, there are several types of non-local model parameters in real-life reservoir
109
history-matching problems, in which case the development of robust non-distance
110
localization schemes is highly desirable. Surprisingly, to the best of our knowledge,
111
publications about this problem in the history-matching literature are non-existent.
112
This work attempts to contribute to fill this lacuna. We present a comparison
113
among five non-distance localization methods from literature, covariance inflation
114
and a deterministic scheme to address the problem of excessive loss of variance.
115
Moreover, we also propose a simple sensitivity analysis procedure which can be used
116
to select or improve the localization coefficients for non-local parameters.
117
This paper is organized as follows: the next section briefly reviews the base data 5
118
assimilation method used in this paper. The section after that describes the test
119
case proposed to compare the methods. Then, we present a separate section for each
120
method investigated. In each of these sections, we provide a brief description of the
121
method followed by the corresponding results. After that, we present a section with
122
a discussion of the overall performance of the methods. The last section of the paper
123
presents the conclusions.
124
2. Ensemble Data Assimilation
125
2.1. Ensemble Smoother with Multiple Data Assimilation
126
Recent papers indicate that iterative forms of the ensemble smoother (ES) are
127
better suited for reservoir history matching than the standard sequential data assim-
128
ilation from EnKF (Chen and Oliver, 2012, 2013; Emerick and Reynolds, 2013a; Luo
129
et al., 2015; Stordal and Elsheikh, 2015; Le et al., 2016; Ma et al., 2017). Among the
130
iterative ES, the ensemble smoother with multiple data assimilation (ES-MDA) (Em-
131
erick and Reynolds, 2013a) has been widely used in history-matching applications.
132
In this paper, we use the standard ES-MDA, although the methods investigated here
133
could be used with other iterative smoothers.
134
ES-MDA was introduced in (Emerick and Reynolds, 2013a) with the objective
135
to improve the history-matching results obtained by ES. In the ES-MDA, the same
136
data are assimilated multiple times with an inflated covariance of the measurement
137
errors. The ES-MDA analysis equation for a vector of model parameters m ∈ RNm
138
can be written as
mk+1 = mkj + Ckmd Ckdd + αk Ce j
−1
dobs + ekj − g mkj
,
(1)
139
for j = 1, . . . , Ne , with Ne denoting the ensemble size. In the above equation, the
140
superscript k stands for the data assimilation index; dobs ∈ RNd is the vector of
141
observations, ekj ∈ RNd is the vector of random perturbations which is obtained by
142
143
144
sampling N (0, αk Ce ), with Ce ∈ RNd ×Nd denoting the data-error covariance matrix and αk denoting the ES-MDA inflation coefficient. g mkj ∈ RNd is the vector of predicted data. Ckmd ∈ RNm ×Nd is the matrix containing the cross-covariance values
6
145
between model parameters and predicted data and Ckdd ∈ RNd ×Nd is the covariance
146
matrix of predicted data. Both matrices are estimated using the ensemble, i.e.,
Ckmd 147
Ne > 1 X g mkj − g (mk ) g mkj − g (mk ) , Ne − 1 j=1
Ne 1 X = mk Ne j=1 j
and g (mk ) =
150
151
(3)
where
mk
149
(2)
and
Ckdd = 148
Ne > 1 X k k k k mj − m = g mj − g (m ) , Ne − 1 j=1
(4)
Ne 1 X g mkj . Ne j=1
(5)
In the standard implementation of ES-MDA, the inflation coefficient αk > 1 is a applied a pre-defined number of times, Na , and the set {αk }N k=1 must satisfy the conP a −1 dition N k=1 αk = 1. This condition was derived by Emerick and Reynolds (2013a)
152
to ensure that ES-MDA is consistent with the standard ES for linear-Gaussian prob-
153
lems. The same condition can also be derived directly from Bayes’ rule as discussed
154
in (Stordal and Elsheikh, 2015; Emerick, 2016). Here, we use the standard choice of
155
αk = Na = 4 (Emerick and Reynolds, 2013a; Emerick, 2016).
156
Defining the Kalman gain matrix as
Kk = Ckmd Ckdd + αk Ce 157
−1
,
(6)
we can rewrite Eq. 1 in a compact form
mk+1 = mkj + Kk dobs + ekj − g mkj j
.
(7)
158
In this paper, the matrix inversion in Eq. 6 is done using truncated singular
159
value decomposition (TSVD). The number of singular values retained in the TSVD is
160
defined by keeping 99% of the sum of the singular values. It is important to note that
161
TSVD acts as a regularization of the inversion helping to stabilize the Kalman gain. 7
162
This procedure is known to help to reduce undesirable variance loss in the ensemble,
163
especially in the case of a large number of redundant measurements. However, in
164
our experience, TSVD alone is not enough to avoid variance underestimation and
165
some sort of localization or inflation scheme are required in realistic settings. On
166
a side note, we usually apply TSVD before the construction of the matrix Ckdd ,
167
because it results in a more efficient method when the number of data points is
168
larger. This procedure is known as subspace inversion (Evensen, 2004). Nevertheless,
169
170
here because we consider a relatively small test case, we apply TSVD directly to Ckdd + αk Ce rescaled by the data variance to avoid loss of relevant information,
171
as discussed in (Wang et al., 2010).
172
2.2. Sampling Errors
173
The limited size of the ensembles employed in practice introduces sampling errors
174
which generates spurious correlations between data and model parameters. As a
175
result, we observe incorrect changes in model parameters resulting in excessive loss
176
of variance. In the extreme cases, even a collapse of the ensemble may be observed.
177
Hamill et al. (2001) investigated the effect of sampling errors in a hypothetical
178
problem with two parameters and one observation. Here, we present a shortened
179
version of their result to highlight some points that are important in the analysis of
180
the results presented in the next sections.
181
Assume that we want to estimate the porosity and permeability of a rock sample,
182
m = [φ κ]> , given a single observation of porosity with variance σe2 . Assume that
183
there is a prior correlation between porosity, φ, and permeability, κ, and that the
184
actual prior covariance matrix has the form Cm =
σφ2
ρσφ σκ
ρσφ σκ
σκ2
,
(8)
185
where σφ and σκ are the prior standard deviation of porosity and permeability,
186
respectively, and ρ is the prior correlation coefficient between φ and κ. Suppose
187
e m where the variances are that instead of Cm , we have an inaccurate version of C
188
correct but the cross-covariances are corrupted with an additive sampling error cε ,
189
i.e., 8
em = C
σφ2
ρσφ σκ + cε
ρσφ σκ + cε
σκ2
.
(9)
190
For a given observation φobs , it is straightforward to show that the maximum a
191
posteriori (MAP) (Tarantola, 2005) estimate of κ is
κmap = κpr +
ρσφ σκ (φobs − φpr ) σφ2 + σe2
(10)
if we use the correct prior covariance and κ emap = κpr +
ρσφ σκ + cε σφ2 + σe2
! (φobs − φpr )
(11)
192
e m . In the above expressions, φpr and κpr are if we use the corrupted covariance C
193
the prior estimates of φ and κ, respectively. We want to analyse the error in the
194
estimate of κ. It follows that the expected squared error has the form ρ2 σφ2 σκ2 E (κ − κ emap )2 = σκ2 − 2 σφ + σe2
1−
cε ρσφ σκ
2 ! .
(12)
195
The derivation of (12) is straightforward and follows from the direct application of
196
the definitions of expectation and variance. Hamill et al. (2001) interpreted the term
197
cε ρσφ σκ
198
points: (i) if
199
clearly an inconsistency (there is a degradation of the estimate of κ by assimilating
200
data); (ii) If the actual correlation coefficient between φ and κ is small, it is more
201
likely that we have a degradation in our estimates due to the sampling error.
as a “relative error” in the covariance. From Eq. 12, we can emphasize two cε ρσφ σκ
> 1 the variance of κ increases by assimilating data, which is
202
The second point is particularly important because weak correlations are harder
203
to estimate with limited samples. In order to illustrate this last statement, we con-
204
ducted a small numerical experiment to estimate coefficient between two parameters,
205
φ and κ, based on an ensemble of Ne = 50 members. We consider two cases where
206
the true correlations are ρ = 0 and ρ = 0.8 and repeat the sampling 100 times.
207
Fig. 1 shows the histograms of the estimates of ρ, which clearly indicates a large
208
variance for the case with ρ = 0.
9
209
2.3. Localization
210
Localization is the standard technique to reduce spurious correlations caused by
211
sampling problems in ensemble-based methods. Localization is done by making the
212
Schur (Hadamard) product (element-wise product of matrices) between a correlation
213
matrix and a covariance matrix. This Schur product has the property of increasing
214
the degrees of freedom available to assimilate data (Aanonsen et al., 2009). Here,
215
we apply the Schur product directly to the Kalman gain matrix because this leads
216
to an efficient implementation for the case with a large number of measurements
217
(Emerick, 2016). This procedure corresponds to our standard implementation for
218
practical applications. In this case, the ES-MDA analysis equation can be written
219
as
mk+1 = mkj + R ◦ Kk j
dobs + ekj − g mkj
.
(13)
220
where ◦ denotes the Schur product and R ∈ RNm ×Nd is the so-called localization
221
matrix.
222
In the typical application of ES-MDA to update porosity and permeability fields,
223
the components of the matrix R are computed assuming a correlation function with
224
compact support with argument defined by the distance between the location of each
225
datum, usually the position of a well, and the parameter location, the position of
226
a gridblock. Hence, the name localization. For non-local parameters, on the other
227
hand, we cannot define distances, but the overall idea of applying the Schur product
228
is still valid. In this case, we need to derive alternative schemes to compute the
229
entries of the matrix R without relaying in the Euclidian distance between data
230
and parameters. It is worth noting that here we emphasize the case of non-local
231
parameters because it is a very common problem in history matching. However,
232
the same situation occurs with non-local data, for example, if one is interested in
233
assimilating data from the total field production as opposed to assimilate data on a
234
well-by-well basis.
10
235
3. Test Problem
236
The test problem consists of a modified version of the well-known PUNQ-S3 case
237
(Floris et al., 2001). The PUNQ-S3 model contains 19 × 28 × 5 gridblocks, of which
238
1761 blocks are active (Fig. 2). The reservoir is supported by a strong analytical
239
aquifer on West and South sides. There are six oil producing wells under oil-rate
240
control operating during the historical period. We consider a production history
241
with a first year of extended well testing, followed by a three-years shut-in period
242
and then four more years of production. There are five infill-drilling wells starting
243
production in the forecast period. The observed data corresponds to measurements
244
of water (WPR) and gas (GPR) production rates and bottom-hole pressure (BHP).
245
The observations were corrupted with a Gaussian noise with zero mean and standard
246
deviation corresponding to 10% of the data value for rate data and 1% for BHP data.
247
In the original PUNQ-S3 case the history matching parameters are the poros-
248
ity and the horizontal and vertical permeability distribution. However, because
249
our goal is to evaluate the performance of techniques for non-local model param-
250
eters, we consider a different set of parameters: five porosity multipliers (MULT-
251
POR1, ..., MULTPOR5) and five permeability multipliers (EXP MULTIPERM1, ...,
252
EXP MULTIPERM5), one for each layer of the model. The parameters for perme-
253
ability multipliers correspond to the exponent of a base-10 power. For example, the
254
permeability multiplier applied to the first layer of the model is 10EXP MULTIPERM1 .
255
We also included one parameter defining the rock compressibility (CCPOR), two
256
parameters to compute variations in the water-oil (DELTA DWOC) and gas-oil
257
(DELTA DGOC) contact depths and two exponents defining the analytical aquifer
258
radius (EXP AQRADIUS1 and EXP AQRADIUS2). Table 1 shows the prior distri-
259
butions for these parameters including the true values used to generate the synthetic
260
production history. Note that we intentionally selected the prior mean for the param-
261
eters EXP AQRADIUS1, EXP AQRADIUS2, DELTA DWOC and DELTA DGOC
262
biased compared to the true values. The parameters were selected such that we
263
have a wide spectrum of data sensitivity. For example, the selected measurements
264
are highly sensitive to permeability multipliers and fluid contacts, while rock com-
265
pressibility and aquifer radius have a moderate sensitivity and porosity multiplies are 11
266
weakly sensitive. Therefore, we expect different levels of variance reduction in these
267
parameters. Moreover we added five insensitive parameters, labeled as DUMMY1,
268
..., DUMMY5. Because these parameters are not related to data, we expect any
269
variance reduction to be due to sampling errors. Hereafter, we refer to the sensitive
270
parameters as “actual” and the five insensitive as “dummy” parameters.
271
In order to compare the methods, we consider ten independently sampled prior
272
ensembles with 100 members each. All cases use the same prior ensembles with
273
ES-MDA with αk = Na = 4 for k = 1, . . . , Na . We consider the following metrics to
274
compare the methods:
275
(i) Average normalized data-mismatch objective function: Ne 1 X Od (m) = Od (mj ) Ne j=1
276
(14)
where
Od (mj ) =
1 (dobs − g(mj ))> C−1 e (dobs − g(mj )) . 2Nd
(15)
277
Here, based on our experience with the test case and the arguments presented
278
in Oliver et al. (2008, Chap. 8), we assume that a good data match correspond to a
279
value of Od (m) approximately equal to or less than one. The total number of data
280
points in the test problem is Nd = 2784.
281
(ii) Root-mean-square error (RMSE): v u 2 Nm u 1 X mi − mtrue,i t RMSE = , Nm i=1 σm,i
(16)
282
where mtrue,i is the true value for the ith model parameter and mi is the correspond-
283
ing ensemble mean. σm,i is the prior standard deviation of the ith model parameter.
284
(iii) Sum of normalized variance (SNV): Nm 1 X var[m0i ] SNV = , Nm i=1 var[mi ]
12
(17)
285
where var[m0i ] and var[mi ] denote the variance of the ith posterior and prior model
286
parameters, respectively. The SNV is used as an approximate measure of the re-
287
duction in uncertainty (Oliver et al., 2008). Note that SNV = 0 means ensemble
288
collapse and SNV = 1 means no uncertainty reduction.
289
As final remark on our test problem, we note that even though sampling errors
290
are the main cause of underestimation of the posterior variances, in practice, we
291
may observe similar problems caused by neglecting or poor treatment of model and
292
measurement errors. These cases are not addressed in this paper. Nevertheless, it
293
is important to note that because we used a controlled synthetic problem, we have
294
no model error and the correct level of measurement errors is known. Therefore, we
295
only have the sampling errors problem.
296
3.1. Reference Case
297
The sampling errors are caused by the limited size of the ensemble. Conceptually,
298
increasing the size of the ensemble mitigates this problem, however, computational
299
efficiency limits the size of the ensemble in practical applications. Nevertheless, for a
300
small problem such as the PUNQ-S3 case it is computationally feasible to consider a
301
very large ensemble. Therefore, we consider a case with standard ES-MDA without
302
localization and an ensemble of Ne = 50000 members as reference for comparisons.
303
The last row of the Table 2 presents the values of Od (m), RMSE and SNV obtained
304
by the reference case. The results in this table show that for the actual model
305
parameters, we observe an average reduction in the ensemble variance of 23%. For
306
the dummy parameters the SNV is 1.00, indicating that the selected size for the
307
ensemble is enough to mitigate significant effects of sampling errors. The value of
308
Od (m) is 0.8, which is less than one, indicating a good data match according to our
309
criterium. The value of the RMSE is 0.89.
310
3.2. No Localization Case
311
We initially consider the data assimilations results with standard ES-MDA with-
312
out any localization or inflation method. Table 2 summarizes the results for this
313
case. The results in this table indicates good data matches with Od (m) inferior to
314
the reference case. The average RMSE is 1.19, which is 34% higher than reference 13
315
value. This result shows a reduction in the ability of the data assimilation to identify
316
the true values of the parameters. However, the main differences are in the values
317
of SNV. For the actual model parameters the average value of SNV is 0.10, which is
318
less than half of the reference result. For the dummy parameters the average SNV
319
is only 0.43. These results show a significant variance underestimation, evidencing
320
the need of localization or inflation.
321
3.3. Pseudo-optimal Localization
322
Furrer and Bengtsson (2007) derived a localization function for general covari-
323
ance structures without relaying on distance between variables. They minimized
324
term-by-term the norm of the difference between the true covariance and the local-
325
ized estimate ignoring the positive-definiteness constraint. The resulting expression
326
depends on the true covariance
ri,j =
c2i,j
, c2i,j + c2i,j + ci,i cj,j /Ne
(18)
327
where ri,j is the (i, j)th entry of the localization matrix R and ci,j is the corre-
328
sponding covariance value. For the cases of interest in this paper, ci,j represents the
329
covariance between the ith model parameter and the jth predicted data, i.e., ci,j is
330
an entry of the matrix Cmd . Furrer and Bengtsson (2007) proposed to replace ci,j by
331
the corresponding ensemble estimate, forming a simple and computationally inex-
332
pensive non-distance dependent localization procedure. In the following, we refer to
333
this procedure as POL, which stands for pseudo-optimal localizaiton. The authors
334
also suggested that sparseness can be introduced by zeroing small values of ci,j .
335
Although the definition of small value for ci,j is problem-dependent, a reasonable
336
choice is to set rij = 0 if √ |ci,j | < ci,i cj,j ,
(19)
337
where > 0 is small user-defined threshold. Our tests indicate that the choice of
338
has a significant impact in the overall performance of the method.
339
Table 3 summarizes the results of POL method with = 10−3 and = 10−1 .
340
The results for the case = 10−3 are more stable in terms of the data match. Note 14
341
that the case with = 10−1 resulted in large values of Od (m) for the fifth and eighth
342
ensembles. In terms of RMSE, both cases resulted in average values smaller than the
343
reference case. The most interesting results are the SNV. The case with = 10−1
344
resulted in an average SNV of 0.98 for the dummy parameters, which is very close
345
to the reference. However, the average SNV of the actual parameters is considerable
346
higher than the reference. The case with = 10−3 resulted in a better value for the
347
average SNV for the actual parameters, but only 0.77 for the dummy parameters.
348
Even though the mean SNV of dummy parameters for the case = 10−1 is close
349
to the unity, it is important to note that for some ensembles, we obtained SNV
350
larger then one, which can be explained by the discussion in the Section 2.2 when
351
the relative error in the covariance is larger than one. Overall, we concluded that
352
= 10−3 is a better choice for our test problem. We also tested the cases with = 0
353
and = 10−2 . Fig. 3 presents boxplots of SNV for different values indicating that
354
the method is relatively stable for ≤ 10−2 .
355
ES-MDA can be considered an iterative method in the sense that the analysis
356
equation is applied multiple times. Before each iteration, the covariance matrices
357
are updated with the current ensemble. Therefore, it is reasonable apply Eq. 18
358
before each iteration to update the localization coefficients. However, our tests indi-
359
cate that it is better to compute the localization matrix R based only on the prior
360
ensemble. The results shown in Table 3 and Fig. 3 were obtained computing R
361
using the prior ensemble. It is not completely clear the reason for this behaviour.
362
However, one possible explanation is that the application of the analysis builds cor-
363
relations in the ensemble members. This effect is referred to as inbreeding in the
364
literature (Houtekamer and Mitchell, 1998; van Leeuwen, 1999). Fig. 4 illustrates
365
this effect showing cross-plots between the parameter DUMMY5 and the predicted
366
BHP from well PRO-4 at time 2921 days for each iteration. The estimated correla-
367
tion coefficient is very small for the prior ensemble. However, after the first iteration,
368
a spurious correlation between DUMMY and BHP appears, which affects the com-
369
putation of the localization coefficient. Moreover, the variance of the predicted data
370
tends to reduce after iterations, this is a natural consequence of the improvement
371
of the data match. Note that the horizontal axes of the plots in Fig. 4 are different 15
372
and that the variance of BHP clearly reduces. It is conceivable that this fact also
373
impacts the calculation of the localization coefficient with Eq. 18. The data variance
374
corresponds to the entry cj,j in this equation. Hence, small values of cj,j tend to
375
increase the value of rj,j .
376
3.4. Bootstrap Localization
377
Zhang and Oliver (2010) presented a modification of the hierarchical filter of
378
Anderson (2007a), where instead of Nh ensembles of Ne members, a single ensemble
379
is resampled with replacement to generate Nb bootstrapped ensembles. This modi-
380
fication has the advantage of avoiding the cost of running additional ensembles. In
381
this method, the Nb bootstrapped ensembles are used only to compute the confi-
382
dence factors; the data assimilation is done with the original ensemble. They use
383
the bootstrapped ensembles to generate Nb Kalman gains, K∗n , and estimate the
384
2 population variance, σK , as i,j
2 σK i,j
Nb 1 X = ([K∗n ]i,j − [K]i,j )2 Nb n=1
(20)
385
where [K∗n ]i,j and [K]i,j denote the (i, j)th entry of the nth bootstrapped and original
386
Kalman gain matrices, respectively.
387
Zhang and Oliver (2010) followed the same procedure presented in Anderson
388
(2007a) to derive an expression for confidence factors for the Kalman gain, which
389
are used as entries of the localization matrix R. Because the resulting procedure
390
sometimes generates negative values, Zhang and Oliver (2010) proposed to add a
391
regularization term and derived the following expression for the confidence factors
ri,j =
1+
2 βi,j
1 , (1 + 1/σr2 )
(21)
392
2 2 where βi,j = σK /[K]2i,j and σr is a weighting factor for regularization of the estimate i,j
393
of ri,j . Here, we refer to this method as bootstrap localization (BL).
394
We tested BL method using different values for σr with Nb = 50 and Nb = 100.
395
We selected the case with Nb = 50 and σr = 0.6, which is the same value suggested
396
by Zhang and Oliver (2010). Table 4 summarizes the results indicating that BL
397
method was able to improve the average values of SNV for both, actual and dummy 16
398
parameters. The mean RMSE value is relatively close to the reference. However,
399
BL failed to mach data for two out of the ten ensembles, showing a possible lack of
400
robustness in the method. Note that the SNV of the two ensembles with failed data
401
match are much higher, increasing the mean values reported in Table 4. Without
402
these two ensemble the mean values of SNV are 0.24 and 0.78 for the actual and
403
dummy parameters, respectively.
404
3.5. SENCORP
405
Another localization scheme available in the literature is the Smoothed Ensemble
406
Correlations Raised to a Power (SENCORP) proposed by Bishop and Hodyss (2007).
407
In this method, spurious correlations are attenuated by raising them to a power. In
408
the following, we summarize the main steps of the SENCORP method adapted for
409
our problem and notation. A complete description of each step can be found in the
410
original paper (Bishop and Hodyss, 2007).
411
412
1. Define the augmented model vector y ∈ RNy , where Ny = Nm + Nd , by including the predicted data, i.e., y=
m g(m)
,
(22)
413
and standardize it by subtracting the mean and dividing its elements by the
414
corresponding standard deviation, i.e.,
e= y
h
y1 −y1 σy,1
where
···
yNy −yNy σy,Ny
i>
Ne 1 X yi = yi,j Ne j=1
and σy,i 415
y2 −y2 σy,2
v u Ne u 1 X =t (yi,j − yi )2 . Ne j=1
(23)
(24)
(25)
2. Compute the ensemble estimate of the covariance matrix N
e 1 X ej y ej> . Cye = y Ne − 1 j=1
17
(26)
416
Note that for a general history-matching problem, the vector y may include
417
rock properties for all reservoir gridblocks. In this case, the matrix Cye may
418
be very large, which may limit the application of the method. The reason
419
for using the augmented vector is to allow the matrix product of step 4. For
420
the problems of interest of this paper, however, we have only a few non-local
421
parameters. Hence, the construction of Cye is not an issue. Also note that the
422
resulting matrix can be divided into four sub-matrices Cye =
Cm e
Cm e ed
Cde m Cde de e
.
(27)
423
For Kalman gain localization, we need only the elements referent to the sub-
424
matrix Cm e . However, we need to carry the entire Cy e to perform the matrix ed
425
products of step 4. At this step, Bishop and Hodyss (2007) also applies a
426
spatial smoothing procedure to this estimated covariance. However, it is not
427
obvious how to define this smoothing procedure for our non-local parameters,
428
so we do not apply this step in our test case.
429
430
3. Raise the matrix to the element-wise power n by a sequence of n Schur products, i.e., C◦n e ◦ Cy e ◦ · · · ◦ Cy e. e = Cy y | {z }
(28)
n times
431
4. Raise the resulting matrix to the power q by a sequence of multiplications C◦n e y
q
◦n ◦n = C◦n e × Cy e × · · · × Cy e . y | {z }
(29)
q times
432
433
5. Renormalize the matrix C◦n e y
q
using
^ q q −1/2 ◦n Cye = S−1/2 C◦n S (30) e y h q i where S = diag C◦n . According to Bishop and Hodyss (2007), the mae y
434
trix product has an smoothing effect, but it boosts both actual and spurious
435
correlations. For this reason, they introduced a final element-wise product
436
(step 6) to attenuate the spurious correlations. 18
437
q ^ 6. Raise the matrix C◦n to the element-wise power p, i.e., e y "
q ^ C◦n e y
#◦p
q ^ q q ^ ^ ◦n ◦n = C◦n ◦ C ◦ · · · ◦ C . e e e y y y | {z }
(31)
p times
"
438
439
440
q ^ ◦n The matrix Cye
#◦p is used for Schur product localization. Again, because
here we"use only #Kalman gain localization, we need only the Nm × Nd sub^q ◦p matrix C◦n to construct the localization matrix R. e ed m
441
The number of matrix element-wise products and matrix multiplications used
442
in SENCORP are defined by the integers n, q and p. The optimal values of these
443
parameters must be determined through experimentation. The final element-wise
444
product parameter p should be chosen to ensure that the SENCORP elements are
445
positive or zero in order to keep the signs of correlations. The fact that we have
446
three free parameters to select, n, q and p, gives a potential flexibility to the method,
447
but it also makes harder to make a robust selection. We tested 28 combinations for
448
the values n, q and p. Based on our tests (not reported here), we selected n = 2,
449
q = 2 and p = 1.
450
Table 5 summarizes the results obtained with SENCORP. The method failed
451
to match data in one ensemble. The mean RMSE is 0.74, which is smaller than
452
the reference case, showing a good performance on revealing the true parameter
453
values. The mean SNV of the actual parameters is 0.27, which is larger than the
454
reference indicating an excessive localization. For the dummy parameters, on the
455
other hand, the mean SNV is 0.76, which is clearly an improvement compared to ES-
456
MDA without localization, but is still smaller than the reference. The localization
457
coefficients with SENCORP were recalculated every MDA iteration. We also tried a
458
case with localization computed based only on the prior ensemble, but this procedure
459
did not improve the results and several ensembles failed to match data properly.
460
3.5.1. Simplified SENCORP
461
462
Bannister (2015) proposed to use a simplified version of SENCORP by applying only steps 1–3, in which case we have only one free parameter to select n, i.e., 19
R = Cm e ed
◦n
.
(32)
463
With a single free parameter, the method becomes easier to tune. Moreover, we do
464
not need to construct the entire Ny ×Ny matrix Cye , which is an advantage compared
465
to the original SENCORP. Hereafter, we refer to this procedure a S-SENCORP.
466
Table 6 shows that S-SENCORP resulted in excessively large values for the SNV of
467
the actual parameter. For the dummy parameters that mean SNV is 0.94, which
468
is close to the correct value. However, because of the values of SNV for the actual
469
parameters, we conclude that this method introduced an excessive localization. This
470
is also in concordance with the fact that the values of Od (m) obtained are larger
471
than one, indicating poor data matches.
472
3.6. Correlation-Based Localization
473
Luo et al. (2018) proposed a localization method that does not rely on the dis-
474
tance between variables. They estimate correlation coefficient between the ith model
475
parameter and jth predicted data, ρi,j as
gj (mn ) − gj (m) n=1 mi,n − mi r =r 2 P 2 . PNe Ne n=1 mi,n − mi n=1 gj (mn ) − gj (m) PNe
ρi,j
476
(33)
Based on the value of ρi,j , Luo et al. (2018) select the localization coefficients using
ri,j
1, if |ρ | > θ i,j = 0, otherwise
(34)
477
where θ > 0 is a threshold value. Luo et al. (2018) described two special cases in
478
the use of Eq. 33 to avoid division by zero: (i) if all values of mi in the ensemble are
479
identical, we set ρi,j = 0, which corresponds to ri,j = 0. (ii) Similarly, if all values of
480
gj (m) in the ensemble are the same, set ρi,j = 0 and ri,j = 0. Luo et al. (2018) also
481
present a procedure for selecting θ for the case with grid-based model parameters
482
and seismic data points. Luo et al. (2018) suggested that a different threshold value
483
could be used for each type of model parameter. For non-local parameters, the
484
authors mention that Eqs. 33 and 34 can still be used, but the choice of θ must be 20
485
provided by the user. We considered a single and constant value of θ for each data
486
assimilation run. We tested the method updating the localization matrix every ES-
487
MDA iteration and using a constant matrix computed based on the prior ensemble.
488
The second procedure resulted in better results, similarly to what it was observed
489
in the POL method. The same observation was reported in Luo et al. (2018). Here,
490
we refer to this method as CBL (correlation-based localization).
491
We run some experiments to determine the best θ in our test problem and the
492
results are summarized in Fig. 5. For θ = 0.1 or higher, we observed a failure on
493
matching data indicating excessive localization. For θ = 0.01 the results approached
494
the case without localization. Our experiments indicate that θ = 0.05 was the best
495
choice. Table 7 sumarizes the results for θ = 0.05. The method failed to match data
496
for the second ensemble. The mean SNV value of actual parameters is close to the
497
reference, but a large variation is observed among the ensembles indicating a lack of
498
robustness in the method. Note, for example that the SNV of the actual parameters
499
is 0.39 for the seventh ensemble, but only 0.18 for the ninth. Moreover, the SNV
500
values for the dummy parameters are larger than one for almost all ensembles, which
501
shows that the method has difficulties dealing with weakly or uncorrelated model
502
parameters.
503
3.7. Covariance Inflation
504
Covariance inflation (CI) (Anderson and Anderson, 1999) is typically applied
505
with EnKF and other square-root filters to compensate for the erroneous variance
506
loss due to sampling error in the numerical weather prediction literature. Here, we
507
apply an inflation factor, γ > 1, after each iteration of ES-MDA. The idea is to
508
inflate the variance without changing the mean, which is accomplished using
mkinf,j
= γk
mkj
−
mk
+ mk ,
(35)
509
for j = 1, . . . , Ne , where the superscript k denotes the ES-MDA data assimilation
510
index. The choice of the inflation factor is problem dependent. There are several
511
works in the literature proposing covariance inflation schemes; see, e.g., (Kotsuki
512
et al., 2015) and references therein. Here, we use the scheme proposed by Evensen 21
513
(2009) adapted for use with ES-MDA. In this procedure, before each ES-MDA iter-
514
ation, we augment the vector of model parameters with a vector of white noise, i.e.,
515
zj ∼ N (0, I). After each ES-MDA update, we compute the inflation factor as
γk = where σz,i
v u u =t
1 PNz
1 Nz
i=1
(36)
σz,i
N
e 1 X (zi,j − zi )2 . Ne − 1 j=1
(37)
516
We applied this covariance inflation scheme in the test case with Nz = 100. Fig. 6
517
shows boxplots of the inflation factors for the four ES-MDA iterations. The inflation
518
factor varies between 1.06 to 1.29. The largest inflations occur after the second ES-
519
MDA iteration. Table 8 shows the results of covariance inflation indicating a clear
520
improvement in the values of SNV compared to the case with no localization without
521
compromising the quality of the data match or the RMSE. The final SNV of the
522
dummy parameters is larger than one for half of the ensembles, but the average
523
is 1.02, which is very close to the correct value. Compared to the other methods
524
investigated in this paper, covariance inflation with the scheme from (Evensen, 2009)
525
has the advantage of not requiring multiple tests to tune the method.
526
3.8. Deterministic ES-MDA
527
Deterministic or square-root schemes avoid the need to perturb the observation
528
vector before analysis. The deterministic EnKF of Sakov and Oke (2008b) is not
529
exactly an square scheme but it was proposed with the same objective, i.e., avoid
530
the sampling errors caused by the process of perturbing the observations. Recently
531
the DEnKF scheme was combined with ES-MDA forming the method DES-MDA
532
(Emerick, 2018). DES-MDA computes the same Kalman gain of ES-MDA (Eq. 6),
533
which is used to update the ensemble mean with
k+1
m
k
k
=m +K
dobs −
g(mk )
.
(38)
534
Unlike the standard ES-MDA, DES-MDA updates the matrix with the ensemble
535
deviations from the mean directly using 22
1 ek ∆Mk+1 = ∆Mk − K ∆Dk , 2
(39)
where ∆M =
h
m1 − m · · ·
mNe − m
i
(40)
and ∆D =
h
g(m1 ) − g(m) · · ·
g(mNe ) − g(m)
i
.
(41)
After that, the ensemble members are computed using Mk+1 = ∆Mk+1 + M
k+1
,
(42)
536
where M = [m · · · m]. Sakov and Oke (2008b) showed that the use of Eq. 39
537
incurs in an extra positive semi-definite term in the estimated posterior covariance.
538
They interpreted this term as an implicit covariance inflation in the method, which
539
may alleviate the underestimation posterior variances.
540
We applied DES-MDA with no localization with the same configuration of ES-
541
MDA, i.e, αk = Na = 4. Table 9 summarizes the results of DES-MDA. The mean
542
SNV obtained are 0.19 and 0.50 for actual and dummy parameters, respectively.
543
These values represent improvements compared to the standard ES-MDA without
544
localization, but we still observe a significant underestimation of posterior variances
545
compared to the reference case. In terms of Od (m) and RMSE the results of DES-
546
MDA are in reasonable agreement with the reference.
547
3.9. Localization Based on Sensitivity Analysis
548
Here, we introduce another procedure to mitigate sampling errors for non-local
549
parameters based on simple one-at-a-time sensitivity analysis (SA). The idea is to
550
run two reservoir simulations for each model parameter, one for the parameter at its
551
minimum value and another at its maximum value. All other parameters are fixed
552
at a base (or mean) value. Than, we compute the sensitivity coefficients using si,j =
gj (mi,max ) − gj (mi,min ) σe,j
2 ,
(43)
553
where si,j corresponds to the sensitivity coefficient for the ith model parameters
554
with respect to the jth predicted data point, gj (mi,max ) and gj (mi,min ) denote the jth 23
555
predicted data from the ith model parameter at the maximum and minimum values,
556
respectively, σe,j denotes the standard deviation of the jth observed data point.
557
Based on the value of si,j , we compute the corresponding entry in the localization
558
matrix, R, as
ri,j 559
1, if s > θ i,j = 0, otherwise
(44)
where θ > 0 is a threshold value.
560
SA is similar to CBL method in the sense that generates a matrix R with only
561
zeros and ones. This procedure is able to identify insensitive parameters, such as
562
the dummy parameters used in the test case, regardless the choice of the threshold
563
θ. Nevertheless, the value of θ impacts in the overall performance of the method.
564
Large values of θ imposes more localization. We tested different values of θ and the
565
results are summarized in Fig. 7. The results in this figure indicate that θ = 10−1
566
is a good choice. Smaller values of θ resulted in too small posterior variance in the
567
actual parameters. For θ = 1, the scheme imposed too much localization resulting
568
in too large value of SNV and in a poor data match (the average objective function
569
was 16.5).
570
Table 10 shows the results for SA localization. The method failed to match data
571
for one ensemble. The SNV of the actual parameters are lower than the reference,
572
indicating insufficient localization, although the results are better than the case
573
without localization. As expected, this procedure correctly identified the insensitive
574
parameters, preserving the variance.
575
The SA scheme is very effective to remove insensitive parameters, but it is not
576
able to compute different localization coefficients for parameters with different levels
577
of sensitivity, because the procedure returns only binary values. This suggests that
578
the SA scheme can be combined with another method, such as the POL to form a
579
more effective localization scheme. In this case, the localization matrix is the result
580
of the Schur product between the two localization matrices, i.e.,
R = RSA ◦ RPOL ,
24
(45)
581
where RSA is the localization matrix resulted from the SA scheme, preferable with
582
a small threshold used only to remove very weak sensitive parameters and RPOL
583
is the localization matrix resulted from POL method. We tested this procedure
584
considering = 10−3 (threshold for POL) and θ = 10−4 (threshold for SA) and the
585
results are summarized in Table 11. We selected a small threshold for SA to ensure
586
that the method will remove only the data points with very small sensitivities with
587
respect to the parameters. Overall, the results are close to the ones obtained with
588
POL method, but the SNV of the dummy parameters are preserved.
589
Compared to the other methods, SA has the disadvantage of requiring additional
590
reservoir simulations, two for each parameter, which may limit the application of the
591
method. It is worth mentioning, however, that these simulations are independent in
592
the sense that they can be executed simultaneously, reducing the total time of the
593
process if the simulations are executed in a cluster of computers.
594
4. Discussion
595
Table 12 presents the average values of SNV for each parameter computed with
596
the ten ensembles considering the best configurations found for each method. For
597
comparisons, the second column of Table 12 shows the SNV obtained by the refer-
598
ence case. Clearly the standard ES-MDA without localization or inflation resulted
599
in underestimation of posterior variance for all parameters. Overall the methods
600
investigated in this paper minimized this problem, but in different levels and, in
601
some cases, the posterior variances were overestimated. Among the parameters, the
602
permeability multipliers are the most influential in the history matching. We observe
603
a severe variance reduction for these parameters in the reference case, with excep-
604
tion for the multiplier of the second layer, in which case the reference SNV is 0.16.
605
The same behaviour was obtained with all methods investigated. Fig. 8 presents
606
boxplots for the parameter EXP MULTIPERM5 showing that all methods resulted
607
in narrow posterior distributions very close to the correct value for the parameter.
608
The methods S-SENCORP and DES-MDA results in slightly overestimated poste-
609
rior distributions for this parameter, but the results seem acceptable. Note that the
610
scale of the boxplots showing the prior is different from the boxplots showing the 25
611
posterior.
612
The results in Table 12 show a different behaviour for the porosity multipliers.
613
The predicted data are clearly less sensitive to these parameters resulting in smaller
614
reductions in the posterior variance in the reference case. Among all methods, co-
615
variance inflation obtained the values of SNV closest to the reference; POL, SA and
616
DES-MDA also obtained improvements compared to the case without localization.
617
The other methods presented a tendency to overestimate the posterior variances for
618
porosity multipliers. One exception, however, is for the porosity multiplier of the
619
second layer. This is the least influential actual parameter in the history match-
620
ing. The SNV of the reference case is 0.86. All methods underestimated this value.
621
Fig. 9 presents the boxplots for this parameter showing that significantly different
622
posterior distributions were obtained for this parameter when we repeated the data
623
assimilation with ten different priors. The two versions of the SENCORP method
624
presented a slightly better performance in this specific aspect. Unfortunately the
625
same methods overestimated the variances substantially for the remaining porosity
626
multipliers, which indicates an excessive localization. For the other actual parame-
627
ters, all methods resulted in reasonable improvements in term of SNV, although we
628
observe some overestimation, specially for the parameter EXP AQRADIUS2.
629
For the dummy parameters, all methods increased the SNV compared the case
630
with no localization. However, the methods POL, BL, both SENCORP implemen-
631
tations and DES-MDA still underestimated the reference value. Covariance inflation
632
resulted in an slightly overestimation, while SA was able to correctly identify that
633
these parameters have no influence in the prediction. CBL resulted is severe overes-
634
timation of the SNV for the dummy parameters. It is worth noting that the results
635
of CBL methods were obtained with a single threshold value (same value for actual
636
and dummy parameters). This value was selected based on the results presented in
637
Fig. 5. However, as noted by (Luo et al., 2018), one could use different truncation
638
thresholds for different types of parameters. For example, based on Fig. 5, we could
639
select θ = 0.05 for the actual parameters and θ ≥ 0.3 for the dummy. In this case,
640
the method would correctly remove (localize) the updates in all dummy parame-
641
ters resulting in SNV = 1. Of course, tuning the method can be very expensive in 26
642
practice and we typically do not known the correct level of variance for comparisons.
643
Fig. 10 shows the boxplots of predicted increment in cumulative oil production,
644
∆Np , for a total production period of 16.5 years. Besides the six original wells,
645
the values of ∆Np include the production of five infill-drilling wells (wells X1 to
646
X5 in Fig. 2). This figure shows the case without localization underestimates the
647
uncertainty in ∆Np . All methods considered in this paper increased the uncertainty
648
range in ∆Np , but in some cases the range was overestimated. In particular, we
649
note that BL resulted in poor estimation of the distribution of ∆Np for two out
650
of ten ensembles. These two ensemble correspond to the ones with failure in the
651
data match, as indicated in Table 4. S-SENCORP was the method with the poorest
652
estimation of the posterior distribution of ∆Np . The range between the 25th and
653
75th percentiles (size of the boxes) is clearly overestimated and the medians (red
654
line in each box) are biased towards smaller values of ∆Np .
655
Fig. 11 shows the Kalman gain for two parameters EXP MULTPERM5 and
656
DUMMY3 with respect to BHP at well PRO-4 and WPR at well PRO-11, respec-
657
tively. These parameters and wells were selected to illustrate some aspects of the
658
behaviour of the methods. In both cases, we show the Kalman gain obtained with
659
the reference (Ne = 50000) and the first ensemble (Ne = 100) for each method.
660
First, we note in Fig. 11a that the Kalman gain predicted with the ensemble of
661
Ne = 100 is systematically smaller than the reference. In this case, localization
662
cannot improve the Kalman gain. In fact, the best in this case would be to use a lo-
663
calization coefficient of one. Among the methods tested, we note that POL, BL and
664
both SENCORP implementations reduced even more the Kalman gain. This effect
665
was more pronounced for BL and S-SENCORP. SA and CBL were able to correctly
666
identify that the localization coefficients should be one. Note that Fig. 11 does not
667
include the case with covariance inflation nor DES-MDA because both methods use
668
the same Kalman gain of the case without localization. The case POL+SA is not
669
shown but it is essentially the same as POL alone. Fig. 11b shows the same type of
670
plot but for a DUMMY parameter. Ideally, we would like to obtain a Kalman gain
671
of zero for this case. It is interesting to observe that even an ensemble with 50000
672
members is not enough to completely eliminate spurious correlations and some small 27
673
nonzero values of the Kalman gain were obtained. This figure shows that all meth-
674
ods reduced the Kalman gain. For both SENCORP implementations and SA the
675
localized Kalman gain was exactly zero. CBL removed most of the spurious Kalman
676
gain, but failed to remove the values between the time steps 70 and 80.
677
5. Conclusions
678
Sampling errors introduced by the limited ensemble size are one of the main limi-
679
tations of ensemble data assimilation in history-matching applications. For gridblock
680
properties, such as porosity and permeability, distance-based localization is the stan-
681
dard method to address this limitation. However, there is very little discussion in
682
the literature on how to mitigate sampling errors for non-local parameters. In con-
683
trast, some of these parameters, such as relative permeability curves and property
684
multipliers are among the most used in real-life history-matching cases. This paper
685
addressed this problem presenting a systematic comparison among the main methods
686
in the literature to alleviate the negative effects of sampling errors. These meth-
687
ods were compared in a synthetic problem and based on the results the following
688
conclusions can be stated:
689
• Sampling errors caused significant reduction in the posterior ensemble variance
690
in the test case evidencing the need of some strategy to address this problem.
691
• Overall all methods investigated in this paper were able to reduce the vari-
692
ance underestimation, but in different levels and, in some cases, at a cost of
693
compromising the ability of matching data properly.
694
• Most of these methods require the selection of internal parameters with impor-
695
tant impact in the performance of the method. In some cases, these parame-
696
ters can be difficult to select, which may limit the application of the method
697
in large-scale problems where repeating the data assimilation multiple times
698
is not feasible.
699
• POL showed a consistent improvement in the data assimilation results com-
700
pared to the case without localization or inflation. The method seems relatively 28
701
robust with respect to the choice of the truncation threshold as long as the se-
702
lected value is not too large. Our tests indicate that ≤ 10−2 is a good choice.
703
The method is simple to implement adding no significant computational cost
704
to the data assimilation.
705
• BL also showed improvements in preserving the posterior variances, but the
706
method seems to be less robust as we observed failure to match data in some
707
ensembles. The method is relatively easy to tune.
708
• Both SENCORP implementations showed a tendency of introducing exces-
709
sive dumping (localization) in the Kalman gain resulting in a tendency to
710
overestimate the posterior variances. The original SENCORP was the most
711
difficult method to tune. We observed a strong impact of the choice of the
712
tuning parameters (integer exponents) of the method. The simplified version
713
of the method is easier to tune, but its performance was inferior to the original
714
method.
715
• CBL also showed to be very dependent on the choice of the tuning parameter
716
(truncation level in the correlations). The results indicate that the method is
717
not robust as we observed significant variations in the results when we repeated
718
the data assimilation with different prior ensembles.
719
• Covariance inflation was the method with best performance in the test prob-
720
lem. The method is very simple to apply adding no relevant computational
721
cost in the data assimilation. Even though the selection of the inflation fac-
722
tor impacts the performance of the method, the adaptive scheme of (Evensen,
723
2009) worked remarkably well is our test case.
724
• DES-MDA improved the results of standard ES-MDA in terms of preserving
725
variance without compromising the data match. However, the method was not
726
enough to avoid a significant level of variance underestimation.
727
• We introduced a simple sensitivity analysis to select localization coefficients for
728
non-local parameters. The procedure is effective to localize weakly sensitive
729
data and can be used in conjunct with other localization schemes. 29
730
731
Acknowledgement The authors would like to thank Petrobras for supporting this research and for
732
the permission to publish this paper.
733
Aanonsen, S. I., Nævdal, G., Oliver, D. S., Reynolds, A. C., and Vall`es, B. Review
734
of ensemble Kalman filter in petroleum engineering. SPE Journal, 14(3):393–412,
735
2009. doi:10.2118/117274-PA.
736
Anderson, J. L.
An ensemble adjustment Kalman filter for data assimila-
737
tion.
Monthly Weather Review, 129(12):2884–2903, 2001.
738
0493(2001)1292.0.CO;2.
doi:10.1175/1520-
739
Anderson, J. L. Exploring the need for localization in ensemble data assimilation
740
using a hierarchical ensemble filter. Physica D: Nonlinear Phenomena, 230(1–2):
741
99–111, 2007a. doi:10.1016/j.physd.2006.02.011.
742
Anderson, J. L.
An adaptive covariance inflation error correction algorithm
743
for ensemble filters.
744
0870.2006.00216.x.
745
Tellus A, 59(2):210–224, 2007b.
Anderson, J. L. and Anderson, S. L.
doi:10.1111/j.1600-
A Monte Carlo implementation of
746
the nonlinear filtering problem to produce ensemble assimilations and fore-
747
casts. Monthly Weather Review, 127(12):2741–2758, 1999. doi:10.1175/1520-
748
0493(1999)1272.0.CO;2.
749
Bannister, R. N. How is the balance of a forecast ensemble affected by adaptive
750
and nonadaptive localization schemes? Monthly Weather Review, 143:3680–3699,
751
2015. doi:10.1175/MWR-D-14-00379.1.
752
Bianco, A., Cominelli, A., Dovera, L., Nævdal, G., and Vall`es, B. History match-
753
ing and production forecast uncertainty by means of the ensemble Kalman fil-
754
ter: A real field application. In Proceedings of the EAGE/EUROPEC Confer-
755
ence and Exhibition, London, U.K., 11–14 June, number SPE-107161-MS, 2007.
756
doi:10.2118/107161-MS.
30
757
Bishop, C. H. and Hodyss, D. Flow-adaptive moderation of spurious ensemble
758
correlations and its use in ensemble-based data assimilation. Quarterly Journal of
759
the Royal Meteorological Society, 133(629):2029–2044, 2007. doi:10.1002/qj.169.
760
Chen, Y. and Oliver, D. S. Ensemble-based closed-loop optimization applied to
761
Brugge field.
SPE Reservoir Evaluation & Engineering, 13(1):56–71, 2010a.
762
doi:10.2118/118926-PA.
763
Chen, Y. and Oliver, D. S. Cross-covariance and localization for EnKF in multi-
764
phase flow data assimilation. Computational Geosciences, 14(4):579–601, 2010b.
765
doi:10.1007/s10596-009-9174-6.
766
Chen, Y. and Oliver, D. S. Ensemble randomized maximum likelihood method
767
as an iterative ensemble smoother. Mathematical Geosciences, 44(1):1–26, 2012.
768
doi:10.1007/s11004-011-9376-z.
769
Chen, Y. and Oliver, D. S. Levenberg-Marquardt forms of the iterative ensemble
770
smoother for efficient history matching and uncertainty quantification. Computa-
771
tional Geosciences, 17:689–703, 2013. doi:10.1007/s10596-013-9351-5.
772
Chen, Y. and Oliver, D. S. History matching of the Norne full-field model with
773
an iterative ensemble smoother. SPE Reservoir Evaluation & Engineering, 17(2),
774
2014. doi:10.2118/164902-PA.
775
Cominelli, A., Dovera, L., Vimercati, S., and Nævdal, G. Benchmark study of
776
ensemble Kalman filter methodology: History matching and uncertainty quantifi-
777
cation for a deep-water oil reservoir. In Proceedings of the International Petroleum
778
Technology Conference, Doha, Qatar, 7–9 December, number IPTC 13748, 2009.
779
doi:10.2523/13748-MS.
780
781
Emerick, A. A. History Matching and Uncertainty Characterization Using Ensemblebased Methods. PhD thesis, The University of Tulsa, 2012.
782
Emerick, A. A. Analysis of the performance of ensemble-based assimilation of pro-
783
duction and seismic data. Journal of Petroleum Science and Engineering, 139:
784
219–239, 2016. doi:10.1016/j.petrol.2016.01.029. 31
785
Emerick, A. A. Deterministic ensemble smoother with multiple data assimilation
786
as an alternative for history matching seismic data. Computational Geosciences,
787
Online First, 2018. doi:10.1007/s10596-018-9745-5.
788
Emerick, A. A. and Reynolds, A. C. History matching a field case using the en-
789
semble Kalman filter with covariance localization. SPE Reservoir Evaluation &
790
Engineering, 14(4):423–432, 2011a. doi:10.2118/141216-PA.
791
Emerick, A. A. and Reynolds, A. C. Combining sensitivities and prior information for
792
covariance localization in the ensemble Kalman filter for petroleum reservoir appli-
793
cations. Computational Geosciences, 15(2):251–269, 2011b. doi:10.1007/s10596-
794
010-9198-y.
795
Emerick, A. A. and Reynolds, A. C. Ensemble smoother with multiple data assimila-
796
tion. Computers & Geosciences, 55:3–15, 2013a. doi:10.1016/j.cageo.2012.03.011.
797
Emerick, A. A. and Reynolds, A. C. History matching of production and seismic data
798
for a real field case using the ensemble smoother with multiple data assimilation. In
799
Proceedings of the SPE Reservoir Simulation Symposium, The Woodlands, Texas,
800
USA, 18–20 February, number SPE-163675-MS, 2013b. doi:10.2118/163675-MS.
801
802
803
Evensen, G. Sampling strategies and square root analysis schemes for the EnKF. Ocean Dynamics, 54(6):539–560, 2004. doi:10.1007/s10236-004-0099-2. Evensen, G.
The ensemble Kalman filter for combined state and param-
804
eter estimation.
IEEE Control Systems Magazine, pages 83–104, 2009.
805
doi:10.1109/MCS.2009.932223.
806
Evensen, G., Hove, J., Meisingset, H. C., Reiso, E., Seim, K. S., and Espelid, Ø.
807
Using the EnKF for assisted history matching of a North Sea reservoir model. In
808
Proceedings of the SPE Reservoir Simulation Symposium, Houston, Texas, 26–28
809
February, number SPE-106184-MS, 2007. doi:10.2118/106184-MS.
810
Floris, F. J. T., Bush, M. D., Cuypers, M., Roggero, F., and Syversveen, A. R.
811
Methods for quantifying the uncertainty of production forecasts: A comparative
812
study. Petroleum Geoscience, 7(SUPP):87–96, 2001. doi:10.1144/petgeo.7.S.S87. 32
813
Furrer, R. and Bengtsson, T. Estimation of high-dimensional prior and posterior
814
covariance matrices in Kalman filter variants. Journal of Multivariate Analysis,
815
98(2):227–255, 2007. doi:10.1016/j.jmva.2006.08.003.
816
Hamill, T. M., Whitaker, J. S., and Snyder, C.
Distance-dependent filter-
817
ing of background error covariance estimates in an ensemble Kalman fil-
818
ter.
819
0493(2001)1292.0.CO;2.
Monthly Weather Review, 129(11):2776–2790, 2001.
doi:10.1175/1520-
820
Haugen, V., Nævdal, G., Natvik, L.-J., Evensen, G., Berg, A. M., and Flornes,
821
K. M. History matching using the ensemble Kalman filter on a North Sea field
822
case. SPE Journal, 13(4):382–391, 2008. doi:10.2118/102430-PA.
823
Houtekamer, P. L. and Mitchell, H. L.
Data assimilation using an ensem-
824
ble Kalman filter technique. Monthly Weather Review, 126(3):796–811, 1998.
825
doi:10.1175/1520-0493(1998)1262.0.CO;2.
826
Houtekamer, P. L. and Mitchell, H. L. A sequential ensemble Kalman filter for
827
atmospheric data assimilation. Monthly Weather Review, 129(1):123–137, 2001.
828
doi:10.1175/1520-0493(2001)1292.0.CO;2.
829
Kotsuki, S., Ota, Y., and Miyoshi, T. Adaptive covariance relaxation methods for
830
ensemble data assimilation: experiments in the real atmosphere. Quarterly Jour-
831
nal of the Royal Meteorological Society, 143:2001–2015, 2015. doi:10.1002/qj.3060.
832
Li, H., Kalnay, E., and Miyoshi, T. Simultaneous estimation of covariance inflation
833
and observation errors within ensemble Kalman filter. Quarterly Journal of the
834
Royal Meteorological Society, 135(639):523–533, 2009. doi:10.1002/qj.371.
835
Le, D.H., Emerick, A.A., and Reynolds, A.C. An adaptive ensemble smoother with
836
multiple data assimilation for assisted history matching. SPE Journal, 21(6):
837
2–195, 2016. doi:10.2118/173214-PA.
838
Luo, X., Bhakta, T., and Nævdal, G.
Correlation-based adaptive localization
839
with applications to ensemble-based 4D-seismic history matching. SPE Journal,
840
Preprint, 2018. doi:10.2118/185936-PA. 33
841
Luo, X., Lorentzen, R., Valestrand, R. and Evensen, G. Correlation-based adap-
842
tive localization for ensemble-based history matching: Applied to the Norne
843
field case study. S PE Norway One Day Seminar, number SPE-191305, 2018.
844
doi:10.2118/191305-MS.
845
Luo, X., Stordal, A., Lorentzen, R. and Nævdal, G. Iterative ensemble smoother as
846
an approximate solution to a regularized minimum-average-cost problem: theory
847
and applications. SPE Journal, 20(5):962–982, 2016. doi:10.2118/173214-PA.
848
Ma, X., Hetz, G., Wang, X., Bi, L., Stern, D. and Hoda, N. A robust iterative
849
ensemble smoother method for efficient history matching and uncertainty quan-
850
tification. S PE Reservoir Simulation Conference, number SPE-182693-MS, 2017.
851
doi:10.2118/182693-MS.
852
Oliver, D. S., Reynolds, A. C., and Liu, N. Inverse Theory for Petroleum Reservoir
853
Characterization and History Matching. Cambridge University Press, Cambridge,
854
UK, 2008.
855
Peters, L., Arts, R., Brouwer, G., Geel, C., Cullick, S., Lorentzen, R. J., Chen, Y.,
856
Dunlop, N., Vossepoel, F. C., Xu, R., Sarma, P., Alhuthali, A. H., and Reynolds,
857
A. Results of the Brugge benchmark study for flooding optimisation and his-
858
tory matching. SPE Reservoir Evaluation & Engineering, 13(3):391–405, 2010.
859
doi:10.2118/119094-PA.
860
Raanes, P. N., Bocquet, M., and Carrassi, A. Adaptive covariance inflation in the
861
ensemble Kalman filter by Gaussian scale mixtures. arXiv, 2018. URL https:
862
//arxiv.org/abs/1801.08474.
863
Sakov, P. and Bertino, L. Relation between two common localisation methods for
864
the EnKF. Computational Geosciences, 15(2):225–237, 2011. doi:10.1007/s10596-
865
010-9202-6.
866
Sakov, P. and Oke, P. R. Implications of the form of the ensemble transformation in
867
the ensemble square root filters. Monthly Weather Review, 136:1042–1053, 2008a.
868
doi:10.1175/2007MWR2021.1. 34
869
Sakov, P. and Oke, P. R. A deterministic formulation of the ensemble Kalman filter:
870
an alternative to ensemble square root filters. Tellus A, 60(2):361–371, 2008b.
871
doi:10.1111/j.1600-0870.2007.00299.x.
872
Skjervheim, J.-A., Evensen, G., Aanonsen, S. I., Ruud, B. O., and Johansen, T.-
873
A. Incorporating 4D seismic data in reservoir simulation models using ensemble
874
Kalman filter. SPE Journal, 12(3):282–292, 2007. doi:10.2118/95789-PA.
875
Stordal, A. S. and Elsheikh, A. H. Iterative ensemble smoothers in the annealed
876
importance sampling framework. Advances in Water Resources, 86:231–239, 2015.
877
doi:10.1016/j.advwatres.2015.09.030.
878
879
Tarantola, A. Inverse Problem Theory and Methods for Model Parameter Estimation. SIAM, Philadelphia, USA, 2005.
880
Tippett, M. K., Anderson, J. L., Bishop, C. H., Hamill, T. M., and Whitaker,
881
J. S. Ensemble square-root filters. Monthly Weather Review, 131:1485–1490,
882
2003. doi:10.1175/1520-0493(2003)1312.0.CO;2.
883
Vall`es, B. and Nævdal, G. Revisiting Brugge case study using a hierarchical en-
884
semble Kalman filter. In Proceedings of the International Petroleum Technol-
885
ogy Conference, Doha, Qatar, 7–9 December, number IPTC-14074-MS, 2009.
886
doi:10.2523/IPTC-14074-MS.
887
van Leeuwen, P. J. Comment on “Data assimilation using an ensemble Kalman filter
888
technique”. Monthly Weather Review, 127(6):1374–1377, 1999. doi:10.1175/1520-
889
0493(1999)1272.0.CO;2.
890
Wang, Y., Li, G., and Reynolds, A. C. Estimation of depths of fluid contacts by
891
history matching using iterative ensemble-Kalman smoothers. SPE Journal, 15
892
(2), 2010. doi:10.2118/119056-PA.
893
Whitaker, J. S. and Hamill, T. M. Ensemble data assimilation without perturbed ob-
894
servations. Monthly Weather Review, 130(7):1913–1924, 2002. doi:10.1175/1520-
895
0493(2002)1302.0.CO;2.
35
896
Zhang, Y. and Oliver, D. S. Improving the ensemble estimate of the Kalman
897
gain by bootstrap sampling.
898
doi:10.1007/s11004-010-9267-8.
Mathematical Geosciences, 42:327–345, 2010.
899
Zhang, Y. and Oliver, D. S. History matching using the ensemble Kalman filter
900
with multiscale parameterization: A field case study. SPE Journal, 16(2):307–
901
317, 2011. doi:10.2118/118879-PA.
36
Table 1: Prior distributions of the actual and dummy parameters
Parameter
Mean
True
Distribution
Min.
Max.
MULTPOR1
1.00
1.00
Triangle
0.50
1.50
MULTPOR2
1.00
1.00
Triangle
0.50
1.50
MULTPOR3
1.00
1.00
Triangle
0.50
1.50
MULTPOR4
1.00
1.00
Triangle
0.50
1.50
MULTPOR5
1.00
1.00
Triangle
0.50
1.50
EXP MULTIPERM1
0.00
0.00
Triangle
−1.00
1.00
EXP MULTIPERM2
0.00
0.00
Triangle
−1.00
1.00
EXP MULTIPERM3
0.00
0.00
Triangle
−1.00
1.00
EXP MULTIPERM4
0.00
0.00
Triangle
−1.00
1.00
EXP MULTIPERM5
0.00
0.00
Triangle
−1.00
1.00
CCPOR
5 × 10−6
5 × 10−6
Triangle
1 × 10−6
5 × 10−5
EXP AQRADIUS1
1.00
1.74
Triangle
0.00
2.00
EXP AQRADIUS2
1.00
1.75
Triangle
0.00
2.00
DELTA DWOC
0.00
1.00
Triangle
−3.00
3.00
DELTA DGOC
0.00
−1.00
Triangle
−3.00
3.00
DUMMY1
0.00
-
Normal
−4.00
4.00
DUMMY2
0.00
-
Normal
−4.00
4.00
DUMMY3
0.00
-
Normal
−4.00
4.00
DUMMY4
0.00
-
Normal
−4.00
4.00
DUMMY5
0.00
-
Normal
−4.00
4.00
37
Table 2: Results for the data assimilation without localization and inflation Ensemble
Od (m) (prior)
Od (m) (post)
RMSE
SNV (actual)
SNV (dummy)
1
551.7
0.3
0.97
0.11
0.44
2
589.2
0.4
1.21
0.10
0.46
3
719.9
0.4
1.19
0.11
0.45
4
645.9
0.4
0.96
0.10
0.43
5
802.6
0.4
1.25
0.09
0.43
6
613.9
0.3
1.39
0.10
0.45
7
721.2
0.5
1.36
0.11
0.41
8
1148.1
0.3
0.99
0.09
0.42
9
1236.7
0.3
1.20
0.10
0.39
10
810.4
0.7
1.32
0.11
0.39
Mean
784.0
0.4
1.19
0.10
0.43
Reference
695.6
0.8
0.89
0.23
1.00
38
Table 3: Results for the POL method = 10−3 Ensemble
Od (m)
RMSE
(post)
= 10−1
SNV
SNV
Od (m)
(actual)
(dummy)
(post)
RMSE
SNV
SNV
(actual)
(dummy)
1
0.4
0.87
0.20
0.79
0.5
0.82
0.27
1.08
2
0.6
0.71
0.22
0.80
0.4
0.64
0.30
0.89
3
0.5
0.74
0.22
0.77
0.9
0.64
0.30
0.97
4
0.4
0.70
0.23
0.73
0.5
0.76
0.34
0.82
5
0.6
1.06
0.24
0.87
2.1
1.02
0.35
1.20
6
0.4
0.98
0.26
0.70
0.4
0.91
0.33
0.89
7
0.5
0.93
0.26
0.78
0.5
0.81
0.36
0.92
8
0.4
0.81
0.27
0.74
6.1
0.77
0.40
0.85
9
0.5
1.07
0.21
0.80
0.5
0.99
0.29
1.02
10
0.6
0.80
0.27
0.78
0.7
0.80
0.41
1.16
Mean
0.5
0.87
0.24
0.77
1.3
0.82
0.34
0.98
Reference
0.8
0.89
0.23
1.00
0.8
0.89
0.23
1.00
39
Table 4: Results for the BL method Ensemble
Od (m) (post)
RMSE
SNV (actual)
SNV (dummy)
1
1.1
0.93
0.27
0.98
2
0.8
0.78
0.28
0.71
3
0.6
0.78
0.22
0.72
4
0.5
1.00
0.24
0.74
5
3.1
0.80
0.40
1.18
6
0.4
1.04
0.17
0.67
7
0.9
1.06
0.26
0.73
8
0.5
0.78
0.22
0.82
9
0.7
1.24
0.23
0.90
10
25.3
0.71
0.41
1.09
Mean
3.4
0.91
0.27
0.85
Reference
0.8
0.89
0.23
1.00
40
Table 5: Results for the SENCORP method Ensemble
Od (m) (post)
RMSE
SNV (actual)
SNV (dummy)
1
0.5
0.89
0.26
0.73
2
0.4
0.61
0.26
0.79
3
3.8
0.74
0.26
0.80
4
0.5
0.64
0.27
0.75
5
0.5
0.75
0.27
0.78
6
0.5
0.81
0.29
0.72
7
0.5
0.86
0.28
0.74
8
0.5
0.69
0.28
0.71
9
0.4
0.75
0.26
0.78
10
0.4
0.67
0.27
0.76
Mean
0.8
0.74
0.27
0.76
Reference
0.8
0.89
0.23
1.00
41
Table 6: Results for the S-SENCORP method Ensemble
Od (m) (post)
RMSE
SNV (actual)
SNV (dummy)
1
1.0
0.79
0.44
0.94
2
1.3
0.61
0.46
0.94
3
0.9
0.68
0.48
0.96
4
1.7
0.72
0.46
0.92
5
1.3
0.71
0.48
0.95
6
1.7
0.78
0.49
0.92
7
1.2
0.79
0.46
0.93
8
2.4
0.76
0.49
0.92
9
1.7
0.78
0.48
0.96
10
2.4
0.74
0.48
0.94
Mean
1.6
0.74
0.47
0.94
Reference
0.8
0.89
0.23
1.00
42
Table 7: Results for the CBL method (θ = 0.05) Ensemble
Od (m) (post)
RMSE
SNV (actual)
SNV (dummy)
1
0.7
0.97
0.20
1.08
2
5.9
0.79
0.21
0.98
3
0.6
0.89
0.26
1.17
4
0.7
0.93
0.22
1.09
5
1.6
1.00
0.22
1.50
6
0.7
0.97
0.29
1.18
7
0.6
0.63
0.39
1.82
8
0.4
0.86
0.29
1.10
9
0.4
1.05
0.18
1.60
10
0.5
1.04
0.26
1.05
Mean
1.2
0.91
0.25
1.26
Reference
0.8
0.89
0.23
1.00
43
Table 8: Results for the covariance inflation Ensemble
Od (m) (post)
RMSE
SNV (actual)
SNV (dummy)
1
0.4
0.88
0.23
1.00
2
0.4
0.85
0.20
1.01
3
0.5
0.73
0.25
1.05
4
0.6
0.82
0.19
1.06
5
0.4
0.90
0.22
1.11
6
0.4
1.01
0.22
0.99
7
0.7
1.05
0.24
0.91
8
0.6
0.64
0.22
0.98
9
0.6
0.81
0.24
0.94
10
0.7
0.85
0.26
1.10
Mean
0.5
0.85
0.23
1.02
Reference
0.8
0.89
0.23
1.00
44
Table 9: Results for DES-MDA Ensemble
Od (m) (post)
RMSE
SNV (actual)
SNV (dummy)
1
0.5
0.83
0.19
0.51
2
0.8
0.88
0.17
0.51
3
0.8
0.87
0.20
0.51
4
0.8
0.78
0.19
0.52
5
1.2
1.04
0.19
0.53
6
1.1
0.97
0.20
0.48
7
0.8
1.03
0.19
0.45
8
0.6
0.77
0.19
0.47
9
0.5
1.02
0.17
0.51
10
1.0
0.98
0.18
0.48
Mean
0.8
0.92
0.19
0.50
Reference
0.8
0.89
0.23
1.00
45
Table 10: Results for the SA localization (θ = 10−1 ) Ensemble
Od (m) (post)
RMSE
SNV (actual)
SNV (dummy)
1
0.4
0.85
0.17
1.00
2
0.4
0.93
0.16
1.00
3
0.5
1.06
0.19
1.00
4
0.5
0.97
0.18
1.00
5
0.4
1.13
0.17
1.00
6
4.7
0.71
0.29
1.00
7
0.6
1.14
0.20
1.00
8
0.4
0.94
0.17
1.00
9
0.5
1.07
0.24
1.00
10
0.6
1.18
0.19
1.00
Mean
0.9
1.00
0.20
1.00
Reference
0.8
0.89
0.23
1.00
46
Table 11: Results for the SA localization (θ = 10−3 ) combined with POL method ( = 10−4 ) Ensemble
Od (m) (post)
RMSE
SNV (actual)
SNV (dummy)
1
0.4
0.84
0.20
1.00
2
0.8
0.76
0.22
1.00
3
0.6
0.78
0.24
1.00
4
0.4
0.61
0.23
1.00
5
0.6
1.08
0.23
1.00
6
0.4
1.01
0.28
1.00
7
0.5
0.84
0.28
1.00
8
0.4
0.77
0.29
1.00
9
0.4
1.15
0.22
1.00
10
0.6
0.88
0.25
1.00
Mean
0.5
0.87
0.24
1.00
Reference
0.8
0.89
0.23
1.00
47
48
1.00
1.00
DUMMY5
1.00
DUMMY2
1.00
1.00
DUMMY1
DUMMY3
0.12
DELTA DGOC
DUMMY4
0.37
0.20
DELTA DWOC
0.20
EXP AQRADIUS2
0.09
0.14
CCPOR
EXP AQRADIUS1
0.39
0.44
0.42
0.43
0.44
0.06
0.10
0.22
0.10
0.00
0.00
0.00
0.07
0.00
0.00
EXP MULTIPERM3
0.00
0.16
EXP MULTIPERM2
0.01
EXP MULTIPERM5
0.01
EXP MULTIPERM1
0.10
0.09
0.17
0.36
0.20
No local.
EXP MULTIPERM4
0.21
0.25
MULTPOR4
MULTPOR5
0.86
0.43
MULTPOR2
0.42
MULTPOR1
MULTPOR3
Reference
Parameter
0.79
0.84
0.73
0.77
0.73
0.14
0.16
0.45
0.14
0.20
0.00
0.00
0.00
0.20
0.01
0.33
0.27
0.53
0.62
0.52
POL
0.85
0.90
0.87
0.80
0.86
0.17
0.21
0.43
0.25
0.20
0.00
0.00
0.00
0.29
0.04
0.33
0.43
0.53
0.64
0.52
BL
0.74
0.76
0.75
0.78
0.76
0.15
0.21
0.51
0.17
0.24
0.00
0.00
0.00
0.27
0.02
0.36
0.31
0.54
0.70
0.55
SENCORP
0.93
0.95
0.94
0.94
0.93
0.37
0.37
0.86
0.29
0.45
0.01
0.00
0.01
0.66
0.08
0.64
0.68
0.85
0.91
0.88
S-SENCORP
Table 12: Mean SNV
1.19
1.38
1.68
1.10
0.93
0.13
0.14
0.45
0.13
0.18
0.00
0.00
0.00
0.22
0.01
0.30
0.32
0.61
0.65
0.63
CBL
0.97
1.04
1.06
1.05
0.96
0.12
0.19
0.48
0.19
0.15
0.00
0.00
0.00
0.18
0.01
0.25
0.20
0.43
0.72
0.47
CI
0.50
0.51
0.51
0.48
0.49
0.10
0.16
0.34
0.20
0.14
0.01
0.01
0.01
0.18
0.04
0.20
0.25
0.35
0.37
0.28
DES-MDA
1.00
1.00
1.00
1.00
1.00
0.08
0.21
0.48
0.13
0.14
0.00
0.00
0.00
0.14
0.01
0.27
0.25
0.41
0.54
0.31
SA
1.00
1.00
1.00
1.00
1.00
0.14
0.17
0.48
0.14
0.20
0.00
0.00
0.00
0.22
0.01
0.30
0.28
0.56
0.63
0.54
POL+SA
0.6
Relative Frequency
0.5
U = 0 (std. dev. = 0.157) U = 0.8 (std. dev. = 0.040)
0.4 0.3 0.2 0.1 0 -0.4
-0.2
0
0.2
0.4
0.6
0.8
1
Correlation Coefficient Figure 1: Histograms of estimated correlation coefficient for two cases: uncorrelated variables (ρ = 0) and strongly correlated variables (ρ = 0.8).
49
Figure 2: PUNQ-S3 model (Floris et al., 2001). The red contour indicates the position of the gas-oil contact and the blue contour the position of the oil-water contact. The black circles shows the oil producing wells operating during the historical period. The white circles shows the positions of infill-drilling wells.
50
1.2 0.4
1.1 SNV(dummy)
SNV(actual)
0.35
0.3
0.25
1 0.9 0.8 0.7
0.2 0
0.001
0.01
0
0.1
ε
0.001
0.01
0.1
ε
(a) Actual
(b) Dummy
Figure 3: Boxplot of SNV obtained by POL with different values of the threshold.
51
4
3
3
2
2
2
1
1
1
0
DUMMY5
4
3
DUMMY5
DUMMY5
4
0
0
−1
−1
−1
−2
−2
−2
−3
−3
−3
ρ=−0.018082 −4 1000
1500
2000
ρ=0.24486
2500 3000 BHP PRO−4
3500
4000
−4 1800
2200
2400 2600 BHP PRO−4
ρ=−0.20896 2800
3000
4
4
3
3
2
2
1
1
0
−4 2400
2450
−1
−2
−2
−3
2500
2550 2600 2650 BHP PRO−4
2700
2750
(c) Iteration #2
0
−1
−3 ρ=0.0012971
−4 2400
3200
(b) Iteration #1
DUMMY5
DUMMY5
(a) Prior ensemble
2000
2450
2500
2550 2600 2650 BHP PRO−4
ρ=0.055806 2700
2750
2800
(d) Iteration #3
−4 2400
2450
2500
2550 2600 2650 BHP PRO−4
2700
2750
2800
(e) Iteration #4
Figure 4: Correlations between the parameter DUMMY5 and the predicted BHP for welll PRO-4 at time 2921 days.
52
2800
0.6
2
SNV(dummy)
SNV(actual)
0.5 0.4 0.3
1.5
1
0.2 0.5
0.1 0.01
0.05
0.10
0.15
0.20
0.25
0.30
0.01
θ
0.05
0.10
0.15
0.20
0.25
0.30
θ
(a) Actual
(b) Dummy
Figure 5: Boxplot of SNV obtained by CBL with different values of the threshold.
53
1.35
1.3
1.25
γ
1.2
1.15
1.1
1.05
1 1
2 3 ES−MDA iteration
Figure 6: Covariance inflation factors.
54
4
0.7
SNV(actual)
0.6 0.5 0.4 0.3 0.2 0.1 0.001
0.01
0.10
1.0
θ
Figure 7: Boxplot of SNV for the actual model parameters obtained with SA localization with different values of θ.
55
1
0.2
0.2
0.5
0.1
0.1
0
0
0
−0.5
−0.1
−0.1
−1
R
1
2
3
4
5
6
7
8
−0.2
9 10
R
(a) Prior
1
2
3
4
5
6
7
8
−0.2
9 10
0.2
0.1
0.1
0.1
0
0
0
−0.1
−0.1
−0.1
1
2
3
4
5
6
7
8
−0.2
9 10
R
1
(d) BL
2
3
4
5
6
7
8
−0.2
9 10
0.2
0.1
0.1
0.1
0
0
0
−0.1
−0.1
−0.1
1
2
3
4
5
6
7
8
−0.2
9 10
(g) CBL
R
1
2
3
4
5
1
0.2
0.1
0.1
0
0
−0.1
−0.1
R
7
8
−0.2
9 10
R
1
(h) Covariance inflation
0.2
−0.2
6
1
2
3
4
5
6
7
8
9 10
−0.2
(j) SA
R
1
3
4
5
6
7
8
9 10
2
3
4
5
6
7
8
9 10
8
9 10
(f) S-SENCORP
0.2
R
R
(e) SENCORP
0.2
−0.2
2
(c) POL
0.2
R
1
(b) No localization
0.2
−0.2
R
2
3
2
3
4
5
6
7
(i) DES-MDA
4
5
6
7
8
9 10
(k) POL+SA
Figure 8: Boxplots of EXP MULTIPERM5. The label R in each plot stands for reference case and the numbers represent each ensemble. The horizontal black line in each plot indicates the true value.
56
1.5
1.5
1.5
1
1
1
0.5
R
1
2
3
4
5
6
7
8
0.5
9 10
R
(a) Prior
1
2
3
4
5
6
7
8
0.5
9 10
1.5
1
1
1
1
2
3
4
5
6
7
8
0.5
9 10
R
1
(d) BL
2
3
4
5
6
7
8
0.5
9 10
1.5
1
1
1
1
2
3
4
5
6
7
8
0.5
9 10
R
1
2
3
4
5
6
7
8
(h) Covariance inflation
1.5
1.5
1
1
R
1
2
3
4
5
6
7
8
9 10
0.5
(j) SA
R
1
2
0.5
9 10
(g) CBL
0.5
3
4
5
6
7
8
9 10
1
2
3
4
5
6
7
8
9 10
8
9 10
(f) S-SENCORP
1.5
R
R
(e) SENCORP
1.5
0.5
2
(c) POL
1.5
R
1
(b) No localization
1.5
0.5
R
3
R
1
2
3
4
5
6
7
(i) DES-MDA
4
5
6
7
8
9 10
(k) POL+SA
Figure 9: Boxplots of MULTPOR2. The label R in each plot stands for reference case and the numbers represent each ensemble. The horizontal black line in each plot indicates the true value.
57
7
4
7
x 10
2.8
7
x 10
2.8
3.5
2.7
2.7
3
2.6
2.6
2.5
2.5
2.5
2
2.4
2.4
1.5
2.3
2.3
1
R
1
2
3
4
5
6
7
8
2.2
9 10
R
1
(a) Prior
3
4
5
6
7
8
2.2
9 10
7
x 10
2.8
2.8 2.7
2.6
2.6
2.6
2.5
2.5
2.5
2.4
2.4
2.4
2.3
2.3
2.3
1
2
3
4
5
6
7
8
2.2
9 10
R
1
(d) BL
2
3
4
5
6
7
8
2.2
9 10
2.8
2.8
2.7
2.7
2.6
2.6
2.6
2.5
2.5
2.5
2.4
2.4
2.4
2.3
2.3
2.3
R
1
2
3
4
5
6
7
8
2.2
9 10
(g) CBL
R
1
2
3
4
5
2.8 2.7
2.6
2.6
2.5
2.5
2.4
2.4
2.3
2.3 R
1
6
7
8
9 10
1
2
3
4
5
6
7
8
9 10
6
7
8
2.2
9 10
8
9 10
x 10
R
1
2
3
4
5
6
7
(i) DES-MDA
7
x 10
2.7
2.2
R
(h) Covariance inflation
7
2.8
5
7
x 10
2.7
2.2
4
(f) S-SENCORP
7
x 10
3
x 10
(e) SENCORP
7
2.8
2
7
x 10
2.7
R
1
(c) POL
2.7
2.2
R
(b) No localization
7
2.8
2
x 10
2
3
4
5
6
7
8
9 10
2.2
(j) SA
x 10
R
1
2
3
4
5
6
7
8
9 10
(k) POL+SA
Figure 10: Boxplots of ∆Np . The label R in each plot stands for reference case and the numbers represent each ensemble. The horizontal black line in each plot indicates the true value.
58
−6
x 10
Kalman Gain
5 4
−4
x 10 4
Reference No Localization POL BL SENCORP Simp. SENCORP CBL SA
Reference No Localization POL BL SENCORP Simp. SENCORP CBL SA
3 Kalman Gain
6
3 2
2 1 0 −1
1
−2
0 0
20
40
60 time
80
100
50
(a) EXP MULTPERMI5 and BHP of PRO-4
60
70
80 90 time
100
110
(b) DUMMY3 and WPR of PRO-11
Figure 11: Estimated Kalman gain for the first ensemble at the first iteration of ES-MDA.
59