Bayesian Multi-Source Modeling with Legacy Data. Sayan Ghoshâ, Isaac Asherâ , Jesper Kristensenâ, You Ling â , Kevin Ryan â, and Liping Wangâ¡. GE Global ...
AIAA SciTech Forum 8–12 January 2018, Kissimmee, Florida 2018 AIAA Non-Deterministic Approaches Conference
10.2514/6.2018-1663
Bayesian Multi-Source Modeling with Legacy Data Sayan Ghosh∗, Isaac Asher†, Jesper Kristensen∗, You Ling †, Kevin Ryan ∗, and Liping Wang‡ GE Global Research Center, 1 Research Circle, Niskayuna, New York. 12309, USA.
Downloaded by You Ling on January 12, 2018 | http://arc.aiaa.org | DOI: 10.2514/6.2018-1663
Abstract As the complexity of engineering systems increases, the cost of computer simulations, experimentation and physical tests for design and development also gets expensive. Due to the significant resources required by these physical and computational experiments on a new design, only sparse data are generally available and are often not sufficient enough to build accurate meta-models. A common approach to reduce the cost and the time cycle of a design process is to use meta-modeling techniques to build simpler and faster mathematical models to carry out the decision-making process for a new design. If datasets are available from legacy systems which belong to a similar family as the new design, one can leverage that information and knowledge to improve the accuracy of models built for the new system. In this work, a Bayesian multi-source modeling technique has been developed which combines models built on data from legacy systems with sparse data for a new design to improve the predictive capability of meta-model for the new design.
I.
Introduction
Engineering systems are becoming more and more complex as requirements (performance, environmental, cost, etc.) are becoming more stringent. This, has a direct impact on the time and cost to design new systems. A typical industrial approach to manage the time and cost, while achieving the performance goal is to develop a derivative design rather than a completely new and unconventional system. The benefit of designing a derivative design is that the engineers can leverage their knowledge and experience of the existing infrastructure. However, even a new derivative design comes with new technology, parameter settings, and operational conditions, which have not been tested in the legacy systems. Therefore, new experiments are required to understand and design the new derivative design. Due to the complexity and high experimental cost of the system, sparse data are generally available for the new designwhich alone are not sufficient to build an accurate model. To overcome this challenge, the goal of the current work is to use fewer data from the new design and leverage the data available from legacy systems to have a better predictive capability and better inform the design and decision-making process. One of the common approaches for handling multiple sources of data is known as multi-fidelity modeling. Many of the existing methods for generating multi-fidelity surrogate models are based on the idea that the high-fidelity experiments can be approximated as a tuned or corrected functions of low-fidelity models (e.g., [1–10]). This approach was generalized by Toropov [11] with three types of tuning: linear and multiplicative functions, correction factors, and the use of low-fidelity model inputs as tuning parameters. For the first two types, the analyst must specify a functional relationship between the low- and high-fidelity functions. The third type requires that the low-fidelity code(s) possess inputs that can serve as tuning parameters. Kriging and Gaussian process regression have also been proposed for the multi-fidelity analysis problem (e.g., [2,10]). These methods use an auto-regressive model to correct low-fidelity predictions, and the response covariance matrix is modified to account for the multiple data sources. In a typical multi-fidelity modeling scenario, it is assumed that sources of data are available from different fidelities. However, data sources coming from ∗ Research
Engineer, Mechanical Systems, GE Global Research, Niskayuna, New York Engineer, Mechanical Systems, GE Global Research, Niskayuna, New York ‡ Technical Operations Leader, Mechanical Systems, GE Global Research, Niskayuna, New York † Lead
1 of 12 American Institute of Aeronautics and Astronautics Copyright © 2018 by General Electric Co. . Published by the American Institute of Aeronautics and Astronautics, Inc., with permission.
Downloaded by You Ling on January 12, 2018 | http://arc.aiaa.org | DOI: 10.2514/6.2018-1663
different legacy systems may have the same fidelity. Additionally, most of the typical multi-fidelity methods become computationally expensive when more using two or more sources. An alternate approach to multi-fidelity modeling is to use ensemble learning [12]. In ensemble learning, two or more related but different analytical models are combined together to improve the accuracy of predictive analytics. One of the commonly used ensemble learning approaches is Bayesian Model Averaging (BMA)[13]. One of the main advantages of BMA is that it accounts for model uncertainty which may arise due to the modeling method as well sparsity of the data used. However, the main issue of BMA is that all the models are built using same data set. That means, for the current problem addressed in this paper, all the legacy data have to be combined together to form a single database. Different models would need to be built using the single database, which is then combined to form an ensemble model. Marginalizing all the legacy data together can lead to few issues. For example, if discrepancies exist in the data between each legacy system, large data noises will be introduced during the modeling process. Also, models build on each legacy system data can have more information than the models build on combined dataset. In this paper, A Bayesian multi-source modeling approach is developed for legacy data that addresses the aforementioned challenges. The approach is similar to Bayesian Model Averaging where different models are built and combined together to build a better predictive model. However, unlike BMA, the models are built for each legacy data set separately. Then, model validity is evaluated for each model as a function of input space, which is then used to combine the model. The details of the methodology are discussed in Section II. The method is demonstrated on three different numerical experiments in Section III. Finally, the conclusion is given in Section IV.
II.
Methodology
The goal is to build a model for a new design with sparse data, which alone is not sufficient to build an accurate model. However, there is a relatively large number of data available from legacy designs which are not exactly same as the new design, but they belong to a similar family. It is assumed that the performance of the new design will not be exactly same as any individual legacy design, but there will be some similarity in the performance trends. Let’s say n∗ number of data are available for the new design with y∗ = [y1∗ , . . . , yn∗ ∗ ]T as outputs and ∗ x = [x∗1 , . . . , x∗n∗ ]T as inputs. Consider that legacy data are available from p number of legacy systems from a similar family. The input for k th data source is given by xk = [x1k , . . . , xnk k ]T , where nk is the number of data points available in the k th data source. The corresponding output for each data point is given as yk = [y1k , . . . , ynk k ]T . It is assumed that the input variables (design, operational, etc.) and the output variables (performance, cost, etc.) are the same for the new system and all the legacy systems. However it is not required that a subset of xk be collocated with x∗ , i.e. it is not required to have x∗j = xjk for any j and k. Let’s define Mk = ηˆklegacy (x) as a meta-model built on the k th legacy dataset. In a typical Bayesian Model Averaging approach, a model for the new data source is given as fˆ∗ (x) =
p X
wk ηˆklegacy (x)
(1)
k=1
where wi is the model validity associated with each legacy data model and can be estimated by various techniques [14]. However there are two main issues in directly using this approach: • There can be large discrepancies between one or more of the legacy systems and the new system, while having a similar trend. This can lead to an inappropriate allocation of model validity. For example, in Figure 1a legacy-1 has a similar trend to the new system, however due to a large discrepancy it’s validity can be low. • The model validity may also vary in input space. For example, in Figure 1b, the validity of legacy-1 is higher for higher values of x while the validity of legacy 2 is higher for lower values of x. To overcome these challenges, the discrepancy of the legacy data is also included in the legacy model as Mk = ηˆklegacy (x) + δˆklegacy (x). Also, it is assumed that the model validity may vary in the design space. The
2 of 12 American Institute of Aeronautics and Astronautics
20
25
15
20 15
10
10 5 5 0
0
-5 -10 0
-5 0.2
0.4
0.6
0.8
-10 0
1
(a) Lower validity to a better model due to high discrepancy
0.2
0.4
0.6
0.8
1
(b) Variation of validity with input space
Downloaded by You Ling on January 12, 2018 | http://arc.aiaa.org | DOI: 10.2514/6.2018-1663
Figure 1: Issues with traditional Bayesian model averaging applied to legacy data
new formulation used in the current work is given as fˆ∗ (x) =
p X
w ˆk (x) ηˆklegacy (x) + δˆklegacy (x)
(2)
k=1
where δˆklegacy is the discrepancy model of the k th legacy model withPdata from new design, and w ˆk is the p model validity which is a function of the input variables such that k=1 w ˆk (x) for any value of x. In the current work, both ηˆklegacy and δˆklegacy are built using a Bayesian Hybrid Modeling (BHM) approach [15]. The model validity w ˆk is estimated by calculating the legacy model likelihood and predictive uncertainty. II.A.
Bayesian Hybrid Modeling
To model the legacy data and the discrepancy, a Bayesian framework proposed by Kennedy and O’Hangan [15] is used in this work and is referred to as a Bayesian Hybrid Model (BHM). In the original work, BHM is used to combine test/experimental data and simulation data to perform a calibration of a computer model and to quantify the discrepancy of a simulation model. Let y(x) represent the outputs of the experimental process and η(x, θ) represent the outputs of computer simulation where x are the input variables and θ are unobservable model parameters or calibration parameters of the simulation model. According Kennedy and O’Hangan’s framework, the experimental observation of outputs y is expressed as: y(xi ) = η(xi , θ∗ ) + δ(xi ) + (xi ),
for i = 1, . . . , n
(3)
where n is the number of experimental observations, θ∗ are the true values of the calibration parameters, δ is the discrepancy between the calibrated simulator η and the experimental observation, and are wellcharacterized observation errors. In the current work, the models of the legacy data are built using the same Bayesian philosophy: y ∗ (xi ) = ηklegacy (xi ) + δklegacy (xi ) + k (xi ),
for i = 1, . . . , n∗
(4)
where y ∗ is the output of new data source, ηklegacy (x) is the output of the k th legacy system, δk (x) is the discrepancy between the k th legacy system and the new system, and k are well-characterized observation errors. It should be noted that the calibration parameters are not used in Equation 4 because the legacy data may not contain these parameters. However, if calibration parameters θk are available, it can be included in Equation 4. The output of a legacy system η(x) and it’s discrepancy δ(x) are modeled as Gaussian Processes (GP) as described by Kennedy and O’Hagan [15] and Higdon et al [16]. The model for a given legacy system η(x) is approximated as GP model with a zero mean and covariance matrix given by,
3 of 12 American Institute of Aeronautics and Astronautics
1 1 exp βη |Xi − Xj |2 + I , for i, j = 1, . . . nk ληz ληs
Σηij =
(5)
where X is the vector of design variables, the parameters ληz and ληs characterize the marginal data variance captured by the model and by the residuals, respectively, β characterizes the strength of dependence of the outputs on the design variables. The outputs of new system y(x) is modeled as a GP model: Σyij =
1 1 exp βy |Xi − Xj |2 + I , for i, j = 1, . . . n∗ λyz λys
(6)
The cross covariance matrix to represent correlation between the legacy system outputs and the new system observations is given as Σηy ij =
1 1 exp βy |Xi − Xj |2 + I , for i = 1, . . . n∗ , j = 1, . . . , nk λyz λys
(7)
Downloaded by You Ling on January 12, 2018 | http://arc.aiaa.org | DOI: 10.2514/6.2018-1663
The discrepancy parameter δ is modeled as Σδij =
1 exp βδ |Xi − Xj |2 , for i, j = 1, . . . n∗ λδz
The likelihood of combined data z = (y T , η T ) is then given as 1 T −1 1 exp − D Σ D L(D|ληz , βη , ληs , λδz , βδ ) = 2 |Σ|1/2 ! Σδ 0 0 Σ 0 new where D = (y T , η T , δ T ) and Σ = 0 Σy Σyη + 0 Σlegacy 0 Σηy Ση The posterior distribution of all the hyper-parameters is given by π(ληz , βη , ληs , λδz , βδ |D) = L(D|ληz , βη , ληs , λδz , βδ )π(ληz )π(βη )π(ληs )π(λδz )π(βδ )
(8)
(9)
(10)
where π(.) on the right-hand side of the equation is the prior distribution of the parameters. The target posterior distribution is evaluated using a Markov Chain Monte Carlo (MCMC) [17] approach. For more details of BHM modeling and its enhancements please refer to [18, 19]. II.B.
Model Validity
Once the ηˆklegacy and δˆklegacy models are built for each legacy dataset, the next step is to estimate the model validity. In this work two metrics are used to estimate model validity. The first one is a likelihood based model validity, while the second one is an uncertainty based model validity. In the likelihood-based model validity, the validity of a legacy model at an available input setting x∗ of the new design is proportional to the probability of the legacy model to predict the output or performance of the new design, as shown in Figure 2. For example, the likelihood of the k th legacy model, Mk at a location x∗ , for which output of new design is known to be y ∗ is given as 1 (y ∗ − µk (x∗ )2 ) likelihood ∗ ∗ wk (x ) ∝ P (y |Mk ) = p exp − (11) 2σk (x∗ )2 2πσk (x∗ )2 where µk (x∗ ) and σk (x∗ ) are the predictive mean and standard deviation of the k th legacy model Mk = ηˆklegacy (x) + δˆklegacy (x) at design specified by x∗ . In the uncertainty-based model validity, the model validity is inversely proportional to the predictive uncertainty of legacy only model ηˆklegacy (x). In other words, the validity of a legacy model Mk reduces as it goes farther from the available legacy data which is indicated by the predictive standard deviation σηlegacy (x) k as shown in Figure 3 and given as wkuncertainty (x) ∝
1 σηlegacy (x) k
4 of 12 American Institute of Aeronautics and Astronautics
(12)
𝑦 𝑤2𝑙𝑖𝑘 (𝑥 ∗) ∝
1 2𝜋𝜎2 𝑥 ∗
2
𝑒
𝑦∗−𝜇2 𝑥 ∗ 2𝜎2 𝑥 ∗ 2
2
𝑀2 (𝑥)
𝑦∗
𝑀1 (𝑥) Data for new design
𝑥
𝑥∗ 𝑤1𝑙𝑖𝑘 (𝑥
∗) ∝
1
2𝜋𝜎1 𝑥 ∗
2
𝑒
𝑦∗−𝜇1 𝑥 ∗ 2𝜎1 𝑥 ∗ 2
2
Downloaded by You Ling on January 12, 2018 | http://arc.aiaa.org | DOI: 10.2514/6.2018-1663
Figure 2: Likelihood based model validity 𝜂Ƹ 𝑘 𝑙𝑒𝑔𝑎𝑐𝑦 (𝑥)
𝑦
Predictive uncertainty 𝑙𝑒𝑔𝑎𝑐𝑦 𝜎𝑘 (𝑥)
Legacy data
𝑤𝑘𝑢𝑛𝑐𝑒𝑟𝑡. 𝑥 ∝
1 𝑙𝑒𝑔𝑎𝑐𝑦 𝜎𝜂𝑘 (𝑥)
𝑥 Figure 3: Uncertainty based model validity
The overall model validity of the k th legacy model at a given design point x is then given as wk (x) = wklikelihood (x)wkuncertainty (x) = κ(x) where κ(x) is proportionality factors given as κ(x) = 1/
P (y new |Mk ) σηlegacy (x) k
(13)
P(y new |Mk ) k=1 σηlegacy (x) . k
Pp
(x) is the byproduct of each legacy BHM modeling and can be estimated It must be noted that σηlegacy k directly using the legacy Gaussian Process model ηˆklegacy during the prediction. However, the model likelihood P (y new |Mk ) is only known at design points x∗ where data for new designs are available and need to be interpolated or extrapolate for other design points. In the current work, a Gaussian Process model of P (y|Mk ) is built for each legacy model to estimate a likelihood-based model validity at any location in the design space.
III.
Results
In this section, three different examples are carried out to demonstrate the benefit of Bayesian multisource modeling. The first test case is a mathematical problem, the second test case is an analytical problem for borehole design and the third is an experimental problem to estimate the crack propagation rate of a new aluminum alloy.
5 of 12 American Institute of Aeronautics and Astronautics
III.A.
Exponential Function
The first test case is a two dimensional exponential function reported by Currin et al. [20] which is treated as a true function and is given by Equation 14. The design space of the input variables is defined by xi ∈ [0, 1], for all i = 1, 2. 2300x31 + 1900x22 + 2092x1 + 60 1 (14) fexp (x) = 1 − exp − 2x2 100x31 + 500x21 + 4x1 + 20
1 [fexp (x1 + 0.05, x2 + 0.05) + fexp (x1 + 0.05, max(0, x2 − 0.05))] 4 1 + [fexp (x1 − 0.05, x2 + 0.05) + fexp (x1 − 0.05, max(0, x2 − 0.05))] 4
fL1 (x) =
fL2 (x) = 16.55 − 10.15x1 − 13.24x2 + 2.6x1 x2 + 4.56x21 + 3.77x22
(15)
(16) New Legacy-1 Legacy-2
1 0.9 0.8
20
True Legacy-1 Legacy-2
0.7 0.6
10 x2
y
Downloaded by You Ling on January 12, 2018 | http://arc.aiaa.org | DOI: 10.2514/6.2018-1663
Two legacy data sources are used to demonstrate the benefit of multi-source modeling in this example. They are generated using two different lower fidelity functions, and are labeled as legacy-1 and legacy-2. The first legacy source is a lower fidelity model used by Xiong et al. [21] and is given by Equation 15. The second legacy data source is built from a low fidelity function generated using a quadratic equation given by Equation 16. The behavior of the true and the legacy functions are shown in Figure 4a. The legacy-1 function is very similar to the true function in entire design space. The legacy-2 function is only accurate for higher values of x1 and does not capture the non-linear behavior for lower values of x1 .
0.5 0.4
0 1
0.3 0.2 0.1
0.5
0
x2
0
0
0.2
0.4
0.6
0.8
1
0
0.2
0.4
0.6
0.8
1
x1
x1
(a) Exponential function and legacy functions
(b) Distribution of training data
Figure 4: True and legacy functions and data distribution for exponential function To train the BHM and multi-source models, 10 randomly selected new designs are generated using the true function within the original design space. 100 additional are generated using the true function and are used for validation. For the legacy-1 data source, 50 designs are generated using Equation 15 within the design space of x1 ∈ [0, 0.5] and x2 ∈ [0, 1], while for the legacy-2 data source, 50 designs are generated using Equation 16 within the design space of x1 ∈ [0.5, 1] and x2 ∈ [0, 1]. The selected data points in input space are shown in Figure 4b. In the first step of multi-source modeling, BHM models of individual legacy source and discrepancy are built using the respective data sources. The surface plot for each legacy BHM model, Mk = ηˆklegacy (x) + δˆklegacy (x), is shown in Figure 5. Next, the model validity of each legacy model is calculated for each data point generated from the true function and the BHM models. The contour plot of normalized model validity for both the legacy models is shown in Figure 6. It can be seen that the model validity of the legacy-1 BHM model is relatively low for higher values of x1 while, the validity of the legacy-2 BHM model is relatively low for lower values of
6 of 12 American Institute of Aeronautics and Astronautics
15
y
10
y
20
BHM Model Test-data Legacy-1 data
5 0 1
10
0 1
0.5
x2
BHM Model Test-data Legacy-2 data
0.5 0
1
0.5
0
x2
0
x1
(a) Legacy-1 Source
1
0.5
0
x1
(b) Legacy-2 Soruce
x1 . This is mainly due to the predictive uncertainty of each legacy model. For example, the data for the legacy-2 model are only available in the region of x1 ∈ [0.5, 1], which causes a higher predictive uncertainty in the model in the region of x1 ∈ [0, 0.5]. However, there are regions around lower values of x1 where the validity of legacy-2 model is still better than the legacy-1 model. This is due to the contribution of likelihood term in model validity, i.e. in this region the likelihood of the legacy-2 model to predict the true function overpowers the predictive uncertainty when compared to the legacy-1 model. 1
1
1
1
0.9
0.9
0.9
0.9
0.8
0.8
0.8
0.8
0.7
0.7
0.7
0.7
0.6
0.6
0.6
0.6
0.5
0.5
0.5
0.5
0.4
0.4
0.4
0.4
0.3
0.3
0.3
0.3
0.2
0.2
0.2
0.2
0.1
0.1
0.1
0.1
0
x2
x2
Downloaded by You Ling on January 12, 2018 | http://arc.aiaa.org | DOI: 10.2514/6.2018-1663
Figure 5: Bayesian hybrid model (BHM) of legacy sources for exponential function
0
0
0
0.2
0.4
0.6
0.8
1
0
0
x1
0.2
0.4
0.6
0.8
1
x1
(a) Legacy-1 Source
(b) Legacy-2 Soruce
Figure 6: Model validity of legacy sources for exponential function Once the multi-source model is built, it is validated with the validation data set for the new design. In Figure 7a, the left plot shows the predicted versus actual plot of each legacy BHM model. As seen in the figures, the legacy-1 model has a higher error and higher predictive uncertainty when compared to the legacy2 model. The right plot of Figure 7a shows the predictions of the multi-source model which combines both the legacy models. The multi-source model has better predictive accuracy than the legacy BHM models. The improvement in prediction by multi-source modeling is also shown by the percentage error box plot shown in Figure 7b. The median, third quartile and the maximum percentage error was found to be better for the multi-source model. III.B.
Borehole Problem
The second test problem used to demonstrate multi-source modeling is the Borehole problem [22] where an analytical function is used to calculate the water flow rate through a borehole. The analytical function is
7 of 12 American Institute of Aeronautics and Astronautics
Legacy BHM Model 14
Multi-Source Model
Legacy-1 Legacy-2
14
18
Multi-Source
16 12
12
10
10
8
12
% error
Predicted
Predicted
14
8
6
6
4
4
2
2
10 8 6 4 2 0
2
4
6
8
10 12 14
2
4
6
True
8
10 12 14
Legacy-1
Legacy-2
Multi-Source
True
(a) Actual vs Predicted plot
(b) Percentage error distribution
Downloaded by You Ling on January 12, 2018 | http://arc.aiaa.org | DOI: 10.2514/6.2018-1663
Figure 7: Prediction and error on validation data for exponential function problem
given in Equation 17. For the true function, parameter α is given as α = [2π, 1.0, 2.0] and is used to generate data for the new design. Additionally, two legacy data sources are generated using the same function but with different parameter values. For the legacy-1 data source, α = [5, 1.5, 2.0] is used while for the legacy-2 data source, α = [6, 0, 1.8] is used. The definition and ranges of all the variables are given in Table 1. f (x) =
α T (H − Hl ) 1 u u α3 LTu ln(r/rw ) α2 + ln(r/r + 2 w )r Kw ) w
Variable rw r Tu Hu Tl Hl L Kw
Tu Tl
Definition Radius of borehole (m) Radius of influence (m) Transmissivity of upper aquifer (m2 /yr) Potentiometric head of upper aquifer (m) Transmissivity of lower aquifer (m2 /yr) Potentiometric head of lower aquifer (m) Length of borehole (m) Hydraulic conductivity of borehole (m/yr)
(17)
Range [0.05, 0.15] [100, 50000] [63070, 115600] [990, 1110] [63.1, 116] [700, 820] [1120, 1680] [9855, 12045]
Table 1: Variable definition and ranges for borehole problem To carry out the multi-source modeling, 23 data points are generated for the new design, 50 data points are generated for the legacy-1 source, and 40 data points are generated for the legacy-2 data source. The distribution of data points in the input design space is shown in Figure 8. One can notice that data is missing in some parts of the design space for each legacy data. This is deliberately done to simulate a scenario typical in real engineering problems where legacy designs are built for different requirements. For validation 200 additional data points were generated using the true function and are predicted using the multi-source model. The left plot of Figure 9a shows the predicted versus actual plot for each legacy BHM model, while the right plot shows the same for the multi-source model. It can be observed that the prediction of the multi-source model is relatively tighter to the 45o line than each legacy model. Figure 9b shows the box plot of percentage error of legacy models and multi-source model for all the validation data. Similar to the previous example, the multi-source model had better predictive capability than each legacy model..
8 of 12 American Institute of Aeronautics and Astronautics
New Legacy-1 0.15 Legacy-2
rw
0.1 4 2
r 104
11 10 9 8 7 1100 1050 1000 110 100 90 80 70 800 750
Tu Hu Tl Hl
1600 1400 1200
L
Downloaded by You Ling on January 12, 2018 | http://arc.aiaa.org | DOI: 10.2514/6.2018-1663
10000 5000
Kw
0.05 0.1
0
2
4
7
104
9 111000
70 90 110
750 8001200 16005000 15000
104
Figure 8: Data distribution of new design and legacy design for borehole problem Legacy BHM Model
Multi-Source Model
200
200 16
180 160
140
140
120
120
Predicted
160
100 80
14 12
% error
180
Predicted
Multi-Source
Legacy-1 Legacy-2
100 80
10 8 6
60
60
40
40
4
20
20
2
0
0 0 0
50
100
True
150
200
0
50
100
150
200
Legacy-1
Legacy-2
Multi-Source
True
(a) Actual vs Predicted plot
(b) Percentage error distribution
Figure 9: Prediction and error on validation data of borehole problem
III.C.
Material Crack Growth Rate Prediction
To demonstrate the multi-source modeling method on an engineering problem, material crack growth rate data of three different aluminum allows are used from a Damage Tolerant Design Handbook [23]. The Al7475 alloy is assumed to be a new design, while Al-7150 and Al-7175 alloys are assumed to be legacy designs. For each of these alloys, crack growth rate (da/dn) data is extracted for various loading frequencies (F req), stress ratios (R), amplitude stress intensity factors (∆K). The data were only extracted for test environment of LAB AIR (Laboratory Air) and L.H.A (Low Humidity Air, < 10%RH) and a material orientation of L-T. For legacy material Al-7150 and Al-7175, 40 and 38 data points are extracted, respectively. For the new design, Al-7475, 24 data points are extracted out of which 10 data points are used for training and 14 data points are used for validation. The extracted data are shown in Figure 10 and the scatter plot of input variables are shown in Figure 11.
9 of 12 American Institute of Aeronautics and Astronautics
Aluminium 7150 Alloy
10-4
Aluminium 7175 Alloy
10-4
10-5
10-5
10-6
10-6 R = 0.1 R = 0.4
R = 0.02 R = 0.08
10-7
10-8
10-7
0
2
4
6
8
10
12
14
16
10-8
18
0
(a) Actual vs Predicted plot
4
6
8
10
12
14
16
(b) Percentage error distribution Aluminium 7475 Alloy
10-4
Downloaded by You Ling on January 12, 2018 | http://arc.aiaa.org | DOI: 10.2514/6.2018-1663
2
10-5
R = 0.1 R = 0.3 R = 0.4
10-6
10-7
4
6
8
10
12
14
16
18
(c) Percentage error distribution
Figure 10: Prediction and error on validation data of borehole problem
20
Freq
10
0.3 R
0.2 0.1
15 10
K
New Legacy-1 Legacy-2
5 0
10
0.1
0.2
0.3
5
10
15
Figure 11: Data distribution of new design and legacy design for material crack growth rate problem To carry out the multi-source modeling, legacy BHM models of crack growth rate and models of model validity are built as a function of frequency, stress ratio, and amplitude stress intensity factor. The multi-
10 of 12 American Institute of Aeronautics and Astronautics
Legacy BHM Model
Multi-Source Model
Legacy-1 Legacy-2
20
Multi-Source
20
50
15
10
40
% error
Predicted
Predicted
15
10
30
20 5
5 10
0
0 0 0
5
10
True
15
20
0
5
10
15
20
Legacy-1
Legacy-2
Multi-Source
True
(a) Actual vs Predicted plot
(b) Percentage error distribution
Downloaded by You Ling on January 12, 2018 | http://arc.aiaa.org | DOI: 10.2514/6.2018-1663
Figure 12: Prediction and error on validation data of material crack growth rate problem
source model is validated with 14 data points from the Al-7475 alloy and were not used for training. The predicted versus actual plot is shown in Figure 12a and the box plot of percentage error is shown in Figure 12b. It is observed that the multi-source modeling method is able to significantly improve the predictive accuracy in this case.
IV.
Conclusion
A Bayesian multi-source modeling approach for legacy data is developed to overcome some of the issues of current multi-fidelity and ensemble modeling techniques. In this approach, models of each legacy system and the corresponding discrepancy with the new design is built using a Bayesian Hybrid Model framework. Models were built for the validity of each legacy system as a function of input space. Then, the ensemble model is used to carry out the predictive analysis for a new design. The method has been demonstrated with two analytical problems and and engineering problem and has been found to improve the accuracy of prediction for all three examples.
References 1 Keane, A. and Nair, P., Computational Approaches for Aerospace Design: The Pursuit of Excellence, John Wiley & Sons, 2005. 2 Kennedy, M., “Predicting the output from a complex computer code when fast approximations are available,” Biometrika, Vol. 87, No. 1, 2000, pp. 1–13. 3 Vitali, R., Haftka, R., and Sankar, B., “Correction response surface approximations for stress intensity factors of a composite stiffened plate,” 39th AIAA/ASME/ASCE/AHS/ASC Structures, Structural Dynamics, and Materials Conference and Exhibit, Structures, Structural Dynamics, and Materials and Co-located Conferences, American Institute of Aeronautics and Astronautics, April 1998. 4 Alexandrov, N. M. and Lewis, R. M., “An Overview of First-Order Model Management for Engineering Optimization,” Optimization and Engineering, Vol. 2, No. 4, Dec. 2001, pp. 413–430. 5 Eldred, M., Giunta, A., and Collis, S., “Second-Order Corrections for Surrogate-Based Optimization with Model Hierarchies,” 10th AIAA/ISSMO Multidisciplinary Analysis and Optimization Conference, Multidisciplinary Analysis Optimization Conferences, American Institute of Aeronautics and Astronautics, Aug. 2004. 6 Haftka, R. T., “Combining global and local approximations,” AIAA Journal, Vol. 29, No. 9, Sept. 1991, pp. 1523–1525. 7 Umakant, J., Sudhakar, K., Mujumdar, P., and Rao, C., “Customized Regression Model for Improving Low Fidelity Analysis Tool,” 11th AIAA/ISSMO Multidisciplinary Analysis and Optimization Conference, Multidisciplinary Analysis Optimization Conferences, American Institute of Aeronautics and Astronautics, Sept. 2006. 8 Berci, M., Toropov, V., Hewson, R., and Gaskell, P., “Metamodelling Based on High and Low Fidelity Model Interaction for UAV Gust Performance Optimization,” 50th AIAA/ASME/ASCE/AHS/ASC Structures, Structural Dynamics, and Materials Conference, Structures, Structural Dynamics, and Materials and Co-located Conferences, American Institute of Aeronautics and Astronautics, May 2009. 9 Qian, P. Z. G. and Wu, C. F. J., “Bayesian Hierarchical Modeling for Integrating Low-Accuracy and High-Accuracy Experiments,” Technometrics, Vol. 50, No. 2, May 2008, pp. 192–204.
11 of 12 American Institute of Aeronautics and Astronautics
Downloaded by You Ling on January 12, 2018 | http://arc.aiaa.org | DOI: 10.2514/6.2018-1663
10 Forrester, A. I. J., S´ obester, A., and Keane, A. J., “Multi-fidelity optimization via surrogate modelling,” Proceedings of the Royal Society A: Mathematical, Physical and Engineering Sciences, Vol. 463, No. 2088, 2007, pp. 3251–3269. 11 Toropov, V. V., “Modelling and Approximation Strategies in Optimization: Global and Mid-Range Approximations, Response Surface Methods, Genetic Programming, Low/High Fidelity Models,” Emerging Methods for Multidisciplinary Optimization, edited by J. Blachut and H. Eschenauer, chap. 5, Vol. 425, CISM Courses and Lectures, Springer-Verlag, New York, 2001, pp. 205–256. 12 Opitz, D. and Maclin, R., “Popular ensemble methods: An empirical study,” Journal of Artificial Intelligence Research, Vol. 11, 1999, pp. 169–198. 13 Hoeting, J. A., Madigan, D., Raftery, A. E., and Volinsky, C. T., “Bayesian model averaging: a tutorial,” Statistical science, 1999, pp. 382–401. 14 Zhang, C. and Ma, Y., Ensemble machine learning, Vol. 1, Springer, 2012. 15 Kennedy, M. C. and O’Hagan, A., “Bayesian calibration of computer models,” Journal of the Royal Statistical Society: Series B (Statistical Methodology), Vol. 63, No. 3, 2001, pp. 425–464. 16 Higdon, D., Kennedy, M., Cavendish, J. C., Cafeo, J. A., and Ryne, R. D., “Combining field data and computer simulations for calibration and prediction,” SIAM Journal on Scientific Computing, Vol. 26, No. 2, 2004, pp. 448–466. 17 Hastings, W. K., “Monte Carlo sampling methods using Markov chains and their applications,” Biometrika, Vol. 57, No. 1, 1970, pp. 97–109. 18 Kumar, N. C., Subramaniyan, A. K., and Wang, L., “Improving high-dimensional physics models through Bayesian calibration with uncertain data,” ASME Turbo Expo 2012: Turbine Technical Conference and Exposition, American Society of Mechanical Engineers, 2012, pp. 407–416. 19 Kumar, N. C., Subramaniyan, A. K., Wang, L., and Wiggs, G., “Calibrating transient models with multiple responses using Bayesian inverse techniques,” ASME Turbo Expo 2013: Turbine Technical Conference and Exposition, American Society of Mechanical Engineers, 2013, pp. V07AT28A007–V07AT28A007. 20 Currin, C., “A Bayesian approach to the design and analysis of computer experiments,” Tech. rep., ORNL Oak Ridge National Laboratory (US), 1988. 21 Xiong, S., Qian, P. Z., and Wu, C. J., “Sequential design and analysis of high-accuracy and low-accuracy computer codes,” Technometrics, Vol. 55, No. 1, 2013, pp. 37–46. 22 Morris, M. D., Mitchell, T. J., and Ylvisaker, D., “Bayesian design and analysis of computer experiments: use of derivatives in surface prediction,” Technometrics, Vol. 35, No. 3, 1993, pp. 243–255. 23 Skinn, D., Gallagher, J. P., Berens, A. P., Huber, P., and Smith, J., “Damage Tolerant Design Handbook. Volume 5. Chapter 8.” Tech. rep., Dayton Univ OH Research Inst, 1994.
12 of 12 American Institute of Aeronautics and Astronautics