Apr 19, 2017 - Emulator. (emulated SCOPE). Which model would you choose? Any difference? 3/53 .... Gaussian Processes Regression (GPR). ⢠Variational ...
Emulation of radiative transfer models (RTMs): new opportunities for spectroscopy data processing Jochem Verrelst, Juan Pablo Rivera, Jordi Muñoz-Marí, Neus Sabater, Jorge Vicent, Jose Moreno and Gustau Camps-Valls Image Processing Laboratory, Univ. of Valencia (Spain)
EARSeL Imaging Spectroscopy Workshop 19 April2017
0.7 0.7
0.5 310
500#
0.6 0.6
9
0.4 8
2.5
0.5 0.5 Reflectance Reflectance Reflectance
7
N
0.3 0.3
25
0.2 4 3
0.2 0.2
1.5 2 0.1
0.1 0.1 0
LAI Cw
0.3 6
0.4 0.4
1
500 500 500
1000 1000 1000
1500 1500 1500 Wavelength (nm) Wavelength (nm) Wavelength (nm)
2000 2000 2000
1000 00
501 0.05 100 2
Cab hc Cdm
2/53
Any difference? 37 min
9s
Which model would you choose?
SCOPE
Emulator
(emulated SCOPE)
3/53
BACKGROUND
Advanced RTMs: generation of a large LUT (>1000#) SCOPE
DART
MODTRAN
40
0.5 0.35
10
0.3
9
0.25
8
0.2
7
35 0.4 30 25 AOT
dif
T
LAI
Reflectance
0.3
0.15
6
0.1
5
0.2
20 15 10
0.1 0.05
0
Hours
5
4
600
800
1000 1200 1400 1600 1800 2000 2200 Wavelength (nm)
days
3
0 60 80 CC
500
1000
1500 Wavelength (nm)
2000
>days
2500
0
100 200 300 AMS
4/53
Advanced RTMs: more realistic but slow • RTMs are widely used in RS science, e.g. for development of new missions and retrieval (inversion). • When choosing an RTM, a trade-of between applicability and realism has to be made: simpler models are easier to apply but less realistic, while advanced models are more realistic but require a large amount of variables to be configured.
Examples of advanced RTMs: • Ray tracing models (e.g. FLIGHT, RAYTRAN) • Voxel models: DART • Soil-Vegetation-Atmosphere-Transfer (SVAT) models: e.g. SCOPE • Atmospheric transfer models: e.g. MODTRAN
• Main drawback of advanced RTMs involves their long processing time: the more advanced, the longer it takes to generate output. • Long processing time makes that advanced RTMs are of little use for operational tasks, e.g., pixel-by-pixel inversion schemes.
5/53
Emulation of RTMs
Emulators are statistical models that are able to approximate the processing of a physical model (e.g. RTM) - at a fraction of the computational cost: making a statistical model of a physical model RTM
Machine learning
6/53
Regression vs. Emulation: Classical use of machine learning in optical RS: 0.7 0.6
Reflectance
0.5 0.4
Variable (e.g. LAI)
0.3 0.2 0.1 0
500
1000
1500 Wavelength (nm)
2000
Statistical method: •
Variable/data-driven, black box, 1 output, portability is questionable
Emulation in optical RS:
Replace RTM: •
Multiple applications, e.g. inversion Radiometric method: Spectral fitting Portable: generally applicable
7/53
Emulation in practice: Spectra 0.7
Variables
regression
10 9
0.6
8 7 6
0.4 LAI
Reflectance
0.5
5
0.3
4
0.2
3
Emulation
2
0.1 0
(e.g. LAI, chlorophyll,…)
1
500
1000
1500 Wavelength (nm)
2000
0 0
50
100
Cab
• In principle any nonlinear, adaptive machine learning regression algorithm (MLRAs) can serve as emulator. MLRAs:
• However, to emulate RTM spectral output, the MLRA should have the capability to reconstruct multiple outputs, i.e. the complete spectrum. Only Few MLRAs possess multi-output capability (e.g. NN). • Multi-output resolved with dimensionality reduction techniques (e.g. PCA). “spectral redundancy” is a blessing
8/53
Processing steps Input (variables + spectra)
Splitting into training/validation
PCA on spectra
MLRA training looping over components
Prediction of components
Prediction of components
Validation
Emulator
Sc = U · X
PCA on spectra
MLRA training looping over components
Reconstruction of spectra
W = (Y + λI)-1 · Sc Sp = Sc · W
Reconstruction of spectra
http://isp.uv.es/software.html
Xr = U⊤ · Sp 9/53
Emulator toolbox With ARTMO’s emulation processing chain any RTM can be converted into an emulator. Input (variables + spectra)
Splitting into training/validation
PCA on spectra
MLRA training looping over components
Prediction of components
Reconstruction of full spectrum
Validation
Emulator
Xr = U⊤ · Sp
(Rivera et al., 2015)
10/53
Emulators great idea… what about accuracy? Various open questions: 1) Role of machine learning regression algorithm? 2) Role of dimensionality reduction (DR) method?
3) Role of LUT size training? 0.6 0.5 Reflectance
4) Role of data type?
0.7
0.4 0.3 0.2 0.1 0
500
1000
1500 Wavelength (nm)
2000
11/53
Is a small LUT of #500 samples sufficiently covering the parameter space? Latin Hypercube Sampling (LHS)
12/53
Experimental setup: emulating SCOPE Experimental setup: • SCOPE: LUT 500# @ 1 nm; 400-2500 nm; 8 variables, R and SIF outputs • 6 machine learning methods tested: RF, SVR, KRR, NN, GPR, VH-GPR • # 10, 20, 30, 40 components tested; 70/30% Training/Validation (T/V) • LUT of # 500, 1000 samples SCOPE: N (1-3) LCC (0-100) Cw (0-0.5) Cdm (0-0.5) Cs (0-0.3 LAI (0-10) Hc (0-2) SMC (0-0.7)
Emulator 0.7 0.6 0.5 Reflectance
0.7 0.6 0.5 Reflectance
• • • • • • • •
Emulated SCOPE
37 min (500#)
0.4 0.3
0.4
0.2
0.3
0.1
0.2
0
0.1 0
500
1000
1500 Wavelength (nm)
500
1000
1500 Wavelength (nm)
2000
2000
13/53
Role of machine learning • • • • • •
Random Forests (RF) Neural Networks (NN) Support vector regression (SVR) Kernel Ridge Regression (KRR) Gaussian Processes Regression (GPR) Variational Heteroscedastic GPR (VHGPR) 14/53
Validation SCOPE R emulation
1) Role of machine learning Relative errors for 30% validation data (#150) with 20 PCA. 16 RF NN SVR GPR
KRR RF
NN KRR
GPR SVR VH-GPR VHGPR
14
NRMSE (%)
12 10 8 6 4 2 0 400
600
800
1000 1200 1400 1600 1800 2000 2200 2400 Wavelength (nm)
Here, NN best performing (