parameters optimization

advertisement
NACLIM Deliverable D32.27
1st Batch-Updates of data sets containing time series of
relevant key ocean quantities
This deliverable consists of collecting and organizing relevant data sets, time series and model outputs from the first
NACLIM period to be used in the model – observation synthesis.
WP No.
3.2
WP title
Impact of Arctic initialization on forecast skill
Work
duration1)
24
Due delivery deadline:
30 October 2014
R= report
Nature of the
deliverable
P= prototype
D= demonstrator
X
O= Other
PU = public
Dissemination
level
X
PP= restricted to other programme participants, including the Commission services
RE= restricted to a group specified by the consortium, including the Commission services
CO= confidential, only for members of the consortium, including the Commission services
1) Work duration = project month
Authors of this report
University of Hamburg (UHAM): Detlef Stammer, Armin Köhl, Reema Agrawal
NACLIM project is financed by the European Commission
through the 7th Framework Programme for Research
Theme 6 Environment
Grant Agreement 308299
Page 1
Index
1. Executive summary ....................................................................................................................................... 3
2. Project objectives .......................................................................................................................................... 3
3. Detailed report on the deliverable .................................................................................................................. 4
4. References .................................................................................................................................................. 15
5. Dissemination and uptake ........................................................................................................................... 15
6. The delivery is delayed:  Yes  No ....................................................................................................... 16
7. Changes made and difficulties encountered, if any ...................................................................................... 16
8. Efforts for this deliverable ............................................................................................................................ 16
9. Sustainability ............................................................................................................................................... 16
10. Dissemination and uptake activities ........................................................................................................... 16
Uptake by the targeted audience ....................................................................................................................... 16
Page 2
1. Executive summary
Work performed
Climate models are tuned using various approaches to bring models closer to real world observations. An
automatic optimization procedure is required to find the best estimate of a set of parameters. Our approach is a
parameter optimization problem in which optimal parameter values are obtained by reducing the model-data
errors (cost function). We use multivariate data assimilation technique for the optimization of parameters of the
THOR coupled assimilation system, which is now established as the CESAM (CEN Earth System Assimilation
Model). A sensitivity analysis of CESAM is carried out. The behaviour of the cost function with respect to
changes in each control parameter is studied, the most sensitive parameters are identified and regional effects
of parameter perturbation are discussed. The results of the sensitivity analysis serve as a guideline for a useful
application of the assimilation procedure and will help interpreting the results. Results are evaluated with
improvements in basic physical properties predicted by model such as surface temperature, precipitation and
net heat flux with respect to observations.
The Simultaneous Perturbation Stochastic Approximation (SPSA) method is used for assimilating annual mean
data (pseudo and ERA-interim reanalysis data). The advantage of using SPSA method is its ease of
implementation, robustness to noise in cost function and cost effectiveness.
Main results achieved
-
Identical twin testing demonstrates that the method reliably finds an estimate of parameters. The errors in
optimized state for example in surface temperature and net heat flux reduced up to 16% and 30%,
respectively.
-
Tuning the model to reanalysis data by adjusting parameters yields a reduction in temperature (all levels)
and net heat flux. The optimized state of CESAM shows an improvement in climate variables such as
temperature at 1000 hPa, sea level pressure, zonal winds at 500 hPa and surface precipitation.
2. Project objectives
With this deliverable, the project has contributed to the achievement of the following objectives (see DOW
Section B.1.1):
No.
1.
2.
3.
4.
5.
6.
Objective
Assessing the predictability and quantifying the uncertainty in forecasts of the North
Atlantic/Arctic Ocean surface state
Assessing the atmospheric predictability related to the North Atlantic/Arctic Ocean
surface state
Monitoring of volume, heat and fresh water transports across key sections in the
North Atlantic
Quantifying the benefit of the different ocean observing system components for the
initialization of decadal climate predictions
Establishing the impact of an Arctic initialization on the forecast skill in the North
Atlantic/European sector
Quantifying the impact of predicted North Atlantic upper ocean state changes on the
Page 3
Yes
X
No
X
X
X
X
X
oceanic ecosystem
Quantifying the impact of predicted North Atlantic upper ocean state changes on
socioeconomic systems in European urban societies
8. Providing recommendations for observational and prediction systems
9. Providing recommendations for predictions of the oceanic ecosystem
10. Disseminating the key results to the climate service community and relevant
endusers/stakeholders
11. Constructing a dataset for sea surface and sea ice surface temperatures in the Arctic
7.
X
X
X
X
X
3. Detailed report on the deliverable
As part of Task 3.2 we performed model optimization though parameter estimation with the THOR coupled
assimilation system, which is now established as the CESAM (CEN Earth System Assimilation Model). During
the THOR Project, the adjoint model was already used to explore the capabilities of the adjoint method for
optimizing models by estimating ocean and atmospheric parameters. Due to the nonlinear dynamics of the
atmospheric component the length of the feasible assimilation window is limited to the period over which the
model can be approximated by a linear model. Although parameter optimization turned out to be feasible for up
to several weeks, another result of the test experiments showed that the parameter optimization with the full
model parameterizations needed to simulate a realistic climate is limited to one day. For climate change
predictions, an important task is to reduce errors in the model and bring the model closer to observations.
Improving the initial conditions results in reducing short term forecast errors, however, their influence on
climate scale predictions is negligible. Since parameters other than the initial conditions influence the climate of
the model on longer time scales than a few days extending the window of assimilation is essential for matching
the climate of the model to the observations by parameter optimization
To cope with these problems we employ here a stochastic optimization algorithm because of its ability to cope
with inherent system noise and highly nonlinear systems. We use stochastic optimization algorithms for
parameter optimization in CESAM. The stochastic optimization algorithm Simultaneous perturbation stochastic
approximation (SPSA) techniques is implemented and after testing used for the tuning the model to the
observed climate by fitting the climatological output from the model to climatological data from ERA-Interim
(Dee et al., 2011).
3.1 Model configuration and control run
For all experiments in this study, the horizontal resolution of CESAM is set to spectral T21 (approximately 5.6◦
grid spacing) with 10 vertical sigma levels and an integration time step of 45 minutes. Monthly averaged
climatological sea ice concentration (sea ice model is switched off), sea surface temperature (SST), glacier
coverage, albedo and soil temperature from ERA40 reanalysis (Uppala et al., 2005) are used as boundary
conditions. Beginning from a cold start, a forward run of CESAM was carried out for 25 years. A restart file was
recorded at the end of 25 years and this restart which corresponds to 1st January is used as initial condition in
Page 4
all the experiments. A control simulation is made from the initial and boundary conditions and a forward run of
CESAM for a period of one year is carried out. Default values of control parameters given in Table 3.1 are
used. This simulation is referred to as the control run and is used as a reference in all the experiments. The
control run is referred to as model’s original state, i.e. the state before parameter optimization.
Table 3.1— List of 15 control parameters used in optimization procedure
3.2 Comparison of model state with ERA-Interim reanalysis data
In order to validate the model before optimization basic climate variables such as temperature at 1000 hPa,
sea level pressure, zonal winds at 500 hPa and surface precipitation are selected and a comparison of the
annual mean control run (1 year period as discussed in section 2.2) with annual mean ERA-Interim reanalysis
(referred to as observations hereafter) is performed. The annual mean observations are constructed from 21
years of data (from 1989 to 2009) and the values are averaged on to the model grid. Figure 3.1 shows the
comparison of annual mean temperature at 1000 hPa (which is near the surface). Overall patterns and range
of temperature distribution in the control run (top panel) agree well with the observations (middle panel).
However from the difference control - observations plot (bottom panel), it can be clearly seen that CESAM in its
original state in general has colder temperature with respect to observations. Exceptions are the regions over
middle east, eastern China, south-west Australia, most parts of central North America and over north west
region over South America. In tropics the difference is around -2 to -4 K, however in polar regions the model is
far too cold (-10 to -15 K), specially over the Arctic Ocean. Addtitional comparisons are performed for e.g.
precipitation, pressure or zonal wind which demonstrated that the model in general produces a realistic climate
Page 5
while it simulates colder surface temperatures, weaker zonal winds over the Southern Ocean and a poorly
resolved ITCZ
Figure 3.1— Annual mean temperature (K) at 1000 hPa from (top) control run, (middle) observations and (bottom)
difference (control-observations). Control run represents the annual mean state resulting from one year run discussed in
section 2.2. Observations represents annual mean of 21 years (1989-2009) from ERA-Interim reanalysis.
Page 6
3.3 Parameter estimation using SPSA technique
The details of SPSA are given by (Spall 1998); however the technique is briefly discussed here. In SPSA the
cost function Y() is minimized using the iterative procedure
k+1 = k − akgk(k)
where gk(k) is a stochastic approximation of ∇Y(k), which denotes the gradient of the cost function with
respect to the parameters  evaluated at previous iterate k. The stochastic gradient gk(k) in SPSA is
calculated by
𝑔𝑘 (𝜃𝑘 ) =
𝑌(𝜃𝑘 + 𝑐𝑘 ∆𝑘 ) − 𝑌(𝜃𝑘 − 𝑐𝑘 ∆𝑘 )
2𝑐𝑘 ∆𝑘
∆𝑘 is a p-dimensional perturbation vector generated by Monte Carlo method where each of the p components
of ∆𝑘 are independently generated from a zero mean probability distribution. Here ∆𝑘 is sampled from a
Bernoulli ± distribution with probability of 0.5 for each ±1 outcome. The gain sequences (ak and ck) are defined
as
𝑎𝑘 =
𝑎
𝑐
, 𝑐𝑘 =
𝛼
(𝐴 + 𝑘 + 1)
(𝑘 + 1)𝛾
The choice of gain sequences (ak and ck) is critical for the performance of SPSA. It is observed that in a high
noise setting it is necessary to pick smaller a and larger c than in a low noise setting. The asymptotically
optimal values of  and  are 1.0 and 1/6, respectively (Fabian 1971;Chin 1997). As a rule of thumb (with the
Bernoulli ±1distribution for the elements of ∆𝑘 ), it is effective to set c at a level approximately equal to standard
deviation of the noise in model cost function Y from equation 1. However from Figure 3.2, it is observed that
model is highly nonlinear and so is the model cost function Y.
3.4 Cost function (Model data misfit)
It is assumed that the difference between data and simulation results (model-data misfit) is only due to
measurement errors and incorrectly prescribed model parameters. The prior information on the control
parameters is quantified by a Gaussian probability density function (PDF) in control space and observational
information by a PDF in the space of observations. The problem of estimation is then solved by minimizing the
cost function Y with respect to the parameters
Y = 0.5[(M(x) − d)TCd −1 (M(x) − d)+(x − x0)TC0−1(x − x0)]
Page 7
Here M(x) denotes the model operated as a mapping of the parameters onto observations used. d is the
observation vector and x0 is model’s background state vector. Cd and C0 represents uncertainities in the
observations and the model respectively (Blessing et al 2013). Due to the chaotic nature of the model, annual
means of model data differences are used. In equation 1, the first term is the model-data misfit and second
term
represents
the
prior
information
(background
term).
Since
the
number
of
observations
(64*32*365*16=11960320) are more than control parameters (15), the influence of background term will be
much smaller than model-data misfit. Considering the negligible contribution of background term, we set the
background term to zero i.e. we are not using prior information on the control variables in this study.
Table 3.2. Variables contributing for cost function computation. The second column list standard deviation (STD) that are
globally averaged values computed from annual averaged data sets spanning (1979-2012)
In this study, the cost function comprises of 16 variables which are listed in Table 3.2 together with their
corresponding globally averaged STD (C0). C0 is computed from ERA-Interim reanalysis (Dee and Uppala et al
2011) using the annual averages of variables over a period of 34 years (1979-2012). For real data
experiments, ERAInterim reanalysis data is used and therefore variances are calculated using ERA-Interim
data.
3.6 Sensitivity Analysis
In order to see the effect of control parameters used in various parameterization procedures on model’s cost
function, 15 control parameters are chosen for SA. These parameters are listed in Table 3.1 along with their
definitions. Cloud and radiation parameters are commonly chosen optimization procedures (Annan et al 2005;
Severijins and Hazeleger 2005) as they are mainly responsible for accurate temperature and precipitation
Page 8
simulations. Recently, (Blessing et al 2013) used diffusion time scale related parameters for the coupled model
in their optimization experiments using 4D var method. In this study, parameters involved in the
parameterization of long wave; short wave radiation; cloud parameters and vertical diffusion time scales have
been used.
The parameters tfrc1 and tfrc2 are Rayleigh friction time scales in level 10 and level 9 respectively. The
parameters tdissz, tdissd and tdisst represents time scale of damping in the atmosphere and are used in
horizontal diffusion parameterizations. The parameters vdiff lamm, vdiff b, vdiff c and vdiff d are used in
calculation of mixing lengths, drag and transfer coefficients. The parameters tpofmt, th2oc, tswr1, tswr2, tswr3
and acllwr are used in cloud parameterizations. A high value of th2oc increases the absorption of radiation in
the atmosphere leading to increase in surface temperature. This results in more evaporation and hence
increases the moisture in air. The small value of parameter tswr1 increases the cloud albedo and leads to
more precipitation. In order to have balance of precipitation and evaporation these two parameters should be
consistent. The parameters tswr2, tswr3 and acllwr are also responsible for cloud back scattering and
absorption coefficient for clouds.
Figure 3.2. Plot of model cost function vs. perturbation applied to model parameters. The cost function is based on annual
mean pseudo data over a period of 1 year model integration. It also shows the range of minimum cost function that can
achieved if parameter perturbations are not precisely zero. 15 parameters that are used in the procedure are defined in
Table 3.1
Page 9
Sensitivity studies give an idea of the range of perturbation for each parameter that corresponds to a certain
change in cost function. The range of true minimum of cost function in case of identical twin experiment can be
identified from sensitivity plot. Even for shorter time scales (less than 7 days) the cost function doesn’t have a
unique minimum (Figure not shown), however, due to the non- linearity of an atmosphere system, for longer
integration time scales, several secondary minima exist.
For the present sensitivity experiments, the parameters are perturbed (one at a time) from -90% to 800% (with
an interval of 10%) of the original parameter value and the model is integrated for a period of 1 year. The
synthetic data used are annual mean values of temperature, precipitation and heat flux. From Figure 1, the
range of cost function values for all parameters perturbed in above mentioned range varies between 0 and
400. Because of the stochastic nature of the cost function, the true minimum, which is zero in case of an
identical twin experiment, cannot be reached; instead the minimum can only be reached by the precise set of
parameter values but any deviation leads to sizable cost function values, this suggests that in identical twin
experiments unless we obtain exact values of original parameters, minimal value of cost function that could be
reached would be between 10 and 20 and further reduction in cost function may not be possible.
3.7 Identical twin Experiments
A default setup is used to generate pseudo data for one year period. The iterative procedure is started using a
perturbed vector of chosen 15 parameters as mentioned in Table 1, known as the control vector. The size of
the perturbation is based upon the sensitivity of cost function to each parameter. As seen in Figure 3.2, the
parameters tfrc1, tfrc2, tdissz, tdissd, tdisst, vdiff b, vdiff c, vdiff d and tpofmt are contributing almost same as
the noise in the cost function and therefore the size perturbation applied to these is smaller (~ 20% of the
original value) as compared to the remaining parameters (~ 200% of the original value). The assimilation is
carried out over 1-year integration time scale. The experiment is successful if we can approximately recover
the default values through the assimilation of the data via reduction in cost function.
The value of a and c in the gain sequence were chosen to be 0.01 and 0.2 respectively. a is used for gain
sequence ak which relates to the inverse of Hessian in the Quasi-Newton algorithm. The value of c is decided
by the variability of the cost function over certain range of perturbations in parameters (Spall 1998). All 15
parameters (normalised to 1) are perturbed simultaneously and using the SPSA method, gradients are
computed and corrections to the parameter are applied.
Page 10
Figure 3.3. Cost function vs. iteration number for selected 15 parameters in the identical twin experiment. Cost function is
computed for a period of 1 year model integration
Figure 3.3 shows the cost function vs. iterations plot. It can be seen that the cost function achieves the
minimum value of 15 in around 120 iterations and then continues to fluctuate about this value further on. As
discussed in section 3.1 this value of cost function lies in the range of an acceptable minimum and it cannot
reduce further. At this stage about 2/3 of the cost corresponds to temperature and 1/3 to the surface fluxes.
Ideally, in identical twin experiments, the reduction in cost function and in Euclidean distance of control vector
to the default values is achieved simultaneously. The parameters are expected to converge to their original
values, however in our case only values for 8 out of 15 parameters were retrieved, while the remaining 7
parameters were not converging to their default values. From the sensitivity experiments curve in Figure 3.2 it
can be seen that the cost function is less sensitive to the parameters, which were not correctly retrieved and
that some parameters were not contributing in the cost function more than the noise level.
3.8 Optimization with ERA-Interim Reanalysis data
The observational data used are from ERA-Interim reanalysis. These datasets are interpolated on model levels
and then averaged to the model horizontal grid. We have used annual means of the temperature (all levels),
large scale precipitation (LSP), convective precipitation (CP), surface sensible heat flux (SSHF), surface latent
heat flux (SLHF), surface solar radiation (SSR) and surface thermal radiation (STR). As expected, with ERAInterim data the cost function has a larger value (order 3 more) as compared to cost function against pseudo
data and therefore the value of a in the algorithm was chosen to be low (~1e-4) for stable results in successive
iterations. The model was integrated for 1 year using default values of parameters and using SPSA, the
gradients with respect to ERA-Interim data were used to apply corrections to parameters iteratively. The cost
function reduces until it reaches a value after which there is no substantial decrease and the cost function
fluctuates around that minimum value. This happens after about 500 iterations and by then the reduction in
Page 11
cost function is around 25%. The main contribution for reduction in cost function is from tfrc1, tpofmt, th2oc,
tswr1 and tswr3 control parameters. Remaining parameters contribution is significantly less as compared to
mentioned parameters.
Figure 3.4. Total cost function (using ERA-Interim renalysis data) vs iteration number for three ensemble members.
The robustness of the experiment results is assessed by an ensemble with three members, each of which start
from a different point in control space. Three ensembles (Ens-1, Ens-2 and Ens-3) using different initial values
of control parameters are carried out with 15 control parameters. In Ens-1, Ens-2 and Ens-3 the value of each
control parameter is increased by 0% (default parameter values), 60% and 100% respectively. Figure 3.4
shows the reduction in cost function with successive model iterations for three ensembles. All three cases
converging to same value of the cost function gives an impression that the technique is robust.
The performance of optimization was judged by means of Root Mean Square Error (RMSE) normalized by the
prior error which is the same as the square root of normalized cost function number of observations (Figure
3.5). The model state resulting after using optimized parameters shows improvement on the original state
when compared to ERA-Interim reanalysis. The total RMSE which was ~ 3.1 in the original run is reduced by ~
16% to 2.6 while for the individual variables the reduction is 5%, 11% and 18% for precipitation, net heat flux
and surface temperature, respectively.
Page 12
Figure 3.5. RMSE of global mean Temperature between ERA-Interim and original and optimized model temperature. Y
axis denotes height of the atmosphere represented as sigma levels with model level 10 corresponding to the bottom and
model level 1 corresponding to the top of the atmosphere
RMSE of global mean temperature profile from original and an optimized state is shown in Figure 3.5. The
errors have reduced at almost all levels barring level 2. The levels of maximum errors experienced significant
reduction. Overall there is ~10-20% of reduction of error in temperature at various levels in the atmosphere.
Page 13
Figure 3.6. Root mean squared differences (RMSD) of ERA-Interim with Original (top panels) and Optimized (bottom
panels) model near surface temperature.
Spatial distribution of errors of surface temperature from original and optimized state is shown in Figure 3.6.
Most of error reduction in surface temperature takes place in the equatorial West Pacific, North Atlantic and
Southern Indian Ocean regions. Net heat flux shows improvement almost in every region with large errors of >
50 Watts/m2 coming down within 10 Watts/m2 in the North Pacific, south-east Africa and also in South Atlantic.
There are some places where optimization has resulted in larger errors, most prominent of these are the
regions around Greenland in the North Atlantic and in the eastern equatorial Indian Ocean.
Page 14
4. References
Annan, J. D., et al. "Parameter estimation in an atmospheric GCM using the ensemble Kalman filter."
Nonlinear processes in geophysics 12.3 (2005): 363-371.
Blessing and co-authors; Testing variational estimation of process parameters and initial conditions of an Earth
System Model, Tellus, 66:22606. DOI: 10.3402/tellusa.v66.22606
Chin, Daniel C. ”Comparative study of stochastic algorithms for system optimization based on gradient
approximations.” Systems, Man, and Cybernetics, Part B: Cybernetics, IEEE Transactions on 27.2
(1997): 244-249.
Dee, D. P., et al. ”The ERA Interim reanalysis: Configuration and performance of the data assimilation system.”
Quarterly Journal of the Royal Meteorological Society 137.656 (2011): 553-597.
Fabian, Vclav. Stochastic approximation. Department of Statistics and Probability, Michigan State University,
1971.
Spall, James C. ”Implementation of the simultaneous perturbation algorithm for stochastic optimization.”
Aerospace and Electronic Systems, IEEE Transactions on 34.3 (1998): 817- 823.
Severijns C. A., and W. Hazeleger. Optimizing parameters in an atmospheric general circulation model.
Journal of climate 18.17 (2005): 3527-3535.
Uppala, Sakari M., et al. ”The ERA40 reanalysis.” Quarterly Journal of the Royal Meteorological Society
131.612 (2005): 2961-3012.
5. Dissemination and uptake
Plan for future publication
In
preparation
OR
submitted?
Title
Main author
All authors
Title of the
periodical or the
series
Is/Will open access[2]
provided to this
publication?
To be
submitted
Parameter
optimization in an
atmospheric GCM
using
Simultaneous
Perturbation
Stochastic
Approximation
(SPSA)
Variational
assimilation of SSMI
sea ice concentration
in a coupled Arctic
sea-ocean model
Agrawal, R., A. Köhl,
D. Stammer
Tellus
Will be open access
Nikolay V. Koldunov,
Armin Köhl, Nuno
Serra, Detlef
Stammer
J. Geophys. Res
Will be open access
To be
submitted
Page 15
6. The delivery is delayed:
 Yes
 No
7. Changes made and difficulties encountered, if any
Difficulties were encountered during the usage of the adjoint method for the estimation of parameters in chaotic
atmospheric models. For this reason, as explained above a different Method the Simultaneous Perturbation
Stochastic Approximation (SPSA) method was employed, which showed success and enabled the estimation
of parameters.
8. Efforts for this deliverable
How many person-months have been used up for this deliverable?
Partner
Person-months
25
25
UHAM
Period covered
1 Nov.2012-30 October 2014
Total
Total estimated effort for this deliverable (DOW) was 25 person-months. The names of the scientist involved
are mentioned on the cover page of this report.
9. Sustainability
Parameter estimation with chaotic climate models is a challenging task and topic of ongoing research. Despite
the momentary switch from the adjoint to the SPSA method, which was needed to cope with the chaotic
dynamics of the atmospheric component of CESAM, we will continue to explore alternative modifications to the
adjoint method.
10. Dissemination and uptake activities
The complete list of dissemination activities is available on the NACLIM website
http://naclim.zmaw.de/Dissemination.2509.0.html and on the ECAS portal, section “List of Dissemination
Activities”.
Uptake by the targeted audience
According to the DOW, your audience for this deliverable is:
X
The general public (PP)
The project partners, including the Commission services (PU)
A group specified by the consortium, including the Commission services (RE)
This reports is confidential, only for members of the consortium, including the Commission services (CO)
How are you going to ensure the uptake of this deliverable by the targeted audience?
The work will be published in in scientific journals accessible to scientist and the general public.
A summary for the general public will be made available on the website as soon as the work is published.
Page 16
Download