Evaluation of snow models: SnowMIP2 and future high resolution evaluation Nick Rutter

advertisement
Evaluation of snow models:
SnowMIP2 and future high
resolution evaluation
Nick Rutter1, Richard Essery2 and
Hans-Peter Marshall3
n.rutter@sheffield.ac.uk
1
2
3
Department of Geography, University of Sheffield, UK
School of Geosciences, University of Edinburgh, UK
Institute of Arctic and Alpine Research, University of Colorado at Boulder
IP3, Waterloo, Canada, 9 November 2007
1
Outline
New collaborator to IP3
Processes and parameterisation themes: forestsnowpack models (working with Richard Essery)
Links between modelling and remote sensing
SnowMIP2
Evaluation of forest-snow process models
Ground-based FMCW radar
IP3, Waterloo, Canada, 9 November 2007: Outline
2
SnowMIP2: Why, what and how?
Current Land Surface Schemes (LSS) in models either neglect or use
highly simplified representations of physical processes controlling the
accumulation and melt of snow in forests
Snow Model Inter-comparison Project 2 (SnowMIP2)
Quantify uncertainty in simulations of forest snow processes
Range of models of varying complexity (not just LSS)
Primarily evaluate the ability of models to estimate SWE
33 models (CRHM and CLASS involved)
5 locations: 3 presented herein (Switzerland, Canada, USA)
2 sites per location: forest and clearing (open)
IP3, Waterloo, Canada, 9 November 2007 : Why, what and how?
3
Model inputs
Meteorological driving data :
Initialisation data:
Precipitation rate (rain and snow)
Soil temperature profile
Incoming SW and LW
Soil moisture
Air Temperature
Wind Speed
Relative humidity
Site specific data:
Calibration data:
Tree height
In-situ snow water equivalent
Effective LAI
(SWE) from Year 1, forest sites
Instrument heights
Snow free albedo
Soil composition
IP3, Waterloo, Canada, 9 November 2007 : Model inputs
4
Model outputs
Energy fluxes
Radiative, turbulent, conductive, advected and phase changes
Mass fluxes
sublimation, evaporation, transpiration, phase change, infiltration, runoff,
unloading, drip
State variables
mass (solid and liquid), temperature
IP3, Waterloo, Canada, 9 November 2007 : Model outputs
5
Alptal, Switzerland (47°N, 8°E)
OPTIONAL CALIBRATION
UNCALIBRATED
UNCALIBRATED ‘CONTROL’
IP3, Waterloo, Canada, 9 November 2007 : Results
6
BERMS, Saskatchewan, Canada (53°N, 104°W)
IP3, Waterloo, Canada, 9 November 2007 : Results
7
Fraser, Colorado, USA (39°N, 105°W)
IP3, Waterloo, Canada, 9 November 2007 : Results
8
Model uncertainties: forest vs open
IP3, Waterloo, Canada, 9 November 2007: Results
9
Model uncertainties: a Canadian perspective
IP3, Waterloo, Canada, 9 November 2007 : Results
10
Model uncertainties: forest vs open
x 2.16
x 2.42
x 1.59
x 1.35
IP3, Waterloo, Canada, 9 November 2007 : Results
x 1.53
x 1.17
11
Model uncertainties: forest vs open
x 1.02
x 1.40
x 1.05
x 5.35
x 3.00
IP3, Waterloo, Canada, 9 November 2007 : Results
x 1.37
12
Model uncertainties: a Canadian perspective
Models ranked by RMSE
1 = lowest RMSE
33 = highest RMSE
Mean rank
4 sites at each location
Consistency of rank (st dev)
Doesn’t explain a lot!
Need to categorise by process
and compare by rank
Currently finding out more to
enable model categorisation:
by process (canopy, snow, soil)
by calibration, data manipulation
and structure
IP3, Waterloo, Canada, 9 November 2007: Results
13
SnowMIP2: Conclusions
33 models successfully participated in SnowMIP2 – thanks to the hard
work and goodwill of the participants
Low correlation coefficients suggests good model performance at forest
sites do not necessarily mean good model performance at open sites
(and vice versa)
One year of calibration data in forest models is not necessarily good
enough for subsequent years
It is easier to model inter-annual variability at open sites than forest
Current work focussed on analysing whether 1) process representation,
2) strength of calibration or 3) canopy complexity makes inter-annual
variability of forest snow hard to model
IP3, Waterloo, Canada, 9 November 2007 : Conclusions
14
FMCW radar: spatial evaluation
6,511 radar measurements over ~2.5km
IP3, Waterloo, Canada, 9 November 2007: FMCW spatial evaluation
15
FMCW radar: temporal evaluation
Evaluation of 1-D models
Evaluation of 2-D models
Evaluation of radiative
transfer models?
IP3, Waterloo, Canada, 9 November 2007: FMCW temporal evaluation
16
Download