Forecasting wind for the renewable energy market Matt Pocernich Research Applications Laboratory National Center for Atmospheric Research pocernic@ucar.edu Why is forecasting wind hard? Turbulence Inherently stochastic process Issue of scales Consequence – attempts at improving a deterministic forecasts have physical limitations. ENVR Workshop - October 2010 4/13/2015 Results of Spectral Decomposition (Rife, Davis, Liu 2004 MWR) ENVR Workshop - October 2010 4/13/2015 Power Needs of the customer – a typical power curve ENVR Workshop - October 2010 4/13/2015 Outline Components of a typical numerical weather prediction system. Ensembles Forecasting system Methods of post processing Verification of wind forecasts Excitement to come ENVR Workshop - October 2010 4/13/2015 Dynamic Integrated Forecast TM System - DICast Ensemble Input + Dynamic Weighting + Bias Correction + Dynamic MOS = Optimized Forecast Performance ENVR Workshop - October 2010 4/13/2015 4-D Continuous Data Assimilation and Forecasts Modified WRF/MM5: Dx/Dt = ... + GW (xobs – xmodel ) All WMO/GTS GOES where x = T, U, V, Q, P1, P2 … Radars W is weight function Weather observations RTFDDA Wind Prof Regional-scale NWP models WRF / MM5 t WRF/MM5 Cold start FDDA Forecast MESONETs Etc. ACARS ENVR Workshop - October 2010 Yuabo Liu et al. Farm Met Wind Energy Prediction - R & D Workshop. 11 -12 May 2010 4/13/2015 © 2010 University Corporation for Atmospheric Research Assimilation of Wind Farm Data Met Tower wind spd/dir Turbine hub wind spd Data QC and processing Other met-tower weather Observations Data combining and reformat All other weather Observations WRF RTFDDA ENVR Workshop - October 2010 Yuabo Liu et al. Wind Energy Prediction - R & D Workshop. 11 -12 May 2010 4/13/2015 © 2010 University Corporation for Atmospheric Research NWP Forecast Solve the equations that describe the evolution of the atmosphere We cannot solve the equations analytically: •Discretize them. Horizontally and vertically •Close the equations by parameterizing small/fast physical processes. Load Forecasting Workshop 22 March 2007 Load Forecasting Workshop 22 March 2007 Ensemble Forecasting – a very vague term Random initial conditions Multi-physics model Multi-model (Poor man’s ensemble) Time lag ensemble ENVR Workshop - October 2010 4/13/2015 same event, designed to characterize uncertainty Observational error Model selection error Parameterization error Wind Speed (mph) Multiple forecasts of the Forecast Hour centers and forecasting companies ENVR Workshop - October 2010 Wind Speed (mph) Run at many weather 4/13/2015 Forecast Hour Challenges with ensembles Tend to be under dispersive (not enough spread.) Calibration for both reliability and sharpness. Some methods include ensembleBMA (Chris Frayley + UW) quantile regression (Hopson) ensemble Kalman Filter (more later from Luca) ENVR Workshop - October 2010 4/13/2015 Ensemble BMA Bias removal of each member using linear regression. Estimates weights and variance for each ensemble member which minimizes continuous rank probability score. Essentially, dresses each ensemble member with a distribution. Traditionally uses Gaussian distribution. For winds, use gamma. Key work by Adrian Raftery, Tilman Gneiting, MacLean Sloughter and Chris Frayley (UW). ENVR Workshop - October 2010 4/13/2015 Example of ensemble BMA forecasts ENVR Workshop - October 2010 4/13/2015 Regime switching Algorithms (From M. Hering) ENVR Workshop - October 2010 4/13/2015 ENVR Workshop - October 2010 4/13/2015 ENVR Workshop - October 2010 4/13/2015 ENVR Workshop - October 2010 4/13/2015 ENVR Workshop - October 2010 4/13/2015 ENVR Workshop - October 2010 4/13/2015 ENVR Workshop - October 2010 4/13/2015 Key Verification Issues The most common verification metrics are mean absolute error, RMSE and Bias. These do not address concerns like ramping events. New statistical forecasts are created every 15 minutes with new physical model runs every 3 hours. We don’t have a developed concept or metrics for consistency. Forecast value – cost/benefits is complicated. Value of weather forecast is used with load forecast. There are humans in the loop. ENVR Workshop - October 2010 4/13/2015 Contingency table statistics The most fundamental verification methods involve statistics derived from a contingency table. This requires forecasts and observations be categorized into discrete bins. Basic contingency table statistics include hit rate, false positive rate, bias, false negative rate and percent correct. Changes in power can be classified in such a way in the following manner. An increase (or decrease) in a forecast accompanied by a “similar” increase (or decrease) in observed power is a good forecast. A change forecast in power, but not observed is a false positive. A change observed, but not forecast is a false negative A forecast of no-change, associated with no change is considered a good, negative forecast. The definition of a good forecast can be modified. Regions do not have to be defined by angular regions. ENVR Workshop - October 2010 4/13/2015 Forecast vs. Observed Changes Agree in magnitude and direction Disagree in magnitude and direction Small values forecast and observed False Positive and False Negative ENVR Workshop - October 2010 4/13/2015 Regions translated into a contingency table Forecast Observed Up Ramp Neutral Down Ramp Up Ramp Neutral Down Ramp ENVR Workshop - October 2010 4/13/2015 Changes in power forecast by short term forecast from in the first 3 hours Forecast Observed Up Down Neutral Ramp Ramp Up Ramp 970 322 228 Neutral 573 149 603 Down Ramp 300 435 696 Percent Correct 42% Gerrity Skill Score 0.24 ENVR Workshop - October 2010 4/13/2015 0 hour lead time, 1- 3 hour duration GSS = 0.21 PC = 48% ENVR Workshop - October 2010 GSS = 0.21 PC = 39% GSS = 0.24 PC = 42% 4/13/2015 Criticisms of this approach From C. Ferro - U.Exeter The classification of the observation as either neutral, down ramp or up ramp depends on the value of the forecast. That seems weird! It must lead to some difficulties in interpreting any analysis of the table. Much easier to define categories using boundaries that are parallel to the observation and forecast axes. My initial reaction to deal with this is not to use contingency tables at all but to model the continuous joint distribution. ENVR Workshop - October 2010 4/13/2015 More data + better data = more fun New instruments – LIDAR and SODAR Better quality observations from existing stations. Improvements in sharing data. High quality networks of tall towers. (BPA) ENVR Workshop - October 2010 4/13/2015 Space – Time processes? ENVR Workshop - October 2010 4/13/2015 Concluding Remarks ENVR Workshop - October 2010 4/13/2015