Bayesian Processor of Ensemble Members: combining the ...

1
Bayesian Processor of Ensemble Members: combining the Bayesian Processor of Output with Bayesian Model Averaging for reliable ensemble forecasting R. Marty 1,2 V. Fortin 2 H. Kuswanto 3 A.-C. Favre 4 E. Parent 5 1 Universit´ e Laval, Qu´ ebec, Canada 2 Environnement Canada, Dorval, Canada 3 Institute Technology of Sepuluh Nopember, Indonesia 4 Grenoble INP/LTHE, Grenoble, France 5 AgroParisTech/INRA, Paris, France Why Calibrate Ensemble Predictions ? I Ensemble forecasts: a mean to assess the uncertainty in meteorological forecasts Unfortunately uncertainty underestimated by current ensemble prediction systems (EPS) Unfortunately ensemble often provided at unsuitable spatial/temporal scales (e.g. for hydrological predictions) Statistical post-processing required to obtain reliable ensemble forecasts at appropriate scales Verification Scores Continuous Ranked Probability Score I Comparison of cumulative distribution functions from ensemble forecasts F, with observation y through the Heaviside function I Negatively oriented (smaller is better) I See details in Hersbach (2000) CRPS{F , y } = Z -∞ {F (u ) - H (u - y )} 2 du Reliability I Statistical consistency between a priori predicted probabilities and a posteriori observed frequencies of the occurrence I Reliability measured by the Reliability Component of the CRPS decomposition (Hersbach, 2000) Observation and Forecast Datasets Canadian Daily Climate Data I Temperature observation from meteorological station in Jean Lesage Intl. Airport (YQB) in Quebec City over the period 1978-2007 I Temperature observed at 00Z I http://www.climat.meteo.gc.ca North American Ensemble Forecasting System I 21 members from the Global Ensemble Prediction System of CMC based on the Global Environmental Multiscale (GEM) model I 21 members from the ensemble configuration of the Global Forecast System (GFS) of NCEP I Forecasts provided on a 1-degree grid I Runs 00Z, lead times +24h to +384h by 24h I from April 2008 to March 2009 observations vs ensemble forecast (Quebec City) Verification of raw ensemble forecasts I GFS: lowest skill for all lead times I GEM+GFS: skill closed to GEM No additional benefit in forecasting by using 42 members I Ensemble predictions less skillful than climatology I Ensemble predictions are not reliable Calibration of ensemble predictions from GEM Skill (CRPS) Reliability (CRPS Rel) (y) log scale Notations and Hypotheses Notations t valid date y t predictand (quantity to predict) S ensemble size X (h) t ensemble forecasts h forecast’s lead time X (h) t = {X (h) t ,s , s =1, 2, ..., S } temporal independance: h is omitted Assumptions I Ensemble members generated in the same way exchangeability I Numerical model well suited for predicting an unobserved (latent) variable ξ t (e.g. gridded temperature) I Latent variable ξ t exchangeable with all ensemble members X t ,s I ξ t contains all the information required to predict y t X t and y t are conditionally independant BMA Component (Raftery et al., 2005) I Law of total probability p (y t |X t )= Z p (y t |ξ t , X t )p (ξ t |X t )d ξ t I As X t and y t are conditionally independant p (y t |X t )= Z p (y t |ξ t )p (ξ t |X t )d ξ t I Since the latent variable is exchangeable with ensemble members p (ξ t |X t ) 1 S S X s =1 δ (ξ t - X t ,s ) BMA framework: Non-parametric approximation of the predictive distribution p (y t |X t ) 1 S S X s =1 p (y t |X t ,s ) BPO Component (Krzysztofowicz, 2004) I Bayes’ rule p (y t |ξ t ) | {z } posterior p (ξ t |y t ) | {z } likelihood p (y t ) | {z } prior I Prior distribution of the predictand = climatology . Temperature: Gaussian distribution . Estimated from the past 30 years (1978–2007) with a moving window of 5 days around the valid date p (y t )= N y t ; μ, σ 2 I Likelihood function . Assuming linear model with Gaussian residuals between the predictand and the latent variable ξ t = α + β y t + t t ∼N (02 |y l 1) . But ξ t not observable ¯ X t = α + β y t + η t η t ∼N (02 ¯ X |y l 1) . Bayesian specification of the linear regression with informative conjugate priors defining 4 hyperparameters . Bayesian point estimators α, β and σ 2 ¯ X |y : a compromise between a priori and least squares estimates . Less information in ¯ X t than in ξ t : σ 2 ¯ X |y is an upper bound of σ 2 |y Optimal variance σ 2 |y estimated by minimizing the CRPS BPO framework: p (y t |ξ t )= N (y t ; μ y |ξ 2 y |ξ ) μ y |ξ = σ 2 β 2 ( ξ t -α β )+ σ 2 |y μ σ 2 β 2 + σ 2 |y σ 2 y |ξ = σ 2 σ 2 |y σ 2 β 2 + σ 2 |y Bayesian Processor of Ensemble Members Combining the BMA and BPO frameworks p (y t |X t ) 1 S S X s =1 N (y t ; μ y |X t ,s 2 y |X t ,s ) μ y |X t ,s = σ 2 β 2 ( X t ,s -α β )+ σ 2 |y μ σ 2 β 2 + σ 2 |y σ 2 y |X t ,s = σ 2 σ 2 |y σ 2 β 2 + σ 2 |y I μ y |X t ,s : weighted average of bias-corrected member X t ,s -α β and of the prior mean from climatology μ I σ 2 y |X t ,s : weighted mixture of residuals variance of the linear model σ 2 |y and of the climatological variance σ 2 I The predictive distributions’s shape depends on the empirical distribution of the ensemble members the predictive distribution is not necessary Gaussian ◦• GEM / calibrated members Predictive distribution Observation Constituent distribution Identification of the Optimum Training Length I training period: 10-days to 50-days joint samples I No significant gain in forecasts’s skill beyond 15 days I Calibrated forecasts less skillful than climatology-based forecasts for longer-range lead times I Optimal training length: 15 days I Short training length: limit effects of seasonality and frequent changes to operational forecasting systems Illustration of the BPEM Method 24-hour ahead forecast valid the 10th of July 2008 Prior distribution I Climatology: observations from the past 30 years (1978–2007) inside a moving window of 5 days around the valid date I No seasonality effect nor climatological trend I Estimated parameters: ˜ μ = 20.64 and ˜ σ =3.51 Likelihood function I Bayesian Linear Regression estimated from the joint sample (ensemble means and observations) over the training window I Hyperparameters estimated from least squares linear regressions, in the same season, using joint samples of closed lead time with the same training length I Estimated parameters: ˜ α = 10.17, ˜ β =0.40 and ˜ σ 2 X |y =2.22 Optimal error variance of the latent variable I Standard error of the latent variable determined by grid search I Minimisation of CRPS within the interval [0.0, 2.22] I Optimal error variance: ˜ σ 2 |y =1.20 ◦• GEM and calibrated ensemble members Empirical distribution of GEM members Predictive distribution Climatological distribution Observation Verification of Calibrated Ensemble Forecasts I Verification performed on the summer season (July and August 2008) I Optimal training length for each calibration method I Forecasts’ skill improved by calibration (up to +192h) with similar pattern for each calibration method I Significant improvment of forecasts’ reliability Conclusions on BPEM ... a new approach to calibrated ensemble forecasts I Based on BMA and BPO frameworks I Capable to generate reliable forecasts I Outperforms slightly both the BMA and BPO approaches as well as a climatology I Short optimal training length: avoid negative impacts of seasonality and of frequent changes to operational forecasting systems I Successfully applied to 7 other stations accross the Quebec Some References Hersbach, H. (2000). Decomposition of the Continuous Ranked Probability Score for ensemble prediction systems. Weather and Forecasting 15 (5), 559–570. Krzysztofowicz, R. (2004). Bayesian Processor of Output: a new technique for probabilistic weather forecasting. In 17th Conference on Probability and Statistics in the Atmospheric Sciences, Volume 4.2, Seattle, Washington, USA. American Meteorological Society. Raftery, A. E., T. Gneiting, F. Balabdaoui, and M. Polakowski (2005). Using Bayesian Model Averaging to calibrate forecast ensembles. Monthly Weather Review 133 (5), 1155–1174. 12-16 November 2012, Toulouse, France International Conference on Ensemble Methods in Geophysical Sciences [email protected]

Transcript of Bayesian Processor of Ensemble Members: combining the ...

Page 1: Bayesian Processor of Ensemble Members: combining the ...

Bayesian Processor of Ensemble Members: combining theBayesian Processor of Output with Bayesian Model

Averaging for reliable ensemble forecastingR. Marty1,2 V. Fortin2 H. Kuswanto3 A.-C. Favre4 E. Parent5

1Universite Laval, Quebec, Canada 2Environnement Canada, Dorval, Canada 3Institute Technology of Sepuluh Nopember, Indonesia4Grenoble INP/LTHE, Grenoble, France 5AgroParisTech/INRA, Paris, France

Why Calibrate Ensemble Predictions ?

I Ensemble forecasts: a mean to assess the uncertainty inmeteorological forecasts

å Unfortunately uncertainty underestimated by current ensembleprediction systems (EPS)

å Unfortunately ensemble often provided at unsuitable spatial/temporalscales (e.g. for hydrological predictions)

å Statistical post-processing required to obtain reliable ensembleforecasts at appropriate scales

Verification Scores

Continuous Ranked Probability Score

I Comparison of cumulativedistribution functions fromensemble forecasts F, withobservation y through theHeaviside function

I Negatively oriented (smaller isbetter)

I See details in Hersbach (2000)

CRPS{F , y} =

∫ ∞−∞{F (u)− H(u − y)}2 du

Reliability

I Statistical consistency between a priori predicted probabilities and aposteriori observed frequencies of the occurrence

I Reliability measured by the Reliability Component of the CRPSdecomposition (Hersbach, 2000)

Observation and Forecast Datasets

Canadian Daily Climate Data

I Temperature observation from meteorological station in Jean LesageIntl. Airport (YQB) in Quebec City over the period 1978-2007

I Temperature observed at 00Z

I http://www.climat.meteo.gc.ca

North American Ensemble Forecasting System

I 21 members from the GlobalEnsemble Prediction Systemof CMC based on the GlobalEnvironmental Multiscale(GEM) model

I 21 members from theensemble configuration of theGlobal Forecast System (GFS)of NCEP

I Forecasts provided on a1-degree grid

I Runs 00Z, lead times +24h to+384h by 24h

I from April 2008 to March2009

observations vs ensemble forecast (Quebec City)

Verification of raw ensemble forecasts

I GFS: lowest skill for all lead times

I GEM+GFS: skill closed to GEM

å No additional benefit in forecasting by using 42 members

I Ensemble predictions less skillful than climatology

I Ensemble predictions are not reliable

å Calibration of ensemble predictions from GEM

Skill (CRPS)

Reliability (CRPS Rel)

+ (y) log scale

Notations and Hypotheses

Notations

t valid date yt predictand (quantity to predict)

S ensemble size X(h)t ensemble forecasts

h forecast’s lead time X(h)t = {X (h)

t,s , s = 1, 2, ..., S}temporal independance: h is omitted

Assumptions

I Ensemble members generated in the same way

å exchangeability

I Numerical model well suited for predicting an unobserved (latent)variable ξt (e.g. gridded temperature)

I Latent variable ξt exchangeable with all ensemble members Xt,sI ξt contains all the information required to predict yt

å Xt and yt are conditionally independant

BMA Component (Raftery et al., 2005)

I Law of total probability

p(yt|Xt) =

∫p(yt|ξt,Xt)p(ξt|Xt)dξt

I As Xt and yt are conditionally independant

p(yt|Xt) =

∫p(yt|ξt)p(ξt|Xt)dξt

I Since the latent variable is exchangeable with ensemble members

p(ξt|Xt) ≈1

S

S∑s=1

δ(ξt − Xt,s)

å BMA framework: Non-parametric approximation of the predictivedistribution

p(yt|Xt) ≈1

S

S∑s=1

p(yt|Xt,s)

BPO Component (Krzysztofowicz, 2004)

I Bayes’ rulep(yt|ξt)︸ ︷︷ ︸posterior

∝ p(ξt|yt)︸ ︷︷ ︸likelihood

p(yt)︸︷︷︸prior

I Prior distribution of the predictand = climatology. Temperature: Gaussian distribution. Estimated from the past 30 years (1978–2007) with a moving

window of 5 days around the valid date

p(yt) = N(yt ;µ, σ

2)

I Likelihood function. Assuming linear model with Gaussian residuals between the

predictand and the latent variable

ξt = α + βyt + εt εt ∼ N (0, σ2ε|y l1)

. But ξt not observable

Xt = α + βyt + ηt ηt ∼ N (0, σ2X |y l1)

. Bayesian specification of the linear regression with informativeconjugate priors defining 4 hyperparameters

. Bayesian point estimators α, β and σ2X |y : a compromise between a

priori and least squares estimates. Less information in Xt than in ξt : σ

2X |y is an upper bound of σ2

ε|yå Optimal variance σ2

ε|y estimated by minimizing the CRPS

å BPO framework:p(yt|ξt) = N (yt ;µy |ξ, σ

2y |ξ)

µy |ξ =σ2β2(ξt−αβ ) + σ2

ε|yµ

σ2β2 + σ2ε|y

σ2y |ξ =

σ2σ2ε|y

σ2β2 + σ2ε|y

Bayesian Processor of Ensemble Members

Combining the BMA and BPO frameworks

p(yt|Xt) ≈1

S

S∑s=1

N (yt ;µy |Xt,s , σ2y |Xt,s)

µy |Xt,s =σ2β2(

Xt,s−αβ ) + σ2

ε|yµ

σ2β2 + σ2ε|y

σ2y |Xt,s =

σ2σ2ε|y

σ2β2 + σ2ε|y

I µy |Xt,s : weighted average of bias-corrected memberXt,s−αβ and of

the prior mean from climatology µ

I σ2y |Xt,s : weighted mixture of residuals variance of the linear model

σ2ε|y and of the climatological variance σ2

I The predictive distributions’s shape depends on the empiricaldistribution of the ensemble members

å the predictive distribution is not necessary Gaussian

◦ • GEM / calibrated members Predictive distribution

Observation Constituent distribution

Identification of the Optimum Training Length

I training period: 10-daysto 50-days joint samples

I No significant gain inforecasts’s skill beyond 15days

I Calibrated forecasts lessskillful thanclimatology-basedforecasts for longer-rangelead times

I Optimal training length:15 days

I Short training length: limit effects of seasonality and frequentchanges to operational forecasting systems

Illustration of the BPEM Method

24-hour ahead forecast valid the 10th of July 2008

Ê Prior distribution

I Climatology: observationsfrom the past 30 years(1978–2007) inside a movingwindow of 5 days around thevalid date

I No seasonality effect norclimatological trend

I Estimated parameters:µ = 20.64 and σ = 3.51

Ë Likelihood function

I Bayesian Linear Regressionestimated from the jointsample (ensemble means andobservations) over thetraining window

I Hyperparameters estimatedfrom least squares linearregressions, in the sameseason, using joint samples ofclosed lead time with thesame training length

I Estimated parameters:α = 10.17, β = 0.40 andσ2X |y = 2.22

Ì Optimal error variance of thelatent variable

I Standard error of the latentvariable determined by gridsearch

I Minimisation of CRPS withinthe interval [0.0, 2.22]

I Optimal error variance:σ2ε|y = 1.20

◦ • GEM and calibrated ensemble members

Empirical distribution of GEM members

Predictive distribution

Climatological distribution

Observation

Verification of Calibrated Ensemble Forecasts

I Verification performed on the summer season (July and August 2008)

I Optimal training length for each calibration method

I Forecasts’ skill improved by calibration (up to +192h) with similarpattern for each calibration method

I Significant improvment of forecasts’ reliability

Conclusions on BPEM

... a new approach to calibrated ensemble forecasts

I Based on BMA and BPO frameworks

I Capable to generate reliable forecasts

I Outperforms slightly both the BMA and BPO approaches as well asa climatology

I Short optimal training length: avoid negative impacts of seasonalityand of frequent changes to operational forecasting systems

I Successfully applied to 7 other stations accross the Quebec

Some References

Hersbach, H. (2000). Decomposition of the Continuous RankedProbability Score for ensemble prediction systems. Weather andForecasting 15 (5), 559–570.

Krzysztofowicz, R. (2004). Bayesian Processor of Output: a newtechnique for probabilistic weather forecasting. In 17th Conference onProbability and Statistics in the Atmospheric Sciences, Volume 4.2,Seattle, Washington, USA. American Meteorological Society.

Raftery, A. E., T. Gneiting, F. Balabdaoui, andM. Polakowski (2005). Using Bayesian Model Averaging tocalibrate forecast ensembles. Monthly Weather Review 133 (5),1155–1174.

12-16 November 2012, Toulouse, France International Conference on Ensemble Methods in Geophysical Sciences [email protected]