<<

AUGUST 2006 NOTES AND CORRESPONDENCE 2279

NOTES AND CORRESPONDENCE

Improving Week-2 Forecasts with Multimodel Reforecast Ensembles

JEFFREY S. WHITAKER AND XUE WEI NOAA–CIRES Climate Diagnostics Center, Boulder, Colorado

FRÉDÉRIC VITART Seasonal Forecasting Group, ECMWF, Reading, United Kingdom

(Manuscript received 1 August 2005, in final form 3 November 2005)

ABSTRACT

It has recently been demonstrated that (MOS) computed from a long retrospective dataset of ensemble “reforecasts” from a single model can significantly improve the skill of probabilistic week-2 forecasts (with the same model). In this study the technique is extended to a multimodel reforecast dataset consisting of forecasts from ECMWF and NCEP global models. Even though the ECMWF model is more advanced than the version of the NCEP model used (it has more than double the horizontal resolution and is about five years newer), the forecasts produced by the multimodel MOS technique are more skillful than those produced by the MOS technique applied to either the NCEP or ECMWF forecasts alone. These results demonstrate that the MOS reforecast approach yields benefits for week-2 forecasts that are just as large for high-resolution state-of-the-art models as they are for relatively low resolution out-of- date models. Furthermore, operational forecast centers can benefit by sharing both retrospective reforecast datasets and real-time forecasts.

1. Introduction than the operationally produced forecasts from NCEP’s Climate Prediction Center when computed using fore- Operational weather prediction centers strive to im- cast-error statistics from the reforecast dataset. prove forecasts by continually improving weather fore- Ensembles of forecasts with a single model suffer cast models, and the analyses used to initialize those from a deficiency in spread partly because they do not models. In a recent study, Hamill et al. (2004) investi- represent the error in the forecast model itself. This gated an alternative approach to improving forecasts, causes probabilistic forecasts derived from such en- namely, using the statistics of past forecast errors to sembles to be overconfident. Statistical postprocessing correct errors in independent forecasts. They used a using reforecast datasets can ameliorate this problem to dataset of retrospective ensemble forecasts (or “refore- some extent, yielding reliable probabilities (e.g., Hamill casts”) with a 1998 version of the operational global et al. 2004). However, the use of multimodel ensembles forecast model from the National Centers for Environ- (ensembles consisting of forecasts from different mod- mental Prediction (NCEP; part of the U.S. National els, perhaps initialized from different analyses) have Weather Service). Probabilistic forecasts of surface also been shown to significantly increase the skill of temperature and precipitation in week-2 (days 8–14) probabilistic forecasts (Krishnamurti et al. 1999; Palmer were not skillful relative to when computed et al. 2004; Rajagopalan et al. 2002; Mylne et al. 2002), directly from the model output, but were more skillful even without statistical postprocessing. In this note, we examine whether the results of Hamill et al. (2004) can be improved upon if the reforecast dataset used to sta- Corresponding author address: Jeffrey Whitaker, NOAA– CIRES Climate Diagnostics Center, 325 Broadway, R/CDC1, tistically correct the forecasts consists of more than one Boulder, CO 80305-3328. model. Here we combine the reforecast dataset de- E-mail: [email protected] scribed in Hamill et al. (2006) using a 1998 version of

© 2006 American Meteorological Society

Unauthenticated | Downloaded 09/25/21 03:38 AM UTC

MWR3175 2280 MONTHLY WEATHER REVIEW VOLUME 134 the NCEP Global Forecast System (GFS) model, with condition. The breeding method and the forecast model a reforecast dataset generated at the European Cen- are the same as that used operationally at NCEP in tre for Medium-Range Weather Forecasts (ECMWF) January 1998. Sea surface conditions were specified to support monthly forecasting (Vitart 2004). The from the NCEP–NCAR reanalysis, and were held fixed ECMWF forecasts are run at roughly 1° grid spacing to their initial values throughout the forecast. Further compared with a 2.5° grid spacing used for the NCEP details describing this dataset are available in Hamill et 1998 model forecasts. al. (2004) and Hamill et al. (2006). In this note we address two questions. The first con- An independent set of reforecasts has been gener- cerns whether the large gains in skill reported by ated at the ECMWF to support operational monthly Hamill et al. (2004) were a consequence of the fact that forecasting (Vitart 2004). The atmospheric component the forecast model was run at a relatively low resolution of the forecast model is a T159 (roughly 1° grid) version and is now nearly 6 yr behind the state-of-the art model. of the ECMWF Integrated Forecast System (IFS) with In other words, do the benefits of statistically correcting 40 levels in the vertical. The oceanic component is the forecasts using reforecasts datasets found in that study Hamburg Ocean Primitive Equation model. Atmo- also apply to newer, higher-resolution models? Second, spheric initial conditions were taken from the ERA-40 we investigate whether the methods described in reanalysis, and ocean conditions were taken from the Hamill et al. (2004) can be applied to multimodel en- ocean data assimilation system used to produce sea- sembles. Specifically, we seek to address the question of sonal forecasts at ECMWF. A five-member ensemble whether statistically corrected multimodel forecasts are was run out to 32 days once every 2 weeks for a 12-yr more skillful than the forecasts generated from the period from 27 March 1990 to 18 June 2002. Initial component models (using the statistics from the com- atmospheric perturbations were generated using the ponent reforecast datasets). same singular vector technique used in ECMWF opera- tions (Buizza and Palmer 1995). Further details describ- ing the ECMWF monthly forecasting system can be 2. Datasets and methodology found in Vitart (2004). a. The analyses A combined reforecast dataset was created by sub- sampling five members from the NCEP CDC refor- Forecasts for 850-hPa temperature are presented in ecasts for just those dates in December, January, and this note. These forecasts have been verified using both February for which the ECMWF forecasts were avail- the NCEP–National Center for Atmospheric Research able (a total of 84). Only week-2 (8–14-day average) (NCAR) reanalysis (Kistler et al. 2001) and the 40-yr forecast means for 850-hPa temperature in the North- ECMWF Re-Analysis (ERA-40; Uppala et al. 2005). ern Hemisphere poleward of 20°N were included. Both of these analyses are on a 2.5° grid, and only those points poleward of 20°N are included. Because the veri- c. Methodology fication statistics computed with the NCEP–NCAR and Three-category probability forecasts for 850-hPa ERA-40 analyses are so similar, all of the results de- temperature in week 2 were produced for the Northern scribed here have been computed using the ERA-40 Hemisphere poleward of 20°N. The three categories reanalyses unless otherwise noted. are the lower, middle, and upper terciles of the clima- tological distribution of analyzed anomalies, so that the b. The forecasts climatological forecast for each category is always 33%. A reforecast dataset was created at the National Oce- The climatological distribution is defined for all winter anic and Atmospheric Administration (NOAA) Cli- (December–February) seasons from 1971 to 2000. The mate Diagnostics Center (CDC), using a T62 resolution climatological mean for each day is computed by (roughly 2.5° grid spacing) version of NCEP’s GFS smoothing the 30-yr mean analyses for each calendar model, which was operational until January 1998. This day with a 31-day running mean. The terciles of the model was run with 28 vertical sigma levels. A 15- climatological distribution are calculated using the ana- member ensemble run out to 15 days lead time is avail- lyzed anomalies over a 31-day window centered on able for every day from 1979 to the present, starting each calendar day. Further details may be found in from 0000 UTC initial conditions. The ensemble initial Hamill et al. (2004). conditions consisted of a control initialized with the Tercile probability forecasts are generated from the NCEP–NCAR reanalysis (Kistler et al. 2001) and a set raw five-member ensemble output for each model by of seven bred pairs of initial conditions (Toth and Kal- simply counting the number of members that fall into nay 1997), centered each day on the reanalysis initial each category at each grid point, and then dividing that

Unauthenticated | Downloaded 09/25/21 03:38 AM UTC AUGUST 2006 NOTES AND CORRESPONDENCE 2281 number by 5. The NCEP 15-member ensemble was subsampled by taking the control and the first two (out of seven) positive and negative perturbations generated by the breeding method (Toth and Kalnay 1997). These uncorrected, or “raw” ensemble probabilities are com- pared with probabilities computed using a technique called logistic regression (Wilks 1995), which is de- scribed in detail in Hamill et al. (2004). In short, as implemented here, the logistic regression predicts the tercile probabilities given the ensemble mean forecast anomaly (relative to the observed climatology). As dis- cussed in Hamill et al. (2004), we have found adding other predictors, such as ensemble spread, into the lo- gistic regression does not improve forecast skill. Soft- ware from the Numerical Algorithms Group (NAG; more information available online at http://www.nag. co.uk) was used to compute the regression coefficients (NAG library routine G02GBF). These coefficients are computed at each grid point using cross validation (i.e., for each of the 84 forecast times, the other 83 forecast– verification pairs are used to compute the regression coefficients). Since the forecasts are separated by 2 weeks, the correlation between adjacent forecasts is quite small and each forecast is essentially independent of the others. Probabilities are computed using logistic regression using each model’s ensemble mean anomaly separately, and then a multimodel forecast probability is produced by using a combined ECMWF–NCEP en- semble mean anomaly as the predictor in the logistic regression. The combined ensemble mean anomaly is computed using ϭ ϩ ϩ Fcombined b0 bNCEPFNCEP bECMWFFECMWF , ͑1͒ where the FNCEP is the NCEP ensemble mean, FECMWF FIG. 1. Reliability diagrams for week-2 tercile probability fore- casts of 850-hPa temperature. A total of 84 forecasts during De- is the ECMWF ensemble mean, Fcombined is the multi- model ensemble mean, b is the weight given to the cember–February 1990–2002 are used, and forecasts are evaluated NCEP for the Northern Hemisphere poleward of 20°N using the ERA-40 NCEP ensemble mean, bECMWF is the weight given to reanalysis for verification. The reliability curve shown is for both the ECMWF ensemble mean, and b0 is a constant term. upper and lower tercile probability forecasts. Inset histograms The coefficients (b0, bNCEP, and bECMWF) are estimated indicate frequency with which extreme tercile probabilities were at each Northern Hemisphere grid point using an itera- issued. Forecasts from (a) the five-member ECMWF model en- tive least squares regression algorithm (NAG library semble and (b) the five-member NCEP model ensemble. routine G02GAF). Instead of combining the ensemble means and using 3. Results the result as a predictor in the logistic regression, a multimodel forecast can be computed by simply using Figure 1 is a reliability diagram summarizing the skill each ensemble mean as a separate predictor in the lo- of the raw, uncorrected week-2 tercile probability fore- gistic regression. We have found that forecasts obtained casts for each model in the Northern Hemisphere pole- with a two-predictor logistic regression are slightly less ward of 20°N. For both the NCEP and ECMWF mod- skillful, so only results using the two-step procedure els, the reliability of the forecasts is poor (the forecasts (using the combined ensemble mean as a single predic- are overconfident) and the ranked probability skill tor in the logistic regression) are presented here. score (RPSS; see Wilks 1995) is near zero, indicating no

Unauthenticated | Downloaded 09/25/21 03:38 AM UTC 2282 MONTHLY WEATHER REVIEW VOLUME 134

FIG. 3. Same as in Fig. 1, but for multimodel ensemble. The NCEP and ECMWF forecast ensemble means are combined using linear regression, and then the combined ensemble mean is used to predict tercile probabilities using logistic regression. The com- bined ensemble is more skillful than the logistic regression cor- rection applied to either the ECMWF or NCEP models alone (Fig. 2).

ϭ perfect model has no skill (BM 0) then an infinite ensemble should have a BSS of 0.166. However, we have found that the RPSS of the NCEP forecast is still negative for an ensemble size of 15 (the total number of ensemble members available). This suggests that model error is primarily responsible for the low skill of the NCEP model forecasts, not the small ensemble size used to generate the probabilities. Although we cannot test the sensitivity of the RPSS to ensemble size with the ECMWF model (since the reforecasts were only FIG. 2. Same as in Fig. 1, but for forecasts corrected with a run with five members), the fact that the reliability dia- logistic regression. grams shown in Fig. 1 are so similar suggests that the same holds true for the ECMWF forecasts. The lack of reliability shown in Fig. 1 is a consequence of the spread skill relative to the climatological forecast of 33% in deficiency in both models—on average the ensemble each category. Richardson (2001) has shown that the spread is a factor of nearly 2 smaller than the ensemble Brier skill score (BSS; which is the same as the RPSS mean error (not shown). for forecasts of a single category) depends on ensemble Applying a logistic regression to the ensemble means size according to produces forecasts that are more reliable and have B ϩ MϪ1 higher skill for both ensembles (Fig. 2). The corrected ϭ M ͑ ͒ Bϱ Ϫ1 , 2 ECMWF model forecasts are significantly more skillful 1 ϩ M than the corrected NCEP model forecasts (RPSS of where BM is the BSS for an ensemble of size M and Bϱ 0.155 for the ECMWF versus 0.113 for the NCEP). The is the BSS for an infinite ensemble. This equation as- relative improvement in the ECMWF forecasts is sumes the ensemble prediction system is perfect (i.e., nearly the same as the NCEP forecasts, demonstrating the verifying analysis is statistically indistinguishable that the value of having a reforecast dataset is just as from a randomly chosen ensemble member). Accord- large for the ECMWF model as it was for the NCEP ing to this formula, if a five-member ensemble with a model. This is despite the fact that the ECMWF model

Unauthenticated | Downloaded 09/25/21 03:38 AM UTC AUGUST 2006 NOTES AND CORRESPONDENCE 2283

FIG. 4. A map of the weights used to combine the ECMWF and NCEP ensemble mean forecasts. These maps correspond to bNCEP and bECMWF in Eq. (1). The mean term b0 is not shown. has the benefit of twice the resolution and five extra ECMWF forecast. This is obviously only possible if a years of model development. The improvement in skill reforecast dataset is available for both models. is limited by the small sample size—if all 25 yr of fore- All of the results discussed in this section have been casts available for the NCEP model are used to com- computed using the ERA-40 reanalysis as the verifying pute the logistic regression, the RPSS increases to 0.15. analysis. The results are not substantively changed if The multimodel forecast, using the combined ensemble the NCEP–NCAR reanalysis is used instead (not mean in the logistic regression, produces a forecast with shown). an average RPSS of 0.17 (Fig. 3), about a 10% improve- ment over the corrected ECMWF forecast alone. The 4. Discussion fact that the multimodel ensemble forecast is more skillful than the ECMWF forecast, even though the The study of Hamill et al. (2004) has been extended ECMWF forecast is, on average, superior to the NCEP to include more than one forecast model. The results forecast, is a consequence of the fact that there are show that the benefits of reforecasts apply equally to places where the NCEP model is consistently more older, lower-resolution forecasts systems as they do to skillful than the ECMWF model (Krishnamurty et al. higher-resolution state-of-the-art forecast systems. In 2000). This is reflected in the weights used to com- this case, logistic regression was used to combine the bine the two ensemble means (Fig. 4). Although the NCEP and ECMWF week-2 forecasts. The resulting ECMWF model receives more weight on average, there multimodel forecast was more skillful than the statisti- are regions where the NCEP model receives a weight of cally corrected ECMWF forecast, which was run at greater than 0.5 (the red regions in the left panel of Fig. more than twice the horizontal resolution of the other 4). In other words, the NCEP model, though inferior on model, a 1998 version of the NCEP global forecast average to the ECMWF model, supplies independent model. These results clearly demonstrate that all opera- information that can be used to improve upon the tional centers can benefit by sharing both reforecast

Unauthenticated | Downloaded 09/25/21 03:38 AM UTC

Fig 4 live 4/C 2284 MONTHLY WEATHER REVIEW VOLUME 134 datasets and real-time forecasts. This is true even if the analysis: Montly means CD-ROM and documentation. Bull. models run by the various centers differ substantially in Amer. Meteor. Soc., 82, 247–268. resolution, as long as they have some skill and provide Krishnamurti, T. N., C. M. Kishtawal, T. E. LaRow, D. R. Bachio- chi, Z. Zhang, E. C. Williford, S. Gadgil, and S. Surendran, independent information to the statistical scheme used 1999: Improved weather and seasonal climate forecasts from to combine the forecasts. For week-2 tercile probability multimodel superensembles. Science, 285, 1548–1550. forecasts of 850-hPa temperature, the benefits of mul- ——, ——, Z. Zhang, T. LaRow, D. Bachiochi, E. Williford, S. timodel ensembles cannot be realized unless all the Gadgil, and S. Surendran, 2000: Multimodel ensemble fore- models have corresponding reforecast datasets with casts for weather and seasonal climate. J. Climate, 13, 4196– 4216. which to estimate regression equations to combine (and Mylne, K. R., R. E. Evans, and R. T. Clark, 2002: Multi-model calibrate) the forecasts. multi-analysis ensembles in quasi-operational medium-range forecasting. Quart. J. Roy. Meteor. Soc., 128, 361–384. Acknowledgments. Fruitful discussions with Tom Palmer, T. N., and Coauthors, 2004: Development of a European Hamill are gratefully acknowledged. The NOAA multimodel ensemble system for seasonal to interannual pre- diction (DEMETER). Bull. Amer. Meteor. Soc., 85, 853–872. “Weather–Climate Connection” program funded this Rajagopalan, B., U. Lall, and S. E. Zebiak, 2002: Categorical cli- project. mate forecasts through regularization and optimal combina- tion of multiple GCM ensembles. Mon. Wea. Rev., 130, 1792– REFERENCES 1811. Richardson, D. S., 2001: Measures of skill and value of ensemble Buizza, R., and T. N. Palmer, 1995: The singular vector structure prediction systems, their interrelationship and the effect of of the atmospheric global circulation. J. Atmos. Sci., 52, 1434– ensemble size. Quart. J. Roy. Meteor. Soc., 127, 2473–2489. 1456. Toth, Z., and E. Kalnay, 1997: at NCEP and Hamill, T. M., J. S. Whitaker, and X. Wei, 2004: Ensemble refor- the breeding method. Mon. Wea. Rev., 125, 3297–3319. ecasting: Improving medium-range forecast skill using retro- Uppala, S. M., and Coauthors, 2005: The ERA-40 reanalysis. spective forecasts. Mon. Wea. Rev., 132, 1434–1447. Quart. J. Roy. Meteor. Soc., 131, 2961–3012. ——, ——, and S. L. Mullen, 2006: Reforecasts: An important Vitart, F., 2004: Monthly forecasting at ECMWF. Mon. Wea. Rev., dataset for improving weather predictions. Bull. Amer. Me- 132, 2761–2779. teor. Soc., 87, 33–46. Wilks, D. S., 1995: Statistical Methods in the Atmospheric Sciences. Kistler, R., and Coauthors, 2001: The NCEP–NCAR 50-Year Re- Academic Press, 467 pp.

Unauthenticated | Downloaded 09/25/21 03:38 AM UTC