<<

FEBRUARY 2006 MARZBAN ET AL. 657

MOS, Perfect Prog, and Reanalysis

CAREN MARZBAN Department of Statistics, University of Washington, Seattle, Washington, and Center for Analysis and Prediction of Storms, University of Oklahoma, Norman, Oklahoma

SCOTT SANDGATHE Applied Physics Laboratory, University of Washington, Seattle, Washington

EUGENIA KALNAY Department of Meteorology, University of Maryland, College Park, College Park, Maryland

(Manuscript received 11 April 2005, in final form 6 July 2005)

ABSTRACT

Statistical postprocessing methods have been successful in correcting many defects inherent in numerical prediction model forecasts. Among them, model output statistics (MOS) and perfect prog have been most common, each with its own strengths and weaknesses. Here, an alternative method (called RAN) is examined that combines the two, while at the same time utilizes the information in reanalysis data. The three methods are examined from a purely formal/mathematical point of view. The results suggest that whereas MOS is expected to outperform perfect prog and RAN in terms of mean squared error, bias, and error variance, the RAN approach is expected to yield more certain and bias-free forecasts. It is suggested therefore that a real-time RAN-based postprocessor be developed for further testing.

1. Introduction whose training set is a blend of “old” and “new” data. Vislocky and Young (1989) even consider perfect prog It is well known that forecasts of numerical weather as a further postprocessor of MOS. prediction (NWP) models have certain defects that can In spite of the development of increasingly more so- be removed by statistically postprocessing their output phisticated versions of MOS and perfect prog, they (Wilks 1995). Two of the more popular postprocessing have certain characteristics that have remained distinct, approaches are model output statistics (MOS) and per- resulting in certain advantages or disadvantages of the fect prog (Glahn and Lowry 1972; Klein et al. 1959), respective approaches. For example, although MOS is both of which are based on the idea of relating model known to remove the bias from NWP forecasts, its de- forecasts to observations through linear regression. A velopment generally requires not-readily-available great many variations on the original ideas have been large datasets involving both model variables and ob- proposed: Vislocky and Fritsch (1997), for example, in- servations. The development of perfect prog, on the clude observations as both predictor and predictand, other hand, is simpler because it requires only obser- and Marzban (2003) additionally allows for nonlinear vations as both predictor and predictand. However, relationships among the various variables. In response perfect prog forecasts are not bias free. Meanwhile, to the necessity of deriving new regression equations Brunet et al. (1988) have shown that perfect prog out- every time an NWP model changes, Wilson and Vallée performs MOS in short-term forecasts. Wilson and (2002) have developed an updateable version of MOS Vallée (2003) found that their updateable MOS outper- forms both perfect prog and MOS, at least in forecast- ing 2-m temperature, 10-m direction and speed, Corresponding author address: Caren Marzban, Center for Analysis and Prediction of Storms, University of Oklahoma, Nor- and probability of precipitation. man, OK 73019. MOS is known to maintain reliability but loses sharp- E-mail: [email protected] ness and converges to climatology for longer forecast

© 2006 American Meteorological Society

Unauthenticated | Downloaded 10/02/21 08:19 PM UTC

MWR3088 658 MONTHLY WEATHER REVIEW VOLUME 134

projections. In ensemble systems, this is disadvanta- geous because an MOS-based ensemble forecast system fails to adequately forecast rare events. For this reason, some operational centers now rely on perfect prog as their method of choice in postprocessing ensemble out- puts (Denis and Verret 2004). On the other hand, an MOS-type approach to postprocessing of ensemble forecasts has been developed (Stephenson et al. 2004), which shows improved forecasts in comparison with in- dividual model forecasts as well as the ensemble mean. In the development of all the variations on MOS and perfect prog, a limitation is the amount of data avail- able for developing (or training) the regression model. A possible solution to this data shortage problem was proposed by Kalnay (2003). There, in appendix C.2, it is suggested to utilize reanalysis data (Kalnay et al. 1996; Kistler et al. 2001) to develop a postprocessor with the

advantages of both MOS and perfect prog, but without FIG. 1. A graphical display of two versions of MOS. MOS1 the weaknesses due to limited training data. In this ar- represents the “conventional” MOS where the predictor and pre- ticle, this method is referred to as RAN (for reanalysis). dictand are contemporaneous, while MOS2 allows for different As described in the next section, RAN also has the times for the two variables. added virtue of separating the loss of information be- tween predictor and predictand into its components— relevant quantities. The right portion displays the man- one due to the inadequacies of the NWP model, and the ner in which the developed regression equations are other due to chaos in the atmosphere itself. utilized for making forecasts. The vertical lines mark The object of this work is to compare MOS, perfect the initialization/analysis time. The variable x denotes a prog, and RAN within the confines of a simple statis- predictor, and y represents the predictand. The sub- tical model that allows for exact/analytic conclusions. A script “o” indicates observation, “m” stands for model, databased comparison will be performed elsewhere. and “f” represents forecast. These definitions are tabu- The approach assumes a single underlying numerical lated in Table 1. As an example, one may think of x as model, and not an ensemble of models. Finally, the the 700-mb height, and y as the surface temperature. postprocessors are for processing model data at a given Which MOS in Fig. 1 is better? To answer that ques- station, and not for all points on a grid. The next section tion, it is reasonable to assume that the best predictor of addresses the three approaches in more detail, and the y is x , and not x . This is so because x can, at best, mathematical framework is presented in section 3. Two o o m m be equal to xo. In other words, if the NWP model is subsections compare the three postprocessors in terms ϭ perfect, then xm xo. Therefore, if the NWP model is of bias, error variance, and uncertainty of the forecasts. perfect, then MOS1 will outperform MOS2, because at The paper ends with conclusions and discussions in sec- every point in time, the best predictor of y is the xo at tion 4. This last section also discusses questions regard- that same time. ing the effect (on the three postprocessors) of changes However, no NWP model is perfect. The value of xm in the NWP models, as well as the consequences of the agrees with x at the initialization/analysis time, but low resolution of the reanalysis data. o

TABLE 1. The definition and examples of the variables. Note the 2. MOS, perfect prog, and the reanalysis method time dependence of the various quantities. First, it is worthwhile to examine MOS more closely. Variable Definition Example Figure 1 displays two schematic versions of MOS: one x (t) Observed predictor Observed 700-mb height where the predictor and predictand are contemporane- o xm(t) Model predictor Model forecast of 700-mb ous (MOS1), and one that allows for different times for height ␶ the two variables (MOS2)—t for the forecast time, and xr( ) Reanalysis predictor Reanalysis value of 700-mb T for the valid time. The left portion of the figures height refers to the development stage of the equations, where yo(T) Observed predictand Observed 2-m temperature y (t, T) Forecast predictand Forecast 2-m temperature linear regression equations are developed to relate the f

Unauthenticated | Downloaded 10/02/21 08:19 PM UTC FEBRUARY 2006 MARZBAN ET AL. 659

1 thereafter diverges from xo. This loss of information/ skill is only partially due to the imperfect nature of the model. There is another loss due to the chaotic nature of the atmosphere itself. That loss reflects itself in the divergence between xo and y as the time difference between them increases. In other words, quite indepen- dently of the model, the skill in forecasting y from xo decreases as a function of the time between them. If the loss of information due to the imperfect nature of the model is larger than that due to chaotic growth, then MOS2 will outperform MOS1. In short, whether MOS1 is better than MOS2 de- pends on the rate at which the mutual information be- tween x and y is lost in the atmosphere relative to the rate at which the model loses information as it propa- gates a state into the future. At one extreme, if the model is perfect, then MOS1 is better, because one would run the model forward all the way to the valid FIG. 2. A graphical display of perfect prog, MOS, and the RAN x time, confident that m at that time has not diverged methods. from xo. At the other extreme, if the model is com- pletely flawed, then y is better predicted from xm if the model is not run at all, for running the model forward that maps xm to xo. This regression model would cap- for any duration would only worsen xm. Given that ture only model deficiencies, since it maps the model most models reside somewhere in the range between value of a quantity to the observed value of that same perfect and completely flawed, it follows that there ex- quantity.2 The second step would then involve mapping ists an “optimal” time lag between the predictor and xo to yo. Since this regression does not involve the predictand. model at all, it captures loss of information due to The same argument applies to perfect prog (see Fig. chaos.3 In practice, one may employ this two-step ap- 2). The assumption of perfect prog is that the NWP proach to produce a forecast for y. In a way, this ϭ model is perfect in the sense that xm xo for all times. method can be considered as a hybrid of MOS and Then one can run the model forward to the valid time perfect prog, since both xo and xm are employed in the (T), and produce a regression forecast (yf) based on the forecasting y. value of xm at that time. As such, it would be sufficient One drawback of this two-step approach is that it to produce the regression equations from contempora- does not allow for x and y to be the same physical neous values of xo and yo at the valid time. But if the quantities, because the second step of the model would model is completely flawed, then the model should not then involve mapping a variable to itself. This is a draw- be run forward at all. It follows, again, that a general- back because one would expect that the model forecast ization of perfect prog where the predictor and pre- for, say, temperature is a better predictor of surface dictand are measured at different times may outper- temperature than any other single predictor. To over- form the conventional perfect prog. Hereafter, “perfect come this problem, one may replace xo with a “reanaly- prog” refers to this more general definition of the sis” value of x. This will allow same-variable regression model. models, for at no point in the development stage will a The arguments presented above for allowing a time variable be regressed onto itself. Meanwhile, the re- lag between predictor and predictand is predicated on the acknowledgment that information is lost in two ways—one due to model deficiency and another due to 2 This step is affected by more than model deficiencies; for the atmosphere itself. This raises the possibility of mod- example, observational and assimilation errors still come into eling the two components, separately. In other words, play. However, for the present analysis, only two sources of “er- as a first step, one may develop a regression equation ror” are considered—one due to an imperfect model, and another due to chaos in the atmosphere. 3 Technically, this regression models the dynamic relation be-

tween xo and yo, and so it captures all contributions unrelated to 1 To be exact, the value of xm never agrees with xo, but their the NWP model. For example, even instrument error may be difference is minimum at the initialization time. captured.

Unauthenticated | Downloaded 10/02/21 08:19 PM UTC 660 MONTHLY WEATHER REVIEW VOLUME 134

analysis value of x (xr) encapsulates knowledge of the The RAN method involves two regression equations, observed value, and so the information in the observa- and one forecast equation: tions is not lost. In this article, this reanalysis-based RAN: x ͑␶͒ ϭ f͓x ͑t͔͒ ϩ error approach is denoted RAN. Note that during the train- r m ͑ ͒ ϭ ͓ ͑␶͔͒ ϩ ing phase, the NWP model is utilized to provide the yo T g xr error best estimate of the reanalysis, followed by a regression ͑ ␶ ͒ ϭ ͓ ͑ ͔͒ ͑ ͒ yf t, , T g͕f xm t ͖, 3 model to provide the best estimate of yo (Kalnay 2003). Moreover, note that the optimal time lag for the where the functions f and g are linear functions whose three postprocessors (MOS, perfect prog, and RAN) parameters are again estimated from data. Note that for ␶ may be mutually different. This is indicated in Figs. 1 RAN, yf depends also on the “reanalysis time,” .Inthe and 2 by marking the time at which the forecast is made following, this ␶ dependence will not be expressed un- (t) at different points along the time line. As such, in less necessary. comparing the three approaches, one must first derive Although the choice of the performance measure is the optimal value of t for each approach, and then com- arbitrary, here only the mean-squared error, as well as pare the models at their respective optimal values of t. its decomposition into bias and error variance, is con- This is an important point and will be further addressed sidered. Therefore, the model parameters are estimated in the next section. by minimizing the mean of the squared errors in Eqs. In summary, better postprocessors may be obtained (1)–(3), and the verification measure is computed as if 1) the predictor is measured at some “optimal” time ͑ ͒ ϭ ͓ ͑ ͒ Ϫ ͑ ͔͒2 ͑ ͒ preceding the time at which the predictand is measured, E t, T yf t, T yo T , 4 and 2) the two-step approach involving the reanalysis where the overbar denotes average. A useful decom- field is employed. Figure 2 illustrates the general situ- position of E is E ϭ V ϩ B2, where V and B are called ation underlying the three statistical forecasting ap- the error variance and bias, respectively, that is, proaches. The subscript “r” represents the reanalysis value of the predictor (see Table 1). Note that the ͑ ͒ ϭ ␴2͓ ͑ ͒ Ϫ ͑ ͔͒ V t, T yo T yf t, T , above arguments for allowing a time lag between pre- ͑ ͒ ϭ ͑ ͒ Ϫ ͑ ͒ ͑ ͒ dictor and predictands applies to each of the two steps B t, T yf t, T yo T , 5 ␶ in RAN, introducing an additional time parameter ( ). where ␴2[x(t)] denotes the variance of x(t). Good fore- The aim of this article is to assess the relative expected casts yield small values of V and B. performance of the three postprocessors displayed in The parametric functions in (1)–(3) are written as Fig. 2. follows: ͑ ͒ ϭ ␣ ͑ ͒ ͑ ͒ ϩ ␤ ͑ ͒ Perf Prog: yo T P t, T xo t P t, T , 3. Setup ͑ ͒ ϭ ␣ ͑ ͒ ͑ ͒ ϩ ␤ ͑ ͒ MOS: yo T M t, T xm t M t, T , The statistical model is developed over some histori- RAN: x ͑␶͒ ϭ ␣ ͑t, ␶͒x ͑t͒ ϩ ␤ ͑t, ␶͒, cal data on predictors at time t, and predictands at time r 1 m 1 ͑ ͒ ϭ ␣ ͑␶ ͒ ͑␶͒ ϩ ␤ ͑␶ ͒ ͑ ͒ T. The perfect prog and MOS regression equations, and yo T 2 , T xr 2 , T . 6 forecasts (y ) can be written as (see Fig. 2) f The minimization of E in (4) yields the well-known ͑ ͒ ϭ ͓ ͑ ͔͒ ϩ ordinary least squares (OLS) estimates: Perfect Prog: yo T f xo t error, ͑ ͒ ϭ ͓ ͑ ͔͒ ͑ ͒ ␣ ͑ ͒ ϭ ␴͓ ͑ ͒ ͑ ͔͒ր␴2͓ ͑ ͔͒ yf t, T f xm t , 1 P t, T xo t , yo T xo t , ␤ ͑ ͒ ϭ ͑ ͒ Ϫ ␣ ͑ ͒ ͑ ͒ ͑ ͒ ϭ ͓ ͑ ͔͒ ϩ P t, T yo T P t, T xo t , MOS: yo T f xm t error, ͑ ͒ ϭ ͓ ͑ ͔͒ ͑ ͒ ␣ ͑t, T ͒ ϭ ␴͓x ͑t͒, y ͑T ͔͒ր␴2͓x ͑t͔͒, yf t, T f xm t , 2 M m o m ␤ ͑t, T ͒ ϭ y ͑T ͒ Ϫ ␣ ͑t, T ͒x ͑t͒, where the function f is usually a linear equation whose M o M m parameters are estimated from data. The same function ␣ ͑ ␶͒ ϭ ␴͓ ͑ ͒ ͑␶͔͒ր␴2͓ ͑ ͔͒ 1 t, xm t , xr xm t , with the estimated parameters is then utilized for mak- ␤ ͑t, ␶͒ ϭ x ͑␶͒ Ϫ ␣ ͑t, ␶͒x ͑t͒, ing the forecasts yf for time T. It is assumed that there 1 r 1 m exists only a single predictor; the analysis can be gen- 2 ␣ ͑␶, T ͒ ϭ ␴͓x ͑␶͒, y ͑T ͔͒ր␴ ͓x ͑␶͔͒, eralized to the multiple predictor case in a straightfor- 2 r o r ␤ ͑␶ ͒ ϭ ͑ ͒ Ϫ ␣ ͑␶ ͒ ͑␶͒ ͑ ͒ ward manner. 2 , T yo T 2 , T xr , 7

Unauthenticated | Downloaded 10/02/21 08:19 PM UTC FEBRUARY 2006 MARZBAN ET AL. 661

where ␴[x(t), y(T)] denotes the covariance between x at where the various times (t, T, ␶) have been supple- time t and y at time T. Note that these OLS estimates mented with an index—P, M, R—denoting the respec- are denoted by the same symbol as the arbitrary pa- tive approaches (perfect prog, MOS, and RAN). These rameters appearing in the regression Eqs. (6). Hence- time parameters represent the respective optimal times forth, any reference to the ␣, ␤ parameters refers to for each method, as discussed in section 2. ␶ ϭ their respective OLS values given in Eq. (7). It is worth mentioning that BR( R, tR, T) 0 only if Substituting the OLS estimates into the forecasts one assumes that the average of xm over the training equations in Eqs. (1)–(3) yields the forecast quantities data is equal to that over the data of the forecasting

in each method: phase. In other words, it is assumed that xm for the left portion of Fig. 2 is the same as that of the right portion. ͑ ͒ ϭ ␣ ͑ ͒ ͑ ͒ ϩ ␤ ͑ ͒ Perf Prog: yf t, T P t, T xm t P t, T , The same is assumed for xr. This is a reasonable as- ͑ ͒ ϭ ␣ ͑ ͒ ͑ ͒ ϩ ␤ ͑ ͒ MOS: yf t, T M t, T xm t M t, T , sumption if the sample employed for training is repre- sentative of the population. ͑ ␶ ͒ ϭ ␣ ͑ ␶ ͒ ͑ ͒ ϩ ␤ ͑ ␶ ͒ ͑ ͒ RAN: yf t, , T R t, , T xm t R t, , T , 8 where a. Comparison of bias and variance ␣ ͑ ␶ ͒ ϭ ␣ ͑ ␶͒␣ ͑␶ ͒ R t, , T 1 t, 2 , T , With a bit of algebra, one can show ␤ ͑t, ␶, T ͒ ϭ y ͑T ͒ Ϫ ␣ ͑t, ␶, T ͒x ͑t͒. ͑9͒ ͑ ͒ Ϫ ͑ ͒ ϭ ͓␳ ͑ ͒ Ϫ ␳ ͑ ͔͒2 R o R m VP tP VM tM P tP M tP To bring transparency to the equations, it is convenient ϩ ͓␳2 ͑ ͒ Ϫ ␳2 ͑ ͔͒ M tM M tP , to introduce the quantities ͑␶ ͒ Ϫ ͑ ͒ ϭ ͓␳ ͑␶ ͒ Ϫ ␳ ͑ ͔͒2 VR R, tR VM tM R R, tR M tR ␴͓x ͑t͔͒ ␳ ͑ ͒ ϭ ͓ ͑ ͒ ͑ ͔͒ m ␴͓ ͑ ͔͒ P t, T r xo t , yo T ␴͓ ͑ ͔͒ yo T , ϩ ͓␳2 ͑t ͒ Ϫ ␳2 ͑t ͔͒, xo t M M M R ␳ ͑ ͒ ϭ ͓ ͑ ͒ ͑ ͔͒␴͓ ͑ ͔͒ ͑␶ ͒ Ϫ ͑ ͒ ϭ ͓␳ ͑␶ ͒ Ϫ ␳ ͑ ͔͒2 M t, T r xm t , yo T yo T , VR R, tR VP tP R R, tR M tR ␳ ͑ ␶ ͒ ϭ ͓ ͑ ͒ ͑␶͔͒ ͓ ͑␶͒ ͑ ͔͒␴͓ ͑ ͔͒ ͑ ͒ Ϫ ͓␳ ͑ ͒ Ϫ ␳ ͑ ͔͒2 R t, , T r xm t , xr r xr , yo T yo T , 10 P tP M tP where r[x, y] is the linear correlation coefficient be- ϩ ͓␳2 ͑ ͒ Ϫ ␳2 ͑ ͔͒ ͑ ͒ M tP M tR , 14 tween x and y, defined as where, for brevity, the dependence of the ␳ quantities ␴͑x, y͒ r͑x, y͒ ϭ . ͑11͒ on the valid time T has been suppressed. This form of ␴͑x͒␴͑y͒ the equations is particularly useful, because the last Given that r(x, y) assesses the linear correlation be- term on the right-hand side of all three equations in- ␳ tween a pair of variables (usually a predictor and a volves only M evaluated at different times. The re- ␳ ␳ ␳ maining terms on the right-hand side are all perfect predictand), the quantities P, M, and R also assess the “correlation,” but in a generalized sense; if the vari- squares, and hence nonnegative. ables are scaled to have unit variance, then the ␳’s are For example, consider the first equation in (14). The determined entirely from the r’s. first term on the right-hand side is nonnegative. The ␳ Then V and B in Eq. (5) for the three methods be- second term is the difference between M at time tM and come time tP. However, by the definition of tM, it is the time at which the mean squared error for MOS is a mini- ͑ ͒ ϭ ␴2͓ ͑ ͔͒ ϩ ␳2 ͑ ͒ VP tP, T yo T P tP, T mum. But since the bias is zero at that time [Eq. (13)], then V , too, is minimized at time t . Finally, based on Ϫ ␳ ͑ ͒␳ ͑ ͒ ͑ ͒ M M 2 P tP, T M tP, T , 12 ␳ ␳ Ն Eq. (12), M is maximized at time tM. In short, M(tM) ␳ ͑ ͒ ϭ ␴2͓ ͑ ͔͒ Ϫ ␳2 ͑ ͒ M(tP), which implies that the right-hand side of the VM tM, T yo T M tM, T , first equation in Eq. (14) is nonnegative. It follows that ͑␶ ͒ ϭ ␴2͓ ͑ ͔͒ ϩ ␳2 ͑␶ ͒ Ն VR R, tR, T yo T R R, tR, T VP(tP) VM(tM), that is, MOS outperforms perfect prog in terms of the variance of the forecast errors. Ϫ 2␳ ͑t , ␶ , T ͒␳ ͑t , T ͒, R R R M R The same argument applies to the second equation in ͑ ͒ ϭ ␣ ͑ ͓͒ ͑ ͒ Ϫ ͑ ͔͒ ␶ Ն BP tP, T P tP, T xo tP xm tP , (12), which in turn, implies that VR( R, tR) VM(tM), that is, MOS outperforms RAN in terms of the variance B ͑t , T ͒ ϭ 0, M M of the forecast errors. ͑␶ ͒ ϭ ͑ ͒ BR R, tR, T 0, 13 Given that the bias of MOS forecasts is zero, it then

Unauthenticated | Downloaded 10/02/21 08:19 PM UTC 662 MONTHLY WEATHER REVIEW VOLUME 134 follows that MOS outperforms both perfect prog and forecasts. In short, if the reanalysis data are larger than RAN in terms of bias and error variance (and mean NWP model data, then RAN forecasts have lower un- squared error).4 certainty than MOS and perfect prog forecasts. There- fore, as far as uncertainty is concerned, RAN outper- b. Comparison of uncertainty forms the alternatives. It is now possible to compute the uncertainty associ- ated with the forecasts, namely, the variance of the 4. Conclusions and discussion forecasts.5 For brevity, the time dependence of the vari- ous quantities is suppressed in this section, because the MOS is expected to outperform perfect prog and central quantity in the analysis of these uncertainties is RAN in terms of bias, error variance, and mean the sample size of the data. Note that the three ap- squared error. However, the uncertainty of MOS fore- proaches may be based on samples of different size. Let casts may be hindered by the limited size of available model data. Perfect prog is less restrictive because its the sample size of the data employed in MOS be NM, development is not limited by the availability of model and that in perfect prog NP. As for RAN, the samples in the two steps of that approach may again have dif- data; however, its forecasts are biased and have higher ferent sizes. In practice, the first step of the approach error variance than MOS forecasts. On the other hand, RAN forecasts are bias free and have lower uncertainty involves model predictor, xm, and as such will have a than MOS and perfect prog. In short, even though sample size of NM. The second step of RAN involves reanalysis data, and for that reason, its sample size is MOS has higher performance (in terms of bias and er- ror variance) than perfect prog and RAN, the latter has denoted NR. The uncertainties can be computed as lower uncertainty associated with its forecasts if its sample size is larger than MOS’s sample size. ␴2͑y ͒ x Ϫ x 2 ␴2͑ ͒ ϭ o ͭ ϩ ͫ m oͬ ͮ These findings are based on several assumptions, in- Perf Prog: yf 1 ␴͑ ͒ , NP xo cluding the following: First, the underlying relations are assumed to be linear. It is this linearity that allows for ␴2͑y ͒ x Ϫ x 2 ␴2͑ ͒ ϭ o ͭ ϩ ͫ m mͬ ͮ analytic conclusions like those presented here. Nonlin- MOS: yf 1 ␴͑ ͒ , NM xm ear models do not lend themselves to such analysis, and Ϫ 2 are better assessed with real data. That task will be 1 xm xm RAN: ␴2͑y ͒ ϭ ␴2͑y ͒ͭ ϩ ͫ ͬ addressed in the future. Second, ordinarily, in regres- f f N ␴͑x ͒ R m sion modeling, the predictors are assumed to be error r2͑x , x ͒ r2͑x , y ͒ free, and only the target is assumed to have errors. In ϫ m r ϩ r o ͑ ͒ ͫ ͬͮ. 15 RAN, however, there are two regression models—one NM NR from a model variable to a reanalysis variable, and a A comparison of MOS and perfect prog uncertainties in second from the reanalysis variable to an observation. Eq. (15) suggests that the latter may in fact have smaller Here, the reanalysis variable is assumed to be error free N Ͼ N uncertainties when P M, at least for some values of during the first step, but with error during the second x m. However, this slight advantage is offset by the fact step. The more realistic analysis wherein both the pre- that perfect prog forecasts are not bias free [Eq. (13)]. dictors and predictands are assumed to have error is As for MOS and RAN, it follows from (15) that, if more complicated [see section 3.4 of Draper and Smith N Ͼ N R M, then the uncertainty in RAN forecasts is (1998)] and will be considered elsewhere. x smaller than that of MOS forecasts, for all values of m. Given all of these assumptions, it is important to em- Therefore, if the second step of RAN (i.e., the step with phasize that the results reported here regarding the reanalysis predictors) is based on a larger dataset than relative performance of the three methods are only “ex- the first step (i.e., the step based on model predictors), pected” and not guaranteed. However, as shown here, then RAN forecasts have lower uncertainty than MOS there exists sufficient theoretical justification to de- velop and test a real-time, RAN-based postprocessor. 4 That MOS is better than the alternatives is related to the One question is whether changes in the numerical Gauss–Markov theorem, which states that the best linear unbi- model call for retraining RAN. In other words, in an ased estimator is the OLS estimator. David Stephenson is ac- operational setting, is one expected to retrain RAN knowledged for pointing out this connection. each time a numerical model undergoes some revision? 5 The variance of the forecasts is to be contrasted with the error variance computed previously. The former is the variance of the The important point to note is that the difficulty in forecasts, while the latter is the variance of the forecast errors training MOS is in the requirement for large datasets (technically, residuals). involving observations. In RAN, however, observations

Unauthenticated | Downloaded 10/02/21 08:19 PM UTC FEBRUARY 2006 MARZBAN ET AL. 663 play a role only in the second step, where they are Acknowledgments. David Stephenson is acknowl- mapped to the reanalysis data, not model data. When a edged for pointing out to us the connection of this work new version of a model is released, only the first step of to the Gauss–Markov theorem. We also thank Frederic RAN must be retrained, but that is a relatively simple Atger, and Bertrand Denis for commenting on the role task (compared to MOS), given that it involves map- of MOS and perfect prog in ensemble-based forecast ping the “new” model data only to the reanalysis data, systems. Partial funding was provided under a coopera- that is, the reanalysis based on the “old” model com- tive agreement between the National Oceanic and Atmo- pared to the “new” model, where a large amount of spheric Administration (NOAA) and the University Cor- data exists on identical grids. Consequently, only a rela- poration for Atmospheric Research (UCAR). The views tively short model rerun is required to retrain. This is expressed herein are those of the authors and do not especially important when dealing with rare phenom- reflect the views of NOAA, its subagencies, or UCAR. ena, where MOS would require exceedingly large train- ing sets mapping the “new” model to the observations. REFERENCES Moreover, it is possible to implement an updateable Brunet, N., R. Verret, and N. Yacowar, 1988: An objective com- approach (Wilson and Vallée 2002) for RAN. As such, parison of model output statistics and perfect prog systems in model changes are expected to have minimal effect on producing numerical weather element forecasts. Wea. Fore- RAN in an operational setting. casting, 3, 273–283. Denis, B., and R. Verret, 2004: Toward a new Canadian medium- Another question is whether or not there is an ad- range perfect-prog temperature forecast system. Preprints, vantage in using the NWP model used for the reanalysis 17th Conf. on Probability and Statistics in Atmospheric Sci- also for producing the forecasts. One might expect that ences, Seattle, WA, Amer. Meteor. Soc., CD-ROM, 1.6. the correlation between the reanalysis (i.e., modeled Draper, N. R., and H. Smith, 1998: Applied Regression Analysis. John Wiley & Sons, 706 pp. “truth”) and an NWP prediction based on the same Glahn, H. R., and D. A. Lowry, 1972: The use of model output underlying dynamics, physics, and numerical approxi- statistics (MOS) in objective weather forecasting. J. Appl. mations should be higher than the correlation based on Meteor., 11, 1203–1211. an entirely different NWP system. If true, this would Kalnay, E., 2003: Atmospheric Modeling, Data Assimilation and constitute a severe restriction to RAN. However, it is Predictability. Cambridge University Press, 341 pp. ——, and Coauthors, 1996: The NCEP/NCAR 40-Year Reanaly- also possible that the benefits of a sufficiently long sis Project. Bull. Amer. Meteor. Soc., 77, 437–471. training period (as is possible in RAN) may outweigh Kistler, R., and Coauthors, 2001: The NCEP–NCAR 50-year re- any drawbacks of using different models for reanalysis analysis: Monthly mean CD-ROM and documentation. Bull. and forecasts. This is especially true if differences be- Amer. Meteor. Soc., 82, 247–267. tween models are “systematic.” This question will be Klein, W. H., B. M. Lewis, and I. Enger, 1959: Objective predic- tion of five-day mean temperature during winter. J. Meteor., further addressed in the future. 16, 672–682. Finally, how does the relatively low resolution of the Marzban, C., 2003: A neural network for post-processing model reanalysis data affect the practical utility of RAN? A output: ARPS. Mon. Wea. Rev., 131, 1103–1111. RAN postprocessor will reproduce a response to Stephenson, D. B., C. A. S. Coelho, F. J. Doblas-Reyes, and M. larger-scale features but not very high resolution fea- Balmaseda, 2004: Forecast assimilation: A unified framework for the combination of multi-model weather and climate pre- tures (small eddies, , etc.). But that is dictions. Tellus A, 57, 253–264. precisely what one would desire in a postprocessor— Vislocky, R. L., and G. S. Young, 1989: The use of perfect prog the ability to calibrate to large-scale features that are forecasts to improve model output statistics forecasts of pre- highly predictable, rather than to “contaminate” the cipitation probability. Wea. Forecasting, 4, 202–209. calibration with less predictable, small-scale features. ——, and J. M. Fritsch, 1997: An automated, observations-based system for short-term prediction of ceiling and . Wea. (It is useful to note here that the NWS operational Forecasting, 12, 31–43. MOS was developed using a very low resolution version Wilks, D. S., 1995: Statistical Methods in the Atmospheric Sciences. NWP model, following this reasoning.) The first step of Academic Press, 467 pp. RAN calibrates the model to the reanalysis, correcting Wilson, L. J., and M. Vallée, 2002: The Canadian updateable the model representation of the large scale. The second model output statistics (UMOS) system: Design and devel- opment tests. Wea. Forecasting, 17, 206–222. step of RAN calibrates the reanalysis to the observa- ——, and ——, 2003: The Canadian updateable model output tions, ensuring that site-specific properties of the ob- statistics (UMOS) system: Validation against perfect prog. servations are appropriately represented. Wea. Forecasting, 18, 288–302.

Unauthenticated | Downloaded 10/02/21 08:19 PM UTC