<<

Why is Forecasting Still Under a ?

Abstract: Weather forecasts are now impressively accurate, with short-term predictions achieving success rates of around 85 per cent. Despite this, forecasting remains the butt of much folklore, such as the notion that taking an umbrella on the advice of a forecast makes less likely to fall. Using probability theory, I show that such folklore has some basis in reality via the so-called "base-rate effect". The public's intuitive recognition of this effect may well explain their continued scepticism about weather forecasts.

INTRODUCTION

Weather forecasting is one of the triumphs of applied mathematics. World-wide data collection, sophisticated numerical models and state-of-the-art computing have now been combined by to forecast the behaviour of this complex and non-linear phenomenon with impressive accuracy. The UK Meteorological Office, widely regarded as one of the best forecasting services in the world, currently reports computer model accuracies of 71 per cent for its 24 hr forecasts of rain, rising to 83 per cent following input from human forecasters1. Many people, however, remain resolutely sceptical of the reliability of weather forecasting: opinion polls show that dissatisfaction with forecasts currently runs at around 15-20 per cent1.

Given the non-linearity of the weather system - and thus the sensitivity of forecasts to the unavoidably imperfect state of meteorological data - some level of dissatisfaction is, of course, inevitable. This of public scepticism might best be tackled by promoting a wider awareness of the implications of chaos for weather forecasting. Another source of dissatisfaction lies in the limit on accuracy imposed by computer technology. Even using capable of tens of gigaflops (thousand million floating point operations a second), today's numerical weather models are still too broad-brush to permit truly "local" forecasts to be made. As a result, it will be some years yet before forecasters are able to end the frustration of predictions proving accurate in one region, yet failing miserably just a few kilometres away. There is, however, a third source of dissatisfaction which appears not to have been widely recognised by meteorologists. At its heart is a probabilistic concept known as the "base rate effect", which ties the value of a forecast to the frequency of the phenomenon being forecast. In what follows, I use probability theory to show how even today's impressively accurate forecasting methods can fall foul of the base-rate effect, with consequences that can seem distinctly paradoxical.

FORECASTS AND BAYES'S THEOREM

The forecasting of any complex non-linear system such as the weather is inevitably a probabilistic process. The aim of the forecaster is thus be to produce predictions that are significantly more reliable than those achieved by random guessing. Mathematically, the forecasting process can be modelled by Bayes's Theorem, which shows how the odds on the occurrence of a specific event - say, a rain-shower - are improved (i.e. increased) in the light of a forecast:

Odds(Event | Forecast) = LR x Odds(Event) (1) where Odds(Event) = Pr(Event)/Pr(~Event) etc., "~" denotes negation, " | " denotes "given", and "LR" is the Likelihood Ratio, defined by

LR = Pr(Forecast | Event) / Pr(Forecast |~Event) (2)

Forecasts based on random guessing are as likely to be right as wrong, and the odds of the event occurring in the light of such a forecast, Odds(Event | Guessing) are thus no higher than the "base-rate", Odds(Event). By (1), guessed forecasts can thus be characterised by a likelihood ratio of unity: they add no information about the chances of the event taking place. If a forecasting technique is to be useful, therefore, it must give LR > 1. The success of forecasting techniques is not usually stated in terms of likelihood ratios. Instead, it is typically given in terms of a somewhat vague concept such as "percentage of accurate forecasts". We can, however, convert such percentages into the corresponding likelihood ratios.

Let R represent the event of rain, F be the event of rain being forecast, and Q the observed probability of the forecast proving correct, i.e. the frequency with which a forecast correctly predicts rain or correctly predicts no rain. As these latter events are mutually exclusive, we have

Q = Pr(F & R) + Pr(~F & ~R) (3)

= Pr(F | R).Pr(R) + Pr(~F |~R).Pr(~R) (4)

Met Office data for rain forecasting2 shows that Pr(F | R) ~ Pr(~F |~R) so that (4) reduces to

Q = Pr(F | R) = Pr(~F | ~R) (5) and the likelihood ratio LR becomes

LR = Pr( F | R) / Pr(F | ~R) = Q /(1 - Q) (6) and so, by (1)

Odds(Rain | Forecast) = Q . Odds(Rain) /(1 - (Q) (7)

The UK Meteorological Office's stated accuracy rate for its 24hr forecasts is 83 per cent; putting Q = 0.83 into (6) implies that for Met Office forecasts we have LR = 4.9. To illustrate the implications of this, we note that the daily probability of rain for England and Wales is about 0.4, thus giving Odds(Rain) of 0.67. By (7), this implies that a forecast of rain made using the Met Office's 83 per cent accurate techniques lead to odds of rain taking place of 4.9 x 0.67 = 3.3; i.e. the forecast can be expected to be correct about 77 per cent of the time.

This highlights two crucial aspects of the interpretation of forecasting. First, our perception of forecast accuracy is not determined solely by the accuracy rate Q. The quantity of real importance to users of forecasting data is the conditional probability Odds(Event | Forecast) - and as (1) shows, this depends crucially on the base-rate for the phenomenon being forecast, Odds(Event). Second, when this fact is taken into account, the probability of a specific forecast proving correct can be significantly lower than the accuracy figure quoted for the forecasting technique. For the Met Office forecasts of rain, for example, the accuracy figure of 83 per cent becomes, after allowance for the UK daily rain base-rate, a conditional probability of rain of 77 per cent. This reduction in perceived accuracy is the so-called "base-rate effect": the ability of a low base-rate to dilute the reliability of an accurate forecasting method. Neglect of the base-rate effect has been shown to have serious implications in fields as diverse as cancer screening3 and DNA profiling4. Its implications for weather forecasting, however, appear to have been largely overlooked. Yet as I now show, the base-rate effect can seriously - and negatively - affect public perception of the reliability of even highly accurate weather forecasts.

BASE-RATES AND UMBRELLA-TAKING

The essence of the base-rate is simply put. If an event is sufficiently rare, then even highly accurate forecasting methods can still fail to raise the chances of the event taking place above 50:50. From (7), this will happen for any phenomenon whose base rate is less than Pr(min) where

Pr(min) < (1 - Q ) (8)

With Q = 0.83, this leads to Pr(min) for Met Office forecasts of 0.17; predictions of weather events with a frequency below this are more likely to be wrong than right -despite the impressively high accuracy of the forecasting technology. For example, consider the well-known problem of deciding whether or not to take an umbrella in the light of a forecast of rain. At first glance, it would seem that Met Office forecasts are well-able to provide reliable advice on which to base such a decision. The daily base-rate for rain in England and Wales is 0.4, which exceeds the critical value of 0.17 by a comfortable margin. However, this probability is not appropriate for the umbrella-carrying problem; what we require is the probability of rain occurring on the hourly timescale relevant to umbrella-taking5: this is 0.08 - a much lower base-rate which does meets the inequality (8). From (7) we then find that the probability that we shall require our umbrella, following even an 83 per cent accurate forecast, is just 0.3. In other words, those in the UK who take an umbrella in response to a forecast will typically find themselves needlessly burdened with it about 2 times out of 3. The situation will be even worse for those living in South-East England, where the hourly probability of rain is only about half the national average. To this extent, the folklore that taking an umbrella reduces the chances of rain falling is borne out. It is not, of course, that the weather "knows" that one is carrying an umbrella. It is simply that placing one's complete faith in the forecast alone fails to take into account the relatively low base-rate of hourly rain in the UK. Indeed, in a recent paper6, I showed that decision theory leads to the conclusion that unless one is quite concerned about getting wet, the optimal decision is never to take an umbrella on walks, even if showers are forecast.

It thus seems that public scepticism of weather forecasting may be an example of where ordinary people have a good, intuitive grasp of the impact of base-rates on their decision-making7: experience has told them that for the relatively short time they are out on a walk or shopping trip, the chances of rain falling are relatively low. Perhaps the Met Office could consider making this more clear in its forecasts, especially during showery weather. Certainly it would be wrong to respond by blaming the current inability of computer models to accurately forecast rain on hourly, rather than a daily, timescales: the low base-rate of hourly rain will still lead to apparently poor forecast reliability even if the Met Office does succeed in predict hourly showers with the same accuracy as its current forecasts. I now turn to a second source of discontent with weather forecasts: their ability to be right most of the time, yet go wrong in the worst possible way.

FORECAST ERRORS AND SELECTIVE MEMORY

When a forecast of rain goes wrong, it can do so in one of two ways: it can either fail to predict rain that subsequently falls, or vice versa. Of the two failure modes, being caught in unpredicted rain is more likely to stay in the memory than the inconvenience of carrying a redundant umbrella. This suggests that the relative probabilities of the two possible failure modes also play a role in determining the public perception of weather forecasting accuracy. To calculate these two probabilities, let R be the event of rain, F be the event of a forecast of rain, and U be the event of taking an umbrella. By Bayes's Theorem, the probability of the first failure mode, i.e. being caught in the rain without an umbrella, is

Pr(R & ~U) = Pr(R & ~U | F).Pr(F) + Pr(R & ~U | ~F).Pr(~F) (9) while the probability of the second failure mode, of having an umbrella when one is not required, is:

Pr(~R & U) = Pr(~R & U | F).Pr(F) + Pr(~R & U | ~F).Pr(~F) (10) The conditional joint probabilities on the rhs of (9) and (10) can be decomposed into

Pr(R & ~U | F) = Pr(R | F & ~U).Pr(~U | F) etc. (11)

Under the reasonable assumption that the probability of rain is independent of our decision to carry an umbrella, then

Pr(R | F & ~U) = Pr(R | F) = Pr(F | R).Pr(R)/Pr(F) etc (12) so we finally obtain

Pr(R & ~U) = Pr(R)[Pr(F | R).Pr(~U | F) + Pr(~F | R).Pr(~U | ~F)] (13)

Pr(~R & U) = Pr(~R)[Pr(F | ~R).Pr(U | F) + Pr(~F |~R).Pr(U | ~F)] (14)

Putting one's trust in a forecast implies that Pr(U | F) = 1, Pr(U | ~F) = 0, and similarly for the negations, so that (13) and (14) reduce to

Pr(R & ~U) = Pr(R). Pr(~F | R) (15)

Pr(~R & U) = Pr(~R).Pr(F | ~R) (16)

Thus, a forecasting technique is more likely to fail in the worst possible modality if

Pr(R) > Pr(F | ~R) / Pr(F | ~R) + Pr(~F |R) (17)

As Pr(F | R) ~ Pr(~F | ~R), forecasting failures will thus be biased towards the worst possible outcome if the base-rate of rain exceeds Pr(R) > 0.5, and the relative probability of the two possible failure mode outcomes is

Pr(R & ~U) / Pr(~R & U) ~ Pr(R) / Pr(~R) (18)

As before, UK forecasts would at first glance seem to be immune from this effect, as the average probability for rain is 0.4, below the critical level. However, there are many regions of the UK, chiefly west of the Pennines, which experience over 220 days of rain a year, corresponding to a base-rate of over 0.6. By (18), this implies that in these areas, most inaccurate rain forecasts will be of the type that leads to an unexpected downpour. Those living in such areas who rely on forecasts are thus more likely to be caught in unpredicted rain when the forecast proves inaccurate. This towards the worst failure mode, combined with selective memory effects, may be a source of scepticism among those living in relatively high-rainfall areas.

CONCLUSION

Despite its familiarity, weather forecasting is one of the most demanding of scientific challenges, and one of the most successful: statistics show that short-term forecasts are now impressively accurate. However, as I have shown, the so-called base-rate effect has the power to undermine confidence in even highly accurate forecasts. Specifically, neglect of the base-rate effect can lead those living in the drier parts of the UK into taking precautions against rain that is in fact unlikely to fall. In contrast, when forecasts go awry in the wetter areas of the UK, the most likely result is an unexpected soaking. Both these manifestations of the base-rate effect may help explain the otherwise puzzling disparity between the claims for high accuracy made by meteorologists, and the persistent scepticism of the public. It appears that, at least as far as weather forecasts are concerned, the public have a better handle on the base-rate effect than the experts.

ACKNOWLEDGEMENTS

It is a pleasure to thank Prof Dennis Lindley and Dr David Balding for their interest, advice and comments on an early draft, and Philip Eden of Philip Eden Weather Consultancy and Andy Yeatman of the Meteorological Office for supplying UK weather and forecasting accuracy data.

References

1. Yeatman, A. (1996) Personal communication

2. Burroughs, W.J. (1991) Watching the World's Weather (C.U.P) p 147

3. Sutherland, S. (1994) Irrationality: The Enemy Within (Penguin) pp173-181

4. Matthews, R.A.J. (1994) "Improving the odds on justice?" New Scientist 16 April, pp 12-13

5. Chandler, T.J. and Gregory, S. (1976) The of the British Isles Longman table 6.1 p 136.

6. Matthews, R.A.J. (1996) "Base-rate errors and rain forecasts" Nature 29 August 382 766

7. Koehler, J. J. (1996) "The base rate fallacy reconsidered" Behavioral and brain vol 19 p 1 - 53