SCIENTIFIC DUE DILIGENCE FOR HUMANITARIAN DISASTER RISK FINANCING A GUIDE FOR DATA SCIENTISTS AND HUMANITARIAN PRACTITIONERS
CLARE HARRIS RACHEL HALDANE ELIZABETH REES OUR PARTNERS SCIENTIFIC DUE DILIGENCE FOR HUMANITARIAN DISASTER RISK FINANCING 03
CONTENTS
KEY DEFINITIONS 04 EXECUTIVE SUMMARY 05 INTRODUCTION TO DISASTER RISK FINANCING AND FORECAST-BASED FINANCING 07 PURPOSE OF THIS GUIDE 08
CHECKPOINTS
01 CREATING A DATA AND DECISION FRAMEWORK: IDENTIFYING OPERATIONAL NEEDS AND THE QUESTIONS THE MODEL SHOULD LOOK TO ANSWER 10 02 COLLECTING AND TESTING THE DATA THROUGH DIFFERENT LENSES – GEOGRAPHICAL, SPATIAL, TEMPORAL, SEVERITY AND IMPACT 12 03 OPTIMISING THE AVAILABLE TESTING DATA: CREATING STRONGER HISTORICAL DATASETS 14 04 UNCERTAINTY TESTING: TESTING AND COMPARING MODELS AND DATA, AND PRESENTING THE RESULTS 16 05 SKILL TESTING 19 06 COMMUNICATING COMPLEXITY CLEARLY 21 07 AGREEING THE OPERATIONAL DATA AND DECISION FRAMEWORK 23 08 COMPLETING THE DESIGN OF THE DRF SYSTEM: APPLYING THE ANALYSIS TO OPERATIONAL HUMANITARIAN DRF DESIGN 25
END NOTE 26
APPENDICES
01 PURPOSE OF THE DROUGHT RISK FINANCE SCIENCE LABORATORY (DRISL) 28 02 ANIMATION 29 03 COLLABORATION AND PARTNERSHIPS 30 04 OTHER WORK 32 05 GLOSSARY 33 04 SCIENTIFIC DUE DILIGENCE FOR HUMANITARIAN DISASTER RISK FINANCING
KEY DEFINITIONS MODEL A model is a simplified representation of reality, often used to help us understand what might happen in the future. Models and modelling are used across scientific, business and financial fields.
In Disaster Risk Financing (DRF), we develop risk models to tell us the likelihood of a damaging event occurring. A risk model might use biophysical indicators of a hazard (e.g. a drought) that have themselves been derived from climate or weather forecast models. The risk model might combine this with information on impacts. It then predicts the chance of an event occurring, and its likely socioeconomic impacts over a particular period.
DATA Refers to both hazard data (e.g. rainfall, temperature) and impact data (e.g. yields, health data).
PROVIDERS In the DRF context, “providers” are the data scientists collecting, analysing and calibrating the data.
USERS Refers to the humanitarian practitioners, and others including at-risk groups, who are utilising the data provided by the scientists to build decision-making systems such as a DRF system.
AT-RISK To build a successful DRF system it is essential that the humanitarian practitioners COMMUNITIES and data scientists work in partnership with the at-risk communities that their risk models seek to represent and support. At-risk groups can act as both providers – of information on which the model is based and validated – and users, working with practitioners to develop and apply the model and utilise the risk information themselves. At-risk communities should always be at the forefront of decision-making and supported to be able to define their own needs and acceptable levels of risk.
BASIS RISK In the finance and insurance industry, ‘basis risk’ means the risk that a trader might accept in terms of a product not performing as it should when hedging positions. In DRF and the use of parametric solutions, the term applies to situations where statistical models are used to predict the outcome of a current or likely event to trigger financing (the payout is not indemnity-based but a predicted or modelled loss). The basis risk lies in the combination of:
l model error—the inherent errors found within a model’s data and calculations; l outcome uncertainties—the uncertainty within which the model operates (the difference between the model results and real-world results); l social miscommunication or misinterpretation of a model and product’s capabilities.1
1 Basis risk in disaster risk financing for humanitarian action: Potential approaches to measuring, monitoring, and managing it, Harris, C. and Cardenes, I., Centre for Disaster Protection Insight paper, Centre for Disaster Protection, London (2020). SCIENTIFIC DUE DILIGENCE FOR HUMANITARIAN DISASTER RISK FINANCING 05
EXECUTIVE SUMMARY DISASTER RISK FINANCING (DRF) HAS HUGE POTENTIAL TO IMPROVE THE HUMANITARIAN SYSTEM BY ENABLING US TO ANTICIPATE AND PROACTIVELY MANAGE DISASTER RISK. DRF IS BASED ON THREE PILLARS:
UNDERSTANDING THE RISKS AND HAZARDS PLANNING THE INTERVENTIONS PRE-POSITIONING THE FUNDING faced in different countries, ensuring in advance of disasters. so action can occur in at-risk communities have access to this anticipation of a crisis. information, and using scientific models to quantify the risks and agree triggers or danger thresholds.
The DRF approach increases the efficiency, speed and predictability of crisis management. Critically, it al- lows humanitarians to act before a risk turns into crisis, reducing costs, lessening impact and saving lives. Because DRF requires actions to be designed and planned before the worst happens, it has the added benefit of creating more participatory space, enhancing transparency and accountability – downstream towards those at risk, and upstream towards the donors. However, any DRF system will only be as effective as the scientific models on which it is based. Scientific modelling is an attempt to predict what might happen in the future; a degree of uncertainty will always be inherent in this process. The difference between what a scientific model predicts and what actually occurs is often referred to as the basis risk. Basis risk exists because our ability to replicate or measure the complexities of natural and human process- es through modelling is limited: a model will always be a simplification of reality. There are three categories of basis risk:2 l model error—the inherent errors found within a model’s data and calculations; l outcome uncertainties—the uncertainty within which the model operates (the difference between the model results and real-world results); l social miscommunication or misinterpretation of a model and product’s capabilities.
It is important that these potential sources of error are managed. See table in Appendix 5, which outlines the risk management steps that can be taken.
2 Basis risk in disaster risk financing for humanitarian action: Potential approaches to measuring, monitoring, and managing it, Harris, C. and Cardenes, I., Centre for Disaster Protection Insight paper, Centre for Disaster Protection, London (2020). 06 SCIENTIFIC DUE DILIGENCE FOR HUMANITARIAN DISASTER RISK FINANCING
Using the worked examples provided by the Drought Risk Finance Science Laboratory (DRiSL) project, we have developed this guide to aid the responsible and effective use of scientific data and modelling in hu- manitarian decision-making. The guide aims to provide an interface between the “providers” of data and models and humanitarian organisations and at-risk groups – the “users” of the models to take decisions – giving each group a better understanding of the other. It supports users to ensure due care is taken in the design of their data and analytics in their DRF systems, and helps providers to better understand what will be needed for their work to be applied responsibly and effectively to achieve humanitarian goals. The guide sets out a pathway of eight checkpoints for users and providers to take together. Working through this sequence of steps collaboratively will ensure the strongest datasets are selected and that basis risk is kept to a minimum. Ultimately, the success of this work depends not just on providers and practitioners working together, but on the ability of humanitarian organisations to connect and collaborate with the at-risk people that their DRF systems seek to support. Effective anticipatory action demands a system-wide approach that integrates at-risk groups, who are at the forefront of both observing and generating risk information and responding to hazards. It is critical that these groups contribute to the quality of the data going into the DRF system, and are then able to see and use its outputs quickly and easily. Establishing a high level of community prepared- ness – the aim at the heart of DRF – can only be achieved through a two-way dialogue with local end-users. The focus of this guide is on the work of data scientists and practitioners on the technical development of DRF systems but it should be read alongside: l Information is Power: Connecting Local Responders to the Risk Information that they need 3 and l Basis risk in disaster risk financing for humanitarian action: Potential approaches to measuring, monitoring, and managing it.4
3 Information is Power: Connecting Local Responders to the Risk Information that they need, Start Network policy brief (2021). 4 Basis risk in disaster risk financing for humanitarian action: Potential approaches to measuring, monitoring, and managing it, Harris, C. and Cardenes, I., Centre for Disaster Protection Insight paper, Centre for Disaster Protection, London (2020). SCIENTIFIC DUE DILIGENCE FOR HUMANITARIAN DISASTER RISK FINANCING 07
INTRODUCTION TO DISASTER RISK FINANCING AND FORECAST-BASED FINANCING
Disaster Risk Financing (DRF) and Forecast-based Financing (FBF) are new forms of humanitarian action that move away from a solely reactive approach to crisis, and instead encourage the humanitarian and development sectors to take a more systematic and robust approach to managing and financing activities to address emerging risks. One key reason why humanitarian action has traditionally arrived late is that it is often triggered by indicators that are based on impact and loss, and is responding to an event that has already occurred.5 Equally, the development sphere has traditionally failed to recognise the dynamism of the context in which longer-term work is implemented. DRF and FBF take a different approach. These techniques focus on monitoring the early risk indicators of a forecasted, impending or imminent crisis, and using those indicators to release financing earlier than would usually be possible. DRF and FBF therefore provide the opportunity both to avert and mitigate the crisis, and to prepare for more timely emergency responses, thereby minimising losses and saving lives. These approaches enable us to act earlier, either before the shock has occurred to enable households to prepare and manage the consequences (“early/anticipatory action”), or very soon after (“early” or “timely response”) to limit the impact. DRF and FBF are similar in that they bring together scientific risk analytics, contingency planning and risk- based funding. They can be developed across both relatively slow-onset and predictable natural hazards, such as drought, flooding, heat and cold waves, and more sudden-onset crises, such as landslides, earthquakes and tsunamis. Conflict and other human-induced crises can also be monitored, and in some cases predicted, allowing the risks to be better managed. Using DRF increases the efficiency, speed, and predictability of crisis manage- ment, and allows humanitarians to act before risks turn into crises, thereby reducing costs and the lives lost in any disaster. However, any DRF system will only be as effective as the scientific models on which it is based. Scientific modelling is an attempt to predict what might happen, and a degree of uncertainty is inherent in any prediction process. The difference between what a scientific model predicts and what actually happens is often referred to as the basis risk (see Key Definitions, page 4, and Appendix 5). This guide advises on how to establish the optimal risk model to trigger funding, whilst identifying the range of data needed within a comparative monitor to manage real-time basis risk. This falls under the first pillar of the DRF system approach (Appendix 5).
5 For additional reasons see: A Dangerous Delay: The cost of late response to early warnings in the 2011 drought in the Horn of Africa, Oxfam (2012). 08 SCIENTIFIC DUE DILIGENCE FOR HUMANITARIAN DISASTER RISK FINANCING
PURPOSE OF THIS GUIDE SUPPORTING DATA SCIENTISTS AND HUMANITARIAN PRACTITIONERS SCIENTIFIC DUE DILIGENCE FOR HUMANITARIAN DISASTER RISK FINANCING 09
This guide was developed by the Drought Risk Finance Science Laboratory (DRiSL) (Appendix 1) to provide a tool for both scientist and humanitarian practitioners to aid responsible and effective use of scientific data and modelling in humanitarian decision-making. The guide aims to act as an interface between providers of data and models, and the humanitarian users of those products. It supports users so they can ensure that due care is taken in the design of DRF systems, and it allows providers to better understand what users require for the responsible application of their work to humanitarian ends. This guide focuses on the early roles and interface between providers and practition- ers in the development of DRF systems. It should be read alongside Information is Power: Connecting Local Responders to the Risk Information that they need6 which focuses on the importance of integrating at-risk communities in the development and application of DRF systems. The guide sets out a pathway of eight checkpoints which data scientists and humanitarian practitioners can follow together. Working through this sequence of steps and questions will ensure that the best possible datasets are selected and that basis risk is reduced (see Key definitions, page 4, for more information on basis risk). The result will be a more robust DRF system founded on the collective knowledge and expertise of both the practitioner and data scientist. Using this guide will help you identify where data failures are most likely to occur, and under what condi- tions. It will help you bring to light sources of uncertainty and limitations so that they can be understood and hopefully, mitigated. The guide also aims to demonstrate why it might be appropriate to integrate a wider suite of data sources into your DRF system for different decision-making purposes. This guide advocates that all learnings gathered during testing and validation of a DRF system are com- municated clearly and openly to all stakeholders so that they can understand and evaluate the system. It is an important principle of DRF that those who design the system are accountable and responsive to those who will ultimately rely on it.
AGREEING THE USE OF THE GUIDE WITH ALL PARTIES
Throughout the process it is critical that the de- cisions taken are reviewed and agreed by both the practitioners and the scientists involved, ensuring roles and responsibilities are clearly communicated at each checkpoint. Also key is the involvement of at-risk communi- ties – from the outset. Decentralised, inclusive and participatory processes should be used to capture on-the-ground forecasting knowledge and empower local responders, make use of local capacity and give agency to those at risk.
6 Information is Power: Connecting Local Responders to the Risk Information that they need, Start Network policy brief (2021). 10 SCIENTIFIC DUE DILIGENCE FOR HUMANITARIAN DISASTER RISK FINANCING
CHECKPOINT 01 CREATING A DATA AND DECISION FRAMEWORK: IDENTIFYING OPERATIONAL NEEDS AND THE QUESTIONS THE MODEL SHOULD LOOK TO ANSWER
At the start of the journey, it is critical that users work with all relevant stakeholders, including at-risk com- munities, to agree a clear impact statement.7 This should set out the point at which the data will be used to trigger funding and the actions and outcomes that funding will seek to achieve.
KEY QUESTIONS FOR THE PRACTITIONERS TO POSE:
ACTIONS? • WHAT ACTIONS WILL BE NEEDED TO MITIGATE THE HAZARD IN QUESTION? • WHAT OUTCOME DO WE SEEK? • WHICH ACTIONS ARE ESSENTIAL AND WHICH ARE ONLY IDEAL OR “NICE TO HAVE”?
INFORMATION? • WHAT INFORMATION DO WE NEED TO KNOW IN ORDER TO TRIGGER THOSE ACTIONS AT THE RIGHT TIME?
WHEN? • WHEN DO WE NEED TO KNOW THIS INFORMATION TO TRIGGER PRE-EMPTIVE AND PROTECTIVE ACTION?
At this point you may well not be an expert in the hazard you are trying to mitigate. Try to think in broad terms about what, in an ideal world, you would like your model to tell you and at which point. You are essen- tially creating a wish list of the practical information you would want to know and when. The information you generate at this stage can be refined later as the availability and skill level of the data is better understood. The specifics, including what can be reliably achieved in what timescale, can be decided at later checkpoints. With this advance thinking done, the users can now work with the providers to explore the system to be built, basing this on three elements: 01. QUESTIONS WHICH QUESTIONS DOES THE DATA SYSTEM NEED TO ANSWER TO ENABLE EFFECTIVE DECISIONS TO BE MADE? This should include working closely with at-risk groups to identify and assess the real-life impacts of the risks you are addressing. Impact forecasting is key to at-risk communities building their ability to anticipate
7 For more information on working with at-risk groups, see: Information is Power: Connecting Local Responders to the Risk Information that they need, Start Network policy brief, 2021. SCIENTIFIC DUE DILIGENCE FOR HUMANITARIAN DISASTER RISK FINANCING 11
and respond to hazards. Examples of questions are: how much rain has fallen and is this normal or above or below normal? Will this impact the crops? Will this level of crop loss effect people’s food security? 02. DECISIONS
WHICH DECISIONS NEED TO BE TAKEN AND WHEN DO QUESTIONS NEED TO BE ANSWERED? Working with at-risk communities, practitioners should investigate to understand the impact the systems need to have to address the local impacts of the hazard. 03. DATA WHAT DATASETS ARE AVAILABLE TO ANSWER THE QUESTIONS? Models and indices will be tested to see if they can reliably answer the questions that have been identified. In some instance, you will begin with a blank slate, in others there may be a proposed system that needs testing and verifying. The provider and users should seek to jointly surface and investigate all of the pos- sible existing data sources.
EXAMPLES: TIPS: l Most drought monitoring programs in l Before getting too deep into the analysis, take Pakistan focus on the summer monsoon time to research your country risks including and yet the secondary winter growing developing your understanding of the risks season (“Rabi”) is also of significant from the perspective of at-risk communities. economic and agricultural importance and l Each hazard will be shaped by different key for food security of the poorest. factors: for drought induced food insecurity l Humanitarian practitioners (users) want to for example, crops, seasons, climates as build a DRF system that can trigger action well as human influence on the system, in advance of the lean season, to protect such as river dams or irrigations, land use, people from food insecurity. To do so, they etc. For example: Zimbabwe has quite a will need to know the number of people uniform climate and rain-fed agriculture; forecasted to be affected – over what area Pakistan makes widespread use of natural and time period – by food insecurity arising and artificial irrigation; Madagascar has a from a poor harvest in the Rabi season. complex climate which varies dramatically over the country, and its available data quality l The practitioners know that Pakistan is can vary. These are the types of things you highly irrigated but do not know the best way need to understand before embarking on your to develop the trigger system/framework analysis as they will influence the outcomes that reflects this. They therefore work with you are seeking to achieve. As a first step, map Reading University (the scientific data out these characteristics fully so they can be providers) who can assess the available included in the data system as it develops. data options and identify those that will work best in this context, providing the l Agreements and responsibilities on this most reliable predictor of the Rabi harvest scoping need to be established and should lay levels, through investigation and evidence. the foundations of the remaining checkpoints. 12 SCIENTIFIC DUE DILIGENCE FOR HUMANITARIAN DISASTER RISK FINANCING
CHECKPOINT 02 COLLECTING AND TESTING THE DATA THROUGH DIFFERENT LENSES – GEOGRAPHICAL, SPATIAL, TEMPORAL, SEVERITY AND IMPACT8
Many variables can be applied in the testing of data and models. It will not be practical to test a DRF system from every possible angle and it is therefore important to identify the most informative variables. The key variables for a humanitarian DRF system are:
WINDOW OF ACTION
The time period for which the system will trigger, including: forecast and anticipation, impact, response and recovery. For some hazards, annual seasonal periods will come into play, and it is likely that the strength of data and skill will vary at different times of the year. SEVERITY AND FREQUENCY OF EVENTS
How good is the data or the models when applied at different return periods (RP)?9 In other words, a model that performs well for 1 in 10 year event (a moderate event), may not work so well for a 1 in 4 year event (a mild event). Which severity levels – mild, moderate of extreme – can be easily identified? GEOGRAPHICAL RESOLUTION
Based on the data and decision-making framework, the resolution or scale of any geographical testing. For example, does the model work best at country level, district level, provincial or another denomination, such as livelihood zone or biome? Some models may perform well at a high level but poorly at a more granular resolution and vice versa.
Testing these factors will be a significant task for the provider and the user should therefore do what they can to narrow the focus and make the analysis more targeted. For example, it may be acceptable to test just three severity scenario return periods (1 in 3 years, 1 in 10 year and 1 in 20 years) rather than half a dozen or more. By the time checkpoints 1 and 2 are complete, the users and providers will have created a clear due dili- gence framework.
8 This refers to the input and output of data, meaning both hazard (e.g. rainfall, temperature) and impact data (e.g. yields, health data). You need to decide what you are trying to predict. 9 The return period is the interval – generally measured in years – between events of a particular magnitude. w
SCIENTIFIC DUE DILIGENCE FOR HUMANITARIAN DISASTER RISK FINANCING 13
EXAMPLES:
l In the DRiSL project, the uncertainty in drought finance/action triggering was assessed for an ‘anticipation window’10 at the end of the local wet season, over the last ~40 years, based on a range of plausible drought metrics (derived from various models and sources including rainfall, soil moisture, Water Requirement Satisfaction Index and Normalised Difference Vegetation Index). l Triggering events were identified in the event that drought indices (during the trigger window, tied to the annual livelihood calendar) exceed pre-defined return period scenarios values.
l In this case, the range of severity scenarios comprised 3, 7, 10 and 20-year return periods.
TIPS:
l The most appropriate metric of a hazard, and lead time and contingency plans/activities. the ability to anticipate it, may be highly region- specific and depend on the climatic regime. l Quality may be more important than quantity, i.e. This means that protocols need to be adjusted shorter periods of more recent data may be more for different climatic regimes, and that hazard informative than longer periods of older and less indicators must be carefully evaluated. detailed data (for example pre-1950 data when science and technology were less advanced). l It is very important to check multiple data sources, or you are likely to miss important l Ensure all users are familiar with the concept of errors, biases or omissions. For example, you return period and the probability of occurrence might suspect that WRSI could be a good (inverse of Return Period) terminology: what each drought indicator but you still check other return period means in terms of reliability or odds possible predictors and compare the skill. that a catastrophe will not occur within a given timeframe. l Discuss and agree on what classifies as a hazard ‘event’. Your definitions will be more l Does the past predict the future? Factors such robust if you involve at-risk groups in this as population growth, increased crop resilience, work - mapping impacts, defining triggers climate change and development, can mean that and early actions, identifying vulnerable return periods for particular magnitude events groups and defining needs. — which are often based on records dating back less than a century — change over time. For this l Examples of indicators of a hazard event could reason, it is important to use the most up-to-date include, for example, the WRSI value dropping records and data that is also labelled clearly. below 60; flood depths being >0m; temperatures exceeding the 90th percentile; and a hurricane l If there are trends in the data that are not of reaching within a mile of a particular land boundary. interest (e.g. a decrease in yield caused by changes in land use rather than by climate), you l Factor in model processing time, especially should de-trend the data or remove it so you where the model relies on another data source can study the influence of other factors (e.g. a to calculate an output. This will affect the large observed trend in production and yield).
10 In this instance, the trigger windows lay towards the end of the main growing season, specifically in February-March. 14 SCIENTIFIC DUE DILIGENCE FOR HUMANITARIAN DISASTER RISK FINANCING
CHECKPOINT 03 OPTIMISING THE AVAILABLE TESTING DATA: CREATING STRONGER HISTORICAL DATASETS
Models are tested by essentially running them backwards and comparing what the model depicts as hav- ing happened to what was actually observed and measuring the difference. Providers and users need to agree which historical datasets, and the period they span, will be used for validation. Often the strongest approach will include use of multiple sets of historical data variables so that different aspects of the model can be tested – from physical hazards through to human impact. Across all hazards, a common problem in risk modelling work is the lack of historical observed data from past events, such as actual rainfall, the number of houses destroyed or the number of deaths. To address this, providers and users should agree what historical data can be used for testing, and where there are opportunities to strengthen the data by filing in the gaps in a justifiable and auditable way. Once agreement is reached, this work should be undertaken. The user can support this work but ultimate responsibility for data testing and collection rests with the provider and this should be included in any contractual agreements. Efforts to enhance the testing data could include, for example, crowdsourcing information on historical drought impact using mobile phone surveys or interviews with the at-risk communities and local elders. There will always be significant limitations to this work, but where there are opportunities to strengthen historical data they should be taken. When working with communities to form more comprehensive his- torical event profiles, it is essential that time and attention are given to communicating what and how in- formation they provide will be used. It is also important to provide a way for those communities to receive the results of the model and risk data testing so they can further validate it. Efforts should be made not to be extractive, but to use the information provided, under checkpoint 3 to begin a dialogue with at-risk communities which can be followed through in the later checkpoint stages. Always make a note of any enhancements made to historic data or of any proxies used. Store both the raw data and the enhanced data so that they can be easily understood and used by others. Enhancements to historical data should also be shared with others who may also be looking to test models against histori- cal outcomes. To be clear, here we are talking only output data (e.g. yields). It is not possible to magically produce any input data (e.g. rainfall) if it was not observed at the time. SCIENTIFIC DUE DILIGENCE FOR HUMANITARIAN DISASTER RISK FINANCING 15
EXAMPLES: TIPS: l In the DRiSL project, the main datasets used l Historical data varies in both its availability and to asses drought impacts were: the Integrated usability. It can be biased, trended, missing, not Phase Classification (IPC); socioeconomic spatially detailed and often does not go back datasets from secondary sources including many years. As a result, factor analysis may not agricultural, health and economic indicators be possible.11 that have a direct or indirect relationship with l Really simple things can trip you up. For example, food security; crowdsourced information from in Madagascar the growing season spans two Interactive Voice Response (IVR) surveys (and calendar years. When sharing data labeled “1991”, from focus group discussions when available); it needs to be made crystal clear to other project and disaster information from EMDAT. This participants whether this means data from the created an historical profile of what was 1990-91 season, or from the 1991-92 season. observed in terms of food insecurity in the period in question. This was then compared l Another common error is categorising a past to incidents of drought and the various hurricane event (using the Saffir-Simpson depictions provided by drought model metrics. Hurricane Wind Scale) incorrectly by using its landfall windspeed instead of its maximum l For flood DRF initiatives, you may want to windspeed, which could have been over open enhance flood maps by integrating any local water. flood mapping or historical high-level data and local knowledge of past flood events and the l It may be useful to weight incomplete or areas impacted (where these exist and you less reliable data sources more lightly, while have permission to use them). still using them as a factor in predictions.
11 Factor analysis is the practice of condensing many variables into just a few so that your research data is easier to work with. 16 SCIENTIFIC DUE DILIGENCE FOR HUMANITARIAN DISASTER RISK FINANCING
CHECKPOINT 04 UNCERTAINTY TESTING: TESTING AND COMPARING MODELS AND DATA, AND PRESENTING THE RESULTS
With the first three checkpoints completed, the data providers can now undertake the first round of testing. Uncertainty tests should provide a comparison between datasets to identify the extent to which they agree or disagree with each other, and how much information some give about the others. This testing does not consider which datasets and models are performing best; the point is simply to com- pare results across multiple data sources to assess the level of agreement. The results will then indicate the uncertainty or level of agreement across the possible model and data choices depicting similar processes, such as rainfall or wind speed. The uncertainty testing may also reveal some years and conditions where there is a high degree of agree- ment and disagreement between models and data, indicating the influence of these external conditions on potential basis risk. For example, El Niño years may affect the natural systems and cause more or less agreement between the datasets.
EXAMPLES: l In the DRiSL project, the uncertainty in DRF ‘triggering’ was associated with the choice of biophysical drought metrics. We assessed 15 different drought metrics including soil moisture (SM) and other commonly used agricultural and meteorological drought indicators. For a given event magnitude, expressed as a particular event return period, the degree of uncertainty in triggering between two biophysical datasets was assessed by calculating the Jaccard score. This is the number of times both datasets trigger at the same time as a proportion of all triggers across the two datasets. A Jaccard score of one means the two datasets agreed in every triggered event, and a score zero means there was no agreement in any triggered events. The 15 biophysical metrics were compared against one another to provide a matrix of Jaccard scores, as shown in the example below for the Masvingo province of Zimbabwe.
7 For more information on working with at-risk groups, see: Information is Power: Connecting Local Responders to the Risk Information that they need, Start Network policy brief, 2021. SCIENTIFIC DUE DILIGENCE FOR HUMANITARIAN DISASTER RISK FINANCING 17
JACCARD SCORE FOR THE MASVINGO PROVINCE, ZIMBABWE DROUGHT MODELS FOR THE 1 IN 3 YEAR, 1 IN 7 YEAR, 1 IN 10 YEAR AND 1 IN 20 YEAR RETURN PERIODS. THE DIAGRAM REFLECTS THE DEGREE OF CER- TAINTY BETWEEN 15 DATASETS (SOME OF WHICH ARE WIDELY USED, AND OTHER POTENTIALS). THE BLUE
SQUARESwidely INDICATE widelyused, and used, POOR other and potentials). otherAGREEMENT; potentials). The blue The THEsquares blue RED squares indicate SQUARES indicate poor agreement; poor SHOW agreement; the STRONG red the squares red squares AGREEMENT. show strong show strong agreementagreement. . widely widelyused, and used, other and potentials). other potentials). The blue The squares blue squares indicate indicate poor agreement; poor agreement; the red the squares red squares show strong show strong agreementagreement. .
There Thereis one ismatrix one matrix for each for event each eventmagnitude magnitude: 1 in 3: year,1 in 3 1year, in 7 year,1 in 7 1year, in 10 1 year in 10 and year 1 andin 20 1 year in 20 year Therereturn Therereturn isperiods. one ismatrixperiods. one The matrix forJaccard The each forJaccard score event each ofscore eventmagnitude triggering of magnitude triggering: agreement1 in 3: agreement1year, in 3 between 1year, in 7 between1year, in any 7 1year, twoin any 10 1datasets yeartwoin 10 datasetsand year (shown 1 andin 20(shown 1on yearin the 20 on yearX the X returnand Y returnandaxes) periods. Y canaxes) periods. Thebe can read Jaccard Thebe from read Jaccard score the from intersectionofscore the triggering intersectionof triggering of agreement rows of agreement and rows betweencolumns. and betweencolumns. any Shades two any Shades datasetsof two blue datasetsof indicate b(shownlue indicate (shown poor on the poor on X the X andagreement; Y andagreement;axes) Y canaxes)shades be canshades read of bered from read ofand red the fromorange and intersection the orange indicate intersection indicate of strong rows of strong agreement.and rows columns. agreement.and columns. Shades Shades of blue of indicate blue indicate poor poor There is oneagreement; matrixagreement; shades for shadeseach of red of eventand red orange and magnitude: orange indicate indicate strong 1 in strong agreement. 3 year, agreement. 1 in 7 year, 1 in 10 year and 1 in 20 year return periods. The The diagramsJaccard The diagrams indicate score indicate that of triggering 1. that Agreement 1. Agreement agreement in triggering in triggering between between equally any equally twovalid datasetsdrought valid drought indicators (shown indicators is oftenon isthe often X and Y axes) can be readThelow from e.g.diagramsThelow Consider e.g.diagramsthe indicateConsider intersection events indicate that events with 1. that Agreementreturn ofwith 1. rows Agreementreturn period andin period triggeringof 1columns. in in triggeringof 3 years1 betweenin 3 (top yearsShades between leftequally (top). Over leftofequally valid blue). the Over drought valid 40 indicate the year drought 40 indicators period year poor indicators period there isagreement; often willthere is often will shades of red and orangelowbe about e.g.lowbe indicateConsider about13 e.g. events. Consider 13 events events.strong The events Jaccardwith The agreement. return Jaccardwith score return period fromscore period oftwo from 1 inwidely oftwo 3 1years inwidely used 3 (topyears metrics used left (top ).metrics Over WRSIleft). the OvervsWRSI NDVI40 the vsyear is NDVI40 about period year is about 0.5,period there such 0.5, willthere such will that onlythat about only abouthalf the half events the events triggered triggered using oneusing metric one metric will also will be also triggered be triggered in the inother the other2. 2. The diagramsbe about beindicate about13 events. 13 that events. The 1. Jaccard AgreementThe Jaccard score fromscore in triggeringtwo from widely two widely used between metrics used metrics equallyWRSI vsWRSI NDVI valid vs isNDVI droughtabout is about0.5, indicatorssuch 0.5, such is often low AgreementthatAgreement only of drought about of drought half triggering the triggering events generally triggered generally declines using declines with one increasing metricwith increasing will RP also - until beRP triggered -RP20 until whenRP20 in thewhenscores other scores become 2. become e.g. Considerthat only events about with half the return events period triggered of using1 in one3 years metric (top will alsoleft). be Overtriggered the in 40 the year other period 2. there will be about AunstablegreementAunstablegreement as ofa result drought as ofa result ofdrought thetriggering ofsmall thetriggering samplesmall generally sample generally size declines of size events. declines of with events. This increasing with suggests This increasing suggests RP that - until triggeringRP that -RP20 until triggering when RP20of more whenscores of extrememore scores become extreme become 13 events. The Jaccard score from two widely used metrics WRSI vs NDVI is about 0.5, such that only unstableeventsunstableevents may as abe mayresult inadvisableas abe resultof inadvisable the of smallunless the sample smallunless a certain sample asize certain drought of size events. drought of index events. This isindex suggestsparticularly This is suggestsparticularly that trusted. triggering that trusted. triggering It also of more Itindicates also of extrememore indicates that extreme that about half the events triggered using one metric will also be triggered in the other 2. Agreement of drought eventsevents may be may inadvisable be inadvisable unless unless a certain a certain drought drought index indexis particularly is particularly trusted. trusted. It also It indicates also indicates that that triggering generally declines with increasing RP - until RP20 when scores become unstable as a result of the small sample size of events. This suggests that triggering of more extreme events may19 be19 inadvisable unless a certain drought index is particularly trusted. It also indicates that above RP20 more19 instances19 of basis risk are likely and so if triggered at this level, additional checks and verification should be considered. 18 SCIENTIFIC DUE DILIGENCE FOR HUMANITARIAN DISASTER RISK FINANCING
TIPS: l Weighting is important. In drought models you need to know the important dekads12 for predicting the success of the crops, but assessing how to weight them is difficult. You could weight evenly over the season, use weights from a vendor, or you could weight based on which dekads have the most rain historically and/or by speaking to the at-risk community. l Ensure that the data being assessed stays aligned with the actions defined in checkpoint 1. For example, there may be much more agreement and lower uncertainty in end-of-season measures, but these are no use for triggering DRF systems at the appropriate time. l In composite measures such as the category of cyclone, the various sub-indicators, such as surge level, wind speed, rainfall volumes, should be compared across the different storm forecast products. Comparing the uncertainty of just the categorisation at various time intervals will be less useful.
12 A dekad is a period of 10 days. SCIENTIFIC DUE DILIGENCE FOR HUMANITARIAN DISASTER RISK FINANCING 19
CHECKPOINT 05 SKILL TESTING
Skill testing involves inputting data from past events into a model to test how well it predicts what actually happened (hence the need for checkpoint 3). The process is also known as “hindcasting” or “back casting” and it reveals the difference between what the model predicts should have happened and the reality of the actual event. The performance of the model is referred to as its “skill” and it gives an indication of how well the model is likely to perform as a predictor of future events. For instance, in rainfall forecasts, you would compare the hindcast against historical rainfall datasets to assess the skill of your forecast. The process should also indicate any instances of low skill and high uncer- tainty and what the drivers of this might be. Specific statistical tests can be used to provide a skill score. The higher the skill score the better performing the data or mod- el is. However, the lack of comprehensive and reliable observed test data means that skill scores should not be taken in isolation as an indicator of the best data. The quality of historical data and the levels of uncertainty and complexity from checkpoint 4 should also be considered.
EXAMPLES:
l The following standard metrics of skill were calculated when validating the Pakistan drought model (which uses NDVI to predict wheat yield and production) against district-level observations of yield and production. (The metrics were calculated for hindcasts performed for each year for which NDVI and yield data were available). To view the full set of results please see the validation report, produced by Professor Emily Black and Dr Ross Maidment, Reading University. I. Pearson correlation coefficient between observed and simulated data: a measure of the closeness of the relationship between the observations and simulations. A score of ‘1’ indicates a perfect linear positive relationship; a score of ‘-1’ indicates a perfect linear negative relationship and a score close to ‘0’ indicates no relationship. II. Mean absolute error (MAE): a measure of the errors between the observed and simulated data points. A mean absolute error of 0 indicates a perfect match; a high mean absolute error indicates severe differences between the observed and simulated data. III. Bias between observed and simulated data: a measure of systematic bias between observed and simulated data. A positive value indicates that the simulations tend to be greater than the observations; a negative value indicates that the simulations tend to be smaller than the observations; a zero value indicates that the simulations are neither systematically greater nor less than the simulations. Note that it is possible to have a large mean absolute error, with a near zero bias. less than the simulations. Note that it is possible to have a large mean absolute error, with a near zero bias.
Example of Pearson correlation coefficient The figures below show the correlation between observed and predicted yield production for the years 1982-2018 (wheat). The hindcasts20 were carried out for the following dates:SCIENTIFIC 15 DUE November, DILIGENCE FOR 15 HUMANITARIAN DISASTER RISK FINANCING December, 15 January, 31 January, 15 February, 28 February, 15 March, 31 March, 15 April, 30 April.
It can be seen that the correlations are low until January and continue to improve during February and March. These results show the model’s ability to capture observed yield production, how it EXAMPLElessless thanless thanless thethan thanthe simulations.OF the simulations. thePEARSON simulations. simulations. Note Note Notethat CORRELATION Note that it that is itthat possible isit possibleis it possibleis possible to hto aveCOEFFICIENT tohave toha avelarge haave large a largemean a large mean mean absolute mean absolute absolute absolute error, error, error, with error, with awith witha a a improves steadily through the early nearpartlesslessnear lesszeronear thanthan nearlesszeroof than bias.zero thethethan thebias.zero the bias. simulations.simulations. the bias.simulations.growing simulations. NoteNote Note season thatNotethat that it itthat is isit possible possibleis it. possible isThe possible toto improvements htohaveave htoave aha ave largelarge a large a meanlargemean mean mean absoluteabsolute absolute absolutethen error,error, error, levelerror, withwith with aawith offa a in nearnearnear zerozeronear zero bias. bias.zero bias. bias. the later part of the season. This is because, for most of the districts, the optimal period of interest ExampleExampleTheExample Example of figures Pearson of ofPearson ofPearson Pearson correlationbelow correlation correlation correlation showcoefficient coefficient coefficient thecoefficient correlation between observed and predicted yield production for the years selected during the modelling processTheExampleExampleThe1982-2018 figureExampleThe f TheExampleiguredoes fsigure of belowoffigures Pearson ofPearsonbelows Pearsonofbelownot sshow(wheat). Pearsonbelow show correlationcorrelation show extendthe correlation show the ccorrelation orrelationthe Thec orrelationthe c coefficientorrelationcoefficient cinto coefficienthindcastsorrelation coefficientbetween between later between between observed wereobservedMarch observed observed carriedand and predicted andand predicted and predicted out April.predicted yield for yield yieldproductionthe yieldproduction productionfollowing production for for the for the dates:for the the 15 November, 15 December, 15 yearsTheTheyearsJanuary,Theyears 1982 ffigureigureyearsThe 1982 figure -1982 2018fssigure 1982- belowbelow2018 s31 - below2018(wheat).s- 2018below January, (wheat). showshow (wheat). show (wheat). Theshow thethe Thethe chindcastsc orrelation Theorrelationthe15 chindcasts Theorrelation hindcasts cFebruary,orrelation hindcasts were betweenbetween were between werecarried between were carried 28 carriedobservedobserved out carriedobservedFebruary, outobserved for out for theandoutand for theand following forpredicted predicted theand 15following predictedthe following predicted March, following dates: yieldyield dates: yield dates: productionyieldproduction15 31dates: production15November, productionMarch,15 November, 15November, November, forfor for the15the15 for the15 April,15 the 15 30 April. December,yearsyearsDecember,yearsDecember, December,years 19821982 198215- -1982 2018201815January,- 201815 January,- 15(wheat).2018(wheat).January, (wheat).January, 31 (wheat). 31January, The31The January, The 31January, hindcastshindcasts TheJanuary, 15hindcasts hindcasts15February, 15 February, werewere 15February, were February, werecarried28carried carried28February, 28carried February, outout 28February, out February, forfor out15 for thethe 15March, for the 15followingfollowingMarch, the 15followingMarch, 31March,following 31March, dates: dates:31 March, dates: 31March, dates:15 March, 1515 15April,15 November,November, 15 April,15November, 15April, 30November, April, 30April 30 April 1515 .30April 15 .April 15. . It can be seen that the correlations are low until January and continue to improve during February and December,December, December, December, 1515 15 January,January, 15January, January, 3131 31 January,January, 31January, January, 1515 15 February,February, 15February, February, 2828 28 February,February, 28February, February, 1515 15 March,March, 15March, March, 3131 31 March,March, 31March, March, 1515 15 April,April, 15April, April, 3030 30 AprilApril 30April. . April . . It can ItMarch. can It be canIt beseen can be seen beThesethatseen seenthat the that thethat correlations resultsthe correlations the correlations correlations showare are low are low areuntilthe low until low Januaryuntil model’s untilJanuary January Januaryand and continue abilityand continue and continue continue to to improveto capturetoimprove toimprove improve during during during observed February during February February February yield production, how it improves andItItandsteadily canMarch.canItand canMarch. andIt bebe March.can be seen seenThMarch. beseen eseThthrough that seenthatThese results thateseTh resultsthe thethatese resultsthe correlationscorrelationsshow results the correlationsshow correlationsshow theearly show the model’s the arearemodel’s the partaremodel’s lowlow model’sare abilitylow untiluntil abilityoflow until ability to Januarythe Januaryuntilability capturetoJanuary tocaptureJanuarygrowing to capture andand capture observedand continuecontinue observedand continue observed season.continueobserved yield toto yield improvetoimprove yieldproduction improveto yieldproduction Theimprove production duringproductionduring duringimprovements, how during, FebruaryFebruaryhow, itFebruaryhow , ithowFebruary it it then level off in the later part improvesandandimprovesofandimproves March.theMarch.improvesand March.steadily season.March. steadily Th Thsteadily eseTh ese steadilythrough eseTh resultsthroughresultsese throughresultsThis through resultsthe showshow the earlyisshow the earlyshow thebecause, thethe early partthe model’s model’s early partthe model’sof part model’s theofpart ofabilitythe ability forgrowing oftheability growing theability mostgrowing toto growing capturetocaptureseason capturetoseason of captureseason. seasonobservedtheTheobserved. observedThe .improvements districts,Theobserved .improvements The yieldimprovementsyield improvementsyield productionyieldproduction productionthe then production then optimal levelthen,, howthenlevelhow, offlevelhow , it itlevel howoffin periodit off in itoff in in of interest selected during the theimprovesimprovesthemodelling laterimprovesthe latertheimproves partlater later partsteadilysteadily of partsteadily theofpart processsteadily ofthe season.throughthrough ofthe season.through the season.through season. This thethedoes Thisthe earlyis earlyThis the because, early isThis not because, ispartearlypart because, ispart extend because,of ofpart for theofthe for most theof growingforgrowing mostthe forgrowing into mostof growing most theof season seasonlaterofthe districts, seasonofthe districts, seasonthe .districts,. March TheThe districts,. theThe improvements.improvements Thethe optimalimprovements the and optimalimprovements the optimal optimalperiodApril. period then thenperiod of thenperiod levelinterestofthenlevel oflevelinterest ofoffinterestleveloff interest off inin off in in selectedthetheselectedtheselected laterlaterselectedthe duringlater partlaterpartduring partduring the of ofduringpart theofthe modelling thetheof season.modellingseason. thethe modellingseason. modellingseason. process ThisThis processThis isprocessis This because, because, isprocessdoes because, isdoes because, doesnot does notforfor extend notfor mostmostextend not for mostextend into extendmostofof into theofthe later into theof districts, districts,laterinto the Marchdistricts,later laterdistricts,March March the theand March the and optimaloptimal April. theand optimal April. and optimal April. period periodApril. period period ofof interestofinterest interestof interest selectedselectedselectedselected duringduring during during thethe the modellingmodelling the modelling modelling processprocess process process doesdoes does notdoesnot not extendextend not extend extend intointo into laterlater into later MarchlaterMarch March March andand and April.April. and April. April. Correlation between MID-NOV MID-DEC MID-JAN MID-FEB observed and predicted yield production, plotted for hindcasts carried out between mid- November – end-April MID-MAR END-MAR MID-APR END-APR for 1982-2018 (wheat)
Correlation between observed and CorrelationpredictedCorrelationCorrelationCorrelation between between betweenyield between observed observed production,observed observed and and predicted and predicted and predicted predicted yieldplotted yield yieldproduction, yieldproduction, production, production,for plotted hindcasts plotted plotted plottedfor for hindcasts for hindcasts for hindcasts carried hindcasts carried carried carried outcarried out out out out betweenCorrelationCorrelationbetweenCorrelationbetweenbetweenCorrelation mid mid- Novemberbetweenbetweenmid- Novemberbetweenmid- Novemberbetween-November observedobserved – observed end – observed end –-April end – and-and Aprilend - andforApril predictedpredicted -and Aprilfor 1982predicted for 1982predicted for -19822018 yield1982-yield2018 -yield 2018(wheat) -production,yieldproduction,2018 (wheat) production, (wheat) production, (wheat) plottedplotted plotted plotted forfor for hindcastshindcasts for hindcasts hindcasts carriedcarried carried carried outout out out between mid-November – end-April betweenbetweenfor between between1982 midmid mid--NovemberNovember -mid2018-November-November (wheat) –– endend – end -–-AprilApril end-April - forAprilfor for 19821982 for 1982- -198220182018-2018- (wheat)2018(wheat) (wheat) (wheat)
TipTips Tips Tips s TipTipTipss Tips s • • Between• Between• BetweenBetween the the provider the provider the provider provider and and user and user and thereuser thereuser there needs there needs needs to needs beto tobesomeone tobe someone be someone someone who who iswho good iswho goodis goodatis good statistics!at atstatistics! atstatistics! statistics! •• • • Some• Between BetweenSome• BetweenSome testsSomeBetween tests testsare thethe testsarethe highly providerprovider arethe highlyprovider are highlyprovider sensitive. highly andsensitive.and sensitive.and user user sensitive.and userGather there thereuserGather thereGather as thereGather needsneeds much asneeds asmuch needs to toasmuch information beto bemuch information beto someonesomeone information besomeone information someone about whowho about who about is theiswho about good goodisthe datasets goodtheis datasets goodattheat datasets statistics!atstatistics! datasets usedstatistics!at statistics!used usedas usedyouas asyou can, asyou can, you can, can, Tips •• TIPS: and• SomeSomeand •check Someand checkandSome teststests check the tests check theare testsaremodel the are model highly highlythe aremodel highly parameters. model highly sensitive.parameters.sensitive. parameters.sensitive. parameters.sensitive. GatherIfGather model GatherIf modelIfGather model asasIf parameters model muchasmuch parameters muchas parameters much informationparametersinformation information (e.g. information (e.g. crop(e.g. aboutcrop(e.g.about calendars cropabout crop calendarsabout thethe calendars the calendars datasetsdatasets theor datasets resolutionor datasets orresolution used usedorresolution used resolution asusedas used) youasyou used) youas used) can,can, you used) can, can, doand anddonotanddo not checkcheckaligndoand not check align not ,check align theyourthe ,align theyour modelmodel, resultsyourthe model, yourresults model results parameters.parameters. could resultsparameters. could parameters. could vastly could vastly IfvastlyIf vary. model modelIfvastly vary.model If vary.Check model parameters vary.parametersCheck parametersCheck how parametersCheck how sensitivehow (e.g.sensitivehow(e.g. sensitive(e.g. crop sensitivecrop(e.g. the crop the calendarscalendarsresultscrop the calendarsresults the calendarsresults are results ororare to resolutionorareresolution theto areresolutionor tothe wayresolution tothe way the you wayused)used) you wayused) you used) you • Between the provider and user therehavedodohave do haveaskednotnot havedonotasked alignneedsalign askednot thalign askede , ,th align yourquestion.your e,th yourquestion. e th, toresultsresultsyourquestion.e resultsquestion. be results could could someone could could vastlyvastly vastly vastly vary.vary. vary. Checkvary.Checkwho Check Check howhow is how sensitive sensitivehowgood sensitive sensitive thetheat the resultsresultsstatistics! the results results areare are toto are thetothe the to wayway the way youyou way you you havehave askedasked ththee question.question. • l• Too• TooBetween• many haveToo manyToohave many variablesasked many askedvariables variablesthe th variablese thcan question.providere can question. muddy can muddy can muddy themuddy and the water the water theuser waters. waterTrys. Trytheres to. Trys useto. Try touse ones needs touse ones use onesthat onesthat are that arethat directly are directly are directlyl directlyrelevantMake relevant relevant relevant (i.e. sure (i.e. do (i.e. do no (i.e.the do tno do not temporal not t and spatial scales • Some tests are highly sensitive.•• Gather as much information about the datasets used as you can, use• TooToouseto •yield Toouse bemanyyieldmanyuseToo yielddatamany someone yieldmany data variablesvariables fordata variables datafor allvariables for crops,all canforcan allcrops, whocan all crops, muddymuddy onlycan crops, muddy only is muddy theonly thethe goodtheonly ones the the waterwater onesthe thewater onesyou at water s onessyou.. areTrystatistics!Try syou. areTry syou tointerestedto. areTry usetointeresteduse are interesteduseto onesones interesteduse ones in) thatonesthat .in) that in). are arethat in). are directlydirectly. are directly directly used relevantrelevant relevant relevant are (i.e.(i.e. (i.e. appropriate. dodo (i.e. do nono dotnot tno t A model’s skill can and check the model parameters.• • When•use useWhen •If useWhen yieldcomparingyieldmodelWhenuse yieldcomparing datacomparingyielddata comparingdata forfordata differentparameters for allalldifferent for allcrops,differentcrops, differentallcrops, types crops, onlyonlytypes onlytypes of the thetypesonly datasetsof the ones(e.g.ofdatasetsones the ofdatasetsones datasetsyouonesyou (f oryoucrop ( arearef exampleoryou ( aref or interestedexample interested( fare orexampleinterestedcalendars example interested in thein in) in)thein DRiSL .in).thein DRiSL .in)the DRiSL project: . orDRiSL project: project:resolutiondecrease project: significantly used) when the prediction •• Some tests are highly sensitive. Gather as much agriculture/health/economic• WhenWhenagriculture/health/economic• Whenagriculture/health/economicagriculture/health/economicWhen comparingcomparing comparing comparing differentdifferent different different typesindicatorstypes typesindicators indicators types ofof indicators datasetsofdatasets to datasetsof theto datasets tothe biophysical ( (tofthefor or biophysical (thef exampleorexamplebiophysical (f exampleorbiophysical example data in indata )the thein datait the)inisdata DRiSLitDRiSL )hard the isitDRiSL ) hardis itDRiSL project: project:to hardis project:harddecidetowindow project:todecide todecide what decide what what is(area what is is is and lead time) increases. do not align, your results could vastly“correct”agriculture/health/economicagriculture/health/economic“correct”informationagriculture/health/economic“correct”“correct”agriculture/health/economic asvary. aas point asa pointasa Check point ofabouta point truthof oftruth oftruth for thehow truth for indicatorsidentifyingindicators for datasetsindicatorsidentifying for identifyingindicatorssensitive identifying to tohumanitarian thetothe humanitarian usedthe to humanitarian biophysicalbiophysical thehumanitarian biophysical the biophysicalas needs. you resultsneeds. datadata needs. data needs.can,)) data it it ) is is it hard ) hardis areit hard is totohard to decidetodecide decideto the decide whatwhat whatway whatisis is youis “correct”“correct”and check asas aa pointpoint the of ofmodel truthtruth forfor parameters. identifyingidentifying humanitarianhumanitarian If model needs.needs. l Selecting the most appropriate historic event have asked the question. • l• Check• Check• “correct”Check imageryCheck“correct” imagery imagery imagery asresolution aasresolution point aresolution point resolution ofs between truthofs between truths between sfor between for theidentifying theidentifying model the model the model and humanitarianmodel and humanitarian observed and observed and observed observed data needs. data needs. mdataarry datam arry m up.arry m up. arry up. up. •• • CheckCheckparameters• CheckCheck imageryimagery imagery imagery resolutionresolution (e.g.resolution resolution cropss betweenbetweens between scalendars between thethe the modelmodel the model modelor andand andresolution observedobserved and observed observed datadata data mmdataarry arrymarry m up.up.arry sectionup. up. is important. Ideally, model calibration • Too many variables can muddy theused) water do snot. Try align, to youruse onesresults that could are vastly directly vary. relevantshould (i.e.be completed 22do 22 22no 22 t using multiple historic events Check how sensitive the results are to the way of varying magnitude.2222 22 22 use yield data for all crops, only theyou ones have youasked are the interested question. in). l It is hard to turn historical data into a binary “event • When comparing different typesl ofToo datasets many variables (for example can muddy in the waters. DRiSL Tryproject: / no event” indicator. How do you combine them agriculture/health/economic indicatorsto use ones to the that biophysical are directly relevantdata) it (i.e. is hard do not to decideinto indiceswhat is in order to apply ranks and return use yield data for all crops, only the ones you periods? How do you decide which biophysical “correct” as a point of truth for identifyingare interested humanitarian in). needs. data source is “good”? In DRiSL, Jaccard scores were used (see checkpoint 4 above) to compare • l When comparing different types of datasets (for Check imagery resolutions between the model and observed data marry up. the number of years in which models agreed example in the DRiSL project: agriculture/health/ divided by the total number of years analysed. economic indicators to the biophysical data) it is The results indicated that, especially at high return hard to decide what is “correct” as a point of truth periods, there was22 a high level of disagreement for identifying humanitarian needs. between models. The level of uncertainty when l Check imagery resolutions between the model picking which year is a “drought” year is high. and observed data marry up. This is disappointing but nonetheless vital to know. SCIENTIFIC DUE DILIGENCE FOR HUMANITARIAN DISASTER RISK FINANCING 21
CHECKPOINT 06 COMMUNICATING COMPLEXITY CLEARLY: PRESENTING RESULTS TO SUPPORT DECISION-MAKING BY NON-TECHNICAL USERS13
Once the model’s levels of uncertainty and skill have been identified, these should be communicated in a simple and logical way so that all stakeholders, including at-risk communities, are aware of the findings and able to provide feedback. This should include communicating the model’s strengths and limitations.
EXAMPLES:
l Heatwave analysis: Annual time-series for yearly occurrences of temperature exceeding 35ºC, 40ºC, 45ºC and 50ºC for the Punjab province (left) and city of Karachi (right). View full technical report here