Andrew W. Wood


2023

DOI bib
Hybrid forecasting: blending climate predictions with AI models
Louise Slater, Louise Arnal, Marie‐Amélie Boucher, Annie Y.-Y. Chang, Simon Moulds, Conor Murphy, Grey Nearing, Guy Shalev, Chaopeng Shen, Linda Speight, Gabriele Villarini, Robert L. Wilby, Andrew W. Wood, Massimiliano Zappa, Louise Slater, Louise Arnal, Marie‐Amélie Boucher, Annie Y.-Y. Chang, Simon Moulds, Conor Murphy, Grey Nearing, Guy Shalev, Chaopeng Shen, Linda Speight, Gabriele Villarini, Robert L. Wilby, Andrew W. Wood, Massimiliano Zappa
Hydrology and Earth System Sciences, Volume 27, Issue 9

Abstract. Hybrid hydroclimatic forecasting systems employ data-driven (statistical or machine learning) methods to harness and integrate a broad variety of predictions from dynamical, physics-based models – such as numerical weather prediction, climate, land, hydrology, and Earth system models – into a final prediction product. They are recognized as a promising way of enhancing the prediction skill of meteorological and hydroclimatic variables and events, including rainfall, temperature, streamflow, floods, droughts, tropical cyclones, or atmospheric rivers. Hybrid forecasting methods are now receiving growing attention due to advances in weather and climate prediction systems at subseasonal to decadal scales, a better appreciation of the strengths of AI, and expanding access to computational resources and methods. Such systems are attractive because they may avoid the need to run a computationally expensive offline land model, can minimize the effect of biases that exist within dynamical outputs, benefit from the strengths of machine learning, and can learn from large datasets, while combining different sources of predictability with varying time horizons. Here we review recent developments in hybrid hydroclimatic forecasting and outline key challenges and opportunities for further research. These include obtaining physically explainable results, assimilating human influences from novel data sources, integrating new ensemble techniques to improve predictive skill, creating seamless prediction schemes that merge short to long lead times, incorporating initial land surface and ocean/ice conditions, acknowledging spatial variability in landscape and atmospheric forcing, and increasing the operational uptake of hybrid prediction schemes.

DOI bib
Hybrid forecasting: blending climate predictions with AI models
Louise Slater, Louise Arnal, Marie‐Amélie Boucher, Annie Y.-Y. Chang, Simon Moulds, Conor Murphy, Grey Nearing, Guy Shalev, Chaopeng Shen, Linda Speight, Gabriele Villarini, Robert L. Wilby, Andrew W. Wood, Massimiliano Zappa, Louise Slater, Louise Arnal, Marie‐Amélie Boucher, Annie Y.-Y. Chang, Simon Moulds, Conor Murphy, Grey Nearing, Guy Shalev, Chaopeng Shen, Linda Speight, Gabriele Villarini, Robert L. Wilby, Andrew W. Wood, Massimiliano Zappa
Hydrology and Earth System Sciences, Volume 27, Issue 9

Abstract. Hybrid hydroclimatic forecasting systems employ data-driven (statistical or machine learning) methods to harness and integrate a broad variety of predictions from dynamical, physics-based models – such as numerical weather prediction, climate, land, hydrology, and Earth system models – into a final prediction product. They are recognized as a promising way of enhancing the prediction skill of meteorological and hydroclimatic variables and events, including rainfall, temperature, streamflow, floods, droughts, tropical cyclones, or atmospheric rivers. Hybrid forecasting methods are now receiving growing attention due to advances in weather and climate prediction systems at subseasonal to decadal scales, a better appreciation of the strengths of AI, and expanding access to computational resources and methods. Such systems are attractive because they may avoid the need to run a computationally expensive offline land model, can minimize the effect of biases that exist within dynamical outputs, benefit from the strengths of machine learning, and can learn from large datasets, while combining different sources of predictability with varying time horizons. Here we review recent developments in hybrid hydroclimatic forecasting and outline key challenges and opportunities for further research. These include obtaining physically explainable results, assimilating human influences from novel data sources, integrating new ensemble techniques to improve predictive skill, creating seamless prediction schemes that merge short to long lead times, incorporating initial land surface and ocean/ice conditions, acknowledging spatial variability in landscape and atmospheric forcing, and increasing the operational uptake of hybrid prediction schemes.

2022

DOI bib
Hydrologic Model Sensitivity to Temporal Aggregation of Meteorological Forcing Data: A Case Study for the Contiguous United States
Ashley E. Van Beusekom, Lauren E. Hay, Andrew Bennett, Young-Don Choi, Martyn Clark, J. L. Goodall, Zhiyu Li, Iman Maghami, Bart Nijssen, Andrew W. Wood
Journal of Hydrometeorology, Volume 23, Issue 2

Abstract Surface meteorological analyses are an essential input (termed “forcing”) for hydrologic modeling. This study investigated the sensitivity of different hydrologic model configurations to temporal variations of seven forcing variables (precipitation rate, air temperature, longwave radiation, specific humidity, shortwave radiation, wind speed, and air pressure). Specifically, the effects of temporally aggregating hourly forcings to hourly daily average forcings were examined. The analysis was based on 14 hydrological outputs from the Structure for Unifying Multiple Modeling Alternatives (SUMMA) model for the 671 Catchment Attributes and Meteorology for Large-Sample Studies (CAMELS) basins across the contiguous United States (CONUS). Results demonstrated that the hydrologic model sensitivity to temporally aggregating the forcing inputs varies across model output variables and model locations. We used Latin hypercube sampling to sample model parameters from eight combinations of three influential model physics choices (three model decisions with two options for each decision, i.e., eight model configurations). Results showed that the choice of model physics can change the relative influence of forcing on model outputs and the forcing importance may not be dependent on the parameter space. This allows for model output sensitivity to forcing aggregation to be tested prior to parameter calibration. More generally, this work provides a comprehensive analysis of the dependence of modeled outcomes on input forcing behavior, providing insight into the regional variability of forcing variable dominance on modeled outputs across CONUS.

DOI bib
New projections of 21st century climate and hydrology for Alaska and Hawaiʻi
Naoki Mizukami, Andrew J. Newman, Jeremy S. Littell, Thomas W. Giambelluca, Andrew W. Wood, E. D. Gutmann, Joseph Hamman, Diana R. Gergel, Bart Nijssen, Martyn Clark, J. R. Arnold
Climate Services, Volume 27

In the United States, high-resolution, century-long, hydroclimate projection datasets have been developed for water resources planning, focusing on the contiguous United States (CONUS) domain. However, there are few statewide hydroclimate projection datasets available for Alaska and Hawaiʻi. The limited information on hydroclimatic change motivates developing hydrologic scenarios from 1950 to 2099 using climate-hydrology impact modeling chains consisting of multiple statistically downscaled climate projections as input to hydrologic model simulations for both states. We adopt an approach similar to the previous CONUS hydrologic assessments where: 1) we select the outputs from ten global climate models (GCM) from the Coupled Model Intercomparison Project Phase 5 with Representative Concentration Pathways 4.5 and 8.5; 2) we perform statistical downscaling to generate climate input data for hydrologic models (12-km grid-spacing for Alaska and 1-km for Hawaiʻi); and 3) we perform process-based hydrologic model simulations. For Alaska, we have advanced the hydrologic model configuration from CONUS by using the full water-energy balance computation, frozen soils and a simple glacier model. The simulations show that robust warming and increases in precipitation produce runoff increases for most of Alaska, with runoff reductions in the currently glacierized areas in Southeast Alaska. For Hawaiʻi, we produce the projections at high resolution (1 km) which highlight high spatial variability of climate variables across the state, and a large spread of runoff across the GCMs is driven by a large precipitation spread across the GCMs. Our new ensemble datasets assist with state-wide climate adaptation and other water planning.

DOI bib
Community Workflows to Advance Reproducibility in Hydrologic Modeling: Separating Model‐Agnostic and Model‐Specific Configuration Steps in Applications of Large‐Domain Hydrologic Models
Wouter Knoben, Martyn Clark, Jerad Bales, Andrew Bennett, Shervan Gharari, Christopher B. Marsh, Bart Nijssen, Alain Pietroniro, Raymond J. Spiteri, Guoqiang Tang, David G. Tarboton, Andrew W. Wood
Water Resources Research, Volume 58, Issue 11

Despite the proliferation of computer-based research on hydrology and water resources, such research is typically poorly reproducible. Published studies have low reproducibility due to incomplete availability of data and computer code, and a lack of documentation of workflow processes. This leads to a lack of transparency and efficiency because existing code can neither be quality controlled nor reused. Given the commonalities between existing process-based hydrologic models in terms of their required input data and preprocessing steps, open sharing of code can lead to large efficiency gains for the modeling community. Here, we present a model configuration workflow that provides full reproducibility of the resulting model instantiations in a way that separates the model-agnostic preprocessing of specific data sets from the model-specific requirements that models impose on their input files. We use this workflow to create large-domain (global and continental) and local configurations of the Structure for Unifying Multiple Modeling Alternatives (SUMMA) hydrologic model connected to the mizuRoute routing model. These examples show how a relatively complex model setup over a large domain can be organized in a reproducible and structured way that has the potential to accelerate advances in hydrologic modeling for the community as a whole. We provide a tentative blueprint of how community modeling initiatives can be built on top of workflows such as this. We term our workflow the “Community Workflows to Advance Reproducibility in Hydrologic Modeling” (CWARHM; pronounced “swarm”).

DOI bib
En-GARD: A Statistical Downscaling Framework to Produce and Test Large Ensembles of Climate Projections
E. D. Gutmann, Joseph Hamman, Martyn Clark, Trude Eidhammer, Andrew W. Wood, J. R. Arnold
Journal of Hydrometeorology, Volume 23, Issue 10

Abstract Statistical processing of numerical model output has been a part of both weather forecasting and climate applications for decades. Statistical techniques are used to correct systematic biases in atmospheric model outputs and to represent local effects that are unresolved by the model, referred to as downscaling. Many downscaling techniques have been developed, and it has been difficult to systematically explore the implications of the individual decisions made in the development of downscaling methods. Here we describe a unified framework that enables the user to evaluate multiple decisions made in the methods used to statistically postprocess output from weather and climate models. The Ensemble Generalized Analog Regression Downscaling (En-GARD) method enables the user to select any number of input variables, predictors, mathematical transformations, and combinations for use in parametric or nonparametric downscaling approaches. En-GARD enables explicitly predicting both the probability of event occurrence and the event magnitude. Outputs from En-GARD include errors in model fit, enabling the production of an ensemble of projections through sampling of the probability distributions of each climate variable. We apply En-GARD to regional climate model simulations to evaluate the relative importance of different downscaling method choices on simulations of the current and future climate. We show that choice of predictor variables is the most important decision affecting downscaled future climate outputs, while having little impact on the fidelity of downscaled outcomes for current climate. We also show that weak statistical relationships prevent such approaches from predicting large changes in extreme events on a daily time scale.

DOI bib
Improving station-based ensemble surface meteorological analyses using numerical weather prediction: A case study of the Oroville Dam crisis precipitation event
Patrick Bunn, Andrew W. Wood, Andrew J. Newman, Hsin-I Chang, Christopher L. Castro, Martyn Clark, J. R. Arnold
Journal of Hydrometeorology

Abstract Surface meteorological analyses serve a wide range of research and applications, including forcing inputs for hydrological and ecological models, climate analysis, and resource and emergency management. Quantifying uncertainty in such analyses would extend their utility for probabilistic hydrologic prediction and climate risk applications. With this motivation, we enhance and evaluate an approach for generating ensemble analyses of precipitation and temperature through the fusion of station observations, terrain information, and numerical weather prediction simulations of surface climate fields. In particular, we expand a spatial regression in which static terrain attributes serve as predictors for spatially distributed 1/16th degree daily surface precipitation and temperature by including forecast outputs from the High-Resolution Rapid Refresh (HRRR) numerical weather prediction model as additional predictors. We demonstrate the approach for a case study domain of California, focusing on the meteorological conditions leading to the 2017 flood and spillway failure event at Lake Oroville. The approach extends the spatial regression capability of the Gridded Meteorological Ensemble Tool (GMET) and also adds cross-validation to the uncertainty estimation component, enabling the use of predictive rather than calibration uncertainty. In evaluation against out-of-sample station observations, the HRRR-based predictors alone are found to be skillful for the study setting, leading to overall improvements in the enhanced GMET meteorological analyses. The methodology and associated tool represent a promising method for generating meteorological surface analyses for both research-oriented and operational applications, as well as a general strategy for merging in situ and gridded observations.

2021

DOI bib
Leveraging ensemble meteorological forcing data to improve parameter estimation of hydrologic models
Hongli Liu, Bryan A. Tolson, Andrew J. Newman, Andrew W. Wood, Hongli Liu, Bryan A. Tolson, Andrew J. Newman, Andrew W. Wood
Hydrological Processes, Volume 35, Issue 11

As continental to global scale high-resolution meteorological datasets continue to be developed, there are sufficient meteorological datasets available now for modellers to construct a historical forcing ensemble. The forcing ensemble can be a collection of multiple deterministic meteorological datasets or come from an ensemble meteorological dataset. In hydrological model calibration, the forcing ensemble can be used to represent forcing data uncertainty. This study examines the potential of using the forcing ensemble to identify more robust parameters through model calibration. Specifically, we compare an ensemble forcing-based calibration with two deterministic forcing-based calibrations and investigate their flow simulation and parameter estimation properties and the ability to resist poor-quality forcings. The comparison experiment is conducted with a six-parameter hydrological model for 30 synthetic studies and 20 real data studies to provide a better assessment of the average performance of the deterministic and ensemble forcing-based calibrations. Results show that the ensemble forcing-based calibration generates parameter estimates that are less biased and have higher frequency of covering the true parameter values than the deterministic forcing-based calibration does. Using a forcing ensemble in model calibration reduces the risk of inaccurate flow simulation caused by poor-quality meteorological inputs, and improves the reliability and overall simulation skill of ensemble simulation results. The poor-quality meteorological inputs can be effectively filtered out via our ensemble forcing-based calibration methodology and thus discarded in any post-calibration model applications. The proposed ensemble forcing-based calibration method can be considered as a more generalized framework to include parameter and forcing uncertainties in model calibration.

DOI bib
Leveraging ensemble meteorological forcing data to improve parameter estimation of hydrologic models
Hongli Liu, Bryan A. Tolson, Andrew J. Newman, Andrew W. Wood, Hongli Liu, Bryan A. Tolson, Andrew J. Newman, Andrew W. Wood
Hydrological Processes, Volume 35, Issue 11

As continental to global scale high-resolution meteorological datasets continue to be developed, there are sufficient meteorological datasets available now for modellers to construct a historical forcing ensemble. The forcing ensemble can be a collection of multiple deterministic meteorological datasets or come from an ensemble meteorological dataset. In hydrological model calibration, the forcing ensemble can be used to represent forcing data uncertainty. This study examines the potential of using the forcing ensemble to identify more robust parameters through model calibration. Specifically, we compare an ensemble forcing-based calibration with two deterministic forcing-based calibrations and investigate their flow simulation and parameter estimation properties and the ability to resist poor-quality forcings. The comparison experiment is conducted with a six-parameter hydrological model for 30 synthetic studies and 20 real data studies to provide a better assessment of the average performance of the deterministic and ensemble forcing-based calibrations. Results show that the ensemble forcing-based calibration generates parameter estimates that are less biased and have higher frequency of covering the true parameter values than the deterministic forcing-based calibration does. Using a forcing ensemble in model calibration reduces the risk of inaccurate flow simulation caused by poor-quality meteorological inputs, and improves the reliability and overall simulation skill of ensemble simulation results. The poor-quality meteorological inputs can be effectively filtered out via our ensemble forcing-based calibration methodology and thus discarded in any post-calibration model applications. The proposed ensemble forcing-based calibration method can be considered as a more generalized framework to include parameter and forcing uncertainties in model calibration.

DOI bib
Flood spatial coherence, triggers, and performance in hydrological simulations: large-sample evaluation of four streamflow-calibrated models
Manuela I. Brunner, Lieke Melsen, Andrew W. Wood, Oldřich Rakovec, Naoki Mizukami, Wouter Knoben, Martyn Clark, Manuela I. Brunner, Lieke Melsen, Andrew W. Wood, Oldřich Rakovec, Naoki Mizukami, Wouter Knoben, Martyn Clark
Hydrology and Earth System Sciences, Volume 25, Issue 1

Abstract. Floods cause extensive damage, especially if they affect large regions. Assessments of current, local, and regional flood hazards and their future changes often involve the use of hydrologic models. A reliable hydrologic model ideally reproduces both local flood characteristics and spatial aspects of flooding under current and future climate conditions. However, uncertainties in simulated floods can be considerable and yield unreliable hazard and climate change impact assessments. This study evaluates the extent to which models calibrated according to standard model calibration metrics such as the widely used Kling–Gupta efficiency are able to capture flood spatial coherence and triggering mechanisms. To highlight challenges related to flood simulations, we investigate how flood timing, magnitude, and spatial variability are represented by an ensemble of hydrological models when calibrated on streamflow using the Kling–Gupta efficiency metric, an increasingly common metric of hydrologic model performance also in flood-related studies. Specifically, we compare how four well-known models (the Sacramento Soil Moisture Accounting model, SAC; the Hydrologiska Byråns Vattenbalansavdelning model, HBV; the variable infiltration capacity model, VIC; and the mesoscale hydrologic model, mHM) represent (1) flood characteristics and their spatial patterns and (2) how they translate changes in meteorologic variables that trigger floods into changes in flood magnitudes. Our results show that both the modeling of local and spatial flood characteristics are challenging as models underestimate flood magnitude, and flood timing is not necessarily well captured. They further show that changes in precipitation and temperature are not always well translated to changes in flood flow, which makes local and regional flood hazard assessments even more difficult for future conditions. From a large sample of catchments and with multiple models, we conclude that calibration on the integrated Kling–Gupta metric alone is likely to yield models that have limited reliability in flood hazard assessments, undermining their utility for regional and future change assessments. We underscore that such assessments can be improved by developing flood-focused, multi-objective, and spatial calibration metrics, by improving flood generating process representation through model structure comparisons and by considering uncertainty in precipitation input.

DOI bib
Flood spatial coherence, triggers, and performance in hydrological simulations: large-sample evaluation of four streamflow-calibrated models
Manuela I. Brunner, Lieke Melsen, Andrew W. Wood, Oldřich Rakovec, Naoki Mizukami, Wouter Knoben, Martyn Clark, Manuela I. Brunner, Lieke Melsen, Andrew W. Wood, Oldřich Rakovec, Naoki Mizukami, Wouter Knoben, Martyn Clark
Hydrology and Earth System Sciences, Volume 25, Issue 1

Abstract. Floods cause extensive damage, especially if they affect large regions. Assessments of current, local, and regional flood hazards and their future changes often involve the use of hydrologic models. A reliable hydrologic model ideally reproduces both local flood characteristics and spatial aspects of flooding under current and future climate conditions. However, uncertainties in simulated floods can be considerable and yield unreliable hazard and climate change impact assessments. This study evaluates the extent to which models calibrated according to standard model calibration metrics such as the widely used Kling–Gupta efficiency are able to capture flood spatial coherence and triggering mechanisms. To highlight challenges related to flood simulations, we investigate how flood timing, magnitude, and spatial variability are represented by an ensemble of hydrological models when calibrated on streamflow using the Kling–Gupta efficiency metric, an increasingly common metric of hydrologic model performance also in flood-related studies. Specifically, we compare how four well-known models (the Sacramento Soil Moisture Accounting model, SAC; the Hydrologiska Byråns Vattenbalansavdelning model, HBV; the variable infiltration capacity model, VIC; and the mesoscale hydrologic model, mHM) represent (1) flood characteristics and their spatial patterns and (2) how they translate changes in meteorologic variables that trigger floods into changes in flood magnitudes. Our results show that both the modeling of local and spatial flood characteristics are challenging as models underestimate flood magnitude, and flood timing is not necessarily well captured. They further show that changes in precipitation and temperature are not always well translated to changes in flood flow, which makes local and regional flood hazard assessments even more difficult for future conditions. From a large sample of catchments and with multiple models, we conclude that calibration on the integrated Kling–Gupta metric alone is likely to yield models that have limited reliability in flood hazard assessments, undermining their utility for regional and future change assessments. We underscore that such assessments can be improved by developing flood-focused, multi-objective, and spatial calibration metrics, by improving flood generating process representation through model structure comparisons and by considering uncertainty in precipitation input.

DOI bib
EMDNA: an Ensemble Meteorological Dataset for North America
Guoqiang Tang, Martyn Clark, Simon Michael Papalexiou, Andrew J. Newman, Andrew W. Wood, Dominique Brunet, Paul H. Whitfield, Guoqiang Tang, Martyn Clark, Simon Michael Papalexiou, Andrew J. Newman, Andrew W. Wood, Dominique Brunet, Paul H. Whitfield
Earth System Science Data, Volume 13, Issue 7

Abstract. Probabilistic methods are useful to estimate the uncertainty in spatial meteorological fields (e.g., the uncertainty in spatial patterns of precipitation and temperature across large domains). In ensemble probabilistic methods, “equally plausible” ensemble members are used to approximate the probability distribution, hence the uncertainty, of a spatially distributed meteorological variable conditioned to the available information. The ensemble members can be used to evaluate the impact of uncertainties in spatial meteorological fields for a myriad of applications. This study develops the Ensemble Meteorological Dataset for North America (EMDNA). EMDNA has 100 ensemble members with daily precipitation amount, mean daily temperature, and daily temperature range at 0.1∘ spatial resolution (approx. 10 km grids) from 1979 to 2018, derived from a fusion of station observations and reanalysis model outputs. The station data used in EMDNA are from a serially complete dataset for North America (SCDNA) that fills gaps in precipitation and temperature measurements using multiple strategies. Outputs from three reanalysis products are regridded, corrected, and merged using Bayesian model averaging. Optimal interpolation (OI) is used to merge station- and reanalysis-based estimates. EMDNA estimates are generated using spatiotemporally correlated random fields to sample from the OI estimates. Evaluation results show that (1) the merged reanalysis estimates outperform raw reanalysis estimates, particularly in high latitudes and mountainous regions; (2) the OI estimates are more accurate than the reanalysis and station-based regression estimates, with the most notable improvements for precipitation evident in sparsely gauged regions; and (3) EMDNA estimates exhibit good performance according to the diagrams and metrics used for probabilistic evaluation. We discuss the limitations of the current framework and highlight that further research is needed to improve ensemble meteorological datasets. Overall, EMDNA is expected to be useful for hydrological and meteorological applications in North America. The entire dataset and a teaser dataset (a small subset of EMDNA for easy download and preview) are available at https://doi.org/10.20383/101.0275 (Tang et al., 2020a).

DOI bib
EMDNA: an Ensemble Meteorological Dataset for North America
Guoqiang Tang, Martyn Clark, Simon Michael Papalexiou, Andrew J. Newman, Andrew W. Wood, Dominique Brunet, Paul H. Whitfield, Guoqiang Tang, Martyn Clark, Simon Michael Papalexiou, Andrew J. Newman, Andrew W. Wood, Dominique Brunet, Paul H. Whitfield
Earth System Science Data, Volume 13, Issue 7

Abstract. Probabilistic methods are useful to estimate the uncertainty in spatial meteorological fields (e.g., the uncertainty in spatial patterns of precipitation and temperature across large domains). In ensemble probabilistic methods, “equally plausible” ensemble members are used to approximate the probability distribution, hence the uncertainty, of a spatially distributed meteorological variable conditioned to the available information. The ensemble members can be used to evaluate the impact of uncertainties in spatial meteorological fields for a myriad of applications. This study develops the Ensemble Meteorological Dataset for North America (EMDNA). EMDNA has 100 ensemble members with daily precipitation amount, mean daily temperature, and daily temperature range at 0.1∘ spatial resolution (approx. 10 km grids) from 1979 to 2018, derived from a fusion of station observations and reanalysis model outputs. The station data used in EMDNA are from a serially complete dataset for North America (SCDNA) that fills gaps in precipitation and temperature measurements using multiple strategies. Outputs from three reanalysis products are regridded, corrected, and merged using Bayesian model averaging. Optimal interpolation (OI) is used to merge station- and reanalysis-based estimates. EMDNA estimates are generated using spatiotemporally correlated random fields to sample from the OI estimates. Evaluation results show that (1) the merged reanalysis estimates outperform raw reanalysis estimates, particularly in high latitudes and mountainous regions; (2) the OI estimates are more accurate than the reanalysis and station-based regression estimates, with the most notable improvements for precipitation evident in sparsely gauged regions; and (3) EMDNA estimates exhibit good performance according to the diagrams and metrics used for probabilistic evaluation. We discuss the limitations of the current framework and highlight that further research is needed to improve ensemble meteorological datasets. Overall, EMDNA is expected to be useful for hydrological and meteorological applications in North America. The entire dataset and a teaser dataset (a small subset of EMDNA for easy download and preview) are available at https://doi.org/10.20383/101.0275 (Tang et al., 2020a).

2020

DOI bib
Flood hazard and change impact assessments may profit from rethinking model calibration strategies
Manuela I. Brunner, Lieke Melsen, Andrew W. Wood, Oldřich Rakovec, Naoki Mizukami, Wouter Knoben, Martyn Clark

Abstract. Floods cause large damages, especially if they affect large regions. Assessments of current, local and regional flood hazards and their future changes often involve the use of hydrologic models. However, uncertainties in simulated floods can be considerable and yield unreliable hazard and climate change impact assessments. A reliable hydrologic model ideally reproduces both local flood characteristics and spatial aspects of flooding, which is, however, not guaranteed especially when using standard model calibration metrics. In this paper we investigate how flood timing, magnitude and spatial variability are represented by an ensemble of hydrological models when calibrated on streamflow using the Kling–Gupta efficiency metric, an increasingly common metric of hydrologic model performance. We compare how four well-known models (SAC, HBV, VIC, and mHM) represent (1) flood characteristics and their spatial patterns; and (2) how they translate changes in meteorologic variables that trigger floods into changes in flood magnitudes. Our results show that both the modeling of local and spatial flood characteristics is challenging. They further show that changes in precipitation and temperature are not necessarily well translated to changes in flood flow, which makes local and regional flood hazard assessments even more difficult for future conditions. We conclude that models calibrated on integrated metrics such as the Kling–Gupta efficiency alone have limited reliability in flood hazard assessments, in particular in regional and future assessments, and suggest the development of alternative process-based and spatial evaluation metrics.

DOI bib
EMDNA: Ensemble Meteorological Dataset for North America
Guoqiang Tang, Martyn Clark, Simon Michael Papalexiou, Andrew J. Newman, Andrew W. Wood, Dominique Brunet, Paul H. Whitfield

Abstract. Probabilistic methods are very useful to estimate the spatial variability in meteorological conditions (e.g., spatial patterns of precipitation and temperature across large domains). In ensemble probabilistic methods, equally plausible ensemble members are used to approximate the probability distribution, hence uncertainty, of a spatially distributed meteorological variable conditioned on the available information. The ensemble can be used to evaluate the impact of the uncertainties in a myriad of applications. This study develops the Ensemble Meteorological Dataset for North America (EMDNA). EMDNA has 100 members with daily precipitation amount, mean daily temperature, and daily temperature range at 0.1° spatial resolution from 1979 to 2018, derived from a fusion of station observations and reanalysis model outputs. The station data used in EMDNA are from a serially complete dataset for North America (SCDNA) that fills gaps in precipitation and temperature measurements using multiple strategies. Outputs from three reanalysis products are regridded, corrected, and merged using the Bayesian Model Averaging. Optimal Interpolation (OI) is used to merge station- and reanalysis-based estimates. EMDNA estimates are generated based on OI estimates and spatiotemporally correlated random fields. Evaluation results show that (1) the merged reanalysis estimates outperform raw reanalysis estimates, particularly in high latitudes and mountainous regions; (2) the OI estimates are more accurate than the reanalysis and station-based regression estimates, with the most notable improvement for precipitation occurring in sparsely gauged regions; and (3) EMDNA estimates exhibit good performance according to the diagrams and metrics used for probabilistic evaluation. We also discuss the limitations of the current framework and highlight that persistent efforts are needed to further develop probabilistic methods and ensemble datasets. Overall, EMDNA is expected to be useful for hydrological and meteorological applications in North America. The whole dataset and a teaser dataset (a small subset of EMDNA for easy download and preview) are available at https://doi.org/10.20383/101.0275 (Tang et al., 2020a).

DOI bib
SCDNA: a serially complete precipitation and temperature dataset for North America from 1979 to 2018
Guoqiang Tang, Martyn Clark, Andrew J. Newman, Andrew W. Wood, Simon Michael Papalexiou, Vincent Vionnet, Paul H. Whitfield
Earth System Science Data, Volume 12, Issue 4

Abstract. Station-based serially complete datasets (SCDs) of precipitation and temperature observations are important for hydrometeorological studies. Motivated by the lack of serially complete station observations for North America, this study seeks to develop an SCD from 1979 to 2018 from station data. The new SCD for North America (SCDNA) includes daily precipitation, minimum temperature (Tmin⁡), and maximum temperature (Tmax⁡) data for 27 276 stations. Raw meteorological station data were obtained from the Global Historical Climate Network Daily (GHCN-D), the Global Surface Summary of the Day (GSOD), Environment and Climate Change Canada (ECCC), and a compiled station database in Mexico. Stations with at least 8-year-long records were selected, which underwent location correction and were subjected to strict quality control. Outputs from three reanalysis products (ERA5, JRA-55, and MERRA-2) provided auxiliary information to estimate station records. Infilling during the observation period and reconstruction beyond the observation period were accomplished by combining estimates from 16 strategies (variants of quantile mapping, spatial interpolation, and machine learning). A sensitivity experiment was conducted by assuming that 30 % of observations from stations were missing – this enabled independent validation and provided a reference for reconstruction. Quantile mapping and mean value corrections were applied to the final estimates. The median Kling–Gupta efficiency (KGE′) values of the final SCDNA for all stations are 0.90, 0.98, and 0.99 for precipitation, Tmin⁡, and Tmax⁡, respectively. The SCDNA is closer to station observations than the four benchmark gridded products and can be used in applications that require either quality-controlled meteorological station observations or reconstructed long-term estimates for analysis and modeling. The dataset is available at https://doi.org/10.5281/zenodo.3735533 (Tang et al., 2020).

DOI bib
Spatial Dependence of Floods Shaped by Spatiotemporal Variations in Meteorological and Land‐Surface Processes
Manuela I. Brunner, Eric Gilleland, Andrew W. Wood, Daniel L. Swain, Martyn Clark
Geophysical Research Letters, Volume 47, Issue 13

Floods often affect large regions and cause adverse societal impacts. Regional flood hazard and risk assessments therefore require a realistic representation of spatial flood dependencies to avoid the overestimation or underestimation of risk. However, it is not yet well understood how spatial flood dependence, that is, the degree of co-occurrence of floods at different locations, varies in space and time and which processes influence the strength of this dependence. We identify regions in the United States with seasonally similar flood behavior and analyze processes governing spatial dependence. We find that spatial flood dependence varies regionally and seasonally and is generally strongest in winter and spring and weakest in summer and fall. Moreover, we find that land-surface processes are crucial in shaping the spatiotemporal characteristics of flood events. We conclude that the regional and seasonal variations in spatial flood dependencies must be considered when conducting current and future flood risk assessments.

DOI bib
Probabilistic Spatial Meteorological Estimates for Alaska and the Yukon
Andrew J. Newman, Martyn Clark, Andrew W. Wood, J. R. Arnold
Journal of Geophysical Research: Atmospheres, Volume 125, Issue 22

It is challenging to develop observationally based spatial estimates of meteorology in Alaska and the Yukon. Complex topography, frozen precipitation undercatch, and extremely sparse in situ observations all limit our capability to produce accurate spatial estimates of meteorological conditions. In this Arctic environment, it is necessary to develop probabilistic estimates of precipitation and temperature that explicitly incorporate spatiotemporally varying uncertainty and bias corrections. In this paper we exploit the recently developed ensemble Climatologically Aided Interpolation (eCAI) system to produce daily historical estimates of precipitation and temperature across Alaska and the Yukon Territory at a 2 km grid spacing for the time period 1980–2013. We extend the previous eCAI method to address precipitation gauge undercatch and wetting loss, which is of high importance for this high-latitude region where much of the precipitation falls as snow. Leave-one-out cross-validation shows our ensemble has little bias in daily precipitation and mean temperature at the station locations, with an overestimate in the daily standard deviation of precipitation. The ensemble is statistically reliable compared to climatology and can discriminate precipitation events across different precipitation thresholds. Long-term mean loss adjusted precipitation is up to 36% greater than the unadjusted estimate in windy areas that receive a large fraction of frozen precipitation, primarily due to wind induced undercatch. Comparing the ensemble mean climatology of precipitation and temperature to PRISM and Daymet v3 shows large interproduct differences, particularly in precipitation across the complex terrain of southeast and northern Alaska.

2019

DOI bib
The potential to reduce uncertainty in regional runoff projections from climate models
Flavio Lehner, Andrew W. Wood, J. A. Vano, David M. Lawrence, Martyn Clark, Justin Mankin
Nature Climate Change, Volume 9, Issue 12

Increasingly, climate change impact assessments rely directly on climate models. Assessments of future water security depend in part on how the land model components in climate models partition precipitation into evapotranspiration and runoff, and on the sensitivity of this partitioning to climate. Runoff sensitivities are not well constrained, with CMIP5 models displaying a large spread for the present day, which projects onto change under warming, creating uncertainty. Here we show that constraining CMIP5 model runoff sensitivities with observed estimates could reduce uncertainty in runoff projection over the western United States by up to 50%. We urge caution in the direct use of climate model runoff for applications and encourage model development to use regional-scale hydrological sensitivity metrics to improve projections for water security assessments.

DOI bib
On the choice of calibration metrics for “high-flow” estimation using hydrologic models
Naoki Mizukami, Oldřich Rakovec, Andrew J. Newman, Martyn Clark, Andrew W. Wood, Hoshin V. Gupta, Rohini Kumar
Hydrology and Earth System Sciences, Volume 23, Issue 6

Abstract. Calibration is an essential step for improving the accuracy of simulations generated using hydrologic models. A key modeling decision is selecting the performance metric to be optimized. It has been common to use squared error performance metrics, or normalized variants such as Nash–Sutcliffe efficiency (NSE), based on the idea that their squared-error nature will emphasize the estimates of high flows. However, we conclude that NSE-based model calibrations actually result in poor reproduction of high-flow events, such as the annual peak flows that are used for flood frequency estimation. Using three different types of performance metrics, we calibrate two hydrological models at a daily step, the Variable Infiltration Capacity (VIC) model and the mesoscale Hydrologic Model (mHM), and evaluate their ability to simulate high-flow events for 492 basins throughout the contiguous United States. The metrics investigated are (1) NSE, (2) Kling–Gupta efficiency (KGE) and its variants, and (3) annual peak flow bias (APFB), where the latter is an application-specific metric that focuses on annual peak flows. As expected, the APFB metric produces the best annual peak flow estimates; however, performance on other high-flow-related metrics is poor. In contrast, the use of NSE results in annual peak flow estimates that are more than 20 % worse, primarily due to the tendency of NSE to underestimate observed flow variability. On the other hand, the use of KGE results in annual peak flow estimates that are better than from NSE, owing to improved flow time series metrics (mean and variance), with only a slight degradation in performance with respect to other related metrics, particularly when a non-standard weighting of the components of KGE is used. Stochastically generated ensemble simulations based on model residuals show the ability to improve the high-flow metrics, regardless of the deterministic performances. However, we emphasize that improving the fidelity of streamflow dynamics from deterministically calibrated models is still important, as it may improve high-flow metrics (for the right reasons). Overall, this work highlights the need for a deeper understanding of performance metric behavior and design in relation to the desired goals of model calibration.

DOI bib
Diagnostic Evaluation of Large‐Domain Hydrologic Models Calibrated Across the Contiguous United States
Oldřich Rakovec, Naoki Mizukami, Rohini Kumar, Andrew J. Newman, Stephan Thober, Andrew W. Wood, Martyn Clark, Luis Samaniego
Journal of Geophysical Research: Atmospheres, Volume 124, Issue 24

This study presents diagnostic evaluation of two large‐domain hydrologic models: the mesoscale Hydrologic Model (mHM) and the Variable Infiltration Capacity (VIC) over the contiguous United States (CONUS). These models have been calibrated using the Multiscale Parameter Regionalization scheme in a joint, multibasin approach using 492 medium‐sized basins across the CONUS yielding spatially distributed model parameter sets. The mHM simulations are used as a performance benchmark to examine performance deficiencies in the VIC model. We find that after calibration to streamflow, VIC generally overestimates the magnitude and temporal variability of evapotranspiration (ET) as compared to mHM as well as the FLUXNET observation‐based ET product, resulting in underestimation of the mean and variability of runoff. We perform a controlled calibration experiment to investigate the effect of varying number of transfer function parameters in mHM and to enable a fair comparison between both models (14 and 48 for mHM vs. 14 for VIC). Results of this experiment show similar behavior of mHM with 14 and 48 parameters. Furthermore, we diagnose the internal functioning of the VIC model by looking at the relationship of the evaporative fraction versus the degree of soil saturation and compare it with that of the mHM model, which has a different model structure, a prescribed nonlinear relationship between these variables and exhibits better model skill than VIC. Despite these limitations, the VIC‐based CONUS‐wide calibration constrained against streamflow exhibits better ET skill as compared to two preexisting independent VIC studies.

2018

DOI bib
DOs and DON'Ts for using climate change information for water resource planning and management: guidelines for study design
J. A. Vano, J. R. Arnold, Bart Nijssen, Martyn Clark, Andrew W. Wood, E. D. Gutmann, Nans Addor, Joseph Hamman, Flavio Lehner
Climate Services, Volume 12

Abstract Water managers are actively incorporating climate change information into their long- and short-term planning processes. This is generally seen as a step in the right direction because it supplements traditional methods, providing new insights that can help in planning for a non-stationary climate. However, the continuous evolution of climate change information can make it challenging to use available information appropriately. Advice on how to use the information is not always straightforward and typically requires extended dialogue between information producers and users, which is not always feasible. To help navigate better the ever-changing climate science landscape, this review is organized as a set of nine guidelines for water managers and planners that highlight better practices for incorporating climate change information into water resource planning and management. Each DOs and DON'Ts recommendation is given with context on why certain strategies are preferable and addresses frequently asked questions by exploring past studies and documents that provide guidance, including real-world examples mainly, though not exclusively, from the United States. This paper is intended to provide a foundation that can expand through continued dialogue within and between the climate science and application communities worldwide, a two-way information sharing that can increase the actionable nature of the information produced and promote greater utility and appropriate use.