2023
Abstract Operational flood forecasting in Canada is a provincial responsibility that is carried out by several entities across the country. However, the increasing costs and impacts of floods require better and nationally coordinated flood prediction systems. A more coherent flood forecasting framework for Canada can enable implementing advanced prediction capabilities across the different entities with responsibility for flood forecasting. Recently, the Canadian meteorological and hydrological services were tasked to develop a national flow guidance system. Alongside this initiative, the Global Water Futures program has been advancing cold regions process understanding, hydrological modeling, and forecasting. A community of practice was established for industry, academia, and decision‐makers to share viewpoints on hydrological challenges. Taken together, these initiatives are paving the way towards a national flood forecasting framework. In this article, forecasting challenges are identified (with a focus on cold regions), and recommendations are made to promote the creation of this framework. These include the need for cooperation, well‐defined governance, and better knowledge mobilization. Opportunities and challenges posed by the increasing data availability globally are also highlighted. Advances in each of these areas are positioning Canada as a major contributor to the international operational flood forecasting landscape. This article highlights a route towards the deployment of capacities across large geographical domains.
Abstract. A simple numerical solution procedure – namely the method of lines combined with an off-the-shelf ordinary differential equation (ODE) solver – was shown in previous work to provide efficient, mass-conservative solutions to the pressure-head form of Richards' equation. We implement such a solution in our model openRE. We developed a novel method to quantify the boundary fluxes that reduce water balance errors without negative impacts on model runtimes – the solver flux output method (SFOM). We compare this solution with alternatives, including the classic modified Picard iteration method and the Hydrus 1D model. We reproduce a set of benchmark solutions with all models. We find that Celia's solution has the best water balance, but it can incur significant truncation errors in the simulated boundary fluxes, depending on the time steps used. Our solution has comparable runtimes to Hydrus and better water balance performance (though both models have excellent water balance closure for all the problems we considered). Our solution can be implemented in an interpreted language, such as MATLAB or Python, making use of off-the-shelf ODE solvers. We evaluated alternative SciPy ODE solvers that are available in Python and make practical recommendations about the best way to implement them for Richards' equation. There are two advantages of our approach: (i) the code is concise, making it ideal for teaching purposes; and (ii) the method can be easily extended to represent alternative properties (e.g., novel ways to parameterize the K(ψ) relationship) and processes (e.g., it is straightforward to couple heat or solute transport), making it ideal for testing alternative hypotheses.
Abstract Stochastic simulations of spatiotemporal patterns of hydroclimatic processes, such as precipitation, are needed to build alternative but equally plausible inputs for water‐related design and management, and to estimate uncertainty and assess risks. However, while existing stochastic simulation methods are mature enough to deal with relatively small domains and coarse spatiotemporal scales, additional work is required to develop simulation tools for large‐domain analyses, which are more and more common in an increasingly interconnected world. This study proposes a methodological advancement in the CoSMoS framework, which is a flexible simulation framework preserving arbitrary marginal distributions and correlations, to dramatically decrease the computational burden and make the algorithm fast enough to perform large‐domain simulations in short time. The proposed approach focuses on correlated processes with mixed (zero‐inflated) Uniform marginal distributions. These correlated processes act as intermediates between the target process to simulate (precipitation) and parent Gaussian processes that are the core of the simulation algorithm. Working in the mixed‐Uniform space enables a substantial simplification of the so‐called correlation transformation functions, which represent a computational bottle neck in the original CoSMoS formulation. As a proof of concept, we simulate 40 years of daily precipitation records from 1,000 gauging stations in the Mississippi River basin. Moreover, we extend CoSMoS incorporating parent non‐Gaussian processes with different degrees of tail dependence and suggest potential improvements including the separate simulation of occurrence and intensity processes, and the use of advection, anisotropy, and nonstationary spatiotemporal correlation functions.
Abstract Stochastic simulations of spatiotemporal patterns of hydroclimatic processes, such as precipitation, are needed to build alternative but equally plausible inputs for water‐related design and management, and to estimate uncertainty and assess risks. However, while existing stochastic simulation methods are mature enough to deal with relatively small domains and coarse spatiotemporal scales, additional work is required to develop simulation tools for large‐domain analyses, which are more and more common in an increasingly interconnected world. This study proposes a methodological advancement in the CoSMoS framework, which is a flexible simulation framework preserving arbitrary marginal distributions and correlations, to dramatically decrease the computational burden and make the algorithm fast enough to perform large‐domain simulations in short time. The proposed approach focuses on correlated processes with mixed (zero‐inflated) Uniform marginal distributions. These correlated processes act as intermediates between the target process to simulate (precipitation) and parent Gaussian processes that are the core of the simulation algorithm. Working in the mixed‐Uniform space enables a substantial simplification of the so‐called correlation transformation functions, which represent a computational bottle neck in the original CoSMoS formulation. As a proof of concept, we simulate 40 years of daily precipitation records from 1,000 gauging stations in the Mississippi River basin. Moreover, we extend CoSMoS incorporating parent non‐Gaussian processes with different degrees of tail dependence and suggest potential improvements including the separate simulation of occurrence and intensity processes, and the use of advection, anisotropy, and nonstationary spatiotemporal correlation functions.
DOI
bib
abs
Differentiable modelling to unify machine learning and physical models for geosciences
Chaopeng Shen,
Alison P. Appling,
Pierre Gentine,
Toshiyuki Bandai,
Hoshin Gupta,
Alexandre M. Tartakovsky,
Marco Baity‐Jesi,
Fabrizio Fenicia,
Daniel Kifer,
Li Li,
Xiaofeng Liu,
Wei Ren,
Yi Zheng,
C. J. Harman,
Martyn P. Clark,
Matthew W. Farthing,
Dapeng Feng,
Kumar Prabhash,
Doaa Aboelyazeed,
Farshid Rahmani,
Yalan Song,
Hylke E. Beck,
Tadd Bindas,
Dipankar Dwivedi,
Kuai Fang,
Marvin Höge,
Chris Rackauckas,
Binayak P. Mohanty,
Tirthankar Roy,
Chonggang Xu,
Kathryn Lawson
Nature Reviews Earth & Environment, Volume 4, Issue 8
Process-based modelling offers interpretability and physical consistency in many domains of geosciences but struggles to leverage large datasets efficiently. Machine-learning methods, especially deep networks, have strong predictive skills yet are unable to answer specific scientific questions. In this Perspective, we explore differentiable modelling as a pathway to dissolve the perceived barrier between process-based modelling and machine learning in the geosciences and demonstrate its potential with examples from hydrological modelling. ‘Differentiable’ refers to accurately and efficiently calculating gradients with respect to model variables or parameters, enabling the discovery of high-dimensional unknown relationships. Differentiable modelling involves connecting (flexible amounts of) prior physical knowledge to neural networks, pushing the boundary of physics-informed machine learning. It offers better interpretability, generalizability, and extrapolation capabilities than purely data-driven machine learning, achieving a similar level of accuracy while requiring less training data. Additionally, the performance and efficiency of differentiable models scale well with increasing data volumes. Under data-scarce scenarios, differentiable models have outperformed machine-learning models in producing short-term dynamics and decadal-scale trends owing to the imposed physical constraints. Differentiable modelling approaches are primed to enable geoscientists to ask questions, test hypotheses, and discover unrecognized physical relationships. Future work should address computational challenges, reduce uncertainty, and verify the physical significance of outputs. Differentiable modelling is an approach that flexibly integrates the learning capability of machine learning with the interpretability of process-based models. This Perspective highlights the potential of differentiable modelling to improve the representation of processes, parameter estimation, and predictive accuracy in the geosciences.
2022
Abstract. Human-controlled reservoirs have a large influence on the global water cycle. While global hydrological models use generic parametrisations to model human dam operations, the representation of reservoir regulation is often still lacking in Earth System Models. Here we implement and evaluate a widely used reservoir parametrisation in the global river routing model mizuRoute, which operates on a vector-based river network resolving individual lakes and reservoirs, and which is currently being coupled to an Earth System Model. We develop an approach to determine the downstream area over which to aggregate irrigation water demand per reservoir. The implementation of managed reservoirs is evaluated by comparing to simulations ignoring inland waters, and simulations with reservoirs represented as natural lakes, using (i) local simulations for 26 individual reservoirs driven by observed inflows, and (ii) global-scale simulations driven by runoff from the Community Land Model. The local simulations show a clear added value of the reservoir parametrisation, especially for simulating storage for large reservoirs with a multi-year storage capacity. In the global-scale application, the implementation of reservoirs shows an improvement in outflow and storage compared to the no-reservoir simulation, but compared to the natural lake parametrisation, an overall similar performance is found. This lack of impact could be attributed to biases in simulated river discharge, mainly originating from biases in simulated runoff from the Community Land Model. Finally, the comparison of modelled monthly streamflow indices against observations highlights that the inclusion of dam operations improves the streamflow simulation compared to ignoring lakes and reservoirs. This study overall underlines the need to further develop and test water management parametrisations, as well as to improve runoff simulations for advancing the representation of anthropogenic interference with the terrestrial water cycle in Earth System Models.
Extreme temperature is a major threat to urban populations; thus, it is crucial to understand future changes to plan adaptation and mitigation strategies. We assess historical and CMIP6 projected trends of minimum and maximum temperatures for the 18 most populated Canadian cities. Temperatures increase (on average 0.3°C/decade) in all cities during the historical period (1979–2014), with Prairie cities exhibiting lower rates (0.06°C/decade). Toronto (0.5°C/decade) and Montreal (0.7°C/decade) show high increasing trends in the observation period. Higher-elevation cities, among those with the same population, show slower increasing temperature rates compared to the coastal ones. Projections for cities in the Prairies show 12% more summer days compared to the other regions. The number of heat waves (HWs) increases for all cities, in both the historical and future periods; yet alarming increases are projected for Vancouver, Victoria, and Halifax from no HWs in the historical period to approximately 4 HWs/year on average, towards the end of 2100 for the SSP5–8.5. The cold waves reduce considerably for all cities in the historical period at a rate of 2 CWs/decade on average and are projected to further reduce by 50% compared to the observed period. • CMIP6 simulations for extreme temperature estimation of the largest Canadian cities. • Prairies' cities exhibit a lower rate of temperature increase compared to the cities in Great lakes in observation period. • Cities in Prairies are projected to have 12% more summer days than the rest of the cities. • The number of heat waves increases significantly, especially for Vancouver, Victoria, and Halifax. • Cold waves are expected to decrease by 50% in future.
• The probable impacts of future climate on ice-jam floods are discussed. • Practical suggestions for modelling ice-jam floods under both past and future climates are provided. • Research opportunities that could lead to further improvements in ice-jam flood modelling and prediction are presented. Ice-jam floods (IJFs) are a key concern in cold-region environments, where seasonal effects of river ice formation and break-up can have substantial impacts on flooding processes. Different statistical, machine learning, and process-based models have been developed to simulate IJF events in order to improve our understanding of river ice processes, to quantify potential flood magnitudes and backwater levels, and to undertake risk analysis under a changing climate. Assessment of IJF risks under future climate is limited due to constraints related to model input data. However, given the broad economic and environmental significance of IJFs and their sensitivity to a changing climate, robust modelling frameworks that can incorporate future climatic changes, and produce reliable scenarios of future IJF risks are needed. In this review paper, we discuss the probable impacts of future climate on IJFs and provide suggestions on modelling IJFs under both past and future climates. We also make recommendations around existing approaches and highlight some data and research opportunities, that could lead to further improvements in IJF modelling and prediction.
Abstract Gridded meteorological estimates are essential for many applications. Most existing meteorological datasets are deterministic and have limitations in representing the inherent uncertainties from both the data and methodology used to create gridded products. We develop the Ensemble Meteorological Dataset for Planet Earth (EM-Earth) for precipitation, mean daily temperature, daily temperature range, and dewpoint temperature at 0.1° spatial resolution over global land areas from 1950 to 2019. EM-Earth provides hourly/daily deterministic estimates, and daily probabilistic estimates (25 ensemble members), to meet the diverse requirements of hydrometeorological applications. To produce EM-Earth, we first developed a station-based Serially Complete Earth (SC-Earth) dataset, which removes the temporal discontinuities in raw station observations. Then, we optimally merged SC-Earth station data and ERA5 estimates to generate EM-Earth deterministic estimates and their uncertainties. The EM-Earth ensemble members are produced by sampling from parametric probability distributions using spatiotemporally correlated random fields. The EM-Earth dataset is evaluated by leave-one-out validation, using independent evaluation stations, and comparing it with many widely used datasets. The results show that EM-Earth is better in Europe, North America, and Oceania than in Africa, Asia, and South America, mainly due to differences in the available stations and differences in climate conditions. Probabilistic spatial meteorological datasets are particularly valuable in regions with large meteorological uncertainties, where almost all existing deterministic datasets face great challenges in obtaining accurate estimates.
Abstract Gridded meteorological estimates are essential for many applications. Most existing meteorological datasets are deterministic and have limitations in representing the inherent uncertainties from both the data and methodology used to create gridded products. We develop the Ensemble Meteorological Dataset for Planet Earth (EM-Earth) for precipitation, mean daily temperature, daily temperature range, and dewpoint temperature at 0.1° spatial resolution over global land areas from 1950 to 2019. EM-Earth provides hourly/daily deterministic estimates, and daily probabilistic estimates (25 ensemble members), to meet the diverse requirements of hydrometeorological applications. To produce EM-Earth, we first developed a station-based Serially Complete Earth (SC-Earth) dataset, which removes the temporal discontinuities in raw station observations. Then, we optimally merged SC-Earth station data and ERA5 estimates to generate EM-Earth deterministic estimates and their uncertainties. The EM-Earth ensemble members are produced by sampling from parametric probability distributions using spatiotemporally correlated random fields. The EM-Earth dataset is evaluated by leave-one-out validation, using independent evaluation stations, and comparing it with many widely used datasets. The results show that EM-Earth is better in Europe, North America, and Oceania than in Africa, Asia, and South America, mainly due to differences in the available stations and differences in climate conditions. Probabilistic spatial meteorological datasets are particularly valuable in regions with large meteorological uncertainties, where almost all existing deterministic datasets face great challenges in obtaining accurate estimates.
DOI
bib
abs
Hydrologic Model Sensitivity to Temporal Aggregation of Meteorological Forcing Data: A Case Study for the Contiguous United States
Ashley E. Van Beusekom,
Lauren Hay,
Andrew Bennett,
Young-Don Choi,
Martyn P. Clark,
J. L. Goodall,
Zhiyu Li,
Iman Maghami,
Bart Nijssen,
Andrew W. Wood
Journal of Hydrometeorology, Volume 23, Issue 2
Abstract Surface meteorological analyses are an essential input (termed “forcing”) for hydrologic modeling. This study investigated the sensitivity of different hydrologic model configurations to temporal variations of seven forcing variables (precipitation rate, air temperature, longwave radiation, specific humidity, shortwave radiation, wind speed, and air pressure). Specifically, the effects of temporally aggregating hourly forcings to hourly daily average forcings were examined. The analysis was based on 14 hydrological outputs from the Structure for Unifying Multiple Modeling Alternatives (SUMMA) model for the 671 Catchment Attributes and Meteorology for Large-Sample Studies (CAMELS) basins across the contiguous United States (CONUS). Results demonstrated that the hydrologic model sensitivity to temporally aggregating the forcing inputs varies across model output variables and model locations. We used Latin hypercube sampling to sample model parameters from eight combinations of three influential model physics choices (three model decisions with two options for each decision, i.e., eight model configurations). Results showed that the choice of model physics can change the relative influence of forcing on model outputs and the forcing importance may not be dependent on the parameter space. This allows for model output sensitivity to forcing aggregation to be tested prior to parameter calibration. More generally, this work provides a comprehensive analysis of the dependence of modeled outcomes on input forcing behavior, providing insight into the regional variability of forcing variable dominance on modeled outputs across CONUS.
DOI
bib
abs
New projections of 21st century climate and hydrology for Alaska and Hawaiʻi
Naoki Mizukami,
Andrew J. Newman,
Jeremy S. Littell,
Thomas W. Giambelluca,
Andrew W. Wood,
E. D. Gutmann,
Joseph Hamman,
Diana R. Gergel,
Bart Nijssen,
Martyn P. Clark,
Jeffrey R. Arnold
Climate Services, Volume 27
In the United States, high-resolution, century-long, hydroclimate projection datasets have been developed for water resources planning, focusing on the contiguous United States (CONUS) domain. However, there are few statewide hydroclimate projection datasets available for Alaska and Hawaiʻi. The limited information on hydroclimatic change motivates developing hydrologic scenarios from 1950 to 2099 using climate-hydrology impact modeling chains consisting of multiple statistically downscaled climate projections as input to hydrologic model simulations for both states. We adopt an approach similar to the previous CONUS hydrologic assessments where: 1) we select the outputs from ten global climate models (GCM) from the Coupled Model Intercomparison Project Phase 5 with Representative Concentration Pathways 4.5 and 8.5; 2) we perform statistical downscaling to generate climate input data for hydrologic models (12-km grid-spacing for Alaska and 1-km for Hawaiʻi); and 3) we perform process-based hydrologic model simulations. For Alaska, we have advanced the hydrologic model configuration from CONUS by using the full water-energy balance computation, frozen soils and a simple glacier model. The simulations show that robust warming and increases in precipitation produce runoff increases for most of Alaska, with runoff reductions in the currently glacierized areas in Southeast Alaska. For Hawaiʻi, we produce the projections at high resolution (1 km) which highlight high spatial variability of climate variables across the state, and a large spread of runoff across the GCMs is driven by a large precipitation spread across the GCMs. Our new ensemble datasets assist with state-wide climate adaptation and other water planning.
The Köppen-Geiger (KG) climate classification has been widely used to determine the climate at global and regional scales using precipitation and temperature data. KG maps are typically developed using a single product; however, uncertainties in KG climate types resulting from different precipitation and temperature datasets have not been explored in detail. Here, we assess seven global datasets to show uncertainties in KG classification from 1980 to 2017. Using a pairwise comparison at global and zonal scales, we quantify the similarity among the seven KG maps. Gauge- and reanalysis-based KG maps have a notable difference. Spatially, the highest and lowest similarity is observed for the North and South Temperate zones, respectively. Notably, 17% of grids among the seven maps show variations even in the major KG climate types, while 35% of grids are described by more than one KG climate subtype. Strong uncertainty is observed in south Asia, central and south Africa, western America, and northeastern Australia. We created two KG master maps (0.5° resolution) by merging the climate maps directly and by combining the precipitation and temperature data from the seven datasets. These master maps are more robust than the individual ones showing coherent spatial patterns. This study reveals the large uncertainty in climate classification and offers two robust KG maps that may help to better evaluate historical climate and quantify future climate shifts.
DOI
bib
abs
Community Workflows to Advance Reproducibility in Hydrologic Modeling: Separating Model‐Agnostic and Model‐Specific Configuration Steps in Applications of Large‐Domain Hydrologic Models
Wouter Knoben,
Martyn P. Clark,
Jerad D. Bales,
Andrew Bennett,
Shervan Gharari,
Christopher B. Marsh,
Bart Nijssen,
Alain Pietroniro,
Raymond J. Spiteri,
Guoqiang Tang,
David G. Tarboton,
A. W. Wood
Water Resources Research, Volume 58, Issue 11
Despite the proliferation of computer-based research on hydrology and water resources, such research is typically poorly reproducible. Published studies have low reproducibility due to incomplete availability of data and computer code, and a lack of documentation of workflow processes. This leads to a lack of transparency and efficiency because existing code can neither be quality controlled nor reused. Given the commonalities between existing process-based hydrologic models in terms of their required input data and preprocessing steps, open sharing of code can lead to large efficiency gains for the modeling community. Here, we present a model configuration workflow that provides full reproducibility of the resulting model instantiations in a way that separates the model-agnostic preprocessing of specific data sets from the model-specific requirements that models impose on their input files. We use this workflow to create large-domain (global and continental) and local configurations of the Structure for Unifying Multiple Modeling Alternatives (SUMMA) hydrologic model connected to the mizuRoute routing model. These examples show how a relatively complex model setup over a large domain can be organized in a reproducible and structured way that has the potential to accelerate advances in hydrologic modeling for the community as a whole. We provide a tentative blueprint of how community modeling initiatives can be built on top of workflows such as this. We term our workflow the “Community Workflows to Advance Reproducibility in Hydrologic Modeling” (CWARHM; pronounced “swarm”).
Long-term groundwater droughts are known to persist over timescales from multiple years up to decades. The mechanisms leading to drought persistence are, however, only partly understood. Applying a unique terrestrial system modeling platform in a probabilistic simulation framework over Europe, we discovered an important positive feedback mechanism from groundwater into the atmosphere that may increase drought persistence at interannual time scales over large continental regions. In the feedback loop, groundwater drought systematically increases net solar radiation via a cloud feedback, which, in turn, increases the drying of the land. In commonly applied climate and Earth system models, this feedback cannot be simulated due to a lack of groundwater memory effects in the representation of terrestrial hydrology. Thus, drought persistence and compound events may be underestimated in current climate projections.
Abstract Statistical processing of numerical model output has been a part of both weather forecasting and climate applications for decades. Statistical techniques are used to correct systematic biases in atmospheric model outputs and to represent local effects that are unresolved by the model, referred to as downscaling. Many downscaling techniques have been developed, and it has been difficult to systematically explore the implications of the individual decisions made in the development of downscaling methods. Here we describe a unified framework that enables the user to evaluate multiple decisions made in the methods used to statistically postprocess output from weather and climate models. The Ensemble Generalized Analog Regression Downscaling (En-GARD) method enables the user to select any number of input variables, predictors, mathematical transformations, and combinations for use in parametric or nonparametric downscaling approaches. En-GARD enables explicitly predicting both the probability of event occurrence and the event magnitude. Outputs from En-GARD include errors in model fit, enabling the production of an ensemble of projections through sampling of the probability distributions of each climate variable. We apply En-GARD to regional climate model simulations to evaluate the relative importance of different downscaling method choices on simulations of the current and future climate. We show that choice of predictor variables is the most important decision affecting downscaled future climate outputs, while having little impact on the fidelity of downscaled outcomes for current climate. We also show that weak statistical relationships prevent such approaches from predicting large changes in extreme events on a daily time scale.
Abstract Surface meteorological analyses serve a wide range of research and applications, including forcing inputs for hydrological and ecological models, climate analysis, and resource and emergency management. Quantifying uncertainty in such analyses would extend their utility for probabilistic hydrologic prediction and climate risk applications. With this motivation, we enhance and evaluate an approach for generating ensemble analyses of precipitation and temperature through the fusion of station observations, terrain information, and numerical weather prediction simulations of surface climate fields. In particular, we expand a spatial regression in which static terrain attributes serve as predictors for spatially distributed 1/16th degree daily surface precipitation and temperature by including forecast outputs from the High-Resolution Rapid Refresh (HRRR) numerical weather prediction model as additional predictors. We demonstrate the approach for a case study domain of California, focusing on the meteorological conditions leading to the 2017 flood and spillway failure event at Lake Oroville. The approach extends the spatial regression capability of the Gridded Meteorological Ensemble Tool (GMET) and also adds cross-validation to the uncertainty estimation component, enabling the use of predictive rather than calibration uncertainty. In evaluation against out-of-sample station observations, the HRRR-based predictors alone are found to be skillful for the study setting, leading to overall improvements in the enhanced GMET meteorological analyses. The methodology and associated tool represent a promising method for generating meteorological surface analyses for both research-oriented and operational applications, as well as a general strategy for merging in situ and gridded observations.
DOI
bib
abs
Evaluating a reservoir parametrization in the vector-based global routing model mizuRoute (v2.0.1) for Earth system model coupling
Inne Vanderkelen,
Shervan Gharari,
Naoki Mizukami,
Martyn P. Clark,
David M. Lawrence,
Sean Swenson,
Yadu Pokhrel,
Naota Hanasaki,
Ann van Griensven,
Wim Thiery
Geoscientific Model Development, Volume 15, Issue 10
Abstract. Human-controlled reservoirs have a large influence on the global water cycle. While global hydrological models use generic parameterizations to model dam operations, the representation of reservoir regulation is still lacking in many Earth system models. Here we implement and evaluate a widely used reservoir parametrization in the global river-routing model mizuRoute, which operates on a vector-based river network resolving individual lakes and reservoirs and is currently being coupled to an Earth system model. We develop an approach to determine the downstream area over which to aggregate irrigation water demand per reservoir. The implementation of managed reservoirs is evaluated by comparing them to simulations ignoring inland waters and simulations with reservoirs represented as natural lakes using (i) local simulations for 26 individual reservoirs driven by observed inflows and (ii) global-domain simulations driven by runoff from the Community Land Model. The local simulations show the clear added value of the reservoir parametrization, especially for simulating storage for large reservoirs with a multi-year storage capacity. In the global-domain application, the implementation of reservoirs shows an improvement in outflow and storage compared to the no-reservoir simulation, but a similar performance is found compared to the natural lake parametrization. The limited impact of reservoirs on skill statistics could be attributed to biases in simulated river discharge, mainly originating from biases in simulated runoff from the Community Land Model. Finally, the comparison of modelled monthly streamflow indices against observations highlights that including dam operations improves the streamflow simulation compared to ignoring lakes and reservoirs. This study overall underlines the need to further develop and test runoff simulations and water management parameterizations in order to improve the representation of anthropogenic interference of the terrestrial water cycle in Earth system models.
2021
Abstract The intent of this paper is to encourage improved numerical implementation of land models. Our contributions in this paper are two-fold. First, we present a unified framework to formulate and implement land model equations. We separate the representation of physical processes from their numerical solution, enabling the use of established robust numerical methods to solve the model equations. Second, we introduce a set of synthetic test cases (the laugh tests) to evaluate the numerical implementation of land models. The test cases include storage and transmission of water in soils, lateral sub-surface flow, coupled hydrological and thermodynamic processes in snow, and cryosuction processes in soil. We consider synthetic test cases as “laugh tests” for land models because they provide the most rudimentary test of model capabilities. The laugh tests presented in this paper are all solved with the Structure for Unifying Multiple Modeling Alternatives model (SUMMA) implemented using the SUite of Nonlinear and DIfferential/Algebraic equation Solvers (SUNDIALS). The numerical simulations from SUMMA/SUNDIALS are compared against (1) solutions to the synthetic test cases from other models documented in the peer-reviewed literature; (2) analytical solutions; and (3) observations made in laboratory experiments. In all cases, the numerical simulations are similar to the benchmarks, building confidence in the numerical model implementation. We posit that some land models may have difficulty in solving these benchmark problems. Dedicating more effort to solving synthetic test cases is critical in order to build confidence in the numerical implementation of land models.
Abstract Stations are an important source of meteorological data, but often suffer from missing values and short observation periods. Gap filling is widely used to generate serially complete datasets (SCDs), which are subsequently used to produce gridded meteorological estimates. However, the value of SCDs in spatial interpolation is scarcely studied. Based on our recent efforts to develop a SCD over North America (SCDNA), we explore the extent to which gap filling improves gridded precipitation and temperature estimates. We address two specific questions: (1) Can SCDNA improve the statistical accuracy of gridded estimates in North America? (2) Can SCDNA improve estimates of trends on gridded data? In addressing these questions, we also evaluate the extent to which results depend on the spatial density of the station network and the spatial interpolation methods used. Results show that the improvement in statistical interpolation due to gap filling is more obvious for precipitation, followed by minimum temperature and maximum temperature. The improvement is larger when the station network is sparse and when simpler interpolation methods are used. SCDs can also notably reduce the uncertainties in spatial interpolation. Our evaluation across North America from 1979 to 2018 demonstrates that SCDs improve the accuracy of interpolated estimates for most stations and days. SCDNA-based interpolation also obtains better trend estimation than observation-based interpolation. This occurs because stations used for interpolation could change during a specific period, causing changepoints in interpolated temperature estimates and affect the long-term trends of observation-based interpolation, which can be avoided using SCDNA. Overall, SCDs improve the performance of gridded precipitation and temperature estimates.
DOI
bib
abs
Open Science: Open Data, Open Models, …and Open Publications?
Martyn P. Clark,
Charles H. Luce,
Amir AghaKouchak,
Wouter Berghuijs,
Cédric H. David,
Qingyun Duan,
Shemin Ge,
Ilja van Meerveld,
Kewei Chen,
M. B. Parlange,
S. W. Tyler
Water Resources Research, Volume 57, Issue 4
This commentary explores the challenges and opportunities associated with a possible transition of Water Resources Research to a publication model where all articles are freely available upon publication (“Gold” open access). It provides a review of the status of open access publishing models, a summary of community input, and a path forward for AGU leadership. The decision to convert to open access is framed by a mix of finances and values. On the one hand, the challenge is to define who pays, and how, and what we can do to improve the affordability of publishing. On the other hand, the challenge is to increase the extent to which science is open and accessible. The next steps for the community include an incisive analysis of the financial feasibility of different cost models, and weighing the financial burden for open access against the desire to further advance open science.
Projections of change in high-flow extremes with global warming vary widely among, and within, large midlatitude river basins. The spatial variability of these changes is attributable to multiple causes. One possible and little-studied cause of changes in high-flow extremes is a change in the synchrony of mainstem and tributary streamflow during high-flow extremes at the mainstem-tributary confluence. We examined reconstructed and simulated naturalized daily streamflow at confluences on the Columbia River in western North America, quantifying changes in synchrony in future streamflow projections and estimating the impact of these changes on high-flow extremes. In the Columbia River basin, projected flow regimes across colder tributaries initially diverge with warming as they respond to climate change at different rates, leading to a general decrease in synchrony, and lower high-flow extremes, relative to a scenario with no changes in synchrony. Where future warming is sufficiently large to cause most subbasins upstream from a confluence to transition toward a rain-dominated, warm regime, the decreasing trend in synchrony reverses itself. At one confluence with a major tributary (the Willamette River), where the mainstem and tributary flow regimes are initially very different, warming increases synchrony and, therefore, high-flow magnitudes. These results may be generalizable to the class of large rivers with large contributions to flood risk from the snow (i.e., cold) regime, but that also receive considerable discharge from tributaries that drain warmer basins.
Hydrometeorological flood generating processes (excess rain, short rain, long rain, snowmelt, and rain-on-snow) underpin our understanding of flood behavior. Knowledge about flood generating processes improves hydrological models, flood frequency analysis, estimation of climate change impact on floods, etc. Yet, not much is known about how climate and catchment attributes influence the spatial distribution of flood generating processes. This study aims to offer a comprehensive and structured approach to close this knowledge gap. We employ a large sample approach (671 catchments across the contiguous United States) and evaluate how catchment attributes and climate attributes influence the distribution of flood processes. We use two complementary approaches: A statistics-based approach which compares attribute frequency distributions of different flood processes; and a random forest model in combination with an interpretable machine learning approach (accumulated local effects [ALE]). The ALE method has not been used often in hydrology, and it overcomes a significant obstacle in many statistical methods, the confounding effect of correlated catchment attributes. As expected, we find climate attributes (fraction of snow, aridity, precipitation seasonality, and mean precipitation) to be most influential on flood process distribution. However, the influence of catchment attributes varies both with flood generating process and climate type. We also find flood processes can be predicted for ungauged catchments with relatively high accuracy (R2 between 0.45 and 0.9). The implication of these findings is flood processes should be considered for future climate change impact studies, as the effect of changes in climate on flood characteristics varies between flood processes.
As droughts have widespread social and ecological impacts, it is critical to develop long-term adaptation and mitigation strategies to reduce drought vulnerability. Climate models are important in quantifying drought changes. Here, we assess the ability of 285 CMIP6 historical simulations, from 17 models, to reproduce drought duration and severity in three observational data sets using the Standardized Precipitation Index (SPI). We used summary statistics beyond the mean and standard deviation, and devised a novel probabilistic framework, based on the Hellinger distance, to quantify the difference between observed and simulated drought characteristics. Results show that many simulations have less than ±10% error in reproducing the observed drought summary statistics. The hypothesis that simulations and observations are described by the same distribution cannot be rejected for more than 80% of the grids based on our H distance framework. No single model stood out as demonstrating consistently better performance over large regions of the globe. The variance in drought statistics among the simulations is higher in the tropics compared to other latitudinal zones. Though the models capture the characteristics of dry spells well, there is considerable bias in low precipitation values. Good model performance in terms of SPI does not imply good performance in simulating low precipitation. Our study emphasizes the need to probabilistically evaluate climate model simulations in order to both pinpoint model weaknesses and identify a subset of best-performing models that are useful for impact assessments.
DOI
bib
abs
Flood spatial coherence, triggers, and performance in hydrological simulations: large-sample evaluation of four streamflow-calibrated models
Manuela Irene Brunner,
Lieke Melsen,
A. W. Wood,
Oldřich Rakovec,
Naoki Mizukami,
Wouter Knoben,
Martyn P. Clark
Hydrology and Earth System Sciences, Volume 25, Issue 1
Abstract. Floods cause extensive damage, especially if they affect large regions. Assessments of current, local, and regional flood hazards and their future changes often involve the use of hydrologic models. A reliable hydrologic model ideally reproduces both local flood characteristics and spatial aspects of flooding under current and future climate conditions. However, uncertainties in simulated floods can be considerable and yield unreliable hazard and climate change impact assessments. This study evaluates the extent to which models calibrated according to standard model calibration metrics such as the widely used Kling–Gupta efficiency are able to capture flood spatial coherence and triggering mechanisms. To highlight challenges related to flood simulations, we investigate how flood timing, magnitude, and spatial variability are represented by an ensemble of hydrological models when calibrated on streamflow using the Kling–Gupta efficiency metric, an increasingly common metric of hydrologic model performance also in flood-related studies. Specifically, we compare how four well-known models (the Sacramento Soil Moisture Accounting model, SAC; the Hydrologiska Byråns Vattenbalansavdelning model, HBV; the variable infiltration capacity model, VIC; and the mesoscale hydrologic model, mHM) represent (1) flood characteristics and their spatial patterns and (2) how they translate changes in meteorologic variables that trigger floods into changes in flood magnitudes. Our results show that both the modeling of local and spatial flood characteristics are challenging as models underestimate flood magnitude, and flood timing is not necessarily well captured. They further show that changes in precipitation and temperature are not always well translated to changes in flood flow, which makes local and regional flood hazard assessments even more difficult for future conditions. From a large sample of catchments and with multiple models, we conclude that calibration on the integrated Kling–Gupta metric alone is likely to yield models that have limited reliability in flood hazard assessments, undermining their utility for regional and future change assessments. We underscore that such assessments can be improved by developing flood-focused, multi-objective, and spatial calibration metrics, by improving flood generating process representation through model structure comparisons and by considering uncertainty in precipitation input.
Predictions of floods, droughts, and fast drought‐flood transitions are required at different time scales to develop management strategies targeted at minimizing negative societal and economic impacts. Forecasts at daily and seasonal scale are vital for early warning, estimation of event frequency for hydraulic design, and long‐term projections for developing adaptation strategies to future conditions. All three types of predictions—forecasts, frequency estimates, and projections—typically treat droughts and floods independently, even though both types of extremes can be studied using related approaches and have similar challenges. In this review, we (a) identify challenges common to drought and flood prediction and their joint assessment and (b) discuss tractable approaches to tackle these challenges. We group challenges related to flood and drought prediction into four interrelated categories: data, process understanding, modeling and prediction, and human–water interactions. Data‐related challenges include data availability and event definition. Process‐related challenges include the multivariate and spatial characteristics of extremes, non‐stationarities, and future changes in extremes. Modeling challenges arise in frequency analysis, stochastic, hydrological, earth system, and hydraulic modeling. Challenges with respect to human–water interactions lie in establishing links to impacts, representing human–water interactions, and science communication. We discuss potential ways of tackling these challenges including exploiting new data sources, studying droughts and floods in a joint framework, studying societal influences and compounding drivers, developing continuous stochastic models or non‐stationary models, and obtaining stakeholder feedback. Tackling one or several of these challenges will improve flood and drought predictions and help to minimize the negative impacts of extreme events.
DOI
bib
abs
The Abuse of Popular Performance Metrics in Hydrologic Modeling
Martyn P. Clark,
Richard M. Vogel,
Jonathan R. Lamontagne,
Naoki Mizukami,
Wouter Knoben,
Guoqiang Tang,
Shervan Gharari,
Jim Freer,
Paul H. Whitfield,
Kevin Shook,
S. Papalexiou
Water Resources Research, Volume 57, Issue 9
The goal of this commentary is to critically evaluate the use of popular performance metrics in hydrologic modeling. We focus on the Nash-Sutcliffe Efficiency (NSE) and the Kling-Gupta Efficiency (KGE) metrics, which are both widely used in hydrologic research and practice around the world. Our specific objectives are: (a) to provide tools that quantify the sampling uncertainty in popular performance metrics; (b) to quantify sampling uncertainty in popular performance metrics across a large sample of catchments; and (c) to prescribe the further research that is, needed to improve the estimation, interpretation, and use of popular performance metrics in hydrologic modeling. Our large-sample analysis demonstrates that there is substantial sampling uncertainty in the NSE and KGE estimators. This occurs because the probability distribution of squared errors between model simulations and observations has heavy tails, meaning that performance metrics can be heavily influenced by just a few data points. Our results highlight obvious (yet ignored) abuses of performance metrics that contaminate the conclusions of many hydrologic modeling studies: It is essential to quantify the sampling uncertainty in performance metrics when justifying the use of a model for a specific purpose and when comparing the performance of competing models.
Abstract. Hydrological models are usually systems of nonlinear differential equations for which no analytical solutions exist and thus rely on approximate numerical solutions. While some studies have investigated the relationship between numerical method choice and model error, the extent to which extreme precipitation like that observed during hurricanes Harvey and Katrina impacts numerical error of hydrological models is still unknown. This knowledge is relevant in light of climate change, where many regions will likely experience more intense precipitation events. In this experiment, a large number of hydrographs is generated with the modular modeling framework FUSE, using eight numerical techniques across a variety of forcing datasets. Multiple model structures, parameter sets, and initial conditions are incorporated for generality. The computational expense and numerical error associated with each hydrograph were recorded. It was found that numerical error (root mean square error) usually increases with precipitation intensity and decreases with event duration. Some numerical methods constrain errors much more effectively than others, sometimes by many orders of magnitude. Of the tested numerical methods, a second-order adaptive explicit method is found to be the most efficient because it has both low numerical error and low computational cost. A basic literature review indicates that many popular modeling codes use numerical techniques that were suggested by this experiment to be sub-optimal. We conclude that relatively large numerical errors might be common in current models, and because these will likely become larger as the climate changes, we advocate for the use of low cost, low error numerical methods.
A vector‐river network explicitly uses realistic geometries of river reaches and catchments for spatial discretization in a river model. This enables improving the accuracy of the physical properties of the modeled river system, compared to a gridded river network that has been used in Earth System Models. With a finer‐scale river network, resolving smaller‐scale river reaches, there is a need for efficient methods to route streamflow and its constituents throughout the river network. The purpose of this study is twofold: (1) develop a new method to decompose river networks into hydrologically independent tributary domains, where routing computations can be performed in parallel; and (2) perform global river routing simulations with two global river networks, with different scales, to examine the computational efficiency and the differences in discharge simulations at various temporal scales. The new parallelization method uses a hierarchical decomposition strategy, where each decomposed tributary is further decomposed into many sub‐tributary domains, enabling hybrid parallel computing. This parallelization scheme has excellent computational scaling for the global domain where it is straightforward to distribute computations across many independent river basins. However, parallel computing for a single large basin remains challenging. The global routing experiments show that the scale of the vector‐river network has less impact on the discharge simulations than the runoff input that is generated by the combination of land surface model and meteorological forcing. The scale of vector‐river networks needs to consider the scale of local hydrologic features such as lakes that are to be resolved in the network.
Abstract. This study employs a stochastic hydrologic modeling framework to evaluate the sensitivity of flood frequency analyses to different components of the hydrologic modeling chain. The major components of the stochastic hydrologic modeling chain, including model structure, model parameter estimation, initial conditions, and precipitation inputs were examined across return periods from 2 to 100 000 years at two watersheds representing different hydroclimates across the western USA. A total of 10 hydrologic model structures were configured, calibrated, and run within the Framework for Understanding Structural Errors (FUSE) modular modeling framework for each of the two watersheds. Model parameters and initial conditions were derived from long-term calibrated simulations using a 100 member historical meteorology ensemble. A stochastic event-based hydrologic modeling workflow was developed using the calibrated models in which millions of flood event simulations were performed for each basin. The analysis of variance method was then used to quantify the relative contributions of model structure, model parameters, initial conditions, and precipitation inputs to flood magnitudes for different return periods. Results demonstrate that different components of the modeling chain have different sensitivities for different return periods. Precipitation inputs contribute most to the variance of rare floods, while initial conditions are most influential for more frequent events. However, the hydrological model structure and structure–parameter interactions together play an equally important role in specific cases, depending on the basin characteristics and type of flood metric of interest. This study highlights the importance of critically assessing model underpinnings, understanding flood generation processes, and selecting appropriate hydrological models that are consistent with our understanding of flood generation processes.
Abstract. Hydrological models are usually systems of nonlinear differential equations for which no analytical solutions exist and thus rely on numerical solutions. While some studies have investigated the relationship between numerical method choice and model error, the extent to which extreme precipitation such as that observed during hurricanes Harvey and Katrina impacts numerical error of hydrological models is still unknown. This knowledge is relevant in light of climate change, where many regions will likely experience more intense precipitation. In this experiment, a large number of hydrographs are generated with the modular modeling framework FUSE (Framework for Understanding Structural Errors), using eight numerical techniques across a variety of forcing data sets. All constructed models are conceptual and lumped. Multiple model structures, parameter sets, and initial conditions are incorporated for generality. The computational cost and numerical error associated with each hydrograph were recorded. Numerical error is assessed via root mean square error and normalized root mean square error. It was found that the root mean square error usually increases with precipitation intensity and decreases with event duration. Some numerical methods constrain errors much more effectively than others, sometimes by many orders of magnitude. Of the tested numerical methods, a second-order adaptive explicit method is found to be the most efficient because it has both a small numerical error and a low computational cost. A small literature review indicates that many popular modeling codes use numerical techniques that were suggested by this experiment to be suboptimal. We conclude that relatively large numerical errors may be common in current models, highlighting the need for robust numerical techniques, in particular in the face of increasing precipitation extremes.
Global sensitivity analysis (GSA) has long been recognized as an indispensable tool for model analysis. GSA has been extensively used for model simplification, identifiability analysis, and diagnostic tests. Nevertheless, computationally efficient methodologies are needed for GSA, not only to reduce the computational overhead, but also to improve the quality and robustness of the results. This is especially the case for process-based hydrologic models, as their simulation time typically exceeds the computational resources available for a comprehensive GSA. To overcome this computational barrier, we propose a data-driven method called VISCOUS, variance-based sensitivity analysis using copulas. VISCOUS uses Gaussian mixture copulas to approximate the joint probability density function of a given set of input-output pairs for estimating the variance-based sensitivity indices. Our method identifies dominant hydrologic factors by recycling existing input-output data, and thus can deal with arbitrary sample sets drawn from the input-output space. We used two hydrologic models of increasing complexity (HBV and VIC) to assess the performance of VISCOUS. Our results confirm that VISCOUS and the conventional variance-based method can detect similar important and unimportant factors. Furthermore, the VISCOUS method can substantially reduce the computational cost required for sensitivity analysis. Our proposed method is particularly useful for process-based models with many uncertain parameters, large domain size, and high spatial and temporal resolution.
Abstract Meteorological data from ground stations suffer from temporal discontinuities caused by missing values and short measurement periods. Gap-filling and reconstruction techniques have proven to be effective in producing serially complete station datasets (SCDs) that are used for a myriad of meteorological applications (e.g., developing gridded meteorological datasets and validating models). To our knowledge, all SCDs are developed at regional scales. In this study, we developed the serially complete Earth (SC-Earth) dataset, which provides daily precipitation, mean temperature, temperature range, dewpoint temperature, and wind speed data from 1950 to 2019. SC-Earth utilizes raw station data from the Global Historical Climatology Network–Daily (GHCN-D) and the Global Surface Summary of the Day (GSOD). A unified station repository is generated based on GHCN-D and GSOD after station merging and strict quality control. ERA5 is optimally matched with station data considering the time shift issue and then used to assist the global gap filling. SC-Earth is generated by merging estimates from 15 strategies based on quantile mapping, spatial interpolation, machine learning, and multistrategy merging. The final estimates are bias corrected using a combination of quantile mapping and quantile delta mapping. Comprehensive validation demonstrates that SC-Earth has high accuracy around the globe, with degraded quality in the tropics and oceanic islands due to sparse station networks, strong spatial precipitation gradients, and degraded ERA5 estimates. Meanwhile, SC-Earth inherits potential limitations such as inhomogeneity and precipitation undercatch from raw station data, which may affect its application in some cases. Overall, the high-quality and high-density SC-Earth dataset will benefit research in fields of hydrology, ecology, meteorology, and climate. The dataset is available at https://zenodo.org/record/4762586 .
Abstract. Probabilistic methods are useful to estimate the uncertainty in spatial meteorological fields (e.g., the uncertainty in spatial patterns of precipitation and temperature across large domains). In ensemble probabilistic methods, “equally plausible” ensemble members are used to approximate the probability distribution, hence the uncertainty, of a spatially distributed meteorological variable conditioned to the available information. The ensemble members can be used to evaluate the impact of uncertainties in spatial meteorological fields for a myriad of applications. This study develops the Ensemble Meteorological Dataset for North America (EMDNA). EMDNA has 100 ensemble members with daily precipitation amount, mean daily temperature, and daily temperature range at 0.1∘ spatial resolution (approx. 10 km grids) from 1979 to 2018, derived from a fusion of station observations and reanalysis model outputs. The station data used in EMDNA are from a serially complete dataset for North America (SCDNA) that fills gaps in precipitation and temperature measurements using multiple strategies. Outputs from three reanalysis products are regridded, corrected, and merged using Bayesian model averaging. Optimal interpolation (OI) is used to merge station- and reanalysis-based estimates. EMDNA estimates are generated using spatiotemporally correlated random fields to sample from the OI estimates. Evaluation results show that (1) the merged reanalysis estimates outperform raw reanalysis estimates, particularly in high latitudes and mountainous regions; (2) the OI estimates are more accurate than the reanalysis and station-based regression estimates, with the most notable improvements for precipitation evident in sparsely gauged regions; and (3) EMDNA estimates exhibit good performance according to the diagrams and metrics used for probabilistic evaluation. We discuss the limitations of the current framework and highlight that further research is needed to improve ensemble meteorological datasets. Overall, EMDNA is expected to be useful for hydrological and meteorological applications in North America. The entire dataset and a teaser dataset (a small subset of EMDNA for easy download and preview) are available at https://doi.org/10.20383/101.0275 (Tang et al., 2020a).
Reservoir expansion over the last century has largely affected downstream flow characteristics. Yet very little is known about the impacts of reservoir expansion on the climate. Here, we implement reservoir construction in the Community Land Model by enabling dynamical lake area changes, while conserving mass and energy. Transient global lake and reservoir extent are prescribed from the HydroLAKES and Global Reservoir and Dam databases. Land-only simulations covering the 20th century with reservoir expansion enabled, highlight increases in terrestrial water storage and decreases in albedo matching the increase in open water area. The comparison of coupled simulations including and excluding reservoirs shows only limited influence of reservoirs on global temperatures and the surface energy balance, but demonstrates substantial responses locally, in particular where reservoirs make up a large fraction of the grid cell. In those locations, reservoirs dampen the diurnal temperature range by up to −1.5 K (for reservoirs covering >15% of the grid cell), reduce temperature extremes, and moderate the seasonal temperature cycle. This study provides a first step towards a coupled representation of reservoirs in Earth System Models.
Abstract Gridded precipitation datasets are used in many applications such as the analysis of climate variability/change and hydrological modelling. Regridding precipitation datasets is common for model coupling (e.g., coupling atmospheric and hydrological models) or comparing different models and datasets. However, regridding can considerably alter precipitation statistics. In this global analysis, the effects of regridding a precipitation dataset are emphasized using three regridding methods (first order conservative, bilinear, and distance weighted averaging). The differences between the original and regridded dataset are substantial and greatest at high quantiles. Differences of 46 mm and 0.13 mm are noted in high (0.95) and low (0.05) quantiles respectively. The impacts of regridding vary spatially for land and oceanic regions; there are substantial differences at high quantiles in tropical land regions, and at low quantiles in polar regions. These impacts are approximately the same for different regridding methods. The differences increase with the size of the grid at higher quantiles and vice versa for low quantiles. As the grid resolution increases, the difference between original and regridded data declines, yet the shift size dominates for high quantiles for which the differences are higher. Whilst regridding is often necessary to use gridded precipitation datasets, it should be used with great caution for fine resolutions (e.g., daily and sub-daily), as it can severely alter the statistical properties of precipitation, specifically at high and low quantiles.
The Arctic has been warming faster than the global average during recent decades, and trends are projected to continue through the twenty-first century. Analysis of climate change impacts across the Arctic using dynamical models has almost exclusively been limited to outputs from global climate models or coarser regional climate models. Coarse resolution simulations limit the representation of physical processes, particularly in areas of complex topography and high land-surface heterogeneity. Here, current climate reference and future regional climate model simulations based on the RCP8.5 scenario over Alaska at 4 km grid spacing are compared to identify changes in snowfall and snowpack. In general, results show increases in total precipitation, large decreases in snowfall fractional contribution over 30% in some areas, decreases in snowpack season length by 50–100 days in lower elevations and along the southern Alaskan coastline, and decreases in snow water equivalent. However, increases in snowfall and snowpack of sometimes greater than 20% are evident for some colder northern areas and at the highest elevations in southern Alaska. The most significant changes in snow cover and snowfall fractional contributions occur during the spring and fall seasons. Finally, the spatial pattern of winter temperatures above freezing has small-scale spatial features tied to the topography. Such areas would not be resolved with coarser resolution regional or global climate model simulations.
DOI
bib
abs
Toward open and reproducible environmental modeling by integrating online data repositories, computational environments, and model Application Programming Interfaces
Young-Don Choi,
Jonathan L. Goodall,
Jeffrey M. Sadler,
Anthony M. Castronova,
Andrew Bennett,
Zhiyu Li,
Bart Nijssen,
Shaowen Wang,
Martyn P. Clark,
Daniel P. Ames,
Jeffery S. Horsburgh,
Yi Hong,
C. Bandaragoda,
M. Seul,
Richard Hooper,
David G. Tarboton
Environmental Modelling & Software, Volume 135
Cyberinfrastructure needs to be advanced to enable open and reproducible environmental modeling research. Recent efforts toward this goal have focused on advancing online repositories for data and model sharing, online computational environments along with containerization technology and notebooks for capturing reproducible computational studies, and Application Programming Interfaces (APIs) for simulation models to foster intuitive programmatic control. The objective of this research is to show how these efforts can be integrated to support reproducible environmental modeling. We present first the high-level concept and general approach for integrating these three components. We then present one possible implementation that integrates HydroShare (an online repository), CUAHSI JupyterHub and CyberGIS-Jupyter for Water (computational environments), and pySUMMA (a model API) to support open and reproducible hydrologic modeling. We apply the example implementation for a hydrologic modeling use case to demonstrate how the approach can advance reproducible environmental modeling through the seamless integration of cyberinfrastructure services. • New approaches are needed to support open and reproducible environmental modeling. • Efforts should focus on integrating existing cyberinfrastructure to build new systems. • Our focus is on integrating repositories, computational environments, and model APIs. • An example implementation is shown using HydroShare, JupyterHub, and pySUMMA. • We demonstrate how the approach fosters reproducibility using a modeling case study.
Abstract The diurnal cycle of solar radiation represents the strongest energetic forcing and dominates the exchange of heat and mass of the land surface with the atmosphere. This diurnal heat redistribution represents a core of land–atmosphere coupling that should be accurately represented in land surface models (LSMs), which are critical parts of weather and climate models. We employ a diagnostic model evaluation approach using a signature-based metric that describes the diurnal variation of heat fluxes. The metric is obtained by decomposing the diurnal variation of surface heat fluxes into their direct response and the phase lag to incoming solar radiation. We employ the output of 13 different LSMs driven with meteorological forcing of 20 FLUXNET sites (PLUMBER dataset). All LSMs show a poor representation of the evaporative fraction and thus the diurnal magnitude of the sensible and latent heat flux under cloud-free conditions. In addition, we find that the diurnal phase of both heat fluxes is poorly represented. The best performing model only reproduces 33% of the evaluated evaporative conditions across the sites. The poor performance of the diurnal cycle of turbulent heat exchange appears to be linked to how models solve for the surface energy balance and redistribute heat into the subsurface. We conclude that a systematic evaluation of diurnal signatures is likely to help to improve the simulated diurnal cycle, better represent land–atmosphere interactions, and therefore improve simulations of the near-surface climate.
2020
Abstract. Floods cause large damages, especially if they affect large regions. Assessments of current, local and regional flood hazards and their future changes often involve the use of hydrologic models. However, uncertainties in simulated floods can be considerable and yield unreliable hazard and climate change impact assessments. A reliable hydrologic model ideally reproduces both local flood characteristics and spatial aspects of flooding, which is, however, not guaranteed especially when using standard model calibration metrics. In this paper we investigate how flood timing, magnitude and spatial variability are represented by an ensemble of hydrological models when calibrated on streamflow using the Kling–Gupta efficiency metric, an increasingly common metric of hydrologic model performance. We compare how four well-known models (SAC, HBV, VIC, and mHM) represent (1) flood characteristics and their spatial patterns; and (2) how they translate changes in meteorologic variables that trigger floods into changes in flood magnitudes. Our results show that both the modeling of local and spatial flood characteristics is challenging. They further show that changes in precipitation and temperature are not necessarily well translated to changes in flood flow, which makes local and regional flood hazard assessments even more difficult for future conditions. We conclude that models calibrated on integrated metrics such as the Kling–Gupta efficiency alone have limited reliability in flood hazard assessments, in particular in regional and future assessments, and suggest the development of alternative process-based and spatial evaluation metrics.
Abstract. Land models are increasingly used in terrestrial hydrology due to their process-oriented representation of water and energy fluxes. Land models can be set up at a range of spatial configurations, often ranging from grid sizes of 0.02 to 2 degrees (approximately 2 to 200 km) and applied at sub-daily temporal resolutions for simulation of energy fluxes. A priori specification of the grid size of the land models typically is derived from forcing resolutions, modeling objectives, available geo-spatial data and computational resources. Typically, the choice of model configuration and grid size is based on modeling convenience and is rarely examined for adequate physical representation in the context of modeling. The variability of the inputs and parameters, forcings, soil types, and vegetation covers, are masked or aggregated based on the a priori chosen grid size. In this study, we propose an alternative to directly set up a land model based on the concept of Group Response Unit (GRU). Each GRU is a unique combination of land cover, soil type, and other desired geographical features that has hydrological significance, such as elevation zone, slope, and aspect. Computational units are defined as GRUs that are forced at a specific forcing resolution; therefore, each computational unit has a unique combination of specific geo-spatial data and forcings. We set up the Variable Infiltration Capacity (VIC) model, based on the GRU concept (VIC-GRU). Utilizing this model setup and its advantages we try to answer the following questions: (1) how well a model configuration simulates an output variable, such as streamflow, for range of computational units, (2) how well a model configuration with fewer computational units, coarser forcing resolution and less geo-spatial information, reproduces a model set up with more computational units, finer forcing resolution and more geo-spatial information, and finally (3) how uncertain the model structure and parameters are for the land model. Our results, although case dependent, show that the models may similarly reproduce output with a lower number of computational units in the context of modeling (streamflow for example). Our results also show that a model configuration with a lower number of computational units may reproduce the simulations from a model configuration with more computational units. Similarly, this can assist faster parameter identification and model diagnostic suites, such as sensitivity and uncertainty, on a less computationally expensive model setup. Finally, we encourage the land model community to adopt flexible approaches that will provide a better understanding of accuracy-performance tradeoff in land models.
Abstract. Probabilistic methods are very useful to estimate the spatial variability in meteorological conditions (e.g., spatial patterns of precipitation and temperature across large domains). In ensemble probabilistic methods, equally plausible ensemble members are used to approximate the probability distribution, hence uncertainty, of a spatially distributed meteorological variable conditioned on the available information. The ensemble can be used to evaluate the impact of the uncertainties in a myriad of applications. This study develops the Ensemble Meteorological Dataset for North America (EMDNA). EMDNA has 100 members with daily precipitation amount, mean daily temperature, and daily temperature range at 0.1° spatial resolution from 1979 to 2018, derived from a fusion of station observations and reanalysis model outputs. The station data used in EMDNA are from a serially complete dataset for North America (SCDNA) that fills gaps in precipitation and temperature measurements using multiple strategies. Outputs from three reanalysis products are regridded, corrected, and merged using the Bayesian Model Averaging. Optimal Interpolation (OI) is used to merge station- and reanalysis-based estimates. EMDNA estimates are generated based on OI estimates and spatiotemporally correlated random fields. Evaluation results show that (1) the merged reanalysis estimates outperform raw reanalysis estimates, particularly in high latitudes and mountainous regions; (2) the OI estimates are more accurate than the reanalysis and station-based regression estimates, with the most notable improvement for precipitation occurring in sparsely gauged regions; and (3) EMDNA estimates exhibit good performance according to the diagrams and metrics used for probabilistic evaluation. We also discuss the limitations of the current framework and highlight that persistent efforts are needed to further develop probabilistic methods and ensemble datasets. Overall, EMDNA is expected to be useful for hydrological and meteorological applications in North America. The whole dataset and a teaser dataset (a small subset of EMDNA for easy download and preview) are available at https://doi.org/10.20383/101.0275 (Tang et al., 2020a).
Abstract. Station-based serially complete datasets (SCDs) of precipitation and temperature observations are important for hydrometeorological studies. Motivated by the lack of serially complete station observations for North America, this study seeks to develop an SCD from 1979 to 2018 from station data. The new SCD for North America (SCDNA) includes daily precipitation, minimum temperature (Tmin), and maximum temperature (Tmax) data for 27 276 stations. Raw meteorological station data were obtained from the Global Historical Climate Network Daily (GHCN-D), the Global Surface Summary of the Day (GSOD), Environment and Climate Change Canada (ECCC), and a compiled station database in Mexico. Stations with at least 8-year-long records were selected, which underwent location correction and were subjected to strict quality control. Outputs from three reanalysis products (ERA5, JRA-55, and MERRA-2) provided auxiliary information to estimate station records. Infilling during the observation period and reconstruction beyond the observation period were accomplished by combining estimates from 16 strategies (variants of quantile mapping, spatial interpolation, and machine learning). A sensitivity experiment was conducted by assuming that 30 % of observations from stations were missing – this enabled independent validation and provided a reference for reconstruction. Quantile mapping and mean value corrections were applied to the final estimates. The median Kling–Gupta efficiency (KGE′) values of the final SCDNA for all stations are 0.90, 0.98, and 0.99 for precipitation, Tmin, and Tmax, respectively. The SCDNA is closer to station observations than the four benchmark gridded products and can be used in applications that require either quality-controlled meteorological station observations or reconstructed long-term estimates for analysis and modeling. The dataset is available at https://doi.org/10.5281/zenodo.3735533 (Tang et al., 2020).
Abstract The Integrated Multi-satellitE Retrievals for Global Precipitation Measurement (IMERG) produces the latest generation of satellite precipitation estimates and has been widely used since its release in 2014. IMERG V06 provides global rainfall and snowfall data beginning from 2000. This study comprehensively analyzes the quality of the IMERG product at daily and hourly scales in China from 2000 to 2018 with special attention paid to snowfall estimates. The performance of IMERG is compared with nine satellite and reanalysis products (TRMM 3B42, CMORPH, PERSIANN-CDR, GSMaP, CHIRPS, SM2RAIN, ERA5, ERA-Interim, and MERRA2). Results show that the IMERG product outperforms other datasets, except the Global Satellite Mapping of Precipitation (GSMaP), which uses daily-scale station data to adjust satellite precipitation estimates. The monthly-scale station data adjustment used by IMERG naturally has a limited impact on estimates of precipitation occurrence and intensity at the daily and hourly time scales. The quality of IMERG has improved over time, attributed to the increasing number of passive microwave samples. SM2RAIN, ERA5, and MERRA2 also exhibit increasing accuracy with time that may cause variable performance in climatological studies. Even relying on monthly station data adjustments, IMERG shows good performance in both accuracy metrics at hourly time scales and the representation of diurnal cycles. In contrast, although ERA5 is acceptable at the daily scale, it degrades at the hourly scale due to the limitation in reproducing the peak time, magnitude and variation of diurnal cycles. IMERG underestimates snowfall compared with gauge and reanalysis data. The triple collocation analysis suggests that IMERG snowfall is worse than reanalysis and gauge data, which partly results in the degraded quality of IMERG in cold climates. This study demonstrates new findings on the uncertainties of various precipitation products and identifies potential directions for algorithm improvement. The results of this study will be useful for both developers and users of satellite rainfall products.
Multi-model climate experiments carried out as part of different phases of the Coupled Model Intercomparison Project (CMIP) are crucial to evaluate past and future climate change. The reliability of models' simulations is often gauged by their ability to reproduce the historical climate across many time scales. This study compares the global mean surface air temperature from 29 CMIP6 models with observations from three datasets. We examine (1) warming and cooling rates in five subperiods from 1880 to 2014, (2) autocorrelation and long-term persistence, (3) models' performance based on probabilistic and entropy metrics, and (4) the distributional shape of temperature. All models simulate the observed long-term warming trend from 1880 to 2014. The late twentieth century warming (1975–2014) and the hiatus (1942–1975) are replicated by most models. The post-1998 warming is overestimated in 90% of the simulations. Only six out of 29 models reproduce the observed long-term persistence. All models show differences in distributional shape when compared with observations. Varying performance across metrics reveals the challenge to determine the "best" model. Thus, we argue that models should be selected, based on case-specific metrics, depending on the intended use. Metrics proposed here facilitate a comprehensive assessment for various applications.
Floods often affect large regions and cause adverse societal impacts. Regional flood hazard and risk assessments therefore require a realistic representation of spatial flood dependencies to avoid the overestimation or underestimation of risk. However, it is not yet well understood how spatial flood dependence, that is, the degree of co-occurrence of floods at different locations, varies in space and time and which processes influence the strength of this dependence. We identify regions in the United States with seasonally similar flood behavior and analyze processes governing spatial dependence. We find that spatial flood dependence varies regionally and seasonally and is generally strongest in winter and spring and weakest in summer and fall. Moreover, we find that land-surface processes are crucial in shaping the spatiotemporal characteristics of flood events. We conclude that the regional and seasonal variations in spatial flood dependencies must be considered when conducting current and future flood risk assessments.
It is challenging to develop observationally based spatial estimates of meteorology in Alaska and the Yukon. Complex topography, frozen precipitation undercatch, and extremely sparse in situ observations all limit our capability to produce accurate spatial estimates of meteorological conditions. In this Arctic environment, it is necessary to develop probabilistic estimates of precipitation and temperature that explicitly incorporate spatiotemporally varying uncertainty and bias corrections. In this paper we exploit the recently developed ensemble Climatologically Aided Interpolation (eCAI) system to produce daily historical estimates of precipitation and temperature across Alaska and the Yukon Territory at a 2 km grid spacing for the time period 1980–2013. We extend the previous eCAI method to address precipitation gauge undercatch and wetting loss, which is of high importance for this high-latitude region where much of the precipitation falls as snow. Leave-one-out cross-validation shows our ensemble has little bias in daily precipitation and mean temperature at the station locations, with an overestimate in the daily standard deviation of precipitation. The ensemble is statistically reliable compared to climatology and can discriminate precipitation events across different precipitation thresholds. Long-term mean loss adjusted precipitation is up to 36% greater than the unadjusted estimate in windy areas that receive a large fraction of frozen precipitation, primarily due to wind induced undercatch. Comparing the ensemble mean climatology of precipitation and temperature to PRISM and Daymet v3 shows large interproduct differences, particularly in precipitation across the complex terrain of southeast and northern Alaska.
It is challenging to develop observationally based spatial estimates of meteorology in Alaska and the Yukon. Complex topography, frozen precipitation undercatch, and extremely sparse in situ observations all limit our capability to produce accurate spatial estimates of meteorological conditions. In this Arctic environment, it is necessary to develop probabilistic estimates of precipitation and temperature that explicitly incorporate spatiotemporally varying uncertainty and bias corrections. In this paper we exploit the recently developed ensemble Climatologically Aided Interpolation (eCAI) system to produce daily historical estimates of precipitation and temperature across Alaska and the Yukon Territory at a 2 km grid spacing for the time period 1980–2013. We extend the previous eCAI method to address precipitation gauge undercatch and wetting loss, which is of high importance for this high-latitude region where much of the precipitation falls as snow. Leave-one-out cross-validation shows our ensemble has little bias in daily precipitation and mean temperature at the station locations, with an overestimate in the daily standard deviation of precipitation. The ensemble is statistically reliable compared to climatology and can discriminate precipitation events across different precipitation thresholds. Long-term mean loss adjusted precipitation is up to 36% greater than the unadjusted estimate in windy areas that receive a large fraction of frozen precipitation, primarily due to wind induced undercatch. Comparing the ensemble mean climatology of precipitation and temperature to PRISM and Daymet v3 shows large interproduct differences, particularly in precipitation across the complex terrain of southeast and northern Alaska.
Widespread flooding can cause major damages and substantial recovery costs. Still, estimates of how susceptible a region is to widespread flooding are largely missing mainly because of the sparseness of widespread flood events in records. The aim of this study is to assess the seasonal susceptibility of regions in the United States to widespread flooding using a stochastic streamflow generator, which enables simulating a large number of spatially consistent flood events. Furthermore, we ask which factors influence the strength of regional flood susceptibilities. We show that susceptibilities to widespread flooding vary regionally and seasonally. They are highest in regions where catchments show regimes with a strong seasonality, that is, the Pacific Northwest, the Rocky Mountains, and the Northeast. In contrast, they are low in regions where catchments are characterized by a weak seasonality and intermittent regimes such as the Great Plains. Furthermore, susceptibility is found to be the highest in winter and spring when spatial flood dependencies are strongest because of snowmelt contributions and high soil moisture availability. We conclude that regional flood susceptibilities emerge in river basins with catchments sharing similar streamflow and climatic regimes.
Abstract Global gridded precipitation products have proven essential for many applications ranging from hydrological modeling and climate model validation to natural hazard risk assessment. They provide a global picture of how precipitation varies across time and space, specifically in regions where ground-based observations are scarce. While the application of global precipitation products has become widespread, there is limited knowledge on how well these products represent the magnitude and frequency of extreme precipitation—the key features in triggering flood hazards. Here, five global precipitation datasets (MSWEP, CFSR, CPC, PERSIANN-CDR, and WFDEI) are compared to each other and to surface observations. The spatial variability of relatively high precipitation events (tail heaviness) and the resulting discrepancy among datasets in the predicted precipitation return levels were evaluated for the time period 1979–2017. The analysis shows that 1) these products do not provide a consistent representation of the behavior of extremes as quantified by the tail heaviness, 2) there is strong spatial variability in the tail index, 3) the spatial patterns of the tail heaviness generally match the Köppen–Geiger climate classification, and 4) the predicted return levels for 100 and 1000 years differ significantly among the gridded products. More generally, our findings reveal shortcomings of global precipitation products in representing extremes and highlight that there is no single global product that performs best for all regions and climates.
Abstract. This paper introduces the Topographically InformEd Regression (TIER) model, which uses terrain attributes in a regression framework to distribute in situ observations of precipitation and temperature to a grid. The framework enables our understanding of complex atmospheric processes (e.g., orographic precipitation) to be encoded into a statistical model in an easy-to-understand manner. TIER is developed in a modular fashion with key model parameters exposed to the user. This enables the user community to easily explore the impacts of our methodological choices made to distribute sparse, irregularly spaced observations to a grid in a systematic fashion. The modular design allows incorporating new capabilities in TIER. Intermediate processing variables are also output to provide a more complete understanding of the algorithm and any algorithmic changes. The framework also provides uncertainty estimates. This paper presents a brief model evaluation and demonstrates that the TIER algorithm is functioning as expected. Several variations in model parameters and changes in the distributed variables are described. A key conclusion is that seemingly small changes in a model parameter result in large changes to the final distributed fields and their associated uncertainty estimates.
2019
Quantifying the behavior and performance of hydrologic models is an important aspect of understanding the underlying hydrologic systems. We argue that classical error measures do not offer a complete picture for building this understanding. This study demonstrates how the information theoretic measure known as transfer entropy can be used to quantify the active transfer of information between hydrologic processes at various timescales and facilitate further understanding of the behavior of these systems. To build a better understanding of the differences in dynamics, we compare model instances of the Structure for Unifying Multiple Modeling Alternatives (SUMMA), the Variable Infiltration Capacity (VIC) model, and the Precipitation Runoff Modeling System (PRMS) across a variety of hydrologic regimes in the Columbia River Basin in the Pacific Northwest of North America. Our results show differences in the runoff of the SUMMA instance compared to the other two models in several of our study locations. In the Snake River region, SUMMA runoff was primarily snowmelt driven, while VIC and PRMS runoff was primarily influenced by precipitation and evapotranspiration. In the Olympic mountains, evapotranspiration interacted with the other water balance variables much differently in PRMS than in VIC and SUMMA. In the Willamette River, all three models had similar process networks at the daily time scale but showed differences in information transfer at the monthly timescale. Additionally, we find that all three models have similar connectivity between evapotranspiration and soil moisture. Analyzing information transfers to runoff at daily and monthly time steps shows how processes can operate on different timescales. By comparing information transfer with correlations, we show how transfer entropy provides a complementary picture of model behavior.
DOI
bib
abs
Twenty-three unsolved problems in hydrology (UPH) – a community perspective
Günter Blöschl,
M. F. Bierkens,
António Chambel,
Christophe Cudennec,
Georgia Destouni,
Aldo Fiori,
J. W. Kirchner,
Jeffrey J. McDonnell,
H. H. G. Savenije,
Murugesu Sivapalan,
Christine Stumpp,
Elena Toth,
Elena Volpi,
Gemma Carr,
Claire Lupton,
José Luis Salinas,
Borbála Széles,
Alberto Viglione,
Hafzullah Aksoy,
Scott T. Allen,
Anam Amin,
Vazken Andréassian,
Berit Arheimer,
Santosh Aryal,
Victor R. Baker,
Earl Bardsley,
Marlies Barendrecht,
Alena Bartošová,
Okke Batelaan,
Wouter Berghuijs,
Keith Beven,
Theresa Blume,
Thom Bogaard,
Pablo Borges de Amorim,
Michael E. Böttcher,
Gilles Boulet,
Korbinian Breinl,
Mitja Brilly,
Luca Brocca,
Wouter Buytaert,
Attilio Castellarin,
Andrea Castelletti,
Xiaohong Chen,
Yangbo Chen,
Yuanfang Chen,
Peter Chifflard,
Pierluigi Claps,
Martyn P. Clark,
Adrian L. Collins,
Barry Croke,
Annette Dathe,
Paula Cunha David,
Felipe P. J. de Barros,
Gerrit de Rooij,
Giuliano Di Baldassarre,
Jessica M. Driscoll,
Doris Duethmann,
Ravindra Dwivedi,
Ebru Eriş,
William Farmer,
James Feiccabrino,
Grant Ferguson,
Ennio Ferrari,
Stefano Ferraris,
Benjamin Fersch,
David C. Finger,
Laura Foglia,
Keirnan Fowler,
Б. И. Гарцман,
Simon Gascoin,
Éric Gaumé,
Alexander Gelfan,
Josie Geris,
Shervan Gharari,
Tom Gleeson,
Miriam Glendell,
Alena Gonzalez Bevacqua,
M. P. González‐Dugo,
Salvatore Grimaldi,
A.B. Gupta,
Björn Guse,
Dawei Han,
David M. Hannah,
A. A. Harpold,
Stefan Haun,
Kate Heal,
Kay Helfricht,
Mathew Herrnegger,
Matthew R. Hipsey,
Hana Hlaváčiková,
Clara Hohmann,
Ladislav Holko,
C. Hopkinson,
Markus Hrachowitz,
Tissa H. Illangasekare,
Azhar Inam,
Camyla Innocente,
Erkan Istanbulluoglu,
Ben Jarihani,
Zahra Kalantari,
Andis Kalvāns,
Sonu Khanal,
Sina Khatami,
Jens Kiesel,
M. J. Kirkby,
Wouter Knoben,
Krzysztof Kochanek,
Silvia Kohnová,
Alla Kolechkina,
Stefan Krause,
David K. Kreamer,
Heidi Kreibich,
Harald Kunstmann,
Holger Lange,
Margarida L. R. Liberato,
Eric Lindquist,
Timothy E. Link,
Junguo Liu,
Daniel P. Loucks,
Charles H. Luce,
Gil Mahé,
Olga Makarieva,
Julien Malard,
Shamshagul Mashtayeva,
Shreedhar Maskey,
Josep Mas-Plá,
Maria Mavrova-Guirguinova,
Maurizio Mazzoleni,
Sebastian H. Mernild,
Bruce Misstear,
Alberto Montanari,
Hannes Müller-Thomy,
Alireza Nabizadeh,
Fernando Nardi,
Christopher M. U. Neale,
Nataliia Nesterova,
Bakhram Nurtaev,
V.O. Odongo,
Subhabrata Panda,
Saket Pande,
Zhonghe Pang,
Georgia Papacharalampous,
Charles Perrin,
Laurent Pfister,
Rafael Pimentel,
María José Polo,
David Post,
Cristina Prieto,
Maria‐Helena Ramos,
Maik Renner,
José Eduardo Reynolds,
Elena Ridolfi,
Riccardo Rigon,
Mònica Riva,
David E. Robertson,
Renzo Rosso,
Tirthankar Roy,
João Henrique Macedo Sá,
Gianfausto Salvadori,
Melody Sandells,
Bettina Schaefli,
Andreas Schumann,
Anna Scolobig,
Jan Seibert,
Éric Servat,
Mojtaba Shafiei,
Ashish Sharma,
Moussa Sidibé,
Roy C. Sidle,
Thomas Skaugen,
Hugh G. Smith,
Sabine M. Spiessl,
Lina Stein,
Ingelin Steinsland,
Ulrich Strasser,
Bob Su,
Ján Szolgay,
David G. Tarboton,
Flavia Tauro,
Guillaume Thirel,
Fuqiang Tian,
Rui Tong,
Kamshat Tussupova,
Hristos Tyralis,
R. Uijlenhoet,
Rens van Beek,
Ruud J. van der Ent,
Martine van der Ploeg,
Anne F. Van Loon,
Ilja van Meerveld,
Ronald van Nooijen,
Pieter van Oel,
Jean‐Philippe Vidal,
Jana von Freyberg,
Sergiy Vorogushyn,
Przemysław Wachniew,
Andrew J. Wade,
Philip J. Ward,
Ida Westerberg,
Christopher White,
Eric F. Wood,
Ross Woods,
Zongxue Xu,
Koray K. Yılmaz,
Yongqiang Zhang
Hydrological Sciences Journal, Volume 64, Issue 10
This paper is the outcome of a community initiative to identify major unsolved scientific problems in hydrology motivated by a need for stronger harmonisation of research efforts. The procedure involved a public consultation through online media, followed by two workshops through which a large number of potential science questions were collated, prioritised, and synthesised. In spite of the diversity of the participants (230 scientists in total), the process revealed much about community priorities and the state of our science: a preference for continuity in research questions rather than radical departures or redirections from past and current work. Questions remain focused on the process-based understanding of hydrological variability and causality at all space and time scales. Increased attention to environmental change drives a new emphasis on understanding how change propagates across interfaces within the hydrological system and across disciplinary boundaries. In particular, the expansion of the human footprint raises a new set of questions related to human interactions with nature and water cycle feedbacks in the context of complex water management problems. We hope that this reflection and synthesis of the 23 unsolved problems in hydrology will help guide research efforts for some years to come.
DOI
bib
abs
How Do Modeling Decisions Affect the Spread Among Hydrologic Climate Change Projections? Exploring a Large Ensemble of Simulations Across a Diversity of Hydroclimates
O. Chegwidden,
Bart Nijssen,
David E. Rupp,
Jeffrey R. Arnold,
Martyn P. Clark,
Joseph Hamman,
Shih‐Chieh Kao,
Yixin Mao,
Naoki Mizukami,
Philip W. Mote,
Ming Pan,
Erik Pytlak,
Mu Xiao
Earth's Future, Volume 7, Issue 6
Methodological choices can have strong effects on projections of climate change impacts on hydrology. In this study, we investigate the ways in which four different steps in the modeling chain influence the spread in projected changes of different aspects of hydrology. To form the basis of these analyses, we constructed an ensemble of 160 simulations from permutations of two Representative Concentration Pathways, 10 global climate models, two downscaling methods, and four hydrologic model implementations. The study is situated in the Pacific Northwest of North America, which has relevance to a diverse, multinational cast of stakeholders. We analyze the effects of each modeling decision on changes in gridded hydrologic variables of snow water equivalent and runoff, as well as streamflow at point locations. Results show that the choice of representative concentration pathway or global climate model is the driving contributor to the spread in annual streamflow volume and timing. On the other hand, hydrologic model implementation explains most of the spread in changes in low flows. Finally, by grouping the results by climate region the results have the potential to be generalized beyond the Pacific Northwest. Future hydrologic impact assessments can use these results to better tailor their modeling efforts.
DOI
bib
abs
Hillslope Hydrology in Global Change Research and Earth System Modeling
Ying Fan,
Martyn P. Clark,
David M. Lawrence,
Sean Swenson,
Lawrence E. Band,
Susan L. Brantley,
P. D. Brooks,
W. E. Dietrich,
Alejandro N. Flores,
Gordon E. Grant,
James W. Kirchner,
D. S. Mackay,
Jeffrey J. McDonnell,
P. C. D. Milly,
Pamela L. Sullivan,
Christina Tague,
Hoori Ajami,
Nathaniel W. Chaney,
Andreas Hartmann,
P. Hazenberg,
J. P. McNamara,
Jon D. Pelletier,
J. Perket,
Elham Rouholahnejad Freund,
Thorsten Wagener,
Xubin Zeng,
R. Edward Beighley,
J. R. Buzan,
Maoyi Huang,
Ben Livneh,
Binayak P. Mohanty,
Bart Nijssen,
Mohammad Safeeq,
Chaopeng Shen,
Willem van Verseveld,
John Volk,
Dai Yamazaki
Water Resources Research, Volume 55, Issue 2
Earth System Models (ESMs) are essential tools for understanding and predicting global change, but they cannot explicitly resolve hillslope‐scale terrain structures that fundamentally organize water, energy, and biogeochemical stores and fluxes at subgrid scales. Here we bring together hydrologists, Critical Zone scientists, and ESM developers, to explore how hillslope structures may modulate ESM grid‐level water, energy, and biogeochemical fluxes. In contrast to the one‐dimensional (1‐D), 2‐ to 3‐m deep, and free‐draining soil hydrology in most ESM land models, we hypothesize that 3‐D, lateral ridge‐to‐valley flow through shallow and deep paths and insolation contrasts between sunny and shady slopes are the top two globally quantifiable organizers of water and energy (and vegetation) within an ESM grid cell. We hypothesize that these two processes are likely to impact ESM predictions where (and when) water and/or energy are limiting. We further hypothesize that, if implemented in ESM land models, these processes will increase simulated continental water storage and residence time, buffering terrestrial ecosystems against seasonal and interannual droughts. We explore efficient ways to capture these mechanisms in ESMs and identify critical knowledge gaps preventing us from scaling up hillslope to global processes. One such gap is our extremely limited knowledge of the subsurface, where water is stored (supporting vegetation) and released to stream baseflow (supporting aquatic ecosystems). We conclude with a set of organizing hypotheses and a call for global syntheses activities and model experiments to assess the impact of hillslope hydrology on global change predictions.
Abstract Land models are increasingly used and preferred in terrestrial hydrological prediction applications. One reason for selecting land models over simpler models is that their physically based backbone enables wider application under different conditions. This study evaluates the temporal variability in streamflow simulations in land models. Specifically, we evaluate how the subsurface structure and model parameters control the partitioning of water into different flow paths and the temporal variability in streamflow. Moreover, we use a suite of model diagnostics, typically not used in the land modeling community to clarify model weaknesses and identify a path toward model improvement. Our analyses show that the typical land model structure, and their functions for moisture movement between soil layers (an approximation of Richards equation), has a distinctive signature where flashy runoff is superimposed on slow recessions. This hampers the application of land models in simulating flashier basins and headwater catchments where floods are generated. We demonstrate the added value of the preferential flow in the model simulation by including macropores in both a toy model and the Variable Infiltration Capacity model. We argue that including preferential flow in land models is essential to enable their use for multiple applications across a myriad of temporal and spatial scales.
DOI
bib
abs
The Community Land Model Version 5: Description of New Features, Benchmarking, and Impact of Forcing Uncertainty
David M. Lawrence,
Rosie A. Fisher,
Charles D. Koven,
Keith W. Oleson,
Sean Swenson,
G. B. Bonan,
Nathan Collier,
Bardan Ghimire,
Leo van Kampenhout,
Daniel Kennedy,
Erik Kluzek,
Fang Li,
Hongyi Li,
Danica Lombardozzi,
William J. Riley,
William J. Sacks,
Mingjie Shi,
Mariana Vertenstein,
William R. Wieder,
Chonggang Xu,
Ashehad A. Ali,
Andrew M. Badger,
Gautam Bisht,
Michiel van den Broeke,
Michael A. Brunke,
Sean P. Burns,
J. R. Buzan,
Martyn P. Clark,
Anthony P Craig,
Kyla M. Dahlin,
Beth Drewniak,
Joshua B. Fisher,
M. Flanner,
A. M. Fox,
Pierre Gentine,
Forrest M. Hoffman,
G. Keppel‐Aleks,
R. G. Knox,
Sanjiv Kumar,
Jan T. M. Lenaerts,
L. Ruby Leung,
William H. Lipscomb,
Yaqiong Lü,
Ashutosh Pandey,
Jon D. Pelletier,
J. Perket,
James T. Randerson,
Daniel M. Ricciuto,
Benjamin M. Sanderson,
A. G. Slater,
Z. M. Subin,
Jinyun Tang,
R. Quinn Thomas,
Maria Val Martin,
Xubin Zeng
Journal of Advances in Modeling Earth Systems, Volume 11, Issue 12
The Community Land Model (CLM) is the land component of the Community Earth System Model (CESM) and is used in several global and regional modeling systems. In this paper, we introduce model developments included in CLM version 5 (CLM5), which is the default land component for CESM2. We assess an ensemble of simulations, including prescribed and prognostic vegetation state, multiple forcing data sets, and CLM4, CLM4.5, and CLM5, against a range of metrics including from the International Land Model Benchmarking (ILAMBv2) package. CLM5 includes new and updated processes and parameterizations: (1) dynamic land units, (2) updated parameterizations and structure for hydrology and snow (spatially explicit soil depth, dry surface layer, revised groundwater scheme, revised canopy interception and canopy snow processes, updated fresh snow density, simple firn model, and Model for Scale Adaptive River Transport), (3) plant hydraulics and hydraulic redistribution, (4) revised nitrogen cycling (flexible leaf stoichiometry, leaf N optimization for photosynthesis, and carbon costs for plant nitrogen uptake), (5) global crop model with six crop types and time‐evolving irrigated areas and fertilization rates, (6) updated urban building energy, (7) carbon isotopes, and (8) updated stomatal physiology. New optional features include demographically structured dynamic vegetation model (Functionally Assembled Terrestrial Ecosystem Simulator), ozone damage to plants, and fire trace gas emissions coupling to the atmosphere. Conclusive establishment of improvement or degradation of individual variables or metrics is challenged by forcing uncertainty, parametric uncertainty, and model structural complexity, but the multivariate metrics presented here suggest a general broad improvement from CLM4 to CLM5.
The concept of using representative hillslopes to simulate hydrologically similar areas of a catchment has been incorporated in many hydrologic models but few Earth system models. Here we describe a configuration of the Community Land Model version 5 in which each grid cell is decomposed into one or more multicolumn hillslopes. Within each hillslope, the intercolumn connectivity is specified, and the lateral saturated subsurface flow from each column is passed to its downslope neighbor. We first apply the model to simulate a headwater catchment and assess the results against runoff and evapotranspiration flux measurements. By redistributing soil water within the catchment, the model is able to reproduce the observed difference between evapotranspiration in the upland and lowland portions of the catchment. Next, global simulations based on hypothetical hillslope geomorphic parameters are used to show the model's sensitivity to differences in hillslope shape and discretization. Differences in evapotranspiration between upland and lowland hillslope columns are found to be largest in arid and semiarid regions, while humid tropical and high‐latitude regions show limited evapotranspiration increases in lowlands relative to uplands.
Increasingly, climate change impact assessments rely directly on climate models. Assessments of future water security depend in part on how the land model components in climate models partition precipitation into evapotranspiration and runoff, and on the sensitivity of this partitioning to climate. Runoff sensitivities are not well constrained, with CMIP5 models displaying a large spread for the present day, which projects onto change under warming, creating uncertainty. Here we show that constraining CMIP5 model runoff sensitivities with observed estimates could reduce uncertainty in runoff projection over the western United States by up to 50%. We urge caution in the direct use of climate model runoff for applications and encourage model development to use regional-scale hydrological sensitivity metrics to improve projections for water security assessments.
It is generally acknowledged in the environmental sciences that the choice of a computational model impacts the research results. In this study of a flood and drought event in the Swiss Thur basin, we show that modeling decisions during the model configuration, beyond the model choice, also impact the model results. In our carefully designed experiment we investigated four modeling decisions in ten nested basins: the spatial resolution of the model, the spatial representation of the forcing data, the calibration period, and the performance metric. The flood characteristics were mainly affected by the performance metric, whereas the drought characteristics were mainly affected by the calibration period. The results could be related to the processes that triggered the particular events studied. The impact of the modeling decisions on the simulations did, however, vary among the investigated sub-basins. In spite of the limitations of this study, our findings have important implications for the understanding and quantification of uncertainty in any hydrological or even environmental model. Modeling decisions during model configuration introduce subjectivity from the modeler. Multiple working hypotheses during model configuration can provide insights on the impact of such subjective modeling decisions.
Abstract. Calibration is an essential step for improving the accuracy of simulations generated using hydrologic models. A key modeling decision is selecting the performance metric to be optimized. It has been common to use squared error performance metrics, or normalized variants such as Nash–Sutcliffe efficiency (NSE), based on the idea that their squared-error nature will emphasize the estimates of high flows. However, we conclude that NSE-based model calibrations actually result in poor reproduction of high-flow events, such as the annual peak flows that are used for flood frequency estimation. Using three different types of performance metrics, we calibrate two hydrological models at a daily step, the Variable Infiltration Capacity (VIC) model and the mesoscale Hydrologic Model (mHM), and evaluate their ability to simulate high-flow events for 492 basins throughout the contiguous United States. The metrics investigated are (1) NSE, (2) Kling–Gupta efficiency (KGE) and its variants, and (3) annual peak flow bias (APFB), where the latter is an application-specific metric that focuses on annual peak flows. As expected, the APFB metric produces the best annual peak flow estimates; however, performance on other high-flow-related metrics is poor. In contrast, the use of NSE results in annual peak flow estimates that are more than 20 % worse, primarily due to the tendency of NSE to underestimate observed flow variability. On the other hand, the use of KGE results in annual peak flow estimates that are better than from NSE, owing to improved flow time series metrics (mean and variance), with only a slight degradation in performance with respect to other related metrics, particularly when a non-standard weighting of the components of KGE is used. Stochastically generated ensemble simulations based on model residuals show the ability to improve the high-flow metrics, regardless of the deterministic performances. However, we emphasize that improving the fidelity of streamflow dynamics from deterministically calibrated models is still important, as it may improve high-flow metrics (for the right reasons). Overall, this work highlights the need for a deeper understanding of performance metric behavior and design in relation to the desired goals of model calibration.
Abstract This study presents a gridded meteorology intercomparison using the State of Hawaii as a testbed. This is motivated by the goal to provide the broad user community with knowledge of interproduct differences and the reasons differences exist. More generally, the challenge of generating station-based gridded meteorological surfaces and the difficulties in attributing interproduct differences to specific methodological decisions are demonstrated. Hawaii is a useful testbed because it is traditionally underserved, yet meteorologically interesting and complex. In addition, several climatological and daily gridded meteorology datasets are now available, which are used extensively by the applications modeling community, thus an intercomparison enhances Hawaiian specific capabilities. We compare PRISM climatology and three daily datasets: new datasets from the University of Hawai‘i and the National Center for Atmospheric Research, and Daymet version 3 for precipitation and temperature variables only. General conclusions that have emerged are 1) differences in input station data significantly influence the product differences, 2) explicit prediction of precipitation occurrence is crucial across multiple metrics, and 3) attribution of differences to specific methodological choices is difficult and limits the usefulness of intercomparisons. Because generating gridded meteorological fields is an elaborate process with many methodological choices interacting in complex ways, future work should 1) develop modular frameworks that allows users to easily examine the breadth of methodological choices, 2) collate available nontraditional high-quality observational datasets for true out-of-sample validation and make them publicly available, and 3) define benchmarks of acceptable performance for methodological components and products.
This study presents diagnostic evaluation of two large‐domain hydrologic models: the mesoscale Hydrologic Model (mHM) and the Variable Infiltration Capacity (VIC) over the contiguous United States (CONUS). These models have been calibrated using the Multiscale Parameter Regionalization scheme in a joint, multibasin approach using 492 medium‐sized basins across the CONUS yielding spatially distributed model parameter sets. The mHM simulations are used as a performance benchmark to examine performance deficiencies in the VIC model. We find that after calibration to streamflow, VIC generally overestimates the magnitude and temporal variability of evapotranspiration (ET) as compared to mHM as well as the FLUXNET observation‐based ET product, resulting in underestimation of the mean and variability of runoff. We perform a controlled calibration experiment to investigate the effect of varying number of transfer function parameters in mHM and to enable a fair comparison between both models (14 and 48 for mHM vs. 14 for VIC). Results of this experiment show similar behavior of mHM with 14 and 48 parameters. Furthermore, we diagnose the internal functioning of the VIC model by looking at the relationship of the evaporative fraction versus the degree of soil saturation and compare it with that of the mHM model, which has a different model structure, a prescribed nonlinear relationship between these variables and exhibits better model skill than VIC. Despite these limitations, the VIC‐based CONUS‐wide calibration constrained against streamflow exhibits better ET skill as compared to two preexisting independent VIC studies.
2018
Abstract Water managers are actively incorporating climate change information into their long- and short-term planning processes. This is generally seen as a step in the right direction because it supplements traditional methods, providing new insights that can help in planning for a non-stationary climate. However, the continuous evolution of climate change information can make it challenging to use available information appropriately. Advice on how to use the information is not always straightforward and typically requires extended dialogue between information producers and users, which is not always feasible. To help navigate better the ever-changing climate science landscape, this review is organized as a set of nine guidelines for water managers and planners that highlight better practices for incorporating climate change information into water resource planning and management. Each DOs and DON'Ts recommendation is given with context on why certain strategies are preferable and addresses frequently asked questions by exploring past studies and documents that provide guidance, including real-world examples mainly, though not exclusively, from the United States. This paper is intended to provide a foundation that can expand through continued dialogue within and between the climate science and application communities worldwide, a two-way information sharing that can increase the actionable nature of the information produced and promote greater utility and appropriate use.
DOI
bib
abs
ESM-SnowMIP: assessing snow models and quantifying snow-related climate feedbacks
Gerhard Krinner,
Chris Derksen,
Richard Essery,
M. Flanner,
Stefan Hagemann,
Martyn P. Clark,
Alex Hall,
Helmut Rott,
Claire Brutel‐Vuilmet,
Hyungjun Kim,
Cécile B. Ménard,
Lawrence Mudryk,
Chad W. Thackeray,
Libo Wang,
Gabriele Arduini,
Gianpaolo Balsamo,
Paul Bartlett,
Julia Boike,
Aaron Boone,
F. Chéruy,
Jeanne Colin,
Matthias Cuntz,
Yongjiu Dai,
Bertrand Decharme,
Jeff Derry,
Agnès Ducharne,
Emanuel Dutra,
Xing Fang,
Charles Fierz,
Josephine Ghattas,
Yeugeniy M. Gusev,
Vanessa Haverd,
Anna Kontu,
Matthieu Lafaysse,
R. M. Law,
David M. Lawrence,
Weiping Li,
Thomas Marke,
Danny Marks,
Martin Ménégoz,
О. Н. Насонова,
Tomoko Nitta,
Michio Niwano,
John W. Pomeroy,
M. S. Raleigh,
Gerd Schaedler,
В. А. Семенов,
Tanya Smirnova,
Tobias Stacke,
Ulrich Strasser,
Sean Svenson,
Dmitry Turkov,
Tao Wang,
Nander Wever,
Hua Yuan,
Wenyan Zhou,
Dan Zhu
Geoscientific Model Development, Volume 11, Issue 12
Abstract. This paper describes ESM-SnowMIP, an international coordinated modelling effort to evaluate current snow schemes, including snow schemes that are included in Earth system models, in a wide variety of settings against local and global observations. The project aims to identify crucial processes and characteristics that need to be improved in snow models in the context of local- and global-scale modelling. A further objective of ESM-SnowMIP is to better quantify snow-related feedbacks in the Earth system. Although it is not part of the sixth phase of the Coupled Model Intercomparison Project (CMIP6), ESM-SnowMIP is tightly linked to the CMIP6-endorsed Land Surface, Snow and Soil Moisture Model Intercomparison (LS3MIP).