scholarly journals Multiple domain evaluation of watershed hydrology models

Author(s):  
Karthik Kumarasamy ◽  
Patrick Belmont

Abstract. Watershed scale models simulating hydrology and water quality have advanced rapidly in sophistication, process representation, flexibility in model structure, and input data. Given the importance of these models to support decision-making for a wide range of environmental issues, the hydrology community is compelled to improve the metrics used to evaluate model performance. More targeted and comprehensive metrics will facilitate better and more efficient calibration and will help demonstrate that the model is useful for the intended purpose. Here we introduce a suite of new tools for model evaluation, packaged as an open-source Hydrologic Model Evaluation (HydroME) Toolbox. Specifically, we demonstrate the use of box plots to illustrate the full distribution of common model performance metrics, such as R2, use of Euclidian distance, empirical Quantile-Quantile (Q-Q) plots and flow duration curves as simple metrics to identify and localize errors in model simulations. Further, we demonstrate the use of magnitude squared coherence to compare the frequency content between observed and modeled streamflow and wavelet coherence to localize frequency mismatches in time. We provide a rationale for a hierarchical selection of parameters to adjust during calibration and recommend that modelers progress from parameters with the most uncertainty to the least uncertainty, namely starting with pure calibration parameters, followed by derived parameters, and finally measured parameters. We apply these techniques in the calibration and evaluation of models of two watersheds, the Le Sueur River Basin (2880 km2) and Root River Basin (4300 km2) in southern Minnesota, USA.

2005 ◽  
Vol 2 (3) ◽  
pp. 639-690 ◽  
Author(s):  
G. P. Zhang ◽  
H. H. G. Savenije

Abstract. Based on the Representative Elementary Watershed (REW) approach, the modelling tool REWASH (Representative Elementary WAterShed Hydrology) has been developed and applied to the Geer river basin. REWASH is deterministic, semi-distributed, physically based and can be directly applied to the watershed scale. In applying REWASH, the river basin is divided into a number of sub-watersheds, so called REWs, according to the Strahler order of the river network. REWASH describes the dominant hydrological processes, i.e. subsurface flow in the unsaturated and saturated domains, and overland flow by the saturation-excess and infiltration-excess mechanisms. Through flux exchanges among the different spatial domains of the REW, surface and subsurface water interactions are fully coupled. REWASH is a parsimonious tool for modelling watershed hydrological response. However, it can be modified to include more components to simulate specific processes when applied to a specific river basin where such processes are observed or considered to be dominant. In this study, we have added a new component to simulate interception using a simple parametric approach. Interception plays an important role in the water balance of a watershed although it is often disregarded. In addition, a refinement for the transpiration in the unsaturated zone has been made. Finally, an improved approach for simulating saturation overland flow by relating the variable source area to both the topography and the groundwater level is presented. The model has been calibrated and verified using a 4-year data set, which has been split into two for calibration and validation. The model performance has been assessed by multi-criteria evaluation. This work is the first full application of the REW approach to watershed rainfall-runoff modelling in a real watershed. The results demonstrate that the REW approach provides an alternative blueprint for physically based hydrological modelling.


2021 ◽  
Author(s):  
Sophia Eugeni ◽  
Eric Vaags ◽  
Steven V. Weijs

<p>Accurate hydrologic modelling is critical to effective water resource management. As catchment attributes strongly influence the hydrologic behaviors in an area, they can be used to inform hydrologic models to better predict the discharge in a basin. Some basins may be more difficult to accurately predict than others. The difficulty in predicting discharge may also be related to the complexity of the discharge signal. The study establishes the relationship between a catchment’s static attributes and hydrologic model performance in those catchments, and also investigates the link to complexity, which we quantify with measures of compressibility based in information theory. </p><p>The project analyzes a large national dataset, comprised of catchment attributes for basins across the United States, paired with established performance metrics for corresponding hydrologic models. Principal Component Analysis (PCA) was completed on the catchment attributes data to determine the strongest modes in the input. The basins were clustered according to their catchment attributes and the performance within the clusters was compared. </p><p>Significant differences in model performance emerged between the clusters of basins. For the complexity analysis, details of the implementation and technical challenges will be discussed, as well as preliminary results.</p>


2015 ◽  
Vol 19 (1) ◽  
pp. 209-223 ◽  
Author(s):  
A. J. Newman ◽  
M. P. Clark ◽  
K. Sampson ◽  
A. Wood ◽  
L. E. Hay ◽  
...  

Abstract. We present a community data set of daily forcing and hydrologic response data for 671 small- to medium-sized basins across the contiguous United States (median basin size of 336 km2) that spans a very wide range of hydroclimatic conditions. Area-averaged forcing data for the period 1980–2010 was generated for three basin spatial configurations – basin mean, hydrologic response units (HRUs) and elevation bands – by mapping daily, gridded meteorological data sets to the subbasin (Daymet) and basin polygons (Daymet, Maurer and NLDAS). Daily streamflow data was compiled from the United States Geological Survey National Water Information System. The focus of this paper is to (1) present the data set for community use and (2) provide a model performance benchmark using the coupled Snow-17 snow model and the Sacramento Soil Moisture Accounting Model, calibrated using the shuffled complex evolution global optimization routine. After optimization minimizing daily root mean squared error, 90% of the basins have Nash–Sutcliffe efficiency scores ≥0.55 for the calibration period and 34% ≥ 0.8. This benchmark provides a reference level of hydrologic model performance for a commonly used model and calibration system, and highlights some regional variations in model performance. For example, basins with a more pronounced seasonal cycle generally have a negative low flow bias, while basins with a smaller seasonal cycle have a positive low flow bias. Finally, we find that data points with extreme error (defined as individual days with a high fraction of total error) are more common in arid basins with limited snow and, for a given aridity, fewer extreme error days are present as the basin snow water equivalent increases.


2014 ◽  
Vol 11 (5) ◽  
pp. 5599-5631
Author(s):  
A. J. Newman ◽  
M. P. Clark ◽  
K. Sampson ◽  
A. Wood ◽  
L. E. Hay ◽  
...  

Abstract. We present a community dataset of daily forcing and hydrologic response data for 671 small- to medium-sized basins across the contiguous United States (median basin size of 336 km2) that spans a very wide range of hydroclimatic conditions. Areally averaged forcing data for the period 1980–2010 was generated for three basin delineations – basin mean, Hydrologic Response Units (HRUs) and elevation bands – by mapping the daily, 1 km gridded Daymet meteorological dataset to the sub-basin and basin polygons. Daily streamflow data was compiled from the United States Geological Survey National Water Information System. The focus of this paper is to (1) present the dataset for community use; and (2) provide a model performance benchmark using the coupled Snow-17 snow model and the Sacramento Soil Moisture Accounting conceptual hydrologic model, calibrated using the Shuffled Complex Evolution global optimization routine. After optimization minimizing daily root mean squared error, 90% of the basins have Nash–Sutcliffe Efficiency scores > 0.55 for the calibration period. This benchmark provides a reference level of hydrologic model performance for a commonly used model and calibration system, and highlights some regional variations in model performance. For example, basins with a more pronounced seasonal cycle generally have a negative low flow bias, while basins with a smaller seasonal cycle have a positive low flow bias. Finally, we find that data points with extreme error (defined as individual days with a high fraction of total error) are more common in arid basins with limited snow, and, for a given aridity, fewer extreme error days are present as basin snow water equivalent increases.


2016 ◽  
Vol 7 (4) ◽  
pp. 813-830 ◽  
Author(s):  
Veronika Eyring ◽  
Peter J. Gleckler ◽  
Christoph Heinze ◽  
Ronald J. Stouffer ◽  
Karl E. Taylor ◽  
...  

Abstract. The Coupled Model Intercomparison Project (CMIP) has successfully provided the climate community with a rich collection of simulation output from Earth system models (ESMs) that can be used to understand past climate changes and make projections and uncertainty estimates of the future. Confidence in ESMs can be gained because the models are based on physical principles and reproduce many important aspects of observed climate. More research is required to identify the processes that are most responsible for systematic biases and the magnitude and uncertainty of future projections so that more relevant performance tests can be developed. At the same time, there are many aspects of ESM evaluation that are well established and considered an essential part of systematic evaluation but have been implemented ad hoc with little community coordination. Given the diversity and complexity of ESM analysis, we argue that the CMIP community has reached a critical juncture at which many baseline aspects of model evaluation need to be performed much more efficiently and consistently. Here, we provide a perspective and viewpoint on how a more systematic, open, and rapid performance assessment of the large and diverse number of models that will participate in current and future phases of CMIP can be achieved, and announce our intention to implement such a system for CMIP6. Accomplishing this could also free up valuable resources as many scientists are frequently "re-inventing the wheel" by re-writing analysis routines for well-established analysis methods. A more systematic approach for the community would be to develop and apply evaluation tools that are based on the latest scientific knowledge and observational reference, are well suited for routine use, and provide a wide range of diagnostics and performance metrics that comprehensively characterize model behaviour as soon as the output is published to the Earth System Grid Federation (ESGF). The CMIP infrastructure enforces data standards and conventions for model output and documentation accessible via the ESGF, additionally publishing observations (obs4MIPs) and reanalyses (ana4MIPs) for model intercomparison projects using the same data structure and organization as the ESM output. This largely facilitates routine evaluation of the ESMs, but to be able to process the data automatically alongside the ESGF, the infrastructure needs to be extended with processing capabilities at the ESGF data nodes where the evaluation tools can be executed on a routine basis. Efforts are already underway to develop community-based evaluation tools, and we encourage experts to provide additional diagnostic codes that would enhance this capability for CMIP. At the same time, we encourage the community to contribute observations and reanalyses for model evaluation to the obs4MIPs and ana4MIPs archives. The intention is to produce through the ESGF a widely accepted quasi-operational evaluation framework for CMIP6 that would routinely execute a series of standardized evaluation tasks. Over time, as this capability matures, we expect to produce an increasingly systematic characterization of models which, compared with early phases of CMIP, will more quickly and openly identify the strengths and weaknesses of the simulations. This will also reveal whether long-standing model errors remain evident in newer models and will assist modelling groups in improving their models. This framework will be designed to readily incorporate updates, including new observations and additional diagnostics and metrics as they become available from the research community.


2016 ◽  
Author(s):  
Veronika Eyring ◽  
Peter J. Gleckler ◽  
Christoph Heinze ◽  
Ronald J. Stouffer ◽  
Karl E. Taylor ◽  
...  

Abstract. The Coupled Model Intercomparison Project (CMIP) has successfully provided the climate community with a rich collection of simulation output from Earth system models (ESMs) that can be used to understand past climate changes and make projections and uncertainty estimates of the future. Confidence in ESMs can be gained because the models are based on physical principles and reproduce many important aspects of observed climate. Scientifically more research is required to identify the processes that are most responsible for systematic biases and the magnitude and uncertainty of future projections so that more relevant performance tests can be developed. At the same time, there are many aspects of ESM evaluation that are well-established and considered an essential part of systematic evaluation but are currently implemented ad hoc with little community coordination. Given the diversity and complexity of ESM model analysis, we argue that the CMIP community has reached a critical juncture at which many baseline aspects of model evaluation need to be performed much more efficiently to enable a systematic, open and rapid performance assessment of the large and diverse number of models that will participate in current and future phases of CMIP. Accomplishing this could also free up valuable resources as many scientists are frequently "re-inventing the wheel" by re-writing analysis routines for well-established analysis methods. A more systematic approach for the community would be to develop evaluation tools that are well suited for routine use and provide a wide range of diagnostics and performance metrics that comprehensively characterize model behaviour as soon as the output is published to the Earth System Grid Federation (ESGF). The CMIP infrastructure enforces data standards and conventions for model output accessible via ESGF, additionally publishing observations (obs4MIPs) and reanalyses (ana4MIPs) for Model Intercomparison Projects using the same data structure and organization. This largely facilitates routine evaluation of the models, but to be able to process the data automatically alongside the ESGF, the infrastructure needs to be extended with processing capabilities at the ESGF data nodes where the evaluation tools can be executed on a routine basis. Efforts are already underway to develop community-based evaluation tools, and we encourage experts to provide additional diagnostic codes that would enhance this capability for CMIP. At the same time, we encourage the community to contribute observations for model evaluation to the obs4MIPs archive. The intention is to produce through ESGF a widely accepted quasi-operational evaluation framework for climate models that would routinely execute a series of standardized evaluation tasks. Over time, as the capability matures, we expect to produce an increasingly systematic characterization of models, which, compared with early phases of CMIP, will more quickly and openly identify the strengths and weaknesses of the simulations. This will also expose whether long-standing model errors remain evident in newer models and will assist modelling groups in improving their models. This framework will be designed to readily incorporate updates, including new observations and additional diagnostics and metrics as they become available from the research community.


2018 ◽  
Vol 11 (5) ◽  
pp. 1873-1886 ◽  
Author(s):  
Julian Koch ◽  
Mehmet Cüneyd Demirel ◽  
Simon Stisen

Abstract. The process of model evaluation is not only an integral part of model development and calibration but also of paramount importance when communicating modelling results to the scientific community and stakeholders. The modelling community has a large and well-tested toolbox of metrics to evaluate temporal model performance. In contrast, spatial performance evaluation does not correspond to the grand availability of spatial observations readily available and to the sophisticate model codes simulating the spatial variability of complex hydrological processes. This study makes a contribution towards advancing spatial-pattern-oriented model calibration by rigorously testing a multiple-component performance metric. The promoted SPAtial EFficiency (SPAEF) metric reflects three equally weighted components: correlation, coefficient of variation and histogram overlap. This multiple-component approach is found to be advantageous in order to achieve the complex task of comparing spatial patterns. SPAEF, its three components individually and two alternative spatial performance metrics, i.e. connectivity analysis and fractions skill score, are applied in a spatial-pattern-oriented model calibration of a catchment model in Denmark. Results suggest the importance of multiple-component metrics because stand-alone metrics tend to fail to provide holistic pattern information. The three SPAEF components are found to be independent, which allows them to complement each other in a meaningful way. In order to optimally exploit spatial observations made available by remote sensing platforms, this study suggests applying bias insensitive metrics which further allow for a comparison of variables which are related but may differ in unit. This study applies SPAEF in the hydrological context using the mesoscale Hydrologic Model (mHM; version 5.8), but we see great potential across disciplines related to spatially distributed earth system modelling.


2005 ◽  
Vol 9 (3) ◽  
pp. 243-261 ◽  
Author(s):  
G. P. Zhang ◽  
H. H. G. Savenije

Abstract. Based on the Representative Elementary Watershed (REW) approach, the modelling tool REWASH (Representative Elementary WAterShed Hydrology) has been developed and applied to the Geer river basin. REWASH is deterministic, semi-distributed, physically based and can be directly applied to the watershed scale. In applying REWASH, the river basin is divided into a number of sub-watersheds, so called REWs, according to the Strahler order of the river network. REWASH describes the dominant hydrological processes, i.e. subsurface flow in the unsaturated and saturated domains, and overland flow by the saturation-excess and infiltration-excess mechanisms. The coupling of surface and subsurface flow processes in the numerical model is realised by simultaneous computation of flux exchanges between surface and subsurface domains for each REW. REWASH is a parsimonious tool for modelling watershed hydrological response. However, it can be modified to include more components to simulate specific processes when applied to a specific river basin where such processes are observed or considered to be dominant. In this study, we have added a new component to simulate interception using a simple parametric approach. Interception plays an important role in the water balance of a watershed although it is often disregarded. In addition, a refinement for the transpiration in the unsaturated zone has been made. Finally, an improved approach for simulating saturation overland flow by relating the variable source area to both the topography and the groundwater level is presented. The model has been calibrated and verified using a 4-year data set, which has been split into two for calibration and validation. The model performance has been assessed by multi-criteria evaluation. This work represents a complete application of the REW approach to watershed rainfall-runoff modelling in a real watershed. The results demonstrate that the REW approach provides an alternative blueprint for physically based hydrological modelling.


2016 ◽  
Author(s):  
Yakov A. Pachepsky ◽  
Gonzalo Martinez ◽  
Feng Pan ◽  
Thorsten Wagener ◽  
Thomas Nicholson

Abstract. The accuracy-based model performance metrics not necessarily reflect the qualitative correspondence between simulated and measured streamflow time series. The objective of this work was to use the information theory-based metrics to see whether they can be used as complementary tool for hydrologic model evaluation and selection. We simulated 10-year streamflow time series in five watersheds located in Texas, North Carolina, Mississippi, and West Virginia. Eight model of different complexity were applied. The information theory based metrics were obtained after representing the time series as strings of symbols where different symbols corresponded to different quantiles of the probability distribution of streamflow. The symbol alphabet was used. Three metrics were computed for those strings – mean information gain that measures the randomness of the signal, effective measure complexity that characterizes predictability and fluctuation complexity that characterizes the presence of a pattern in the signal. The observed streamflow time series has smaller information content and larger complexity metrics than the precipitation time series. Watersheds served as information filters and and streamflow time series were less random and more complex than the ones of precipitation. This is reflected by the fact that the watershed acts as the information filter in the hydrologic conversion process from precipitation to streamflow. The Nash Sutcliffe efficiency metric increased as the complexity of models increased, but in many cases several model had this efficiency values not statistically significant from each other. In such cases, ranking models by the closeness of the information theory based parameters in simulated and measured streamflow time series can provide an additional criterion for the evaluation of hydrologic model performance.


Sign in / Sign up

Export Citation Format

Share Document