Saman Razavi

and 35 more

The notion of convergent and transdisciplinary integration, which is about braiding together different knowledge systems, is becoming the mantra of numerous initiatives aimed at tackling pressing water challenges. Yet, the transition from rhetoric to actual implementation is impeded by incongruence in semantics, methodologies, and discourse among disciplinary scientists and societal actors. This paper confronts these disciplinary barriers by advocating a synthesis of existing and missing links across the frontiers distinguishing hydrology from engineering, the social sciences and economics, Indigenous and place-based knowledge, and studies of other interconnected natural systems such as the atmosphere, cryosphere, and ecosphere. Specifically, we embrace ‘integrated modeling’, in both quantitative and qualitative senses, as a vital exploratory instrument to advance such integration, providing a means to navigate complexity and manage the uncertainty associated with understanding, diagnosing, predicting, and governing human-water systems. While there are, arguably, no bounds to the pursuit of inclusivity in representing the spectrum of natural and human processes around water resources, we advocate that integrated modeling can provide a focused approach to delineating the scope of integration, through the lens of three fundamental questions: a) What is the modeling ‘purpose’? b) What constitutes a sound ‘boundary judgment’? and c) What are the ‘critical uncertainties’ and how do they propagate through interconnected subsystems? More broadly, we call for investigating what constitutes warranted ‘systems complexity’, as opposed to unjustified ‘computational complexity’ when representing complex natural and human-natural systems, with particular attention to interdependencies and feedbacks, nonlinear dynamics and thresholds, hysteresis, time lags, and legacy effects.

Banamali Panigrahi

and 5 more

Machine learning (ML) is increasingly considered the solution to environmental problems where only limited or no physico-chemical process understanding is available. But when there is a need to provide support for high-stake decisions, where the ability to explain possible solutions is key to their acceptability and legitimacy, ML can come short. Here, we develop a method, rooted in formal sensitivity analysis (SA), that can detect the primary controls on the outputs of ML models. Unlike many common methods for explainable artificial intelligence (XAI), this method can account for complex multi-variate distributional properties of the input-output data, commonly observed with environmental systems. We apply this approach to a suite of ML models that are developed to predict various water quality variables in a pilot-scale experimental pit lake. A critical finding is that subtle alterations in the design of an ML model (such as variations in random seed for initialization, functional class, hyperparameters, or data splitting) can lead to entirely different representational interpretations of the dependence of the outputs on explanatory inputs. Further, models based on different ML families (decision trees, connectionists, or kernels) seem to focus on different aspects of the information provided by data, although displaying similar levels of predictive power. Overall, this underscores the importance of employing ensembles of ML models when explanatory power is sought. Not doing so may compromise the ability of the analysis to deliver robust and reliable predictions, especially when generalizing to conditions beyond the training data.