Knowledge-informed deep learning for hydrological model calibration: an application to Coal Creek Watershed in Colorado

. Deep learning (DL)-assisted inverse mapping has shown promise in hydrological model calibration by directly estimating parameters from observations. However, the increasing computational demand for running the state-of-the-art hydrological model limits sufﬁcient ensemble runs for its calibration. In this work, we present a novel knowledge-informed deep learning method that can efﬁciently conduct the calibration using a few hundred realizations. The method involves two steps. First, we determine decisive model parameters from a complete parameter set based on the mutual information (MI) between model responses and each parameter computed by a limited number of realizations ( ∼ 50). Second, we perform more ensemble runs (e.g., several hundred) to generate the training sets for the inverse mapping, which selects informative model responses for estimating each parameter using MI-based parameter sensitivity. We applied this new DL-based method to calibrate a process-based integrated hydrological model, the Advanced Terres-trial Simulator (ATS), at Coal Creek Watershed,

Abstract.Deep learning (DL)-assisted inverse mapping has shown promise in hydrological model calibration by directly estimating parameters from observations.However, the increasing computational demand for running the state-of-theart hydrological model limits sufficient ensemble runs for its calibration.In this work, we present a novel knowledgeinformed deep learning method that can efficiently conduct the calibration using a few hundred realizations.The method involves two steps.First, we determine decisive model parameters from a complete parameter set based on the mutual information (MI) between model responses and each parameter computed by a limited number of realizations ( ∼ 50).Second, we perform more ensemble runs (e.g., several hundred) to generate the training sets for the inverse mapping, which selects informative model responses for estimating each parameter using MI-based parameter sensitivity.We applied this new DL-based method to calibrate a processbased integrated hydrological model, the Advanced Terrestrial Simulator (ATS), at Coal Creek Watershed, CO.The calibration is performed against observed stream discharge (Q) and remotely sensed evapotranspiration (ET) from the water year 2017 to 2019.Preliminary MI analysis on 50 realizations resulted in a down-selection of 7 out of 14 ATS model parameters.Then, we performed a complete MI analysis on 396 realizations and constructed the inverse mapping from informative responses to each of the selected parameters using a deep neural network.Compared with calibration using observations covering all time steps, the new inverse mapping improves parameter estimations, thus enhancing the performance of ATS forward model runs.The Nash-Sutcliffe efficiency (NSE) of streamflow predictions increases from 0.53 to 0.8 when calibrating against Q alone.Using ET observations, on the other hand, does not show much improvement on the performance of ATS modeling mainly due to both the uncertainty of the remotely sensed product and the insufficient coverage of the model ET ensemble in capturing the observation.By using observed Q only, we further performed a multiyear analysis and show that Q is best simulated (NSE > 0.8) by including in the calibration the dry-year flow dynamics that show more sensitivity to subsurface characteristics than the other wet years.Moreover, when continuing the forward runs till the end of 2021, the calibrated models show similar simulation performances during this evaluation period as the calibration period, demonstrating the ability of the estimated parameters in capturing climate sensitivity.Our success highlights the importance of leveraging data-driven knowledge in DL-assisted hydrological model calibration.

Introduction
Calibrating a hydrological model is critical in accurately capturing the hydrological dynamics of the simulated watershed, which in turn improves the understanding of the corresponding terrestrial water cycle (Singh and Frevert, 2002).While the increasing complexity and spatiotemporal resolution of the hydrological models enable a better representation of the watershed dynamics (Kollet and Maxwell, 2006;Coon et al., 2019;Wang and Kumar, 2022), running these models is com-Published by Copernicus Publications on behalf of the European Geosciences Union.P. Jiang et al.: Knowledge-informed calibration using deep learning putationally expensive (Clark et al., 2017) even with existing high-performance computing resources.This computational burden significantly impedes the efficient and accurate calibration of integrated hydrological models.
Balancing the trade-off between computational cost and calibration accuracy is necessary when adopting traditional model calibration methods (Kavetski et al., 2018).Newtontype optimization methods (Jorge and Stephen, 2006;Qin et al., 2018) are known for their fast iteration convergence but usually only achieve local optimum.On the other hand, the stochastic methods, such as the shuffled complex evolution algorithm (Duan et al., 1992), the dynamically dimensioned search algorithm (Tolson and Shoemaker, 2007), and the ensemble Kalman filter (Reichle et al., 2002;Moradkhani et al., 2005;Evensen, 2009;Sun and Sun, 2015), are capable of providing better global optimum at the cost of high computational demand.One alternative is to use a surrogate model that provides fast emulations to replace the physical model during calibration so that one might save the computational budget while achieving a reasonable calibration result.Mo et al. (2019) employed a dense convolutional encoderdecoder network as the emulator for a two-dimensional contaminant transport model to estimate the conductivity field using an iterative local updating ensemble smoother.Similar subsurface characterization was also performed by Wang et al. (2021) who developed a theory-guided neural network as the surrogate of a flow model which was coupled with an iterative ensemble smoother to estimate the subsurface characteristics.In light of the dimensionality reduction of the model states, Dagon et al. (2020) calibrated biophysical parameters using a global optimizer on a surrogate that emulates the principle components of the outputs of community land models.Jiang and Durlofsky (2021) adopt a recurrent encoder-decoder network as the data-space inversion parameterization to reduce the dimensionality of the model states and parameters and used an ensemble smoother with multiple data assimilation to update the low-dimension latent variables.Despite the successes of using surrogates in calibration, how to develop an accurate and trustworthy emulator can vary from case to case and, in fact, is still a long-standing challenge (McGovern et al., 2022).
Recently, deep learning (DL)-assisted inverse mapping shows promise in addressing inverse problems and has seen early successes in hydrology (Cromwell et al., 2021;Mudunuru et al., 2021;Tsai et al., 2021), petroleum engineering (Razak et al., 2021), and geophysics (Yang and Ma, 2019;Wang et al., 2022).By employing a well-trained DL model (Goodfellow et al., 2016), this approach maps model parameters from model states/outputs/responses such that once trained, the mapping can directly infer the parameters based on observations.The inverse mapping outperforms the traditional calibration approaches in the following ways.First, DL models can better capture the highly nonlinear relationships encoded in the model than ensemble-based methods, which primarily rely on the linear estimation the-ory through the Kalman filter (Evensen, 2009;Moradkhani et al., 2005;Reichle et al., 2002;Sun and Sun, 2015).Yang and Ma (2019) developed a convolutional neural-networkbased inverse mapping that outperforms the traditional full waveform inversion adopting the adjoint-state optimization method in estimating seismic velocity from seismic data.Cromwell et al. (2021) also demonstrate the improved performance of DL-assisted inverse mapping over an ensemble smoother in estimating subsurface permeability used in a watershed model based on the Advanced Terrestrial Simulator (ATS).Second, training DL models may potentially use fewer realizations than the traditional methods such as iterative calibration methods that usually require several thousands of realizations to achieve the model optimization convergence.Mudunuru et al. (2021) show that DL-assisted inverse mapping using 1000 realizations outperforms dynamically dimensioned search algorithm (Tolson and Shoemaker, 2007) that has to leverage 5000 realizations in calibrating multivariate parameters of models based on the Soil and Water Assessment Tool (SWAT).Third, the calibration workflow is simpler given that ensemble simulations do not have to be fully coupled with the inverse mapping.Traditional calibration methods require sophisticated workflows (White et al., 2020;Jiang et al., 2021) to manipulate the model restart (e.g., ensemble Kalman filter), model rerunning (e.g., gradient-based and ensemble-based methods), and the communications between hydrological model and calibration tools, which can be time-consuming.Meanwhile, such an integrated workflow tool is not necessary for developing inverse mapping because model ensemble runs and DL training are now two separate steps.This decoupling of ensemble runs and DL training allows us to use high-performance computing resources to calibrate hydrological models efficiently.
Despite its success, the current DL-assisted inverse mapping is often designed to take all observed states in estimating hydrological model parameters.However, some observational values can be uninformative, or even misinformative, to estimate parameters (Loritz et al., 2018), thus impeding the mapping performance.While the underlying assumption is that the trained DL model can "automatically" delineate the accurate relationship between parameters and observed responses, the limited realizations (e.g., a few hundred) would potentially restrain the DL model from being well trained (Moghaddam et al., 2020).Further, when using all observed responses as inputs, the potentially large amount of trainable weights of the DL model can make the model training hard and induce the overfitting of the model (Ying, 2019), thus calling for more realizations used in training.Lately, several studies proposed new inverse mapping methods that indirectly address this issue by using dimensionality reduction and differential programming.Razak et al. ( 2021) developed a latent-space inversion that performs the inverse mapping from the model responses to parameters in their reduced spaces through an auto-encoder.The dimensionality reduction by using an auto-encoder is supposed to not only lower the original high-dimensional data space but also indirectly distill the most relevant dynamics to the parameters.Tsai et al. (2021) leveraged differential programming such that the loss function of an inverse mapping is designed to directly minimize the difference between observations and the responses predicted by a differentiable version of the physical model using the estimated parameters.In doing so, the uninformative responses are automatically given less importance in the loss.Nevertheless, both studies reduce the irrelevant information in implicit ways through complicating the DL-based mapping which, due to its "black-box" nature, does not explicitly show to what extent an observation is relevant to a parameter.Also, by adding another layer of complexity, the inverse mapping can potentially be hard to build.For instance, the current solutions to develop a differentiable physical model rely on either a surrogate or rewriting the model using differentiable parameters (Karniadakis et al., 2021), both of which are research challenges that go beyond addressing the inverse problem itself.
The emergence of knowledge-informed DL provides a new opportunity to resolve the uninformative or misinformative issue by explicitly encoding the complex relationship between the inputs and outputs in the DL model (Willard et al., 2020).Knowledge-informed DL includes, but is not limited to, the following three methods: (1) a physics-guided loss function, (2) hybrid modeling, and (3) a physics-guided design of architecture.A physics-guided loss function embeds the mathematical relation between inputs and outputs in the loss function, known as physics-informed deep learning (Karniadakis et al., 2021) and has seen some early successes in earth science.For instance, Jia et al. (2019) leveraged an energy conservation loss in developing a physicsguided recurrent neural network to simulate lake temperature.Hybrid modeling, on the other hand, directly integrates the physical model with the DL model, which often serves as a surrogate for its computationally intense counterpart in the physical model (Kurz, 2021).An example can be coupling a DL-based emulator for turbulent heat fluxes with a process-based hydrological model framework (Bennett and Nijssen, 2021).Lastly, the physics-guided design of architecture explicitly designs the neural networks consistently with prior knowledge.The widely used convolutional (Atlas et al., 1987) and recurrent (Rumelhart and McClelland, 1986) neural networks fall into this category due to their specific network structures to learn the spatial and temporal relationships, respectively.Other related studies include relating intermediate physical variables to hidden neurons (Daw et al., 2020), explicitly learning nonlinear dynamics through the neural operator (Kovachki et al., 2021), and encoding domain knowledge obtained from nonparametric physics-based kernels into the neural network (Sadoughi and Hu, 2019).Compared with the other two types of knowledge-informed DL models, which are usually limited to particular physical dynamics, the physics-guided design of architecture is more generic regarding both the processes of gaining prior knowledge and designing a correspondent neural network.
One important piece of domain knowledge is the pairwise relation between model parameters and responses: that is, how relevant a parameter is to a model response at a given time step.Understanding such a pairwise relationship is essential to select the most relevant model responses to estimate each parameter when building the inverse mapping.To this end, global sensitivity analysis (Razavi and Gupta, 2015;Sarrazin et al., 2016) is a suitable tool due to its capability of quantifying the contribution of uncertainty from model inputs and parameters to model outputs and has been extensively applied in earth system modeling (Hall et al., 2009;Harper et al., 2011;Anderson et al., 2014;Guse et al., 2014;Dai et al., 2017).Through a sensitivity analysis study on SWAT modeling (Jiang et al., 2022), mutual information (MI; Cover and Thomas, 2006) has shown the promise of using a few hundred model realizations to provide similar sensitivity results as the popular Sobol' sensitivity analysis (Sobol, 2001) that usually relies on several thousand realizations.As a result, MI is well suited to unravel the relation between model parameters and responses given a few hundred realizations of a state-of-the-art fully integrated hydrological model.
This study aims to develop a novel knowledge-informed DL method for model calibration by using a few hundred realizations.We leverage MI-based global sensitivity analysis to uncover the dependencies between parameters and observed responses, which are then used to guide the selection of crucial responses as the inputs of DL-assisted inverse mapping.We applied this method in estimating multiple parameters of a fully integrated hydrological model, ATS (Coon et al., 2019), at the Coal Creek Watershed, a snow-dominated alpine basin located in Colorado, US.Multiple water years of hydrological observations are used in both the ATS model's calibration and evaluation.We further performed a multiyear analysis to investigate the significance of wet and dry years in model calibration.Our study highlights the importance of domain knowledge in uncovering the dependencies among variables of interest before hydrological model calibration.

Study site
The Coal Creek Watershed is located in the western part of the larger East Taylor Watershed in Colorado (Fig. 1a).The majority of the discharge flows through Coal Creek from the west to the east.The watershed is an HUC12 (hydrologic unit code) watershed encompassing around 53.2 km 2 of the drainage area (HUC12 ID: 140200010204).According to the Köppen classification system (Köppen and Geiger, 1930), this high alpine watershed is classified as warm summer and humid continental climate with a significant snow process dominating the hydrological cycle.Based on the long-term Daymet forcing dataset (Thornton et al., 2021), the watershed receives ∼ 530 mm of snowfall annually, dominating its annual precipitation (∼ 850 mm).This watershed exhibits strong variations in topography with elevations ranging from 2706 to 3770 m, where the primary land covers are evergreen forest (62.6 %) and shrub (20.5 %).Hydrological observations are available through (1) a USGS gaging station (station number 09111250) that records daily discharge (Q) observations at the watershed outlet, and (2) a remote sensing product of the Moderate Resolution Imaging Spectroradiometer (MODIS) 8 d composite evapotranspiration (ET) at 500 m resolution.Figure 1 shows the time series of Q and watershed-averaged MODIS ET during 1 October 2016 to 31 December 2021, which are used as observations for calibrating and evaluating the ATS model.

ATS model setup
ATS is a fully distributed hydrologic model that integrates surface and subsurface flow dynamics (Coon et al., 2019).The surface hydrological process is characterized by a twodimensional diffusion wave approximation of the Saint-Venant equation.A three-dimensional Richards equation is used to represent the subsurface flow.The model adopts the Priestley-Taylor equation to simulate evapotranspiration (ET) from various processes (e.g., snow and plant transpiration), which are coupled with the surface-subsurface hydrological cycle.
We leveraged an existing ATS setup at the Coal Creek Watershed (Shuai et al., 2022).The Watershed Workflow package (Coon and Shuai, 2022) was used to delineate the mesh, the surface land covers, and the subsurface characteristics of the watershed.The resulting mesh consists of 171 760 cells, formed by a two-dimensional triangle surface mesh followed by 19 terrain-following subsurface layers (Fig. 1a).The surface mesh contains 8588 triangular cells with varying sizes that range from ∼5000 m 2 near the stream network to ∼ 50 000 m 2 away from the stream network.On the surface, the National Land Cover Database was used to delineate the land cover types.In the subsurface, the 19 layers add up to 28 m and contain (1) six soil layers in the top 2 m, (2) 12 geological layers in the middle, and (3) one bedrock layer in the bottom of the simulation domain.The maximum depth to bedrock (28 m) was determined by SoilGrids (Shangguan et al., 2017).The subsurface characteristics of the soil and geological layers are retrieved from the National Resources Conservation Service (NRCS) Soil Survey Geographic (SSURGO) soils database and GLobal HYdrogeology MaPS (GLHYMPS) 2.0 (Huscroft et al., 2018), respectively.The k-means clustering algorithm (Likas et al., 2003) was used to group the soil and geological types based on the default permeability values from SSURGO and GLHYMPS, which leads to five soil types and four geological types shown in Fig. 1c and d, respectively.Each clustered soil or geological type is associated with a specific set of subsurface characteristics (such as permeability), which are assigned to the corresponding grouped grid cells.These subsurface characteristics are important in controlling flow dynamics and can be estimated from hydrological observations.To ensure that the model achieved a physically appropriate initial state, two spin-ups were performed sequentially, including (1) a cold spinup that ran the model for 1000 years by using constant rainfall and led to steady-state model outputs (e.g., converged total amount of subsurface water storage), and (2) a warm spinup that was initialized by the steady-state spinup result and performed a transient simulation for 10 years (i.e., 1 October 2004-1 October 2014) under the Daymet forcing.Please refer to Shuai et al. (2022) for the detailed model setup and spinup.
We select a preliminary set of 14 model parameters to be calibrated, which can be categorized into ET, snow, river channel, and subsurface characteristics.The ET parameters include two coefficients used by the Priestley-Taylor equation (Priestley and Taylor, 1972) in calculating the potential ET of snow and transpiration, respectively (i.e., priestley_taylor_alpha-snow and priestley_taylor_alpha-transpiration).The snow parameters are the snow melting rate (snowmelt_rate) and the temperature determining the snow melting (snowmelt_degree_diff).The river channel characteristic is the Manning's coefficient (manning_n), which describes the roughness of the surface channel.The subsurface characteristics include the major soil and geological permeability (i.e., perm_s1, perm_s2, perm_s3, perm_s4, perm_s5, perm_g1, perm_g2, perm_g3, and perm_g4).A detailed description of these parameters can be found in Table A1.

Knowledge-informed model calibration using deep learning
We develop a new methodology to calibrate ATS using knowledge-informed DL, as shown in Fig. 2. The key idea is to leverage a data-driven approach to identify the sensitive model response as the inputs to the DL-assisted inverse mapping for estimating each parameter.Here, we use the MI as the sensitivity analysis tool due to its capability to uncover nonlinear relationships.Derived from Shannon's entropy (Cover and Thomas, 2006), MI quantifies the shared information between two variables: a model response Y and a model parameter X as follows:  (Cromwell et al., 2021).In this study, we follow a similar strategy of Jiang et al. (2022) to estimate p using 10 evenly divided bins along each dimension and perform SST tests to filter out any nonsignificant MI value with a significance level of 95 % based on 100 bootstrap samples.In other words, the computed MI is set to zero if the statistical significance test fails.By using MI-based sensitivity analysis, our calibration method involves the following four steps (see Fig. 2): (1) narrowing down the parameters to be calibrated using a preliminary MI analysis, (2) computing the parameter sensitivity using a full MI analysis, (3) developing knowledge-guided DL inverse mapping, and (4) estimating parameters with observational error.The details of each step are as follows.

Step 1: narrowing down the parameters to be calibrated
The objective of this step is to further reduce the computational requirement by identifying the parameters that are most relevant to the responses through a coarse-resolution sensitivity analysis.To this end, we first perform a preliminary MI analysis on a sizable preliminary parameter set.MI is computed between the model response at each time step and each parameter based on Eq. (1).Rather than getting an accurate sensitivity result, this preliminary analysis aims to provide an overview of parameter sensitivity and thus is performed on a small number of realizations to save computational cost.At Coal Creek, we generated 50 realizations based on a total of 14 parameters listed in Table A1 and performed the corresponding ATS runs to compute MI.The Sobol' sequence (Sobol, 1967)  coverage of the parameter ranges.This preliminary MI analysis would allow filtering out the parameters that show little sensitivity to the model responses, thus reducing the number of parameters to be calibrated.This filtering process is performed based on whether a parameter demonstrates sufficient sensitivity across the simulation period.In this study, we selected the parameter whose proportion of the nonzero MI is larger than 5 % of the overall time steps for the following full sensitivity analysis.

Step 2: computing the parameter sensitivity
We then conduct a complete MI-based sensitivity analysis on the prescreened parameters in Step 1 by using a more significant number of realizations.The analysis accurately quantifies the sensitivity of model responses to each parameter, thereby uncovering their dependency.The analysis also verifies whether the parameters prescreened in Step 1 are "truly" sensitive and can be further excluded if insensitive.At this watershed, we generated 400 ATS realizations by sampling from a reduced parameter space at Step 1 using the Sobol' sequence, among which 396 ATS runs were completed and are used for computing MI.

Step 3: developing knowledge-informed DL inverse mapping
In this step, we develop the DL-based inverse mapping by leveraging the knowledge about the sensitivities between model responses and parameters, learned from the complete MI analysis in Step 2. While the original inverse mapping directly estimates the model parameters from the model responses using a tuned DL model (Fig. 3a), the complete sen-  3b.Consequently, this "filtering" step is expected to improve the performance of inverse mapping in parameter estimation by reducing the impact of noisy or unimportant model responses.

Step 4: parameter estimation with observation errors
Once an inverse mapping is trained, we first estimate the parameters using the corresponding observations and then use the estimated parameters to perform an ATS simulation.We evaluate the calibration results using different inverse mappings by comparing the simulated responses with the observations.We further perform uncertainty analysis to analyze how sensitive the DL-based parameter estimation is to the observation errors.To this end, we generate 100 realizations of noisy observations, denoted as o n , such that o n = o+ ×o×r, where o is the vector of the original observations, r is the random vector with the same size as o and is drawn from a standard normal distribution, and is the standard deviation of the random vector r and is usually taken as 1/3 of a given observation error.Following Cromwell et al. (2021), is set to 0.0166 for a 5 % observation error in this study.Given an inverse mapping, these noisy observations are used to estimate the ensemble parameter estimations and perform the corresponding ATS ensemble simulation.
Here, we separate the entire observation period in Fig. 1b into model calibration and evaluation periods in order to assess the adaptability of the estimated parameters to an uncalibrated period.To this end, we calibrate ATS only using the simulations during the water year 2017 to water year 2019 and used the remaining observations (till 31 December 2021) for model evaluation.The ensemble runs used for sensitivity analysis and inverse modeling are performed during the calibration period.The calibrated ATS forward runs were then performed on both periods and compared against the observations in Fig. 1b.We assess the performances of the calibrated models on both periods by using two scale-independent metrics: the Nash-Sutcliffe efficiency (NSE; Nash and Sutcliffe, 1970) and the modified Kling-Gupta efficiency (mKGE; Kling et al., 2012).

Deep learning-based inverse mapping development
For comparison purposes, we developed both the original inverse mapping and our proposed knowledge-informed version for parameter estimation.While a separate neural network is developed for estimating each parameter by using knowledge-informed inverse mapping (Fig. 3b), the original inverse mapping estimates all parameters using one neural network and is developed by following the same strategy in Cromwell et al. (2021) and Mudunuru et al. (2021) (Fig. 3a).Further, to assess the impact of different responses in calibration, we developed three types of inverse mappings that take various model responses (1) using both Q and ET, (2) using only Q, and (3) using only ET.Note that in the case of knowledge-informed inverse mapping using both observations, we take the union of the Q and ET that have nonzero mutual information as the inputs of the neural network.Additionally, a multiyear analysis was performed by training inverse mappings using Q of different combinations of observed years to evaluate both the impacts of the dry versus wet years and the number of observed years used in calibration.
All the inverse mappings developed in this study are listed in Table 1.Each mapping was developed using a multilayer perceptron (MLP) model as follows.The input of an MLP is an array concatenating the responses to be assimilated within a given calibration period.The output is the model parameter(s).Let us denote the number of input neurons, outhttps://doi.org/10.5194/hess-27-2621-2023 Hydrol.Earth Syst.Sci., 27, 2621-2644, 2023 P. Jiang et al.: Knowledge-informed calibration using deep learning put neurons, and hidden layers as N i , N o , and N l , respectively.N i depends on the type of inverse mapping (with or without being knowledge guided), the selections of the response variable(s), and the number of calibration years, varying from ∼ 100 using 1 year of Q to 1785 using all 3 years of Q and ET.N o equals either one (i.e., estimating each parameter using knowledge-informed DL calibration) or the number of all the parameters (i.e., using inverse mapping without mutual information).Given N i , N o , and N l , we adopt the arithmetic sequence to determine the number of neurons at each hidden layer N h,l = N i − N i −N o N l × l (where 1 ≤ l ≤ N l and • is the floor function).In doing so, the information from a sequence of observed responses can be gradually propagated to estimate the parameters.We use the leaky rectified linear unit (ReLu) as the nonlinear activation at the end of each layer.Based on the order of the Sobol' sequences, we sequentially split the 396 realizations into 300, 50, and 46 for train, validation, and test sets, respectively, such that each set is able to cover the full range of the parameter ensemble as much as possible.We trained each MLP using mean square error (MSE) as the loss function over 1000 epochs with a batch size of 32.The Adam optimization algorithm, a stochastic gradient descent approach, was used to train the neural network.We performed hyperparameter tuning on each MLP using grid search to find the optimal result by varying the number of hidden layers N l = [1, 3, 5, 7, 9, 10] and the learning rate l r = [1×10 −5 , 1×10 −4 , 1×10 −3  ].The performances of these mappings are further evaluated on the two magnitude-independent metrics, NSE and mKGE.To have consistent comparisons between mappings with and without being knowledge guided, both metrics are computed for the estimation of each parameter based on the test dataset (note that while the original mapping estimates all seven parameters as a whole during the training, we calculate the two metrics for each parameter separately during the postprocessing).

Results and discussions
In this section, we present the ATS calibration result at the Coal Creek Watershed using the proposed knowledgeinformed DL described in Sect.2.3.We first demonstrate the MI-based sensitivity analysis result, which is facilitated by a preliminary analysis using fewer realizations to narrow down the parameters to be calibrated.Then, we report the result of the trained DL-assisted inverse mapping performance using the simulations from the 3-year calibration period on the test dataset.By using Q alone, our result shows (1) the improved ATS simulation calibrated by knowledge-informed DL inverse mapping over the traditional mapping, and (2) the consistent performances of model forward runs between the calibration and evaluation periods.However, we also identify the ET observation extrapolation issue that impedes the calibration performance.Last, we present the multiyear analysis using observed Q only to estimate the parameters regarding (1) the performance of the simulated Q driven by the estimated parameters, and (2) the impact of the observational error on the parameter estimations using inverse mapping.

Parameter prescreening using a preliminary MI analysis
We calculated the MI for each daily Q and ET with each of the parameters listed in Table A1. Figure 4 shows the estimation of MI by using the 50 ATS runs in the water year 2017 (note that 20 ATS runs did not complete after the end of the water year 2017).We then ranked the parameters based on the temporally averaged MI across the analyzed time period, as plotted on the right panel of each MI heat map.The MI heat maps show the varying sensitivities of different parameters over time.For instance, while Q (Fig. 4a) is mainly sensitive to the permeability (e.g., perm_g1) during low-flow seasons, snow melting parameters (e.g., snowmelt_degree_diff) play more critical roles in high-flow periods than others.On the other hand, ET (Fig. 4b) is mostly controlled by the two Priestley-Taylor alpha coefficients in low and high flows, respectively.Based on the proportion of nonzero MI over all the time steps (see Fig. A1), we find that Q is mostly sensitive to (using a threshold of 5 %) perm_s3, perm_s4, perm_g1, perm_g4, snowmelt_rate, snowmelt_degree_diff, and priestley_taylor_alpha_transpiration; and ET is mostly sensitive to priestley_taylor_alpha_transpiration, priest-ley_taylor_alpha_snow, perm_s3, perm_g1, and perm_g4.Consequently, we narrow down the parameters to be calibrated by taking the union of the two sets of parameters that show sensitivities to either Q or ET (also highlighted in Table A1).

The full MI analysis
We performed the ATS ensemble simulations on 400 realizations of the reduced parameter set and computed the MI heat maps on the completed 396 realizations during the calibration period (i.e., water years 2017-2019) shown in Fig. 5.By using more realizations, this complete MI analysis shows a better delineation of parameter sensitivity than the preliminary analysis due to its convergence on MI estimation (see the convergence of the parameter rankings in Fig. A2).The convergence on a few hundred realizations is consistent with another MI-based sensitivity analysis study using the Soil and Water Assessment Tool (SWAT) (Jiang et al., 2022).Further, the MI-based parameter ranking suggests that compared with the preliminary analysis, the full analysis (1) improves the MI estimations (e.g., perm_s3) and ( 2) identifies the insensitive parameters (e.g., perm_s4) that are falsely considered sensitive due to the limited samples in the preliminary analysis (see Fig. 6).The main permeability in the soil layer (i.e., perm_s3), for example, now shows higher and more temporally coherent sensitivity to Q (Fig. 5a).On the other hand, perm_s4, which shows some sensitivity in the preliminary analysis, turns out to be insensitive to both Q and ET with almost zero MI at each time step.Since both Q and ET share little MI with perm_s4, we exclude it from the parameters to be calibrated.In the following, we leveraged this MI result to identify the sensitive responses (i.e., with nonzero MI) to be used as the inputs of the knowledge-informed DL inverse mapping.

Physical knowledge obtained by MI analysis
The sensitivity analysis reveals the seasonal importance of these watershed characteristics to the hydrological fluxes in this area (Fig. 5).During the low-flow period (September through March of next year), Q is mostly controlled by the subsurface permeability (i.e., perm_g1, perm_s3, and perm_s4) which regulates both the infiltration and the groundwater movement.Transpiration also plays a role in driving the low-flow dynamics through the Priestley-Taylor coefficient (e.g., priestley_taylor_alpha_transpiration).During the high-flow period (March through September), the snow melting process turns out to be the most critical factor in contributing to the large runoffs, which complies with the prior knowledge about the dominance of the snow process in this watershed.Likewise, the total ET is by and large attributed to a variety of evaporation and transpiration.Snow evaporation is the main component of the total ET in both late autumn and winter when the snow melting rarely happens.
On the other hand, in warmer and high-flow seasons, transpiration becomes the dominant contributor to the total ET.The seasonable pattern of the sensitivity of each parameter not only uncovers the hydrological process in the watershed but also serves as the basis to select the most informative model responses to estimate each model parameter.

Performances of deep learning-based inverse mappings
The developed inverse mappings demonstrate limited overfitting issues.Figure 7 plots the training and validation loss over epochs of the seven parameters, each of which is estimated by the knowledge-informed inverse mapping using the corresponding 3 years of sensitive streamflows (i.e., miqonly-3yrs).It can be observed from the figure that both losses quickly decrease with epochs with little discrepancies.Particularly, the parameters sharing higher mutual information with streamflows show faster convergences of the loss function and do not have an overfitting problem (e.g., perm_s3 and snowmelt_degree_diff; see Fig. 6a).The discrepancy between training and validation losses gets slightly larger for less sensitive parameters (e.g., perm_g4) where streamflow is less informative in parameter estimation.In- deed, informative model responses can provide better parameter estimations, thus reducing the overfitting impact.The limited impact of overfitting is also evident from the NSE and mKGE bar plots of the training, validation, and test sets of all the inverse mappings (see Figs. A3 and A4), where most mappings have similar performances on parameter estimations among the three sets.

The improved parameter estimation using knowledge-informed inverse mapping
We further compare the performances of the inverse mappings with and without being knowledge informed, represented by blue and green colors, respectively, in Fig. 8. Blank, cross, and circle textures stand for the mapping taking as inputs only discharge (qonly), only evapotranspiration (etonly), and both (qet), respectively.Overall, most inverse mappings are trained well with both mKGE and NSE greater than 0.8, with a few exceptions (e.g., the two snow melting coefficients, perm_g1, and perm_g4 using etonly-3yr and mietonly-3yr).The inferior estimations of the four parameters are due to their minimal MI shared with ET (Fig. 5) such that ET dynamics are insufficient to inform the two parameters.It can be observed from Fig. 8 that a knowledge-informed inverse mapping (i.e., the blueish/black bars) generally outperforms and has higher mKGE/NSE than its counterpart that uses all the observed time steps as inputs (i.e., the greenish bars).Noticeably, the NSE of knowledge-informed mapping increases when estimating the two Priestley-Taylor coefficients, the two snow melting parameters (except that estimated by mi-etonly-3yrs), and perm_s3, regardless of which model response is used in calibration.The extent of how sen-sitive the two responses are to a parameter also plays a role in the performance of the inverse mappings.The two Priestley-Taylor coefficients, which are the two most sensitive parameters to ET, are better estimated by using ET than using only Q dynamics.On the other hand, using Q yields superior performance over using ET in estimating permeability and the snow melting coefficients.As a result, when both Q and ET are used, the knowledge-guided inverse mappings (the black bars) turn out to be the best calibration tool for most of the parameters.
3.3 Forward runs of calibrated ATS using 3-year Q, ET, and Q-ET We estimated the parameters using each of the inverse mappings (see Table A2) and performed the corresponding ATS forward runs.Figure 9 shows the Q and ET observations (the black lines) as well as the calibrated simulations during calibration (the blue lines) and evaluation (the cyan lines) periods, with their one-to-one scatter plots shown on the side.Overall, the parameter estimated by using knowledgeinformed inverse mapping improves the calibrated Q simulation.When using only Q for calibration, mKGE increases from 0.65 (qonly-3yrs) to 0.80 (mi-qonly-3yrs) in the model calibration period.

Adaptability of the calibrated model in the evaluation period
For both Q and ET, both NSE and mKGE of the evaluation period (the cyan lines) are astonishingly close, if not identical, to that of the calibration period (the blue lines).Whenever the calibrated model shows improvement using https://doi.org/10.5194/hess-27-2621-2023Hydrol.Earth Syst.Sci., 27, 2621-2644, 2023 the knowledge-informed inverse mapping (such as the comparison between qonly-3yrs and mi-qonly-3yrs), we can observe the corresponding improvement in the evaluation period.Such consistent performance between the two periods suggests the robustness of the estimated parameters to climate sensitivity.

The extrapolation issue of ET observations
While using knowledge-informed inverse mapping improves the calibrated Q, ET simulations deteriorate, with NSE decreasing from ∼ 0.6 (etonly-3yrs) to ∼ 0.5 (mi-etonly-3yrs).This surprising result is probably attributed to both the extrapolation issue of ET observations and the high uncertainty of the remote sensing product.Compared with the ensemble simulation of Q (Fig. 5a) that captures most observed Q, a majority of ET observations exceed the range of the ATS ensemble of ET during the low-ET period each year (i.e., wet seasons or September through May next year; see Fig. 5b).
While it is possible that the defined sampling ranges of the two Priestley-Taylor coefficients in Table A1 are too limited to provide sufficient variations of ET dynamics, the uncertainty of the MODIS ET product also plays a role here (Khan et al., 2018;Xu et al., 2019).Xu et al. (2019) show that the MODIS ET product has much poorer performance and higher uncertainty in the Colorado Basin than in most of the remaining areas in the United States.The large uncertainty of this remote sensing product probably results from the increasing error in the satellite data caused by the cloudier sky in the mountainous region (Senay et al., 2013), particularly during the dry seasons (i.e., May through September) (Xu et al., 2019).In other words, although the ET ensemble gives better coverage on the observations in the dry seasons than the wet seasons (Fig. 5b), this could be due to the underestimation of the MODIS ET in the dry period with high ET such that the mismatch between the ET ensemble and the observed ET could probably be more significant.One consequence of these ET "outliers" is the inability of the inverse mapping to reasonably estimate the parameters.The resulting estimations of priestley_taylor_alphasnow, whose sensitivities with ET mainly occur during the low-ET period, greatly surpass its maximum sample threshold (i.e., 1.2; see Table A1) and range from 1.8 to 2.1 (see Table A2).These unreliable parameter estimations make the knowledge guidance less valid.Another evidence of the adverse impact of the extrapolation issue is the inferior simulation on Q when both Q and ET are used in calibration (e.g., qet-3yrs and mi-qet-3yrs).While using knowledge-informed inverse mapping still increases the NSE to 0.47 (mi-qet-3yrs) from 0.19 (qet-3yrs), both are worse than that of calibrating  1. Panels (a)-(d) are the simulated discharge (Q) using qonly-3yrs, mi-qonly-3yrs, qet-3yrs, and mi-qet-3yrs, respectively.Panels (e)-(h) are the simulated evapotranspiration (ET) using etonly-3yrs, mi-etonly-3yrs, qet-3yrs, and mi-qet-3yrs, respectively.(Blue and cyan colors represent the ATS forward runs at the model calibration and evaluation period, separately.)against Q alone (i.e., qonly-3yrs and mi-qonly-3yrs).This extrapolation issue underscores the significance of defining the acceptable parameter sampling ranges and understanding the observation uncertainty before calibration.We thus performed the remaining analysis using only Q for calibration to avoid the ET extrapolation impact.

Forward runs of calibrated ATS using multiyear Q
To investigate how wet/dry water years impact the model calibration, we further employed the knowledge-informed inverse mapping to calibrate ATS on a different numbers of years of the Q ensemble.To this end, we used 1 year, 2 years, and all 3 years of Q to develop the inverse mapping for model calibration.Figure 10 plots the calibrated ATS forward runs of Q, along with the observations.First, the performances of the calibration and evaluation periods are closely identical to each other, again illustrating the applicability of the calibrated model in an unseen time period.Second, we find that increasing the number of observed years in calibration does not necessarily improve the performance of ATS simulation on discharge.This is, in fact, consistent with the performance of the corresponding inverse mapping on the test dataset (see Fig. A5).In other words, using only 1 year of observations can yield a similar calibration result to using multiple years at this watershed.This can be attributed to the similar seasonal cycle of the whole-year discharge dynamics such that multiyear dynamics do not necessarily add more information to improve the calibration.This multiyear analysis underscores the potential of using fewer years than the number of available observed years in model calibration, which can save the computational time for one ensemble run and lead to more ensemble simulations given a fixed computational budget.

The significance of dry-year dynamics
Despite the similar calibration results between using 1 and multiple years, we find that by including the dry year (i.e., the second year), the calibration greatly improved the simulation of Q over the scenarios using only wet-year dynamics.Using only the dry year (i.e., second year or mi-qonly-1yr-2) in calibration generates the best-simulated Q with NSE and  1. Panels (a)-(c) are the simulated discharge (Q) using each of the 3 water years to calibrate (i.e., mi-qonly-1yr-1, mi-qonly-1yr-2, and mi-qonly-1yr-3).Panels (d)-(f) are the simulated discharge (Q) using 2 of the 3 water years to calibrate (i.e., mi-qonly-2yrs-12, mi-qonly-2yrs-23, and mi-qonly-2yrs-13).Panel (g) is the simulated discharge using all 3 years to calibrate (i.e., mi-qonly-3yrs).(Blue and cyan colors represent the ATS forward runs at the model calibration and evaluation period, separately.)mKGE above 0.8 for both calibration and evaluation periods.It outperforms the simulated Q using either of the other 2 wet years (i.e., mi-qonly-1yr-1 and mi-qonly-1yr-3) with the two metrics varying from 0.68 to 0.8.Particularly, mi-qonly-1yr-3, that uses the third and wettest year in calibration, generates the worst simulation of Q among the three inverse mappings using 1-year dynamics.For the inverse mappings using 2 years, the simulated Q of the mappings including the dry year (i.e., mi-qonly-2yrs-12 and mi-qonly-2yrs-23) are better than that of using the 2 wet years (i.e., mi-qonly-2yrs-13).We also observe that mi-qonly-2yrs-12 outperforms miqonly-2yrs-23 in the simulation of Q, which probably results from the better performance of using the first year (mi-qonly-1yr-1) only than using the wetter third year (mi-qonly-1yr-3).And using all 3 years (i.e., mi-qonly-3yrs) also guarantees reasonably good simulations of Q with both metrics equal to or slightly above 0.8.
The enhanced calibration performance by including the dry-year dynamics is probably due to the improved dependencies between discharge and subsurface features during low-flow periods.This can be observed in the yearly averaged MI bar plots in Fig. 11.With reduced sensitivities on the snow parameters, the dry year shows that the dominant permeability in the soil layer (i.e., perm_s3) proves to be the most sensitive parameter to discharge (with averaged MI 0.2).Meanwhile, the averaged MI of perm_s3 decreases to only around 0.1 in the third year when the performance of the corresponding calibrated Q (i.e., mi-qonly-1yr-3) is the worst during both calibration and evaluation periods.In fact, all the inverse mappings using the dry year consistently estimate higher perm_s3 (i.e., 10.9log 10 (m 2 )) than the other estimates ranging from −11.4 to −12.2log 10 (m 2 ) (see Table A3).The higher soil permeability estimated by the dry year has a better capability in draining the surface water and thus reduces the outlet discharge spikes during late spring and early autumn, thus yielding the simulations of Q more consistently with the observations plotted in Fig. 10.
Our finding on the significance of dry-year discharge in model calibration indirectly supports some recent studies.Pool et al. (2019) found that high flow provides limited infor- mation to calibrate models in snow-dominated catchments.This is mainly because there are fewer discharge fluctuations during a snow-melting or high-flow period than rainfallfed catchments (Viviroli and Seibert, 2015).The decreased role of high flow, in turn, enhances the importance of the low-flow period in calibration, particularly in dry years.Indeed, in this watershed, we do observe stronger diurnal discharge fluctuations during the low-flow period of the dry year (i.e., WY2018) than the other 2 wetter years (see Fig. A6), which facilitates the better calibration result using observations from the dry year.

Impact of observation errors
For each of the seven inverse mappings, we set 5 % observational error and generated 100 sets of estimated parameters by using 100 realizations of noisy observed discharges.Figure 12 shows the bar plots of these ensemble estimations of parameters.Overall, most estimated parameters (i.e., priestley_taylor_alpha-transpiration, snowmelt_degree_diff, perm_s3, and perm_g1) show little variability, indicating the robustness of the trained inverse mappings.Although there is higher variability for the other parameters (i.e., priestley_taylor_alpha-snow, snowmelt_rate, and perm_g4), these parameters share less MI with Q than the others (Fig. 11).Therefore, the larger variations of these less sensitive parameters have little impact on the corresponding ATS forward runs.Indeed, as shown in Fig. 13, the ATS ensemble runs using inverse mapping mi-qonly-3yrs demonstrate little variations of the ensemble runs of the discharges compared with the corresponding run without observational error, which verifies the robustness of these inverse mappings against the observational error.

Conclusions
We developed a novel model calibration methodology that leverages MI-based sensitivity analysis to guide the architecture of the DL-assisted inverse mapping based on only a few hundred realizations.A case study in the Coal Creek Water-shed shows that the calibrated ATS using such knowledgeinformed DL simulates the discharge well with both mKGE and NSE up to 0.8, outperforming its counterpart that does not adopt the knowledge guidance and only achieves NSE and mKGE around 0.5-0.6 (Fig. 9a and b).
The proposed hierarchical way of sensitivity analysis efficiently utilizes the available limited computational resource through a combination of a prescreening analysis and then a full analysis.Although the prescreening using 50 model runs does not theoretically exclude a false negative case that a sensitive parameter is classified as insensitive, the statistical significance test is able to improve the estimation of mutual information in Fig. 4, thus facilitating narrowing down an "accurate" list of parameters to be estimated.Based on the shortened parameter list, a full sensitivity analysis is successfully performed using nearly 400 model runs and provides physically meaningful results on the dependency between the parameters and model responses in Fig. 5.
Despite the improved streamflow simulations, we observe the adverse impact of the observation outliers when calibrating ATS against a remote sensing ET product (Fig. 9eh).These outliers deteriorate the performance of the inverse mapping in parameter estimation, thereby worsening the calibrated ATS forward runs.Based on the inverse mappings taking ET as inputs, the estimated priestley_taylor_alpha-snow, which is the most sensitive to ET (Fig. 5b), greatly exceeds the range of the parameter sample range (Tables A1 and A2).While the uncertainty of MODIS ET (Khan et al., 2018;Xu et al., 2019) also contributes to this extrapolation issue, this result underscores the significance of suitably defining the parameter sample range to assure the uncertainty of ensemble simulations covers the observations.
We further find that using 1 or 2 years of observations in calibration yields similar or even better results than that of using 3 years.This encouraging result highlights the importance of including abnormal year data for model calibration.This would significantly reduce the computational cost of each model run, increasing the number of model ensemble realizations used in developing calibration techniques.While earlier studies found that several years of discharge observahttps://doi.org/10.5194/hess-27-2621-2023 Hydrol.Earth Syst.Sci., 27, 2621-2644, 2023  tions are necessary to achieve a reasonable model calibration (Sorooshian et al., 1983;Yapo et al., 1996;Perrin et al., 2007) by using either a semi-distributed or a bucket model, we suspect that employing a fully integrated and high-resolution model like ATS can greatly reduce the errors due to the improved physical representation in the model, thus requiring longer observation periods for calibration.Hence, future work can focus on developing a systematical approach in identifying the most "important" observed period in model calibration.
The recently emerged knowledge-guided DL is swiftly gaining popularity in earth science.This study demonstrates one of its applications in calibrating a computationally expensive hydrological model.The developed methodology can be readily adopted in other watersheds to calibrate a different model.One potential future work is to develop a unified inverse modeling framework for multiple basins, where the atmospheric forcings and basin characteristics can also be used as the inputs of the inverse mappings in addition to the realization-dependent model responses.With the increasing complexity of earth system models, we believe such knowledge-guided DL calibration can pave the way for efficient yet effective model calibration without increasing significant computational demand.Table A2.Estimated parameters from the inverse mappings using 3 years of Q, ET, and Q and ET listed in Table 1.(Note that perm_g1 is not estimated by mi-etonly-3yrs because ET is not sensitive to perm_g1 at the analyzed 3 water years shown in Fig. 5.) Inverse mapping qonly-3yrs etonly-3yrs qet-3yrs mi-qonly-3yrs mi-etonly-3yrs mi-qet-3yrs Table A3.Estimated parameters from the inverse mappings using multiple years of Q listed in Table 1.

Figure 1 .
Figure 1.The Coal Creek Watershed and the setup of the Advanced Terrestrial Simulator (ATS).(a) The river network, the digital elevation model (DEM), and the surface mesh of the watershed.(b) The time series of USGS streamflow observations (station number 09111250) at the watershed outlet, and the Moderate Resolution Imaging Spectroradiometer (MODIS) 8 d composite evapotranspiration (ET) averaged across the watershed, where the observations before 1 October 2019 are used for model calibration, and the remaining observations till 31 December 2021 are used for evaluating the climate sensitivity of the estimated model parameters.(c, d) The delineated soil and geological layers, respectively.

Figure 2 .
Figure 2. Diagram of deep learning (DL) inverse mapping development including four steps:(1) performing a preliminary mutual information (MI) analysis using 50 model runs to narrow down the parameters to be estimated, (2) performing a full MI analysis on 396 model runs to correctly delineate the sensitivity between each parameter and each observed response, (3) developing DL inverse mappings with and without being knowledge informed, and (4) estimating parameters from observations with and without observation errors.

Figure 3 .
Figure3.Illustration of two inverse mappings.While inverse mapping (a) estimates all parameters from all observed responses using a deep learning model (e.g., fully connected neural net)(Cromwell et al., 2021;Mudunuru et al., 2021), the proposed knowledge-informed inverse mapping (b) estimates each parameter using only responses that share significant MI with the parameter.

Figure 4 .
Figure 4.A preliminary mutual information (MI)-based sensitivity analysis using 50 realizations of the Advanced Terrestrial Simulator (ATS) on (a) discharge (Q) and (b) evapotranspiration (ET).The top panel of each subplot shows the ATS ensemble (gray) and the corresponding observations (black).(Note that only water year 2016 is completed by all 50 runs).

Figure 5 .
Figure 5.A full MI-based sensitivity analysis using 396 realizations of the Advanced Terrestrial Simulator (ATS) on (a) discharge (Q) and (b) evapotranspiration (ET).The top panel of each subplot shows the ATS ensemble (gray) and the corresponding observations (black).

Figure 6 .
Figure 6.Sensitivity rankings based on mutual information (MI) on (a) the discharge (Q) and (b) evapotranspiration (ET) of both the preliminary and full MI analyses.

Figure 7 .
Figure 7.The mean square error (MSE) loss functions of the training and validation data over 1000 epochs for the seven parameters, each of which is estimated by a knowledge-guided inverse mapping (mi-qonly-3yrs) by using the corresponding sensitive streamflow during the model calibration period (i.e., water years 2017 through 2019).

Figure 8 .
Figure 8. Parameter estimation performance of the developed deep learning (DL) inverse mappings on the test dataset using the model responses in the calibration period with regards to (a) the modified Kling-Gupta efficiency (mKGE) and (b) the Nash-Sutcliffe efficiency (NSE).Green and light blue represent the mappings without and with being knowledge informed, respectively.Blank, cross, and circle textures are used to represent the mapping using discharge only (qonly), evapotranspiration only (etonly), and both (qet), respectively.

Figure 11 .
Figure11.Yearly averaged mutual information (MI) between the parameters and discharge (Q) computed from the full MI analysis result in Fig.5a.Left to right: the dry year (i.e., water year (WY) 2018), the first wet year (i.e., WY2017), the second wet year (i.e., WY2019), and all the 3 years.

Figure 12 .
Figure 12.Box plots of ensemble parameters estimated knowledge-informed inverse mappings of multiple years of discharge (Q) using 100 realizations of noisy discharge observations with 5 % observation error, categorized into (a) evapotranspiration (ET) parameters, (b) snow parameters, and (c) permeability parameters.

Figure 13 .
Figure 13.Ensemble forward runs of the Advanced Terrestrial Simulator (ATS) on discharge using 100 ensemble parameters estimated by mi-qonly-3yr through noisy discharge observations with 5 % observation error.

Figure A3 .Figure A4 .
Figure A3.Parameter estimation performance of the developed deep learning (DL) inverse mapping on the train, validation, and test datasets using 3 years of discharges (Q) and evapotranspiration (ET) with regards to (right) the modified Kling-Gupta efficiency (mKGE) and (left) the Nash-Sutcliffe efficiency (NSE).

Figure A5 .
Figure A5.Performance of the developed deep learning (DL) inverse mapping on the test dataset using multiyear discharges (Q) with regards to (a) the modified Kling-Gupta efficiency (mKGE) and (b) the Nash-Sutcliffe efficiency (NSE).

Figure A6 .
Figure A6.Logarithmic discharge observation from water year 2017 through water year 2019 at the USGS gage station 09111250.The top panel shows the corresponding snow (yellow) and rain (blue) precipitation.

Table 1 .
All inverse mappings with and without being knowledge informed developed for calibrating the Advanced Terrestrial Simulator (ATS) at the Coal Creek Watershed.

Table A1 .
List of the parameters used in the preliminary or full mutual information analysis.