<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing with OASIS Tables v3.0 20080202//EN" "https://jats.nlm.nih.gov/nlm-dtd/publishing/3.0/journalpub-oasis3.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:oasis="http://docs.oasis-open.org/ns/oasis-exchange/table" xml:lang="en" dtd-version="3.0" article-type="research-article">
  <front>
    <journal-meta><journal-id journal-id-type="publisher">TC</journal-id><journal-title-group>
    <journal-title>The Cryosphere</journal-title>
    <abbrev-journal-title abbrev-type="publisher">TC</abbrev-journal-title><abbrev-journal-title abbrev-type="nlm-ta">The Cryosphere</abbrev-journal-title>
  </journal-title-group><issn pub-type="epub">1994-0424</issn><publisher>
    <publisher-name>Copernicus Publications</publisher-name>
    <publisher-loc>Göttingen, Germany</publisher-loc>
  </publisher></journal-meta>
    <article-meta>
      <article-id pub-id-type="doi">10.5194/tc-17-2811-2023</article-id><title-group><article-title>Modelling point mass balance for the glaciers of the Central European Alps using machine learning techniques</article-title><alt-title>Estimating mass balance using machine learning</alt-title>
      </title-group><?xmltex \runningtitle{Estimating mass balance using machine learning}?><?xmltex \runningauthor{R. Anilkumar et al.}?>
      <contrib-group>
        <contrib contrib-type="author" corresp="yes" rid="aff1 aff2">
          <name><surname>Anilkumar</surname><given-names>Ritu</given-names></name>
          <email>ritu.anilkumar@nesac.gov.in</email>
        </contrib>
        <contrib contrib-type="author" corresp="no" rid="aff2">
          <name><surname>Bharti</surname><given-names>Rishikesh</given-names></name>
          
        </contrib>
        <contrib contrib-type="author" corresp="no" rid="aff1">
          <name><surname>Chutia</surname><given-names>Dibyajyoti</given-names></name>
          
        </contrib>
        <contrib contrib-type="author" corresp="no" rid="aff1">
          <name><surname>Aggarwal</surname><given-names>Shiv Prasad</given-names></name>
          
        </contrib>
        <aff id="aff1"><label>1</label><institution>North Eastern Space Applications Centre, Department of Space, Umiam, Ri Bhoi, Meghalaya, India</institution>
        </aff>
        <aff id="aff2"><label>2</label><institution>Department of Civil Engineering, Indian Institute of Technology Guwahati, Guwahati, Assam, India</institution>
        </aff>
      </contrib-group>
      <author-notes><corresp id="corr1">Ritu Anilkumar (ritu.anilkumar@nesac.gov.in)</corresp></author-notes><pub-date><day>13</day><month>July</month><year>2023</year></pub-date>
      
      <volume>17</volume>
      <issue>7</issue>
      <fpage>2811</fpage><lpage>2828</lpage>
      <history>
        <date date-type="received"><day>11</day><month>October</month><year>2022</year></date>
           <date date-type="rev-request"><day>10</day><month>November</month><year>2022</year></date>
           <date date-type="rev-recd"><day>8</day><month>June</month><year>2023</year></date>
           <date date-type="accepted"><day>12</day><month>June</month><year>2023</year></date>
      </history>
      <permissions>
        <copyright-statement>Copyright: © 2023 Ritu Anilkumar et al.</copyright-statement>
        <copyright-year>2023</copyright-year>
      <license license-type="open-access"><license-p>This work is licensed under the Creative Commons Attribution 4.0 International License. To view a copy of this licence, visit <ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">https://creativecommons.org/licenses/by/4.0/</ext-link></license-p></license></permissions><self-uri xlink:href="https://tc.copernicus.org/articles/17/2811/2023/tc-17-2811-2023.html">This article is available from https://tc.copernicus.org/articles/17/2811/2023/tc-17-2811-2023.html</self-uri><self-uri xlink:href="https://tc.copernicus.org/articles/17/2811/2023/tc-17-2811-2023.pdf">The full text article is available as a PDF file from https://tc.copernicus.org/articles/17/2811/2023/tc-17-2811-2023.pdf</self-uri>
      <abstract><title>Abstract</title>

      <p id="d1e114">Glacier mass balance is typically estimated using a range of in situ measurements, remote sensing measurements, and physical and temperature index modelling techniques. With improved data collection and access to large datasets, data-driven techniques have recently gained prominence in modelling natural processes. The most common data-driven techniques used today are linear regression models and, to some extent, non-linear machine learning models such as artificial neural networks. However, the entire host of capabilities of machine learning modelling has not been applied to glacier mass balance modelling. This study used monthly meteorological data from ERA5-Land to drive four machine learning models: random forest (ensemble tree type), gradient-boosted regressor (ensemble tree type), support vector machine (kernel type), and artificial neural networks (neural type). We also use ordinary least squares linear regression as a baseline model against which to compare the performance of the machine learning models. Further, we assess the requirement of data for each of the models and the requirement for hyperparameter tuning. Finally, the importance of each meteorological variable in the mass balance estimation for each of the models is estimated using permutation importance. All machine learning models outperform the linear regression model. The neural network model depicted a low bias, suggesting the possibility of enhanced results in the event of biased input data. However, the ensemble tree-based models, random forest and gradient-boosted regressor, outperformed all other models in terms of the evaluation metrics and interpretability of the meteorological variables. The gradient-boosted regression model depicted the best coefficient of determination value of <inline-formula><mml:math id="M1" display="inline"><mml:mn mathvariant="normal">0.713</mml:mn></mml:math></inline-formula> and a root mean squared error of <inline-formula><mml:math id="M2" display="inline"><mml:mn mathvariant="normal">1.071</mml:mn></mml:math></inline-formula> m w.e. The feature importance values associated with all machine learning models suggested a high importance of meteorological variables associated with ablation. This is in line with predominantly negative mass balance observations. We conclude that machine learning techniques are promising in estimating glacier mass balance and can incorporate information from more significant meteorological variables as opposed to a simplified set of variables used in temperature index models.</p>
  </abstract>
    </article-meta>
  </front>
<body>
      

<sec id="Ch1.S1" sec-type="intro">
  <label>1</label><title>Introduction</title>
      <?pagebreak page2812?><p id="d1e140">We can visualize glaciers as interactive climate-response systems, with their response described by changes in glacial mass over a given period (e.g. <xref ref-type="bibr" rid="bib1.bibx76" id="altparen.1"/>). Several studies have reported the impact of climate change on glacier mass at a global and regional scale (e.g. <xref ref-type="bibr" rid="bib1.bibx37 bib1.bibx29" id="altparen.2"/>), with repercussions including and not limited to glacial outburst floods and diminishing water supplies. Thus, understanding the response of glacier mass balance to climate change is crucial. Glacier mass balance is most commonly measured via (i) the direct glaciological method, where point measures of gain or loss of glacial ice are obtained and extrapolated for the entire glacier (e.g. <xref ref-type="bibr" rid="bib1.bibx35 bib1.bibx68 bib1.bibx54" id="altparen.3"/>); (ii) the geodetic method, where the change in surface elevation between two time instances for the same portion of the glacier is estimated (e.g. <xref ref-type="bibr" rid="bib1.bibx55 bib1.bibx70 bib1.bibx69 bib1.bibx4 bib1.bibx78" id="altparen.4"/>); and (iii) the indirect remote sensing method, where measured mass balance is correlated with the equilibrium line altitude (ELA) values or accumulation area ratio (AAR) values for time series data (e.g. <xref ref-type="bibr" rid="bib1.bibx7 bib1.bibx16" id="altparen.5"/>). In addition to observational data, simple temperature-index-based or sophisticated physics-based energy balance models (e.g. <xref ref-type="bibr" rid="bib1.bibx19" id="altparen.6"/>) have also been developed. Energy balance models compute all energy fluxes at the glacier surface and require measurements of input variables such as meteorological and other inputs at the glacier scale  (e.g. <xref ref-type="bibr" rid="bib1.bibx20 bib1.bibx62" id="altparen.7"/>). As these models are driven by the physical laws governing energy balance, they provide reliable estimates of glacier mass balance. However, the substantial requirement for ground data to force the model, the sizeable number of parameters to calibrate, and the computational complexity associated with running the model make it cumbersome to use for large areas. Temperature index models use empirical formulations between temperature and melt (e.g. <xref ref-type="bibr" rid="bib1.bibx56" id="altparen.8"/>). The simplicity afforded by these models permits extension to large scales effectively. However, using only temperature and precipitation as inputs can lead to oversimplification. Further, the degree day factors (DDFs) considered in temperature index models are often invariant. But studies such as <xref ref-type="bibr" rid="bib1.bibx19" id="text.9"/>, <xref ref-type="bibr" rid="bib1.bibx45" id="text.10"/>, and <xref ref-type="bibr" rid="bib1.bibx30" id="text.11"/> have observed a decreasing trend in DDF, particularly at higher elevations. <xref ref-type="bibr" rid="bib1.bibx30" id="text.12"/> also report the sensitivity of the DDF under the influence of the changing climate, particularly to solar radiation and albedo.</p>
      <p id="d1e181">With increasing data points available, a new set of data-driven techniques has gained prominence in various domains of Earth sciences. For example, weather prediction (for a review, see <xref ref-type="bibr" rid="bib1.bibx63" id="altparen.13"/>), climate downscaling (e.g. <xref ref-type="bibr" rid="bib1.bibx59" id="altparen.14"/>), and hydrology (e.g. <xref ref-type="bibr" rid="bib1.bibx65" id="altparen.15"/>) have used data-driven models, particularly machine learning (ML) and deep learning (DL) models. Cryospheric studies, too, have adopted the use of deep learning in several prediction problems (see review in <xref ref-type="bibr" rid="bib1.bibx40" id="altparen.16"/>). Applications of deep learning in glaciology range from automatic glacier mapping (e.g. <xref ref-type="bibr" rid="bib1.bibx42 bib1.bibx80" id="altparen.17"/>) to ice thickness measurements (e.g. <xref ref-type="bibr" rid="bib1.bibx75 bib1.bibx31 bib1.bibx25" id="altparen.18"/>), calving front extraction (e.g. <xref ref-type="bibr" rid="bib1.bibx84 bib1.bibx48" id="altparen.19"/>), snow cover mapping (e.g. <xref ref-type="bibr" rid="bib1.bibx53 bib1.bibx32 bib1.bibx23" id="altparen.20"/>), snow depth extraction (e.g. <xref ref-type="bibr" rid="bib1.bibx74 bib1.bibx86" id="altparen.21"/>), and sea and river ice delineation (e.g. <xref ref-type="bibr" rid="bib1.bibx12 bib1.bibx38" id="altparen.22"/>). The use of ML and DL in glacier mass balance estimation is significantly lower. Initial data-driven studies used multivariate linear regression to estimate glacier mass balance from temperature and precipitation <xref ref-type="bibr" rid="bib1.bibx27" id="paren.23"/>. Subsequently, several papers have used linear regression methods for varying inputs such as temperature and pressure <xref ref-type="bibr" rid="bib1.bibx41" id="paren.24"/>, positive degree days, precipitation, temperature, and longwave radiation <xref ref-type="bibr" rid="bib1.bibx36" id="paren.25"/>. Recent studies continue to use linear regression for modelling glacier mass balance. For example, <xref ref-type="bibr" rid="bib1.bibx43" id="text.26"/> used linear regression to study the effect of local, regional, and global parameters on glacier mass balance; <xref ref-type="bibr" rid="bib1.bibx10" id="text.27"/> used linear regression to incorporate the effects of elevation models in the estimation of summer and winter mass balance measurements. <xref ref-type="bibr" rid="bib1.bibx67" id="text.28"/> were the first to use neural networks to estimate glacier mass balance. <xref ref-type="bibr" rid="bib1.bibx5" id="text.29"/> used a least absolute shrinkage and selection operator (LASSO) regression, a linear model, and a non-linear neural network model to simulate glacier mass balance. <xref ref-type="bibr" rid="bib1.bibx67" id="text.30"/>, <xref ref-type="bibr" rid="bib1.bibx73" id="text.31"/>, and <xref ref-type="bibr" rid="bib1.bibx5 bib1.bibx6" id="text.32"/> are some of the few studies reporting consistently better performance of non-linear models over linear models. These studies have largely used neural networks. However, a gamut of ML techniques such as ensemble-based and kernel-based techniques exist which have largely been under-utilized for the purpose of modelling glacier mass balance. This limited utilization of ML models is potentially due to the unavailability of large ground truth datasets required for training the ML models and the perceived black-box nature of ML techniques. We aim to address this by assessing the performance of different ML models for varying training dataset sizes. Further, we aim to shed light on the interpretability of ML models by using permutation importance to explain the relative importance of the input meteorological variables. The interpretability of machine learning models is largely dependent on the input variables provided. Existing data-driven models typically use a subset of topographic and meteorological variables. For example, <xref ref-type="bibr" rid="bib1.bibx27" id="text.33"/> uses temperature, precipitation, and cyclonic/anti-cyclonic activity; <xref ref-type="bibr" rid="bib1.bibx67" id="text.34"/> use precipitation and temperature; and <xref ref-type="bibr" rid="bib1.bibx44" id="text.35"/> use temperature, precipitation, and streamflow. To the extent of the authors' knowledge, no ML-based study has attempted to use a complete set of meteorological variables associated with the energy balance equation. We expand upon this and assess the monthly contributions of each of these meteorological variables in the estimation of glacier mass balance.</p>
      <p id="d1e256">Through this study, we assess the ability of ML models to estimate annual point mass balance. We use an example of each of the following classes of ML models: ensemble regression tree based, kernel based, neural network based, and linear models. Under ensemble regression tree based, we chose one example of boosted and unboosted models. Specifically, we compare the performance of the random forest (RF), gradient-boosted regressor (GBR), support vector machine (SVM), and artificial neural network (ANN) models against a linear regression (LR) model. We also assess the performance for varying dataset sizes, as real-world measurements are limited. Finally, to explain the role of the input features in each of the ML models, we use permutation importance described further in <xref ref-type="bibr" rid="bib1.bibx1" id="text.36"/>. The input features for the models are the monthly mean of 14 meteorological variables associated with the energy balance equation. We obtained the meteorological data from the ERA5-Land reanalysis dataset <xref ref-type="bibr" rid="bib1.bibx50" id="paren.37"/>. The target data used for training the ML models are obtained from the Fluctuations of Glaciers database <xref ref-type="bibr" rid="bib1.bibx77 bib1.bibx83" id="paren.38"/> over the second-order region Alps defined by the Randolph Glacier Inventory under first-order region 11: Central Europe <xref ref-type="bibr" rid="bib1.bibx60" id="paren.39"/>. Section <xref ref-type="sec" rid="Ch1.S2"/> of the article further describes each of these datasets. In this section, we also elucidate the preprocessing steps associated with an ML approach and outline the methodology followed. In Sects. <xref ref-type="sec" rid="Ch1.S3"/> and <xref ref-type="sec" rid="Ch1.S4"/>, we compare the performance of each of the models for various configurations of data availability. We also delve into the interpretability of the models from a feature importance perspective. The specific point we investigate as a part of this study can be summarized as follows:
<list list-type="order"><list-item>
      <p id="d1e280">understand the utility of ML models in the estimation of glacier mass balance using limited real-world datasets</p></list-item><list-item>
      <p id="d1e284">identify specific use cases for different classes of ML models (ensemble tree based, kernel based, neural network based, and linear regression) pertaining to data availability, evaluation metrics, and explainability</p></list-item><list-item>
      <p id="d1e288">investigate the ability of ML models to unravel the underlying physical processes</p></list-item><list-item>
      <p id="d1e292">explain the relative importance of meteorological variables contributing to the mass balance estimation on a monthly basis over the year.</p></list-item></list></p>
</sec>
<?pagebreak page2813?><sec id="Ch1.S2">
  <label>2</label><title>Data and methods</title>
<sec id="Ch1.S2.SS1">
  <label>2.1</label><title>Machine learning modelling</title>
      <p id="d1e310">ML modelling is a data-driven set of modelling techniques. Here, we used a supervised learning framework for regression, where inputs are in the form of monthly meteorological variables, and targets are in the form of point measurements of glacier mass balance. The actual point mass balance measurements are the target data vital to tuning the model parameters. We do this parameter tuning by designing a loss function defining the variation between the actual mass balance measurements, i.e. the target data, and the point mass balance estimates, i.e. the model's output. We start with random initialization of model parameters and fine-tune the parameters to minimize the loss function. For each of the ML models used in the study, we used the mean squared error (MSE) as the loss function. Further, we obtained the features of importance by assessing permutation importance. Figure <xref ref-type="fig" rid="Ch1.F1"/> depicts the complete workflow used for the study. The Supplement files include runs of such experiments that impact all the ML models in an equivalent manner.</p>

      <?xmltex \floatpos{t}?><fig id="Ch1.F1"><?xmltex \currentcnt{1}?><?xmltex \def\figurename{Figure}?><label>Figure 1</label><caption><p id="d1e317">Flowchart of the methodology.</p></caption>
          <?xmltex \igopts{width=236.157874pt}?><graphic xlink:href="https://tc.copernicus.org/articles/17/2811/2023/tc-17-2811-2023-f01.png"/>

        </fig>

      <p id="d1e326">The RF model is an ensemble-based algorithm where the base learner used is a decision (regression or classification) tree <xref ref-type="bibr" rid="bib1.bibx9" id="paren.40"/>. It relies on the principle of bootstrap aggregating or bagging (proposed by <xref ref-type="bibr" rid="bib1.bibx8" id="altparen.41"/>) for the generation of multiple training datasets to be used by each base learner <xref ref-type="bibr" rid="bib1.bibx15" id="paren.42"/>. To illustrate this, assume there are <inline-formula><mml:math id="M3" display="inline"><mml:mrow><mml:msub><mml:mi>N</mml:mi><mml:mi mathvariant="normal">data</mml:mi></mml:msub></mml:mrow></mml:math></inline-formula> samples in the training dataset <inline-formula><mml:math id="M4" display="inline"><mml:mi>D</mml:mi></mml:math></inline-formula>, and a new dataset <inline-formula><mml:math id="M5" display="inline"><mml:mover accent="true"><mml:mi>D</mml:mi><mml:mo stretchy="false" mathvariant="normal">^</mml:mo></mml:mover></mml:math></inline-formula> is generated by sampling <inline-formula><mml:math id="M6" display="inline"><mml:mrow><mml:msub><mml:mi>N</mml:mi><mml:mi mathvariant="normal">data</mml:mi></mml:msub></mml:mrow></mml:math></inline-formula> samples with repetition. In addition to the generation of bootstrapped datasets, the decision trees are generated using a random subset of input features at every impure node of the tree instead of a complete set of features that standard regression trees use.</p>
      <p id="d1e379">Like the RF model, the GBR model is an ensemble-based algorithm where aggregated base learners of decision (classification or regression) trees provide an estimate. However, it differs from the RF model because it uses boosting instead of bagging to construct ensembles. In boosting-based ensembles, base learners are typically weak learners, and the design of subsequent learners is such that the overall error reduces <xref ref-type="bibr" rid="bib1.bibx52 bib1.bibx18" id="paren.43"/>.</p>
      <p id="d1e385">The SVM model is a powerful ML tool that relies on Cover's theorem. The theorem suggests that data that might not be linearly separable in a lower dimensional space can be linearly separable when transformed into a higher dimensional space. In the context of classification, the SVM model uses a kernel to transform the data into a higher dimensional space <xref ref-type="bibr" rid="bib1.bibx13" id="paren.44"/> where linear separability is feasible in the form of a hyperplane and decision boundaries. For this purpose, we use kernels such as polynomial kernel and radial basis function kernel <xref ref-type="bibr" rid="bib1.bibx72" id="paren.45"/>. In the case of regression, the hyperplane represents the best-fit line. Thus, unlike empirical risk minimization, where the difference between the actual and predicted model is optimized, the SVM model for regression uses structural risk minimization by identifying the best-fit line.</p>
      <?pagebreak page2814?><p id="d1e394"><xref ref-type="bibr" rid="bib1.bibx47" id="text.46"/> proposed the NN models as mathematical representations of biological neuron interconnections. <xref ref-type="bibr" rid="bib1.bibx28" id="text.47"/> showed that neural networks with as few as a single hidden layer with a sufficiently large number of neurons, when used with a non-constant unbounded activation function, can function as universal function approximators. Presently, several applications <xref ref-type="bibr" rid="bib1.bibx64 bib1.bibx49 bib1.bibx24" id="paren.48"/> using multiple-layered NN models demonstrate that NNs can infer abstract relationships between features. NN models use weighted combinations of input features in tandem with non-linearities provided by activation functions such as sigmoid, tanh, and rectified linear unit (ReLU), resulting in the model output. The weights of the NN model are the model parameters obtained by optimization of the loss function.</p>
</sec>
<sec id="Ch1.S2.SS2">
  <label>2.2</label><title>Preparation of features and target data</title>
      <p id="d1e413">The most crucial component in ML modelling is the availability of target data to train the model. The target data used for training should be representative of the entire population. Hence, we chose the Fluctuations of Glaciers (FoG) database <xref ref-type="bibr" rid="bib1.bibx77 bib1.bibx83" id="paren.49"/> that contains measured point mass balance information (46 356 data points) globally. The study area is the Randolph Glacier Inventory (RGI) version 6 <xref ref-type="bibr" rid="bib1.bibx60" id="paren.50"/> second-order region Alps under the first-order region 11: Central Europe. This consisted of 15 727 glacier mass balance point measurements. We performed a first-level preprocessing where we considered only annual mass balance measurements (10 102 data points) and measurements from 1950 (9595 data points) onward. We then performed an outlier removal where we considered only those points within 2 standard deviations of the median. This was to avoid the effects of noisy data. We finally used 9166 data points to apply our model.</p>
      <p id="d1e422">The second aspect is the input features used by the model to make predictions. The network of weather stations is sparse over much of the Alpine terrain; hence, reanalysis datasets are recommended <xref ref-type="bibr" rid="bib1.bibx26" id="paren.51"/>. We used the ERA5-Land reanalysis dataset <xref ref-type="bibr" rid="bib1.bibx50" id="paren.52"/>. This dataset was chosen primarily due to its comparatively high spatial resolution. This is in line with the findings of <xref ref-type="bibr" rid="bib1.bibx39" id="text.53"/> and <xref ref-type="bibr" rid="bib1.bibx11" id="text.54"/> that suggest that datasets with higher spatial resolution effectively represent the orographic drag and mountain valley circulation, which in turn results in improved performance for orographically complex terrain. The choice of variables reflected the contribution of the same to the energy balance equation that drives mass balance modelling from a physical standpoint. We considered the following 14 variables for the modelling: the temperature at 2 m, snow density, snow temperature, surface net solar radiation, total precipitation, forecast albedo, surface pressure, surface net solar radiation downwards, snowfall, surface net thermal radiation, snowmelt, surface sensible heat flux, snow depth, and surface latent heat flux (for details, see <xref ref-type="bibr" rid="bib1.bibx51" id="altparen.55"/>). We consider these meteorological variables because of their effect and representation of the accumulation and ablation process and define the variables expected to represent accumulation processes as accumulation variables (e.g. snowfall, forecast albedo) and melt processes as ablation variables (e.g. temperature, solar radiation). The monthly mean of each of the accumulation and ablation variables was considered. Thus, we have 168 total input parameters. For each of these variables, we extracted the data using the nearest neighbour algorithm, using latitude, longitude, and year of the glacier mass balance measurement from the FoG database. Thus, the final dataset has 168 input features and 9166 data points.</p>
      <p id="d1e440">We then normalized the data points using a min–max scaling to ensure the absence of user-conceived bias in the model. We have split the dataset using a random split, where 70 % of the total dataset is used for training the model and 30 % is used for testing the model performance. The training split is used in a 3-fold cross-validation process for tuning the hyperparameters, as described further in Sect. <xref ref-type="sec" rid="Ch1.S2.SS3"/>. Finally, we rescaled the model's predictions to assess the model metrics, such as root mean squared error (RMSE), mean absolute error (MAE), normalized mean squared error (nRMSE), and normalized mean absolute error (nMAE) in the measured point mass balance units.</p>
</sec>
<sec id="Ch1.S2.SS3">
  <label>2.3</label><title>Hyperparameter selection and fine-tuning</title>
      <p id="d1e453">In typical ML workflows, we split the complete dataset (set of features and target data) into training, validation, and testing. We fit the model to the data using the training subset, tune the hyperparameters using the validation subset, and report the independent performance metrics using the testing subset. In our case, we used a 70 %–30 % split for training and testing.
We have considered a hyperparameter grid with all combinations of values that each hyperparameter can take (see Table <xref ref-type="table" rid="Ch1.T1"/>). Rather than using a fixed ratio subset for validation, as was the case with the testing, we divided the training data subset into three equal folds. Two folds are randomly selected as the training set, and the third fold is used for validation. The validation score is noted, and the process is then repeated for the other fold combinations. The mean validation score for each hyperparameter setting obtained from the grid is used for the selection of the optimal hyperparameters.
We compute the validation score as the negative of the RMSE after scaling the target data to a range between 0 and 1. Thus, a more negative validation score results in a more significant error.</p>

<?xmltex \floatpos{t}?><table-wrap id="Ch1.T1" specific-use="star"><?xmltex \currentcnt{1}?><label>Table 1</label><caption><p id="d1e461">Grid of settings used for hyperparameter tuning of each of the models.</p></caption><oasis:table frame="topbot"><oasis:tgroup cols="3">
     <oasis:colspec colnum="1" colname="col1" align="justify" colwidth="4cm"/>
     <oasis:colspec colnum="2" colname="col2" align="justify" colwidth="4cm"/>
     <oasis:colspec colnum="3" colname="col3" align="justify" colwidth="7cm"/>
     <oasis:thead>
       <oasis:row rowsep="1">
         <oasis:entry colname="col1">Machine learning model</oasis:entry>
         <oasis:entry colname="col2">Hyperparameter</oasis:entry>
         <oasis:entry colname="col3">Values</oasis:entry>
       </oasis:row>
     </oasis:thead>
     <oasis:tbody>
       <oasis:row rowsep="1">
         <oasis:entry colname="col1">Random forest</oasis:entry>
         <oasis:entry colname="col2">Number of trees</oasis:entry>
         <oasis:entry colname="col3">10, 20, 50, 100</oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1"/>
         <oasis:entry colname="col2">Number of trees</oasis:entry>
         <oasis:entry colname="col3">50, 100, 200</oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1">Gradient-boosted regressor</oasis:entry>
         <oasis:entry colname="col2">Subsampling</oasis:entry>
         <oasis:entry colname="col3">0.7, 1.0</oasis:entry>
       </oasis:row>
       <oasis:row rowsep="1">
         <oasis:entry colname="col1"/>
         <oasis:entry colname="col2">Maximum depth</oasis:entry>
         <oasis:entry colname="col3">3, 5, 10</oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1"/>
         <oasis:entry colname="col2">Cost</oasis:entry>
         <oasis:entry colname="col3">0.1, 1, 10, 20</oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1">Support vector machine</oasis:entry>
         <oasis:entry colname="col2">Kernels</oasis:entry>
         <oasis:entry colname="col3">Sigmoid, radial basis function, polynomial</oasis:entry>
       </oasis:row>
       <oasis:row rowsep="1">
         <oasis:entry colname="col1"/>
         <oasis:entry colname="col2">Degree (polynomial kernel)</oasis:entry>
         <oasis:entry colname="col3">2, 3, 4, 5</oasis:entry>
       </oasis:row>
       <oasis:row>
         <oasis:entry colname="col1">Artificial neural network</oasis:entry>
         <oasis:entry colname="col2">Number of layers and nodes</oasis:entry>
         <oasis:entry colname="col3">1: 10, 50, 100, 200, 300, 400, 500 <?xmltex \hack{\hfill\break}?>2: (100, 50), (200, 100), (400, 200), (200, 400) <?xmltex \hack{\hfill\break}?>3: (400, 200, 100), (500, 200, 100), (200, 100, 50), (100, 50, 10) <?xmltex \hack{\hfill\break}?>4: (200, 300, 400, 500), (300, 200, 100, 50), (200, 100, 50, 10)</oasis:entry>
       </oasis:row>
     </oasis:tbody>
   </oasis:tgroup></oasis:table><?xmltex \gdef\@currentlabel{1}?></table-wrap>

      <p id="d1e592">For the RF model, we tuned the number of trees. We maintained the maximum depth as indefinite, leading to tree expansion until all nodes were pure. We considered all features to obtain the best split, ensuring minimum bias. As computation for absolute error is slow at each split, we used the squared error as the splitting criterion. This ensured the minimization of the variance after each split. For the GBR model,<?pagebreak page2815?> we tuned the number of trees, maximum depth of each tree (which affects the randomness in the choice of features in each tree), and subsampling ratio (for stochastic gradient boosting). Larger values of maximum depth, such as the indeterminate depth of the RF model, are not used as GBR functions with weak learners to increase the randomness. The SVM model hyperparameter fine-tuning involved kernel selection and a choice of the regularization parameter. Further, in the case of polynomial kernels, the degree of the polynomial was also tuned. For the NN model, we used a fully connected feedforward network where the hyperparameters of the number of layers and number of neurons in a layer were tuned. The activation function ReLU was used to incorporate non-linearity. We used the Adam <xref ref-type="bibr" rid="bib1.bibx34" id="paren.56"/> optimizer to minimize the loss function. The training process was performed for 500 iterations with early stopping in the event of convergence before completing the iterations. The NN models for each set of hyperparameters converged before the completion of the 500 iterations.</p>
</sec>
<sec id="Ch1.S2.SS4">
  <label>2.4</label><title>Performance evaluation</title>
      <p id="d1e606">The testing dataset evaluation metrics used to assess the models' performances are the coefficient of determination (<inline-formula><mml:math id="M7" display="inline"><mml:mrow><mml:msup><mml:mi>R</mml:mi><mml:mn mathvariant="normal">2</mml:mn></mml:msup></mml:mrow></mml:math></inline-formula>) which represents the percentage deviation between the target and model predictions, the RMSE which represents the absolute deviations between the target, and the model predictions. Lower <inline-formula><mml:math id="M8" display="inline"><mml:mrow><mml:msup><mml:mi>R</mml:mi><mml:mn mathvariant="normal">2</mml:mn></mml:msup></mml:mrow></mml:math></inline-formula> values suggest that the model does not represent the targets well. Values close to 1 indicate a strong linear correlation. Lower RMSE values are preferable, as this quantifies the variance between the targets and predicted values. Additionally, we report the slope and additive bias using reduced major axis (RMA) regression. We used RMA regression slope and bias to ensure symmetry about the <inline-formula><mml:math id="M9" display="inline"><mml:mrow><mml:mi>y</mml:mi><mml:mo>=</mml:mo><mml:mn mathvariant="normal">1</mml:mn></mml:mrow></mml:math></inline-formula> line. This is preferable, as there exist uncertainties in both target data and outputs.</p>
      <p id="d1e643">ML models are heavily reliant on the availability of training data. To understand the effect of data availability on the model performance, we perform an experiment on varying the training sizes. We split the original dataset into subsets of iteratively increasing sizes. We partition each subset into training and testing partitions using a <inline-formula><mml:math id="M10" display="inline"><mml:mrow><mml:mn mathvariant="normal">70</mml:mn><mml:mo>:</mml:mo><mml:mn mathvariant="normal">30</mml:mn></mml:mrow></mml:math></inline-formula> ratio. For each subset, we trained all the models using the training partition and computed the evaluation metrics over the testing partition.</p>
</sec>
<sec id="Ch1.S2.SS5">
  <label>2.5</label><title>Feature importance</title>
      <p id="d1e667">The feature importance is represented using permutation importance described in <xref ref-type="bibr" rid="bib1.bibx1" id="text.57"/>. Here, we disregard individual features from the model at each iteration and recorded the reduction in evaluation score. This is repeated for each input feature. We normalize the obtained permutation importance for each model and express the importance of each input meteorological variable as a percentage. A comparative analysis of the obtained feature importance is performed on percentage importance associated with the accumulation months (November to March) and the ablation months (June–September) is summed and graphically represented for each model in Fig. <xref ref-type="fig" rid="Ch1.F6"/>.</p>
</sec>
</sec>
<sec id="Ch1.S3">
  <label>3</label><title>Results</title>
      <p id="d1e684">This section describes the major outcomes of the study categorized as the role of dataset size for the effective training of each ML model (see Fig. <xref ref-type="fig" rid="Ch1.F2"/>), the performance, and the feature importance associated with each ML model. Figure <xref ref-type="fig" rid="Ch1.F3"/> represents the comparative performance of each of the models in terms of the accuracy metrics RMSE, <inline-formula><mml:math id="M11" display="inline"><mml:mrow><mml:msup><mml:mi>R</mml:mi><mml:mn mathvariant="normal">2</mml:mn></mml:msup></mml:mrow></mml:math></inline-formula>, slope, and additive bias. A scatter plot of modelled point mass balance and target data is represented in Fig. <xref ref-type="fig" rid="Ch1.F4"/>. Figure <xref ref-type="fig" rid="Ch1.F5"/>a, b, c, and d represent<?pagebreak page2816?> the hyperparameter tuning associated with the models. The feature importance for all input variables summed over the ablation and accumulation months is represented in Fig. <xref ref-type="fig" rid="Ch1.F6"/>.</p>

      <?xmltex \floatpos{t}?><fig id="Ch1.F2" specific-use="star"><?xmltex \currentcnt{2}?><?xmltex \def\figurename{Figure}?><label>Figure 2</label><caption><p id="d1e711"><bold>(a)</bold> Histogram depicting the distribution of the glacier mass balance measurements used for the study. <bold>(b)</bold> Box and whisker plot depicting the training and testing MAE (in mm w.e.) and <inline-formula><mml:math id="M12" display="inline"><mml:mi>r</mml:mi></mml:math></inline-formula> values for varying the size of the training dataset for each of the models. The box represents the quartiles 1 to 3, and the whiskers represent the rest of the distribution ignoring outliers. <bold>(c)</bold> Modelwise training mean absolute error (in mm w.e.) for varying the size of the training dataset size. <bold>(d)</bold> Modelwise testing mean absolute error (in mm w.e.) for varying the size of the training dataset size. Note, the training dataset size is expressed as a percentage of the largest size of the training dataset, i.e. 6416 data points.</p></caption>
        <?xmltex \igopts{width=398.338583pt}?><graphic xlink:href="https://tc.copernicus.org/articles/17/2811/2023/tc-17-2811-2023-f02.png"/>

      </fig>

      <?xmltex \floatpos{t}?><fig id="Ch1.F3" specific-use="star"><?xmltex \currentcnt{3}?><?xmltex \def\figurename{Figure}?><label>Figure 3</label><caption><p id="d1e740">Training and testing performance of each of the models: random forest (RF), gradient-boosted regression (GBR), support vector machine (SVM), artificial neural network (ANN), and linear regression (LR) depicted using the performance metrics <bold>(a)</bold> root mean squared error, <bold>(b)</bold> coefficient of determination, <bold>(c)</bold> slope, and <bold>(d)</bold> additive bias.</p></caption>
        <?xmltex \igopts{width=398.338583pt}?><graphic xlink:href="https://tc.copernicus.org/articles/17/2811/2023/tc-17-2811-2023-f03.png"/>

      </fig>

      <?xmltex \floatpos{t}?><fig id="Ch1.F4" specific-use="star"><?xmltex \currentcnt{4}?><?xmltex \def\figurename{Figure}?><label>Figure 4</label><caption><p id="d1e764">Testing scatter plot depicting the performance for each of the models: random forest (RF), gradient-boosted regression (GBR), support vector machine (SVM), artificial neural network (ANN), and linear regression (LR).</p></caption>
        <?xmltex \igopts{width=341.433071pt}?><graphic xlink:href="https://tc.copernicus.org/articles/17/2811/2023/tc-17-2811-2023-f04.png"/>

      </fig>

      <?xmltex \floatpos{t}?><fig id="Ch1.F5" specific-use="star"><?xmltex \currentcnt{5}?><?xmltex \def\figurename{Figure}?><label>Figure 5</label><caption><p id="d1e775">Hyperparameter tuning for the <bold>(a)</bold> GBR model varying the number of trees, maximum depth of each tree, and subsampling fraction; <bold>(b)</bold> SVM model varying the penalty parameter and kernel as well as degree in the case of the polynomial kernel; <bold>(c)</bold> NN model varying the number of neurons in a single hidden layer; and <bold>(d)</bold> NN model varying the number of hidden layers. The validation score used is the negative scaled RMSE, which is the negative of the normalized RMSE values that can easily be used to rank the hyperparameter settings.</p></caption>
        <?xmltex \igopts{width=398.338583pt}?><graphic xlink:href="https://tc.copernicus.org/articles/17/2811/2023/tc-17-2811-2023-f05.png"/>

      </fig>

      <?xmltex \floatpos{t}?><fig id="Ch1.F6" specific-use="star"><?xmltex \currentcnt{6}?><?xmltex \def\figurename{Figure}?><label>Figure 6</label><caption><p id="d1e798">Percentage importance of all features summed over the accumulation and ablation season for the models: random forest (RF), gradient-boosted regression (GBR), support vector machine (SVM), artificial neural network (ANN), and linear regression (LR).</p></caption>
        <?xmltex \igopts{width=426.791339pt}?><graphic xlink:href="https://tc.copernicus.org/articles/17/2811/2023/tc-17-2811-2023-f06.png"/>

      </fig>

<sec id="Ch1.S3.SS1">
  <label>3.1</label><title>Role of training dataset size</title>
      <p id="d1e814">The number of samples required for training the ML models depends upon the complexity of the model. Thus, each of the models used in this study is variably sensible to the number of training samples. We use the evaluation metrics of RMSE and the correlation coefficient to assess the requirement of training samples for each of the models. Figure <xref ref-type="fig" rid="Ch1.F2"/> depicts the training and testing metrics varying with the size of the training dataset. The training metrics do not show significant change after 20 %–30 % of the training dataset size for the LR, RF, GBR, and SVM models and after 40 % for the NN model. This illustrates the larger number of trainable parameters, resulting in the requirement of larger datasets for artificial neural networks for training. The testing performance of each of the models does not show significant change for training dataset sizes larger than 50 %. We observe that, while a downward trend is evident with the addition of new data, the rate of improvement is slower.</p>
      <p id="d1e819">It is interesting to note that RF, GBR, and LR models see an increase in training MAE as opposed to a consistent decrease in testing MAE with increasing training samples. This depicts the tendency of these models to overfit the training samples in the case of smaller datasets. This is evident when observing the order of variation in the training and testing evaluation metric for smaller datasets; e.g. GBR depicts a training MAE of <inline-formula><mml:math id="M13" display="inline"><mml:mn mathvariant="normal">357</mml:mn></mml:math></inline-formula> mm w.e. and a testing MAE of <inline-formula><mml:math id="M14" display="inline"><mml:mn mathvariant="normal">1183</mml:mn></mml:math></inline-formula> mm w.e. at 10 % training dataset size and training MAE of <inline-formula><mml:math id="M15" display="inline"><mml:mn mathvariant="normal">659</mml:mn></mml:math></inline-formula> mm w.e. and a testing MAE of <inline-formula><mml:math id="M16" display="inline"><mml:mn mathvariant="normal">774</mml:mn></mml:math></inline-formula> mm w.e. at 100 % training dataset size. Thus, care must be taken when using RF and GBR for smaller datasets, as they are susceptible to overfitting. The performance of the LR model deteriorates for training, and testing performance is also poor. This is not due to overfitting but due to the inability of the model to explain the complex relationship between the inputs and the target. NN requires larger datasets for the training of the model. Figure 2b depicts the superior performance of RF, GBR, and SVM in the event of limited dataset availability. However, we have seen that RF and GBR show a marked increase in training MAE with increasing training samples, which suggests overfitting to limited datasets. Thus, SVM is more robust to smaller datasets.</p>
</sec>
<sec id="Ch1.S3.SS2">
  <label>3.2</label><title>Performance of RF modelling</title>
      <p id="d1e858">The best-performing RF model resulted in a testing RMSE value of <inline-formula><mml:math id="M17" display="inline"><mml:mn mathvariant="normal">1083</mml:mn></mml:math></inline-formula> mm w.e. and an <inline-formula><mml:math id="M18" display="inline"><mml:mrow><mml:msup><mml:mi>R</mml:mi><mml:mn mathvariant="normal">2</mml:mn></mml:msup></mml:mrow></mml:math></inline-formula> value of <inline-formula><mml:math id="M19" display="inline"><mml:mn mathvariant="normal">0.71</mml:mn></mml:math></inline-formula>. The testing MAE value is <inline-formula><mml:math id="M20" display="inline"><mml:mn mathvariant="normal">782</mml:mn></mml:math></inline-formula> mm w.e., and the testing nRMSE and nMAE values are <inline-formula><mml:math id="M21" display="inline"><mml:mn mathvariant="normal">0.55</mml:mn></mml:math></inline-formula> and <inline-formula><mml:math id="M22" display="inline"><mml:mn mathvariant="normal">0.40</mml:mn></mml:math></inline-formula> respectively. The training RMSE value is <inline-formula><mml:math id="M23" display="inline"><mml:mn mathvariant="normal">934</mml:mn></mml:math></inline-formula> mm w.e., MAE value is <inline-formula><mml:math id="M24" display="inline"><mml:mn mathvariant="normal">672</mml:mn></mml:math></inline-formula> mm w.e., nRMSE is <inline-formula><mml:math id="M25" display="inline"><mml:mn mathvariant="normal">0.48</mml:mn></mml:math></inline-formula>, nMAE is <inline-formula><mml:math id="M26" display="inline"><mml:mn mathvariant="normal">0.34</mml:mn></mml:math></inline-formula>, and <inline-formula><mml:math id="M27" display="inline"><mml:mrow><mml:msup><mml:mi>R</mml:mi><mml:mn mathvariant="normal">2</mml:mn></mml:msup></mml:mrow></mml:math></inline-formula> value is <inline-formula><mml:math id="M28" display="inline"><mml:mn mathvariant="normal">0.80</mml:mn></mml:math></inline-formula>. We observe that hyperparameter tuning is not important, and no major variations were observed upon changing the number of estimators. The slope of RF was closest to 1, with a value of <inline-formula><mml:math id="M29" display="inline"><mml:mn mathvariant="normal">0.752</mml:mn></mml:math></inline-formula> for the training samples and 0.744 for the testing samples. Both the training and testing additive bias were negative, suggesting the model underestimated the point mass balance (Fig. <xref ref-type="fig" rid="Ch1.F3"/>).</p>
      <p id="d1e964">Feature importance analysis using permutation importance considering the 17 (10 % of all features) most essential features indicates that the RF model is highly influenced by downward solar radiation in January; net solar radiation in July; downward thermal radiation in June; temperature at 2 m in June; forecast albedo in February and December; snow depth in January and  July; snow density and snowmelt in July; sensible heat flux in December, January, March, and May; latent heat flux in August; and surface pressure in June and July. Permutation importance for the RF model summed over the accumulation months had the highest importance scores for sensible heat flux followed by downward solar radiation and forecast albedo. Each of these variables depict a summed percentage importance between 6 %–9 %. Snow depth and pressure are also important, with a summed percentage importance between 3 %–6 %. For the ablation months, only pressure is observed to have a summed percentage importance greater than 6 %. Sensible heat flux, net solar radiation, latent heat flux, snow depth, forecast albedo, snow density, and temperature at 2 m display a summed percentage importance between 3 %–6 %.</p>
</sec>
<sec id="Ch1.S3.SS3">
  <label>3.3</label><title>Performance of GBR modelling</title>
      <p id="d1e976">Tuning the maximum depth permitted for each weak learner tree was important in estimating the best model, and varying the number of weak learner trees during hyperparameter tuning improved performance in the case of smaller depths of the weak learners. Deeper tree structures did not significantly change the model's performance upon changing the number of trees. Stochastic gradient boosting (subsampling at 0.7) resulted in reduced performance. The hyperparameter combination of the best-performing GBR model is 100 trees with a maximum depth of five nodes (Fig. <xref ref-type="fig" rid="Ch1.F5"/>a). The best-performing GBR model resulted in a testing RMSE value of <inline-formula><mml:math id="M30" display="inline"><mml:mn mathvariant="normal">1071</mml:mn></mml:math></inline-formula> mm w.e. and an <inline-formula><mml:math id="M31" display="inline"><mml:mrow><mml:msup><mml:mi>R</mml:mi><mml:mn mathvariant="normal">2</mml:mn></mml:msup></mml:mrow></mml:math></inline-formula> value of <inline-formula><mml:math id="M32" display="inline"><mml:mn mathvariant="normal">0.71</mml:mn></mml:math></inline-formula>. The testing MAE value is <inline-formula><mml:math id="M33" display="inline"><mml:mn mathvariant="normal">774</mml:mn></mml:math></inline-formula> mm w.e., and the testing nRMSE and nMAE are <inline-formula><mml:math id="M34" display="inline"><mml:mn mathvariant="normal">0.55</mml:mn></mml:math></inline-formula> and <inline-formula><mml:math id="M35" display="inline"><mml:mn mathvariant="normal">0.39</mml:mn></mml:math></inline-formula> respectively. The training RMSE value is <inline-formula><mml:math id="M36" display="inline"><mml:mn mathvariant="normal">759</mml:mn></mml:math></inline-formula> mm w.e.,  MAE value is <inline-formula><mml:math id="M37" display="inline"><mml:mn mathvariant="normal">659</mml:mn></mml:math></inline-formula> mm w.e., nRMSE is <inline-formula><mml:math id="M38" display="inline"><mml:mn mathvariant="normal">0.39</mml:mn></mml:math></inline-formula>, nMAE is <inline-formula><mml:math id="M39" display="inline"><mml:mn mathvariant="normal">0.34</mml:mn></mml:math></inline-formula>, and <inline-formula><mml:math id="M40" display="inline"><mml:mrow><mml:msup><mml:mi>R</mml:mi><mml:mn mathvariant="normal">2</mml:mn></mml:msup></mml:mrow></mml:math></inline-formula> value is <inline-formula><mml:math id="M41" display="inline"><mml:mn mathvariant="normal">0.80</mml:mn></mml:math></inline-formula>.</p>
      <p id="d1e1075">The most important meteorological inputs for the GBR model are snowfall in July; downward solar radiation in January and December; forecast albedo in December, January, February, March, and May; sensible heat flux in January, March, May, November, and December; temperature at 2 m in June and August; snow depth in June; and surface pressure in August. Note the marked importance associated with ablation meteorological variables and the months associated<?pagebreak page2817?> with ablation. Permutation importance expressed as a percentage and summed over the accumulation months depicts the most importance to forecast albedo, followed by sensible heat flux, with both variables depicting a summed percentage importance greater than 10 %. Among other meteorological variables, downward solar radiation, net solar radiation, and snow depth in the accumulation months are also important. The ablation months depict higher summed importance values, with forecast albedo in these months prominent. Sensible heat flux, latent heat flux, surface pressure, snowfall, snow depth, and temperature at 2 m above the surface are also important.</p>
</sec>
<sec id="Ch1.S3.SS4">
  <label>3.4</label><title>Performance of SVM modelling</title>
      <?pagebreak page2818?><p id="d1e1086">The SVM model depicted large fluctuations in the validation score with changes in the hyperparameters. This is represented in Fig. <xref ref-type="fig" rid="Ch1.F5"/>b. We considered the hyperparameters of the kernel, degree (for polynomial kernel), and regularization (penalty) factor. The sigmoid kernel resulted in evaluation metrics markedly poorer than the radial basis function (RBF) kernel and polynomial kernels. The sigmoid kernel was excluded from the graphical representation of the validation score to emphasize the variations observed in the other kernels. The polynomial kernel at larger degrees consistently performed better than the RBF kernel in the case of regularization tuning lower than 1. For larger regularization parameters, the RBF kernels demonstrated better performance. The best-performing model in this study is the RBF kernel (penalty factor: 10.0). Figure <xref ref-type="fig" rid="Ch1.F5"/>b depicts the results of hyperparameter tuning for the SVM kernel. The testing RMSE values for the model are <inline-formula><mml:math id="M42" display="inline"><mml:mn mathvariant="normal">1085</mml:mn></mml:math></inline-formula> mm w.e., and the <inline-formula><mml:math id="M43" display="inline"><mml:mrow><mml:msup><mml:mi>R</mml:mi><mml:mn mathvariant="normal">2</mml:mn></mml:msup></mml:mrow></mml:math></inline-formula> value is <inline-formula><mml:math id="M44" display="inline"><mml:mn mathvariant="normal">0.70</mml:mn></mml:math></inline-formula>. The testing MAE value is <inline-formula><mml:math id="M45" display="inline"><mml:mn mathvariant="normal">836</mml:mn></mml:math></inline-formula> mm w.e., and the testing nRMSE and nMAE are <inline-formula><mml:math id="M46" display="inline"><mml:mn mathvariant="normal">0.56</mml:mn></mml:math></inline-formula> and <inline-formula><mml:math id="M47" display="inline"><mml:mn mathvariant="normal">0.43</mml:mn></mml:math></inline-formula> respectively. The training RMSE value is <inline-formula><mml:math id="M48" display="inline"><mml:mn mathvariant="normal">727</mml:mn></mml:math></inline-formula> mm w.e.,  MAE value is <inline-formula><mml:math id="M49" display="inline"><mml:mn mathvariant="normal">727</mml:mn></mml:math></inline-formula> mm w.e., nRMSE is <inline-formula><mml:math id="M50" display="inline"><mml:mn mathvariant="normal">0.37</mml:mn></mml:math></inline-formula>, nMAE is <inline-formula><mml:math id="M51" display="inline"><mml:mn mathvariant="normal">0.37</mml:mn></mml:math></inline-formula>, and <inline-formula><mml:math id="M52" display="inline"><mml:mrow><mml:msup><mml:mi>R</mml:mi><mml:mn mathvariant="normal">2</mml:mn></mml:msup></mml:mrow></mml:math></inline-formula> value is <inline-formula><mml:math id="M53" display="inline"><mml:mn mathvariant="normal">0.76</mml:mn></mml:math></inline-formula>.</p>
      <p id="d1e1187">The permutation importance associated with the sensible heat flux in March is most important, as is the sensible heat flux associated with April, May, June, and December. Latent heat flux in August and October is important. Snowfall in October and snow density for the months of November, December, and January are important. The temperature at 2 m above the surface in June and July; downward solar radiation in December; and forecast albedo in August, October, and December are important. Summing the percentage importance over the accumulation and ablation months, we observe that sensible heat flux in the accumulation months is most important, followed by snow density and downward solar radiation. These three variables depict a summed percentage importance of more than 6 %. The temperature at 2 m a.g.l. and forecast albedo depict an importance between 3 %–6 % for the accumulation months. For the ablation months, sensible heat flux continues to depict a summed percentage importance of more than 6 %. Latent heat flux, snow density, forecast albedo, and temperature at 2 m above the surface also depict a summed percentage importance between 3 %–6 %.</p>
</sec>
<sec id="Ch1.S3.SS5">
  <label>3.5</label><title>Performance of NN modelling</title>
      <p id="d1e1198">The NN model performance is highly susceptible to hyperparameter selection. We varied the number of hidden layers in the network and the number of neurons in each hidden layer. Figure <xref ref-type="fig" rid="Ch1.F5"/>c and d depict the variation in performance of the model for each of these cases. On the left is the variation in the number of neurons for a single hidden layer. A larger number of hidden neurons permits more combinations<?pagebreak page2819?> of the inputs that can affect the targets. The improved performance with the increasing size of neurons illustrates the role of the complexity of the model in estimating mass balance. Increasing the number of layers also affects the performance of the NN model, with the best performance obtained using two hidden layers. This further emphasizes the importance of incorporating non-linear elements in estimating point mass balance. A larger number of hidden layers did not significantly improve performance, as the larger number of parameters demanded a larger training dataset to avoid overfitting and to complete the training. The testing RMSE values for the best-performing model are <inline-formula><mml:math id="M54" display="inline"><mml:mn mathvariant="normal">1096</mml:mn></mml:math></inline-formula> mm w.e. and <inline-formula><mml:math id="M55" display="inline"><mml:mrow><mml:msup><mml:mi>R</mml:mi><mml:mn mathvariant="normal">2</mml:mn></mml:msup></mml:mrow></mml:math></inline-formula> value is <inline-formula><mml:math id="M56" display="inline"><mml:mn mathvariant="normal">0.70</mml:mn></mml:math></inline-formula>. The testing MAE value is <inline-formula><mml:math id="M57" display="inline"><mml:mn mathvariant="normal">836</mml:mn></mml:math></inline-formula> mm w.e., and the testing nRMSE and nMAE are <inline-formula><mml:math id="M58" display="inline"><mml:mn mathvariant="normal">0.56</mml:mn></mml:math></inline-formula> and <inline-formula><mml:math id="M59" display="inline"><mml:mn mathvariant="normal">0.43</mml:mn></mml:math></inline-formula> respectively. The training RMSE value is <inline-formula><mml:math id="M60" display="inline"><mml:mn mathvariant="normal">773</mml:mn></mml:math></inline-formula> mm w.e.,  MAE value is <inline-formula><mml:math id="M61" display="inline"><mml:mn mathvariant="normal">773</mml:mn></mml:math></inline-formula> mm w.e., nRMSE is <inline-formula><mml:math id="M62" display="inline"><mml:mn mathvariant="normal">0.39</mml:mn></mml:math></inline-formula>, nMAE is <inline-formula><mml:math id="M63" display="inline"><mml:mn mathvariant="normal">0.39</mml:mn></mml:math></inline-formula>, and <inline-formula><mml:math id="M64" display="inline"><mml:mrow><mml:msup><mml:mi>R</mml:mi><mml:mn mathvariant="normal">2</mml:mn></mml:msup></mml:mrow></mml:math></inline-formula> value is <inline-formula><mml:math id="M65" display="inline"><mml:mn mathvariant="normal">0.76</mml:mn></mml:math></inline-formula>.</p>
      <p id="d1e1297">The most important meteorological variables in terms of the percentage permutation importance for the NN model<?pagebreak page2820?> are the sensible heat flux for March, April, and May; latent heat flux in July; surface pressure in February; net solar radiation in May and September; downward solar radiation in December; and forecast albedo in July. The snow density in December and the snow depth in January, February, April, July, September, October, and December are important. We see that snow depth across the year dominates the important meteorological inputs for this model. Upon summing the percentage importance for the accumulation and ablation months, we observe that snow depth is the most important for both accumulation and ablation months. Snow density, pressure, sensible heat flux, and downward solar radiation are also important in the accumulation months, with a summed percentage importance value between 3 %–6 %. For the ablation months, net solar radiation is also important. Snow density, forecast albedo, latent heat flux, and sensible heat flux are also important, with summed percentage importance values between 3 %–6 %.</p>
</sec>
<?pagebreak page2821?><sec id="Ch1.S3.SS6">
  <label>3.6</label><title>Performance of LR modelling</title>
      <p id="d1e1308">The testing RMSE values for the LR model are <inline-formula><mml:math id="M66" display="inline"><mml:mn mathvariant="normal">1248</mml:mn></mml:math></inline-formula> mm w.e. and <inline-formula><mml:math id="M67" display="inline"><mml:mrow><mml:msup><mml:mi>R</mml:mi><mml:mn mathvariant="normal">2</mml:mn></mml:msup></mml:mrow></mml:math></inline-formula> value is <inline-formula><mml:math id="M68" display="inline"><mml:mn mathvariant="normal">0.58</mml:mn></mml:math></inline-formula>, and the training RMSE values are <inline-formula><mml:math id="M69" display="inline"><mml:mn mathvariant="normal">1197</mml:mn></mml:math></inline-formula> mm w.e. and <inline-formula><mml:math id="M70" display="inline"><mml:mrow><mml:msup><mml:mi>R</mml:mi><mml:mn mathvariant="normal">2</mml:mn></mml:msup></mml:mrow></mml:math></inline-formula> value is <inline-formula><mml:math id="M71" display="inline"><mml:mn mathvariant="normal">0.61</mml:mn></mml:math></inline-formula> (Fig. <xref ref-type="fig" rid="Ch1.F3"/>). The testing MAE value is <inline-formula><mml:math id="M72" display="inline"><mml:mn mathvariant="normal">941</mml:mn></mml:math></inline-formula> mm w.e., and the nRMSE and nMAE are <inline-formula><mml:math id="M73" display="inline"><mml:mn mathvariant="normal">0.64</mml:mn></mml:math></inline-formula> and <inline-formula><mml:math id="M74" display="inline"><mml:mn mathvariant="normal">0.48</mml:mn></mml:math></inline-formula> respectively. The training MAE value is <inline-formula><mml:math id="M75" display="inline"><mml:mn mathvariant="normal">935</mml:mn></mml:math></inline-formula> mm w.e., nRMSE is <inline-formula><mml:math id="M76" display="inline"><mml:mn mathvariant="normal">0.61</mml:mn></mml:math></inline-formula>, and nMAE is <inline-formula><mml:math id="M77" display="inline"><mml:mn mathvariant="normal">0.48</mml:mn></mml:math></inline-formula>.</p>
      <p id="d1e1407">Snow depth over most of the year is the most important feature for the model, with surface pressure also playing an important role. Other features do not depict as high an importance value. However, relative importance varies across the months.</p>
</sec>
</sec>
<sec id="Ch1.S4">
  <label>4</label><title>Discussion</title>
<sec id="Ch1.S4.SS1">
  <label>4.1</label><title>Comparison of model performance and associated errors</title>
      <?pagebreak page2822?><p id="d1e1426">The performance
of each of the models was evaluated using an independent test dataset. The GBR model resulted in the best testing performance MAE, RMSE, and <inline-formula><mml:math id="M78" display="inline"><mml:mrow><mml:msup><mml:mi>R</mml:mi><mml:mn mathvariant="normal">2</mml:mn></mml:msup></mml:mrow></mml:math></inline-formula> values, outperforming the RF model and SVM and NN models. Neural networks resulted in better bias performance. RF, GBR, SVM, and NN significantly improve upon the LR model's metrics. The ability of all non-linear models to outperform the linear model is further depicted in each model's scatter plot (Fig. <xref ref-type="fig" rid="Ch1.F4"/>). This is in agreement with similar studies in other domains, such as <xref ref-type="bibr" rid="bib1.bibx33" id="text.58"/>, who showed that tree-based models such as RF were preferable to LR models for the bias correction of snow water equivalent, and <xref ref-type="bibr" rid="bib1.bibx58" id="text.59"/>, who depicted the efficacy of non-linear models in estimation of streamflow when compared to linear models.</p>
      <p id="d1e1448">The performance of all models is affected by the uncertainties associated with the input features and targets. Inherent errors exist in point mass balance estimates, as heterogeneity is not captured sufficiently by the available measurements <xref ref-type="bibr" rid="bib1.bibx82 bib1.bibx71" id="paren.60"/>. Of the 727 locations with uncertainty estimation performed, we note a mean uncertainty of 62 mm w.e., which can adversely impact performance evaluation. The uncertainty estimates for the remaining point locations are unknown; hence, their impact is not constrained. In this study, we did not consider the effect of topography and debris cover for the models. This can lead to inflated RMSE values.</p>
      <p id="d1e1454">Further, the use of input meteorological reanalysis data can result in bias, especially in locations without sufficient ground stations <xref ref-type="bibr" rid="bib1.bibx81 bib1.bibx22" id="paren.61"/>. Specifically for the use of ERA5-Land data in complex terrain, <xref ref-type="bibr" rid="bib1.bibx79" id="text.62"/> report that while ERA5-Land represents the intra-annual variations in precipitation characteristics, there is a positive bias in the precipitation variables. Similarly, in the case of temperature, <xref ref-type="bibr" rid="bib1.bibx85" id="text.63"/> show through correlation and RMSE analysis that while the ERA5-Land  dataset captures the temperature trends effectively, the magnitude of the values is not well represented. Thus, we suggest using a bias correction step such as that proposed by <xref ref-type="bibr" rid="bib1.bibx14" id="text.64"/> in the case of RF, GBR, and SVM models. Moreover, the reanalysis data do not fully reflect point scale data, as they have a coarse resolution. <xref ref-type="bibr" rid="bib1.bibx39" id="text.65"/> depict the impact of resolution in simulating drivers of local weather in complex terrain and show that coarser resolutions do not account for orographic drag. Approaches such as using a scaling factor or lapse rates have been attempted in studies (e.g. <xref ref-type="bibr" rid="bib1.bibx57 bib1.bibx46" id="altparen.66"/>). However, these studies largely utilize precipitation and temperature as inputs, the scaling of which with elevation is fairly straightforward. Choosing appropriate scaling factors for other meteorological variables that drive glacier mass balance (e.g. sensible and latent heat fluxes, albedo) is not intuitive. We note that the effects of the larger scale of the input variable will persist in the model. However, these effects will be consistent across all the models. Thus, the effect of the input variable scale is represented by the uncertainty of all models, and a relative analysis of the performance of models will remain well founded.</p>
</sec>
<sec id="Ch1.S4.SS2">
  <label>4.2</label><title>Role of training dataset availability</title>
      <p id="d1e1484">The testing performance improves by increasing the number of training samples. We observe that for a larger number of data points, marginal improvement is observed upon increasing the number of samples further. The reduction in the rate of improvement for all models suggests that all models have been successfully trained. However, the marginal improvements observed suggest that a potential improvement in model performance is possible when including more data samples. The RF and GBR models overfit the training samples in the case of smaller datasets. The NN model training and testing metrics depict improved performance with training size. The NN model had the most trainable parameters and hence is the most data intensive. A larger number of training samples is essential for models with a larger number of trainable parameters. The training performance of the LR model deteriorates with increasing training samples. While the graph (LR model of Fig. <xref ref-type="fig" rid="Ch1.F2"/>) appears similar to the RF and GBR training graphs, the relatively close training and testing metric values suggest that overfitting is not the likely cause. Rather, it suggests that the model cannot explain the non-linear relationship between the inputs and the target.</p>
      <p id="d1e1489">Further, Fig. <xref ref-type="fig" rid="Ch1.F2"/> represents each model's variation in training and testing evaluation metrics. Each model was trained and tested over each dataset size. For each model, the box plots are generated by utilizing the outcome of the models developed using varying training dataset sizes. The training performance, as expected, is better than the testing performance, as the model parameters are tuned to fit this dataset. The range of values is more extensive for the testing errors as a result of overfitting in the case of smaller datasets. In such cases, the use of the SVM model yields better results.</p>
</sec>
<sec id="Ch1.S4.SS3">
  <label>4.3</label><title>Unravelling the physics using machine-learning-derived feature importance</title>
      <p id="d1e1502">Assuming a winter accumulation-type glacier, we expect the months of November to March to be dominated by accumulation processes and June to September to be dominated by ablation processes. Analysis of the permutation importance (by percentage) of the features of each model was studied month-wise based on a physical understanding of which season-specific features will be most important. Figure <xref ref-type="fig" rid="Ch1.F6"/> represents the summed feature importance for each input variable in the accumulation and ablation months. We sum the percentage importance rather than the feature importance values to permit comparison between models. We expect temperature (2 m) for ablation seasons to be significant compared to temperatures in the accumulation season. This is not well reflected when using the LR model. While all the ML models show the reduced importance of temperature in the accumulation months, it is most pronounced in the case of the RF and GBR models. A similar trend is expected for the downward thermal radiation and snowmelt. Here too the LR model<?pagebreak page2823?> does not reflect the expected outcome. All ML models depict reduced importance in the accumulation months, with a pronounced reduction observed in the RF and GBR models. In the case of snowmelt, all ML models and the LR model follow the expected response. Snow depth throughout the year is important when considering snow density. We expect the depth in the ablation months to be important. All models portray this except the SVM model. We observe that the LR model relies heavily on snow depth to estimate the mass balance. The SVM model reports the exaggerated importance of snow density in the accumulation months. While we expect more importance regarding precipitation terms such as total precipitation and snowfall in the accumulation months, we do not observe this for any model. The LR model did show a weak reduction in the importance of total precipitation and snowfall. However, the ML models showed only a weak reduction or a weak increase in importance. This is possibly a result of the scale of the meteorological variables used not sufficiently representing the influence of orographic water vapour transport that results in precipitation <xref ref-type="bibr" rid="bib1.bibx39 bib1.bibx11" id="paren.67"/>.</p>
      <p id="d1e1510">Net solar radiation and albedo are important ablation components.  Albedo over snow-covered regions is higher than that of exposed ice or firn. At higher elevations and in summer months, we expect lower albedo values. Thus, variations in albedo are significant. In the case of ERA5-Land, the forecast albedo variable represents both the direct and diffuse radiation incident on the surface, with values dependent on the land cover type. It is calculated using a weight applied to the albedo in the UV–visible and infrared spectral regions. The albedo of snow and ice land covers differs in the UV–visible and infrared spectral regions. This makes forecast albedo more important than broadband albedo, which depends only on the surface net solar radiation and the surface solar radiation downwards. The expected importance of the albedo is observed in the RF, GBR, NN, and SVM model. LR models, in contrast, depict very low importance of albedo for the accumulation months. Thus, we see that the ML models represent the importance of the ablation features well. This is in agreement with the predominantly negative mass balance observed in in situ measurements.</p>
      <p id="d1e1513">We can observe that the importance associated with the meteorological variables is not dominated solely by total precipitation and temperature, as with temperature index models. Thus, ML modelling can represent the contributions of a complete set of variables with lesser complexity and ease of use than physical models. This also emphasizes the requirement for ML models to use all meteorological variables of interest, as opposed to a subset of them. This is the case with studies such as <xref ref-type="bibr" rid="bib1.bibx5" id="text.68"/>. Further, our results agree with the studies conducted by <xref ref-type="bibr" rid="bib1.bibx67" id="text.69"/> and <xref ref-type="bibr" rid="bib1.bibx6" id="text.70"/> in that artificial neural networks capture the complexity of the mass balance estimation using non-linear relationships between inputs. However, we propose that other ML models, notably ensemble tree-based methods, can be used as an equivalent to improved estimates in the case of fewer real-world data samples for training. This has also been observed in other studies (e.g. <xref ref-type="bibr" rid="bib1.bibx3" id="altparen.71"/>) For this case, feature importance derived using permutation importance for the ensemble-based models, RF and GBR, represented the expected role of meteorological variables in determining feature importance. The evaluation metrics also emphasize the performance of these models.</p>
</sec>
<sec id="Ch1.S4.SS4">
  <label>4.4</label><title>Relevance to future studies</title>
      <p id="d1e1536">With the emergence of artificial intelligence techniques, a number of studies have employed deep learning algorithms for numerous applications. A majority of these studies use neural networks to incorporate non-linearity in the modelling of various Earth observation applications. However, a host of ML techniques exist which remain under-utilized. This is being studied in the ML community (e.g. <xref ref-type="bibr" rid="bib1.bibx17" id="altparen.72"/>, studied 179 classification models), and it has been observed that for tabular datasets, tree-based models remain state of the art <xref ref-type="bibr" rid="bib1.bibx66 bib1.bibx21" id="paren.73"/> for both classification and regression problems for medium-sized datasets (training samples under 10 000). Our study also depicts the improved performance of GBR models, which aligns with these recent findings. While it largely follows the assumptions made by <xref ref-type="bibr" rid="bib1.bibx21" id="text.74"/>, we demonstrate the case of regression with heterogeneous and interdependent input features, and a voided assumption of the identical and independent distribution of samples also depict a better performance by ensemble tree-based models. With glacier mass balance datasets being typically medium-sized datasets with correlated input features, we recommend that studies aiming to use ML for modelling the Earth system consider the ensemble-based techniques. Many ensemble-based techniques exist, including bagging as used by RF and boosting as used by AdaBoost and GBR. Further, studies  that combine ensemble trees models with deep learning are also being used effectively (e.g. <xref ref-type="bibr" rid="bib1.bibx66" id="altparen.75"/>, used XGBoost in tandem with an ensemble of deep models). <xref ref-type="bibr" rid="bib1.bibx5" id="text.76"/> utilize a leave-one-year-out and leave-one-glacier-out mode of testing the performance of the model. This is in line with <xref ref-type="bibr" rid="bib1.bibx61" id="text.77"/>, who suggest that spatially and temporally structured datasets would benefit from a manually designed blocking strategy. As the testing and validation splits will result in similar effects in all the models, performing the grouped splitting does not provide immense value to this study. However, for cases where a single model is to be used to estimate glacier mass balance, the leave-one-glacier-out and leave-one-year-out techniques are useful.</p>
      <?pagebreak page2824?><p id="d1e1558">An aspect not considered in this study is a transfer learning approach to the ML modelling, where glacier mass balance datasets from other locations can be used to pre-train the neural network and generate an initialization of weights to be tuned by the dataset of the region of interest (see <?xmltex \hack{\mbox\bgroup}?><xref ref-type="bibr" rid="bib1.bibx2" id="altparen.78"/><?xmltex \hack{\egroup}?>). In line with utilizing datasets from other locations, another aspect to consider with glacier mass balance datasets is the generalizability of the models. Understanding which machine learning model can be used for local, regional, and global analysis is important and will be a useful study to take up. Feature importance associated with the local, regional, and global analysis will also provide new insights into the changes in the glacier mass balance at these scales. An important factor to note is that through this study, we have considered annual mass balance measurements as opposed to seasonal measurements due to the paucity of sufficient datasets to train a multi-parameter machine learning model fully. The role of ablation and accumulation variables will be better represented in the case of seasonal measurements and is an avenue to explore through future studies.</p>
</sec>
</sec>
<sec id="Ch1.S5" sec-type="conclusions">
  <label>5</label><title>Conclusions</title>
      <p id="d1e1576">In this study, we constructed four ML models to estimate point glacier mass balance for the RGI first-order region 11: Central Europe. We used the ERA5-Land reanalysis meteorological data to train the models against point measurements of glacier mass balance obtained from the FoG database. In addition to the NN model, which is being increasingly utilized for glacier mass balance estimation, we used other classes of ML models, such as ensemble tree-based models, RF and GBR, and the kernel-based model, SVM. We compared these ML models with an LR model commonly used for mass balance modelling. Care must be taken to tune the hyperparameters for the GBR, NN, and SVM models. We observe that for these models, hyperparameter tuning was beneficial for improving the estimates of glacier mass balance. For smaller datasets, ensemble models such as RF and GBR depict overfitting. The NN model requires more data samples for effective training. The SVM model can effectively be used in the case of a smaller number of data samples, which is characteristic of real-world datasets. The LR model is consistently unable to capture the complexity of the data and underperforms. For larger datasets, ensemble models such as RF and GBR perform slightly better in terms of <inline-formula><mml:math id="M79" display="inline"><mml:mrow><mml:msup><mml:mi>R</mml:mi><mml:mn mathvariant="normal">2</mml:mn></mml:msup></mml:mrow></mml:math></inline-formula> and RMSE. However, NN models depict the least bias. The meteorological variables obtained from reanalysis datasets are associated with high bias. Using NN and LR models permits us to use them directly. For other models, bias correction should be incorporated in the preprocessing. Representation of real-world features is also performed more effectively by RF and GBR models. These models indicate the importance of ablation features dominating the mass balance estimates. This is expected, as the mass balance measurements are primarily negative. Further, feature importance suggests that features such as forecast albedo, sensible heat flux, latent heat flux, and net solar radiation also play a pivotal role in estimating point mass balance. Thus, inclusion of these additional variables might be of importance for future studies.</p>
</sec>

      
      </body>
    <back><notes notes-type="codedataavailability"><title>Code and data availability</title>

      <p id="d1e1594">The data used for the study are the monthly mean ERA5-Land reanalysis product for input features (<ext-link xlink:href="https://doi.org/10.24381/cds.e2161bac" ext-link-type="DOI">10.24381/cds.e2161bac</ext-link>, <xref ref-type="bibr" rid="bib1.bibx50" id="altparen.79"/>) and point mass balance measurements from the Fluctuation of Glaciers database (<ext-link xlink:href="https://doi.org/10.5904/wgms-fog-2021-05" ext-link-type="DOI">10.5904/wgms-fog-2021-05</ext-link>, <xref ref-type="bibr" rid="bib1.bibx77" id="altparen.80"/>) for the target data. The code for processing the data and applying all models used in this study is available at <uri>https://github.com/RituAnilkumar/pt-gmb-ml</uri> (last access: 7 July 2023).</p>
  </notes><app-group>
        <supplementary-material position="anchor"><p id="d1e1612">The supplement related to this article is available online at: <inline-supplementary-material xlink:href="https://doi.org/10.5194/tc-17-2811-2023-supplement" xlink:title="zip">https://doi.org/10.5194/tc-17-2811-2023-supplement</inline-supplementary-material>.</p></supplementary-material>
        </app-group><notes notes-type="authorcontribution"><title>Author contributions</title>

      <p id="d1e1621">RA, RB, and DC were involved in the design of the study. RA wrote the code for the study and produced the figures, tables, and first draft of the article using inputs from all authors. RB, DC, and SPA proofread and edited the article. RA performed the first level of analysis, which was augmented by inputs from RB, DC, and SPA.</p>
  </notes><notes notes-type="competinginterests"><title>Competing interests</title>

      <p id="d1e1627">The contact author has declared that none of the authors has any competing interests.</p>
  </notes><notes notes-type="disclaimer"><title>Disclaimer</title>

      <p id="d1e1633">Publisher’s note: Copernicus Publications remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.</p>
  </notes><ack><title>Acknowledgements</title><p id="d1e1639">We acknowledge the contribution of the journal editors, particularly Emily Collier, for the thorough article handling. We thank Jordi Bolibar and the anonymous reviewer, whose detailed suggestions and inputs have substantially improved the quality of the article.
We also acknowledge the engaging discussions with peers, most notably Aniket Chakraborty, who always lent a patient ear and sound suggestions to roadblocks along the way.</p></ack><notes notes-type="reviewstatement"><title>Review statement</title>

      <p id="d1e1644">This paper was edited by Emily Collier and reviewed by Jordi Bolibar and one anonymous referee.</p>
  </notes><ref-list>
    <title>References</title>

      <ref id="bib1.bibx1"><?xmltex \def\ref@label{{Altmann et~al.(2010)Altmann, Toloşi, Sander, and
Lengauer}}?><label>Altmann et al.(2010)Altmann, Toloşi, Sander, and
Lengauer</label><?label altman2010?><mixed-citation>Altmann, A., Toloşi, L., Sander, O., and Lengauer, T.: Permutation
importance: a corrected feature importance measure, Bioinformatics, 26,
1340–1347, <ext-link xlink:href="https://doi.org/10.1093/bioinformatics/btq134" ext-link-type="DOI">10.1093/bioinformatics/btq134</ext-link>, 2010.</mixed-citation></ref>
      <ref id="bib1.bibx2"><?xmltex \def\ref@label{{{Anilkumar} et~al.(2022){Anilkumar}, {Bharti}, and
{Chutia}}}?><label>Anilkumar et al.(2022)Anilkumar, Bharti, and
Chutia</label><?label Anilkumar2022?><mixed-citation>Anilkumar, R., Bharti, R., and Chutia, D.: Point Mass Balance Regression using Deep Neural Networks: A Transfer Learning Approach, EGU General Assembly 2022, Vienna, Austria, 23–27 May 2022, EGU22-5317, <ext-link xlink:href="https://doi.org/10.5194/egusphere-egu22-5317" ext-link-type="DOI">10.5194/egusphere-egu22-5317</ext-link>, 2022.</mixed-citation></ref>
      <ref id="bib1.bibx3"><?xmltex \def\ref@label{{Bair et~al.(2018)Bair, Abreu~Calfa, Rittger, and Dozier}}?><label>Bair et al.(2018)Bair, Abreu Calfa, Rittger, and Dozier</label><?label bair2018?><mixed-citation>Bair, E. H., Abreu Calfa, A., Rittger, K., and Dozie<?pagebreak page2825?>r, J.: Using machine learning for real-time estimates of snow water equivalent in the watersheds of Afghanistan, The Cryosphere, 12, 1579–1594, <ext-link xlink:href="https://doi.org/10.5194/tc-12-1579-2018" ext-link-type="DOI">10.5194/tc-12-1579-2018</ext-link>, 2018.</mixed-citation></ref>
      <ref id="bib1.bibx4"><?xmltex \def\ref@label{{Bash et~al.(2018)Bash, Moorman, and Gunther}}?><label>Bash et al.(2018)Bash, Moorman, and Gunther</label><?label bash2018detecting?><mixed-citation>Bash, E. A., Moorman, B. J., and Gunther, A.: Detecting Short-Term Surface Melt
on an Arctic Glacier Using UAV Surveys, Remote Sensing, 10, 1547,
<ext-link xlink:href="https://doi.org/10.3390/rs10101547" ext-link-type="DOI">10.3390/rs10101547</ext-link>, 2018.</mixed-citation></ref>
      <ref id="bib1.bibx5"><?xmltex \def\ref@label{{Bolibar et~al.(2020)Bolibar, Rabatel, Gouttevin, Galiez, Condom, and
Sauquet}}?><label>Bolibar et al.(2020)Bolibar, Rabatel, Gouttevin, Galiez, Condom, and
Sauquet</label><?label bolibar2020deep?><mixed-citation>Bolibar, J., Rabatel, A., Gouttevin, I., Galiez, C., Condom, T., and Sauquet, E.: Deep learning applied to glacier evolution modelling, The Cryosphere, 14, 565–584, <ext-link xlink:href="https://doi.org/10.5194/tc-14-565-2020" ext-link-type="DOI">10.5194/tc-14-565-2020</ext-link>, 2020.</mixed-citation></ref>
      <ref id="bib1.bibx6"><?xmltex \def\ref@label{{Bolibar et~al.(2022)Bolibar, Rabatel, Gouttevin, Zekollari, and
Galiez}}?><label>Bolibar et al.(2022)Bolibar, Rabatel, Gouttevin, Zekollari, and
Galiez</label><?label bolibar2022nonlinear?><mixed-citation>Bolibar, J., Rabatel, A., Gouttevin, I., Zekollari, H., and Galiez, C.:
Nonlinear sensitivity of glacier mass balance to future climate change
unveiled by deep learning, Nat. Commun., 13, 409,
<ext-link xlink:href="https://doi.org/10.1038/s41467-022-28033-0" ext-link-type="DOI">10.1038/s41467-022-28033-0</ext-link>, 2022.</mixed-citation></ref>
      <ref id="bib1.bibx7"><?xmltex \def\ref@label{{Braithwaite(1984)}}?><label>Braithwaite(1984)</label><?label braithwaite1984can?><mixed-citation>Braithwaite, R. J.: Can the Mass Balance of a Glacier be Estimated from its
Equilibrium-Line Altitude?, J. Glaciol., 30, 364–368,
<ext-link xlink:href="https://doi.org/10.3189/S0022143000006237" ext-link-type="DOI">10.3189/S0022143000006237</ext-link>, 1984.</mixed-citation></ref>
      <ref id="bib1.bibx8"><?xmltex \def\ref@label{{Breiman(1996)}}?><label>Breiman(1996)</label><?label breiman1996bagging?><mixed-citation>Breiman, L.: Bagging predictors, Machine Learning, 24, 123–140,
<ext-link xlink:href="https://doi.org/10.1007/BF00058655" ext-link-type="DOI">10.1007/BF00058655</ext-link>, 1996.</mixed-citation></ref>
      <ref id="bib1.bibx9"><?xmltex \def\ref@label{{Breiman(2001)}}?><label>Breiman(2001)</label><?label breiman2001random?><mixed-citation>Breiman, L.: Random forests, Machine Learning, 45, 5–32,
<ext-link xlink:href="https://doi.org/10.1023/A:1010933404324" ext-link-type="DOI">10.1023/A:1010933404324</ext-link>, 2001.</mixed-citation></ref>
      <ref id="bib1.bibx10"><?xmltex \def\ref@label{{Carturan et~al.(2009)Carturan, Cazorzi, and
Dalla~Fontana}}?><label>Carturan et al.(2009)Carturan, Cazorzi, and
Dalla Fontana</label><?label carturan_cazorzi_dalla?><mixed-citation>Carturan, L., Cazorzi, F., and Dalla Fontana, G.: Enhanced estimation of
glacier mass balance in unsampled areas by means of topographic data, Ann. Glaciol., 50, 37–46, <ext-link xlink:href="https://doi.org/10.3189/172756409787769519" ext-link-type="DOI">10.3189/172756409787769519</ext-link>, 2009.</mixed-citation></ref>
      <ref id="bib1.bibx11"><?xmltex \def\ref@label{{Chen et~al.(2021)Chen, Sharma, Zhou, Yang, Li, Niu, Hu, and
Khadka}}?><label>Chen et al.(2021)Chen, Sharma, Zhou, Yang, Li, Niu, Hu, and
Khadka</label><?label CHEN2021105365?><mixed-citation>Chen, Y., Sharma, S., Zhou, X., Yang, K., Li, X., Niu, X., Hu, X., and Khadka,
N.: Spatial performance of multiple reanalysis precipitation datasets on the
southern slope of central Himalaya, Atmos. Res., 250, 105365,
<ext-link xlink:href="https://doi.org/10.1016/j.atmosres.2020.105365" ext-link-type="DOI">10.1016/j.atmosres.2020.105365</ext-link>, 2021.</mixed-citation></ref>
      <ref id="bib1.bibx12"><?xmltex \def\ref@label{{Chi and Kim(2017)}}?><label>Chi and Kim(2017)</label><?label chi2017prediction?><mixed-citation>Chi, J. and Kim, H.-C.: Prediction of Arctic Sea Ice Concentration Using a
Fully Data Driven Deep Neural Network, Remote Sensing, 9, 1305,
<ext-link xlink:href="https://doi.org/10.3390/rs9121305" ext-link-type="DOI">10.3390/rs9121305</ext-link>, 2017.</mixed-citation></ref>
      <ref id="bib1.bibx13"><?xmltex \def\ref@label{{Cortes and Vapnik(1995)}}?><label>Cortes and Vapnik(1995)</label><?label cortes1995support?><mixed-citation>Cortes, C. and Vapnik, V.: Support-vector networks, Machine Learning, 20,
273–297, <ext-link xlink:href="https://doi.org/10.1007/BF00994018" ext-link-type="DOI">10.1007/BF00994018</ext-link>, 1995.</mixed-citation></ref>
      <ref id="bib1.bibx14"><?xmltex \def\ref@label{{Cucchi et~al.(2020)Cucchi, Weedon, Amici, Bellouin, Lange,
M\"{u}ller~Schmied, Hersbach, and Buontempo}}?><label>Cucchi et al.(2020)Cucchi, Weedon, Amici, Bellouin, Lange,
Müller Schmied, Hersbach, and Buontempo</label><?label cucchi2020?><mixed-citation>Cucchi, M., Weedon, G. P., Amici, A., Bellouin, N., Lange, S., Müller Schmied, H., Hersbach, H., and Buontempo, C.: WFDE5: bias-adjusted ERA5 reanalysis data for impact studies, Earth Syst. Sci. Data, 12, 2097–2120, <ext-link xlink:href="https://doi.org/10.5194/essd-12-2097-2020" ext-link-type="DOI">10.5194/essd-12-2097-2020</ext-link>, 2020.</mixed-citation></ref>
      <ref id="bib1.bibx15"><?xmltex \def\ref@label{{Dietterich(2000)}}?><label>Dietterich(2000)</label><?label dietterich2000experimental?><mixed-citation>Dietterich, T. G.: An experimental comparison of three methods for constructing
ensembles of decision trees: Bagging, boosting, and randomization, Machine
Learning, 40, 139–157, <ext-link xlink:href="https://doi.org/10.1023/A:1007607513941" ext-link-type="DOI">10.1023/A:1007607513941</ext-link>, 2000.</mixed-citation></ref>
      <ref id="bib1.bibx16"><?xmltex \def\ref@label{{Dobhal et~al.(2021)Dobhal, Pratap, Bhambri, and
Mehta}}?><label>Dobhal et al.(2021)Dobhal, Pratap, Bhambri, and
Mehta</label><?label dobhal2021mass?><mixed-citation>Dobhal, D., Pratap, B., Bhambri, R., and Mehta, M.: Mass balance and
morphological changes of Dokriani Glacier (1992–2013), Garhwal Himalaya,
India, Quaternary Science Advances, 4, 100033,
<ext-link xlink:href="https://doi.org/10.1016/j.qsa.2021.100033" ext-link-type="DOI">10.1016/j.qsa.2021.100033</ext-link>, 2021.</mixed-citation></ref>
      <ref id="bib1.bibx17"><?xmltex \def\ref@label{{Fern\'{a}ndez-Delgado et~al.(2014)Fern\'{a}ndez-Delgado, Cernadas,
Barro, and Amorim}}?><label>Fernández-Delgado et al.(2014)Fernández-Delgado, Cernadas,
Barro, and Amorim</label><?label fernandez2014?><mixed-citation>
Fernández-Delgado, M., Cernadas, E., Barro, S., and Amorim, D.: Do We Need
Hundreds of Classifiers to Solve Real World Classification Problems?, J.
Mach. Learn. Res., 15, 3133–3181, 2014.</mixed-citation></ref>
      <ref id="bib1.bibx18"><?xmltex \def\ref@label{{Friedman(2001)}}?><label>Friedman(2001)</label><?label friedman2001greedy?><mixed-citation>Friedman, J. H.: Greedy function approximation: a gradient boosting machine,
Ann. Stat., 29, 1189–1232, <ext-link xlink:href="https://doi.org/10.1214/aos/1013203451" ext-link-type="DOI">10.1214/aos/1013203451</ext-link>, 2001.</mixed-citation></ref>
      <ref id="bib1.bibx19"><?xmltex \def\ref@label{{Gabbi et~al.(2014)Gabbi, Carenzo, Pellicciotti, Bauder, and
Funk}}?><label>Gabbi et al.(2014)Gabbi, Carenzo, Pellicciotti, Bauder, and
Funk</label><?label gabbi_carenzo_pellicciotti_bauder_funk_2014?><mixed-citation>Gabbi, J., Carenzo, M., Pellicciotti, F., Bauder, A., and Funk, M.: A
comparison of empirical and physically based glacier surface melt models for
long-term simulations of glacier response, J. Glaciol., 60,
1140–1154, <ext-link xlink:href="https://doi.org/10.3189/2014JoG14J011" ext-link-type="DOI">10.3189/2014JoG14J011</ext-link>, 2014.</mixed-citation></ref>
      <ref id="bib1.bibx20"><?xmltex \def\ref@label{{Gerbaux et~al.(2005)Gerbaux, Genthon, Etchevers, Vincent, and
Dedieu}}?><label>Gerbaux et al.(2005)Gerbaux, Genthon, Etchevers, Vincent, and
Dedieu</label><?label gerbaux2005surface?><mixed-citation>Gerbaux, M., Genthon, C., Etchevers, P., Vincent, C., and Dedieu, J.: Surface
mass balance of glaciers in the French Alps: distributed modeling and
sensitivity to climate change, J. Glaciol., 51, 561–572,
<ext-link xlink:href="https://doi.org/10.3189/172756505781829133" ext-link-type="DOI">10.3189/172756505781829133</ext-link>, 2005.</mixed-citation></ref>
      <ref id="bib1.bibx21"><?xmltex \def\ref@label{{Grinsztajn et~al.(2022)Grinsztajn, Oyallon, and
Varoquaux}}?><label>Grinsztajn et al.(2022)Grinsztajn, Oyallon, and
Varoquaux</label><?label grinsztajn2022treebased?><mixed-citation>Grinsztajn, L., Oyallon, E., and Varoquaux, G.: Why do tree-based models still
outperform deep learning on tabular data?, in: Advances in Neural Information Processing Systems, edited by:  Koyejo, S.,  Mohamed, S.,  Agarwal, A., Belgrave, D.,  Cho, K., and  Oh, A., Curran Associates, Inc., vol. 35, 507–520, <uri>https://proceedings.neurips.cc/paper_files/paper/2022/file/0378c7692da36807bdec87ab043cdadc-Paper-Datasets_and_Benchmarks.pdf</uri> (last access: 7 July 2023), 2022.</mixed-citation></ref>
      <ref id="bib1.bibx22"><?xmltex \def\ref@label{{Guidicelli et~al.(2023)Guidicelli, Huss, Gabella, and
Salzmann}}?><label>Guidicelli et al.(2023)Guidicelli, Huss, Gabella, and
Salzmann</label><?label guidicelli22?><mixed-citation>Guidicelli, M., Huss, M., Gabella, M., and Salzmann, N.: Spatio-temporal reconstruction of winter glacier mass balance in the Alps, Scandinavia, Central Asia and western Canada (1981–2019) using climate reanalyses and machine learning, The Cryosphere, 17, 977–1002, <ext-link xlink:href="https://doi.org/10.5194/tc-17-977-2023" ext-link-type="DOI">10.5194/tc-17-977-2023</ext-link>, 2023.</mixed-citation></ref>
      <ref id="bib1.bibx23"><?xmltex \def\ref@label{{Guo et~al.(2020)Guo, Chen, Liu, and Zhao}}?><label>Guo et al.(2020)Guo, Chen, Liu, and Zhao</label><?label guo2020extraction?><mixed-citation>Guo, X., Chen, Y., Liu, X., and Zhao, Y.: Extraction of snow cover from
high-resolution remote sensing imagery using deep learning on a small
dataset, Remote Sens. Lett., 11, 66–75,
<ext-link xlink:href="https://doi.org/10.1080/2150704X.2019.1686548" ext-link-type="DOI">10.1080/2150704X.2019.1686548</ext-link>, 2020.</mixed-citation></ref>
      <ref id="bib1.bibx24"><?xmltex \def\ref@label{{Haq et~al.(2014)Haq, Jain, and Menon}}?><label>Haq et al.(2014)Haq, Jain, and Menon</label><?label haq2014gangotri?><mixed-citation>Haq, M. A., Jain, K., and Menon, K.: Modelling of Gangotri glacier thickness
and volume using an artificial neural network, Int. J.
Remote Sens., 35, 6035–6042, <ext-link xlink:href="https://doi.org/10.1080/01431161.2014.943322" ext-link-type="DOI">10.1080/01431161.2014.943322</ext-link>, 2014.</mixed-citation></ref>
      <ref id="bib1.bibx25"><?xmltex \def\ref@label{{Haq et~al.(2021)Haq, Azam, and Vincent}}?><label>Haq et al.(2021)Haq, Azam, and Vincent</label><?label haq2021efficiency?><mixed-citation>Haq, M. A., Azam, M. F., and Vincent, C.: Efficiency of artificial neural
networks for glacier ice-thickness estimation: a case study in western
Himalaya, India, J. Glaciol., 67, 671–684,
<ext-link xlink:href="https://doi.org/10.1017/jog.2021.19" ext-link-type="DOI">10.1017/jog.2021.19</ext-link>, 2021.</mixed-citation></ref>
      <ref id="bib1.bibx26"><?xmltex \def\ref@label{{Hersbach et~al.(2020)Hersbach, Bell, Berrisford, Hirahara, Horányi,
Muñoz-Sabater, Nicolas, Peubey, Radu, Schepers, Simmons, Soci, Abdalla,
Abellan, Balsamo, Bechtold, Biavati, Bidlot, Bonavita, De~Chiara, Dahlgren,
Dee, Diamantakis, Dragani, Flemming, Forbes, Fuentes, Geer, Haimberger,
Healy, Hogan, Hólm, Janisková, Keeley, Laloyaux, Lopez, Lupu, Radnoti,
de~Rosnay, Rozum, Vamborg, Villaume, and Thépaut}}?><label>Hersbach et al.(2020)Hersbach, Bell, Berrisford, Hirahara, Horányi,
Muñoz-Sabater, Nicolas, Peubey, Radu, Schepers, Simmons, Soci, Abdalla,
Abellan, Balsamo, Bechtold, Biavati, Bidlot, Bonavita, De Chiara, Dahlgren,
Dee, Diamantakis, Dragani, Flemming, Forbes, Fuentes, Geer, Haimberger,
Healy, Hogan, Hólm, Janisková, Keeley, Laloyaux, Lopez, Lupu, Radnoti,
de Rosnay, Rozum, Vamborg, Villaume, and Thépaut</label><?label hersbach2020?><mixed-citation>Hersbach, H., Bell, B., Berrisford, P., Hirahara, S., Horányi, A.,
Muñoz-Sabater, J., Nicolas, J., Peubey, C., Radu, R., Schepers, D., Simmons,
A., Soci, C., Abdalla, S., Abellan, X., Balsamo, G., Bechtold, P., Biavati,
G., Bidlot, J., Bonavita, M., De Chiara, G., Dahlgren, P., Dee, D.,
Diamantakis, M., Dragani, R., Flemming, J., Forbes, R., Fuentes, M., Geer,
A., Haimberger, L., Healy, S., Hogan, R. J., Hólm, E., Janisková, M.,
Keeley, S., Laloyaux, P., Lopez, P., Lupu, C., Radnoti, G., de Rosnay, P.,
Rozum, I., Vamborg, F., Villaume, S., and Thépaut, J.-N.: The ERA5 global
reanalysis, Q. J. Roy. Meteor. Soc., 146,
1999–2049, <ext-link xlink:href="https://doi.org/10.1002/qj.3803" ext-link-type="DOI">10.1002/qj.3803</ext-link>, 2020.</mixed-citation></ref>
      <ref id="bib1.bibx27"><?xmltex \def\ref@label{{Hoinkes(1968)}}?><label>Hoinkes(1968)</label><?label hoinkes_1968?><mixed-citation>Hoinkes, H. C.: Glacier Variation and Weather, J. Glaciol., 7,
3–18, <ext-link xlink:href="https://doi.org/10.3189/S0022143000020384" ext-link-type="DOI">10.3189/S0022143000020384</ext-link>, 1968.</mixed-citation></ref>
      <ref id="bib1.bibx28"><?xmltex \def\ref@label{{Hornik(1991)}}?><label>Hornik(1991)</label><?label HORNIK1991251?><mixed-citation>Hornik, K.: Approximation capabilities of multilayer feedforward networks,
Neural Networks, 4, 251–257,
<ext-link xlink:href="https://doi.org/10.1016/0893-6080(91)90009-T" ext-link-type="DOI">10.1016/0893-6080(91)90009-T</ext-link>, 1991.</mixed-citation></ref>
      <ref id="bib1.bibx29"><?xmltex \def\ref@label{{Huss et~al.(2008)Huss, Farinotti, Bauder, and
Funk}}?><label>Huss et al.(2008)Huss, Farinotti, Bauder, and
Funk</label><?label huss2008modelling?><mixed-citation>Huss, M., Farinotti, D., Bauder, A., and Funk, M.: Modelling runoff from highly
glacierized alpine drainage basins in a changing climate, Hydrol.
Process., 22, 3888–3902, <ext-link xlink:href="https://doi.org/10.1002/hyp.7055" ext-link-type="DOI">10.1002/hyp.7055</ext-link>, 2008.</mixed-citation></ref>
      <ref id="bib1.bibx30"><?xmltex \def\ref@label{{Ismail et~al.(2023)Ismail, Bogacki, Disse, Sch\"{a}fer, and
Kirschbauer}}?><label>Ismail et al.(2023)Ismail, Bogacki, Disse, Schäfer, and
Kirschbauer</label><?label ismail2023?><mixed-citation>Ismail, M. F., Bogacki, W., Disse, M., Schäfer, M., and Kirschbauer, L.: Estimating degree-day factors of snow based on energy flux components, The Cryosphere, 17, 211–231, <ext-link xlink:href="https://doi.org/10.5194/tc-17-211-2023" ext-link-type="DOI">10.5194/tc-17-211-2023</ext-link>, 2023.</mixed-citation></ref>
      <ref id="bib1.bibx31"><?xmltex \def\ref@label{{Jouvet et~al.(2021)Jouvet, Cordonnier, Kim, Lüthi, Vieli, and
Aschwanden}}?><label>Jouvet et al.(2021)Jouvet, Cordonnier, Kim, Lüthi, Vieli, and
Aschwanden</label><?label jouvet2021deep?><mixed-citation>Jouvet,<?pagebreak page2826?> G., Cordonnier, G., Kim, B., Lüthi, M., Vieli, A., and Aschwanden, A.:
Deep learning speeds up ice flow modelling by several orders of magnitude,
J. Glaciol., 68, 651–664, <ext-link xlink:href="https://doi.org/10.1017/jog.2021.120" ext-link-type="DOI">10.1017/jog.2021.120</ext-link>, 2021.</mixed-citation></ref>
      <ref id="bib1.bibx32"><?xmltex \def\ref@label{{Kan et~al.(2018)Kan, Zhang, Zhu, Xiao, Wang, Tian, and
Tan}}?><label>Kan et al.(2018)Kan, Zhang, Zhu, Xiao, Wang, Tian, and
Tan</label><?label kan2018snow?><mixed-citation>Kan, X., Zhang, Y., Zhu, L., Xiao, L., Wang, J., Tian, W., and Tan, H.: Snow
cover mapping for mountainous areas by fusion of MODIS L1B and geographic
data based on stacked denoising auto-encoders, Computers, Materials &amp; Continua, 57, 49–68, <ext-link xlink:href="https://doi.org/10.32604/cmc.2018.02376" ext-link-type="DOI">10.32604/cmc.2018.02376</ext-link>, 2018.</mixed-citation></ref>
      <ref id="bib1.bibx33"><?xmltex \def\ref@label{{King et~al.(2020)King, Erler, Frey, and Fletcher}}?><label>King et al.(2020)King, Erler, Frey, and Fletcher</label><?label king2020?><mixed-citation>King, F., Erler, A. R., Frey, S. K., and Fletcher, C. G.: Application of machine learning techniques for regional bias correction of snow water equivalent estimates in Ontario, Canada, Hydrol. Earth Syst. Sci., 24, 4887–4902, <ext-link xlink:href="https://doi.org/10.5194/hess-24-4887-2020" ext-link-type="DOI">10.5194/hess-24-4887-2020</ext-link>, 2020.</mixed-citation></ref>
      <ref id="bib1.bibx34"><?xmltex \def\ref@label{{Kingma and Ba(2014)}}?><label>Kingma and Ba(2014)</label><?label kingma2014adam?><mixed-citation>Kingma, D. P. and Ba, J.: Adam: A method for stochastic optimization, arXiv
[preprint], <ext-link xlink:href="https://doi.org/10.48550/arxiv.1412.6980" ext-link-type="DOI">10.48550/arxiv.1412.6980</ext-link>, 22 December 2014.</mixed-citation></ref>
      <ref id="bib1.bibx35"><?xmltex \def\ref@label{{Kuhn et~al.(1999)Kuhn, Dreiseitl, Hofinger, Markl, Span, and
Kaser}}?><label>Kuhn et al.(1999)Kuhn, Dreiseitl, Hofinger, Markl, Span, and
Kaser</label><?label kuhn1999measurements?><mixed-citation>Kuhn, M., Dreiseitl, E., Hofinger, S., Markl, G., Span, N., and Kaser, G.:
Measurements and models of the mass balance of hintereisferner, Geogr.
Ann. A, 81, 659–670,
<uri>https://onlinelibrary.wiley.com/doi/abs/10.1111/1468-0459.00094</uri> (last access: 7 July 2023), 1999.</mixed-citation></ref>
      <ref id="bib1.bibx36"><?xmltex \def\ref@label{{Lefauconnier and Hagen(1990)}}?><label>Lefauconnier and Hagen(1990)</label><?label lefauconnier1990glaciers?><mixed-citation>Lefauconnier, B. and Hagen, J.: Glaciers and Climate in Svalbard: Statistical
Analysis and Reconstruction of the Brøggerbreen Mass Balance for the Last 77
Years, Ann. Glaciol., 14, 148–152, <ext-link xlink:href="https://doi.org/10.3189/S0260305500008466" ext-link-type="DOI">10.3189/S0260305500008466</ext-link>,
1990.</mixed-citation></ref>
      <ref id="bib1.bibx37"><?xmltex \def\ref@label{{{Le Meur} et~al.(2007){Le Meur}, Gerbaux, Schäfer, and
Vincent}}?><label>Le Meur et al.(2007)Le Meur, Gerbaux, Schäfer, and
Vincent</label><?label le2007disappearance?><mixed-citation>Le Meur, E., Gerbaux, M., Schäfer, M., and Vincent, C.: Disappearance of an
Alpine glacier over the 21st Century simulated from modeling its future
surface mass balance, Earth Planet. Sc. Lett., 261, 367–374,
<ext-link xlink:href="https://doi.org/10.1016/j.epsl.2007.07.022" ext-link-type="DOI">10.1016/j.epsl.2007.07.022</ext-link>, 2007.</mixed-citation></ref>
      <ref id="bib1.bibx38"><?xmltex \def\ref@label{{Li et~al.(2017)Li, Wang, Wang, Zhang, Fu, and Wang}}?><label>Li et al.(2017)Li, Wang, Wang, Zhang, Fu, and Wang</label><?label li2017gaofen?><mixed-citation>Li, J., Wang, C., Wang, S., Zhang, H., Fu, Q., and Wang, Y.: Gaofen-3 sea ice
detection based on deep learning, in: 2017 Progress in Electromagnetics
Research Symposium – Fall (PIERS – FALL), Nanyang Technological Univeiisty, Singapore, 19–22 November 2017, 933–939,
<ext-link xlink:href="https://doi.org/10.1109/PIERS-FALL.2017.8293267" ext-link-type="DOI">10.1109/PIERS-FALL.2017.8293267</ext-link>, 2017.</mixed-citation></ref>
      <ref id="bib1.bibx39"><?xmltex \def\ref@label{{Lin et~al.(2018)Lin, Chen, Yang, and Ou}}?><label>Lin et al.(2018)Lin, Chen, Yang, and Ou</label><?label lin2018impact?><mixed-citation>
Lin, C., Chen, D., Yang, K., and Ou, T.: Impact of model resolution on
simulating the water vapor transport through the central Himalayas:
implication for models’ wet bias over the Tibetan Plateau, Clim.
Dynam., 51, 3195–3207, 2018.</mixed-citation></ref>
      <ref id="bib1.bibx40"><?xmltex \def\ref@label{{Liu(2021)}}?><label>Liu(2021)</label><?label liu2021review?><mixed-citation>Liu, L.: A Review of Deep Learning for Cryospheric Studies, chap. 17,
John Wiley and Sons, Ltd, 258–268,
<ext-link xlink:href="https://doi.org/10.1002/9781119646181.ch17" ext-link-type="DOI">10.1002/9781119646181.ch17</ext-link>, 2021.</mixed-citation></ref>
      <ref id="bib1.bibx41"><?xmltex \def\ref@label{{Lliboutry(1974)}}?><label>Lliboutry(1974)</label><?label lliboutry1974multivariate?><mixed-citation>Lliboutry, L.: Multivariate Statistical Analysis of Glacier Annual Balances,
J. Glaciol., 13, 371–392, <ext-link xlink:href="https://doi.org/10.3189/S0022143000023169" ext-link-type="DOI">10.3189/S0022143000023169</ext-link>, 1974.</mixed-citation></ref>
      <ref id="bib1.bibx42"><?xmltex \def\ref@label{{Lu et~al.(2021)Lu, Zhang, Shangguan, and Yang}}?><label>Lu et al.(2021)Lu, Zhang, Shangguan, and Yang</label><?label lu2021novel?><mixed-citation>Lu, Y., Zhang, Z., Shangguan, D., and Yang, J.: Novel Machine Learning Method
Integrating Ensemble Learning and Deep Learning for Mapping Debris-Covered
Glaciers, Remote Sensing, 13, 2595, <ext-link xlink:href="https://doi.org/10.3390/rs13132595" ext-link-type="DOI">10.3390/rs13132595</ext-link>, 2021.</mixed-citation></ref>
      <ref id="bib1.bibx43"><?xmltex \def\ref@label{{Manciati et~al.(2014)Manciati, Villacís, Taupin, Cadier,
Galárraga-Sánchez, and Cáceres}}?><label>Manciati et al.(2014)Manciati, Villacís, Taupin, Cadier,
Galárraga-Sánchez, and Cáceres</label><?label manciati2014?><mixed-citation>Manciati, C., Villacís, M., Taupin, J.-D., Cadier, E., Galárraga-Sánchez,
R., and Cáceres, B.: Empirical mass balance modelling of South American
tropical glaciers: case study of Antisana volcano, Ecuador, Hydrolog.
Sci. J., 59, 1519–1535, <ext-link xlink:href="https://doi.org/10.1080/02626667.2014.888490" ext-link-type="DOI">10.1080/02626667.2014.888490</ext-link>, 2014.</mixed-citation></ref>
      <ref id="bib1.bibx44"><?xmltex \def\ref@label{{Masiokas et~al.(2016)}}?><label>Masiokas et al.(2016)</label><?label masiokas2015reconstructing?><mixed-citation>Masiokas, M. H., Christie, D. A., Le Quesne, C., Pitte, P., Ruiz, L., Villalba, R., Luckman, B. H., Berthier, E., Nussbaumer, S. U., González-Reyes, Á., McPhee, J., and Barcaza, G.: Reconstructing the annual mass balance of the Echaurren Norte glacier (Central Andes, 33.5<inline-formula><mml:math id="M80" display="inline"><mml:msup><mml:mi/><mml:mo>∘</mml:mo></mml:msup></mml:math></inline-formula> S) using local and regional hydroclimatic data, The Cryosphere, 10, 927–940, <ext-link xlink:href="https://doi.org/10.5194/tc-10-927-2016" ext-link-type="DOI">10.5194/tc-10-927-2016</ext-link>, 2016.</mixed-citation></ref>
      <ref id="bib1.bibx45"><?xmltex \def\ref@label{{Mattews and Hodgkins(2016)}}?><label>Mattews and Hodgkins(2016)</label><?label matthews_hodgkins_2016?><mixed-citation>Mattews, T. and Hodgkins, R.: Interdecadal variability of degree-day factors on
Vestari Hagafellsjökull (Langjökull, Iceland) and the importance of
threshold air temperatures, J. Glaciol., 62, 310–322,
<ext-link xlink:href="https://doi.org/10.1017/jog.2016.21" ext-link-type="DOI">10.1017/jog.2016.21</ext-link>, 2016.</mixed-citation></ref>
      <ref id="bib1.bibx46"><?xmltex \def\ref@label{{Maussion et~al.(2019)Maussion, Butenko, Champollion, Dusch, Eis,
Fourteau, Gregor, Jarosch, Landmann, Oesterle, Recinos, Rothenpieler, Vlug,
Wild, and Marzeion}}?><label>Maussion et al.(2019)Maussion, Butenko, Champollion, Dusch, Eis,
Fourteau, Gregor, Jarosch, Landmann, Oesterle, Recinos, Rothenpieler, Vlug,
Wild, and Marzeion</label><?label maussion2019?><mixed-citation>Maussion, F., Butenko, A., Champollion, N., Dusch, M., Eis, J., Fourteau, K., Gregor, P., Jarosch, A. H., Landmann, J., Oesterle, F., Recinos, B., Rothenpieler, T., Vlug, A., Wild, C. T., and Marzeion, B.: The Open Global Glacier Model (OGGM) v1.1, Geosci. Model Dev., 12, 909–931, <ext-link xlink:href="https://doi.org/10.5194/gmd-12-909-2019" ext-link-type="DOI">10.5194/gmd-12-909-2019</ext-link>, 2019.</mixed-citation></ref>
      <ref id="bib1.bibx47"><?xmltex \def\ref@label{{McCulloch and Pitts(1943)}}?><label>McCulloch and Pitts(1943)</label><?label mcculloch1943logical?><mixed-citation>McCulloch, W. S. and Pitts, W.: A logical calculus of the ideas immanent in
nervous activity, B. Math. Biophys., 5, 115–133,
<ext-link xlink:href="https://doi.org/10.1007/BF02478259" ext-link-type="DOI">10.1007/BF02478259</ext-link>, 1943.</mixed-citation></ref>
      <ref id="bib1.bibx48"><?xmltex \def\ref@label{{Mohajerani et~al.(2021)Mohajerani, Jeong, Scheuchl, Velicogna,
Rignot, and Milillo}}?><label>Mohajerani et al.(2021)Mohajerani, Jeong, Scheuchl, Velicogna,
Rignot, and Milillo</label><?label mohajerani2021automatic?><mixed-citation>Mohajerani, Y., Jeong, S., Scheuchl, B., Velicogna, I., Rignot, E., and
Milillo, P.: Automatic delineation of glacier grounding lines in differential
interferometric synthetic-aperture radar data using deep learning, Scientific
Reports, 11, 4992, <ext-link xlink:href="https://doi.org/10.1038/s41598-021-84309-3" ext-link-type="DOI">10.1038/s41598-021-84309-3</ext-link>, 2021.</mixed-citation></ref>
      <ref id="bib1.bibx49"><?xmltex \def\ref@label{{Moya~Quiroga et~al.(2013)Moya~Quiroga, Mano, Asaoka, Kure, Udo, and
Mendoza}}?><label>Moya Quiroga et al.(2013)Moya Quiroga, Mano, Asaoka, Kure, Udo, and
Mendoza</label><?label hess-17-1265-2013?><mixed-citation>Moya Quiroga, V., Mano, A., Asaoka, Y., Kure, S., Udo, K., and Mendoza, J.: Snow glacier melt estimation in tropical Andean glaciers using artificial neural networks, Hydrol. Earth Syst. Sci., 17, 1265–1280, <ext-link xlink:href="https://doi.org/10.5194/hess-17-1265-2013" ext-link-type="DOI">10.5194/hess-17-1265-2013</ext-link>, 2013.</mixed-citation></ref>
      <ref id="bib1.bibx50"><?xmltex \def\ref@label{{Muñoz Sabater(2019)}}?><label>Muñoz Sabater(2019)</label><?label MunozSabater2019?><mixed-citation>Muñoz Sabater, J.: ERA5-Land hourly data from 1950 to present, Copernicus Climate Change Service (C3S) Climate Data Store (CDS) [data set], <ext-link xlink:href="https://doi.org/10.24381/cds.e2161bac" ext-link-type="DOI">10.24381/cds.e2161bac</ext-link>, 2019.</mixed-citation></ref>
      <ref id="bib1.bibx51"><?xmltex \def\ref@label{{Mu\~{n}oz Sabater et~al.(2021)Mu\~{n}oz Sabater, Dutra,
Agust\'{\i}-Panareda, Albergel, Arduini, Balsamo, Boussetta, Choulga,
Harrigan, Hersbach, Martens, Miralles, Piles, Rodr\'{\i}guez-Fern\'{a}ndez,
Zsoter, Buontempo, and Th\'{e}paut}}?><label>Muñoz Sabater et al.(2021)Muñoz Sabater, Dutra,
Agustí-Panareda, Albergel, Arduini, Balsamo, Boussetta, Choulga,
Harrigan, Hersbach, Martens, Miralles, Piles, Rodríguez-Fernández,
Zsoter, Buontempo, and Thépaut</label><?label era5lbase?><mixed-citation>Muñoz-Sabater, J., Dutra, E., Agustí-Panareda, A., Albergel, C., Arduini, G., Balsamo, G., Boussetta, S., Choulga, M., Harrigan, S., Hersbach, H., Martens, B., Miralles, D. G., Piles, M., Rodríguez-Fernández, N. J., Zsoter, E., Buontempo, C., and Thépaut, J.-N.: ERA5-Land: a state-of-the-art global reanalysis dataset for land applications, Earth Syst. Sci. Data, 13, 4349–4383, <ext-link xlink:href="https://doi.org/10.5194/essd-13-4349-2021" ext-link-type="DOI">10.5194/essd-13-4349-2021</ext-link>, 2021.</mixed-citation></ref>
      <ref id="bib1.bibx52"><?xmltex \def\ref@label{{Natekin and Knoll(2013)}}?><label>Natekin and Knoll(2013)</label><?label natekin2013gradient?><mixed-citation>Natekin, A. and Knoll, A.: Gradient boosting machines, a tutorial, Front.
Neurorobotics, 7, 21, <ext-link xlink:href="https://doi.org/10.3389/fnbot.2013.00021" ext-link-type="DOI">10.3389/fnbot.2013.00021</ext-link>, 2013.</mixed-citation></ref>
      <ref id="bib1.bibx53"><?xmltex \def\ref@label{{Nijhawan et~al.(2019)Nijhawan, Das, and Raman}}?><label>Nijhawan et al.(2019)Nijhawan, Das, and Raman</label><?label nijhawan2019hybrid?><mixed-citation>Nijhawan, R., Das, J., and Raman, B.: A hybrid of deep learning and
hand-crafted features based approach for snow cover mapping, Int.
J. Remote Sens., 40, 759–773, <ext-link xlink:href="https://doi.org/10.1080/01431161.2018.1519277" ext-link-type="DOI">10.1080/01431161.2018.1519277</ext-link>,
2019.</mixed-citation></ref>
      <ref id="bib1.bibx54"><?xmltex \def\ref@label{{Pratap et~al.(2016)Pratap, Dobhal, Bhambri, Mehta, and
Tewari}}?><label>Pratap et al.(2016)Pratap, Dobhal, Bhambri, Mehta, and
Tewari</label><?label pratap2016four?><mixed-citation>Pratap, B., Dobhal, D. P., Bhambri, R., Mehta, M., and Tewari, V. C.: Four
decades of glacier mass balance observations in the Indian Himalaya, Reg.
Environ. Change, 16, 643–658,
<ext-link xlink:href="https://doi.org/10.1007/s10113-015-0791-4" ext-link-type="DOI">10.1007/s10113-015-0791-4</ext-link>, 2016.</mixed-citation></ref>
      <ref id="bib1.bibx55"><?xmltex \def\ref@label{{Rabatel et~al.(2016)Rabatel, Dedieu, and Vincent}}?><label>Rabatel et al.(2016)Rabatel, Dedieu, and Vincent</label><?label rabatel2016spatio?><mixed-citation>Rabatel, A., Dedieu, J. P., and Vincent, C.: Spatio-temporal changes in
glacier-wide mass balance quantified by optical remote sensing on 30 glaciers
in the French Alps for the period 1983–2014, J. Glaciol., 62,
1153–1166, <ext-link xlink:href="https://doi.org/10.1017/jog.2016.113" ext-link-type="DOI">10.1017/jog.2016.113</ext-link>, 2016.</mixed-citation></ref>
      <ref id="bib1.bibx56"><?xmltex \def\ref@label{{Radi\'{c} and Hock(2011)}}?><label>Radić and Hock(2011)</label><?label radic2011regionally?><mixed-citation>Radić, V. and Hock, R.: Regionally differentiated contribution of mountain
glaciers and ice caps to future sea-level rise, Nat. Geosci,, 4, 91–94,
<ext-link xlink:href="https://doi.org/10.1038/ngeo1052" ext-link-type="DOI">10.1038/ngeo1052</ext-link>, 2011.</mixed-citation></ref>
      <ref id="bib1.bibx57"><?xmltex \def\ref@label{{Radi\'{c} et~al.(2014)Radi\'{c}, Bliss, Beedlow, Hock, Miles, and
Cogley}}?><label>Radić et al.(2014)Radić, Bliss, Beedlow, Hock, Miles, and
Cogley</label><?label radic2014regional?><mixed-citation>
Radić, V., Bliss, A., Beedlow, A. C., Hock, R., Miles, E., and Cogley,
J. G.: Regional and global projections of twenty-first century glacier mass
changes in response to climate scenarios from global climate models, Clim.
Dynam., 42, 37–58, 2014.</mixed-citation></ref>
      <?pagebreak page2827?><ref id="bib1.bibx58"><?xmltex \def\ref@label{{Rasouli et~al.(2012)Rasouli, Hsieh, and Cannon}}?><label>Rasouli et al.(2012)Rasouli, Hsieh, and Cannon</label><?label RASOULI2012?><mixed-citation>Rasouli, K., Hsieh, W. W., and Cannon, A. J.: Daily streamflow forecasting by
machine learning methods with weather and climate inputs, J.
Hydrol., 414-415, 284–293,
<ext-link xlink:href="https://doi.org/10.1016/j.jhydrol.2011.10.039" ext-link-type="DOI">10.1016/j.jhydrol.2011.10.039</ext-link>, 2012.</mixed-citation></ref>
      <ref id="bib1.bibx59"><?xmltex \def\ref@label{{Rasp et~al.(2018)Rasp, Pritchard, and Gentine}}?><label>Rasp et al.(2018)Rasp, Pritchard, and Gentine</label><?label rasp2018deep?><mixed-citation>Rasp, S., Pritchard, M. S., and Gentine, P.: Deep learning to represent subgrid
processes in climate models, P. Natl. Acad. Sci. USA,
115, 9684–9689, <ext-link xlink:href="https://doi.org/10.1073/pnas.1810286115" ext-link-type="DOI">10.1073/pnas.1810286115</ext-link>, 2018.</mixed-citation></ref>
      <ref id="bib1.bibx60"><?xmltex \def\ref@label{{RGI(2017)}}?><label>RGI(2017)</label><?label rgiDB?><mixed-citation>RGI: Randolph Glacier Inventory (RGI) – A Dataset of Global Glacier Outlines:
Version 6.0. Technical Report, <ext-link xlink:href="https://doi.org/10.7265/N5-RGI-60" ext-link-type="DOI">10.7265/N5-RGI-60</ext-link>, 2017.</mixed-citation></ref>
      <ref id="bib1.bibx61"><?xmltex \def\ref@label{{Roberts et~al.(2017)Roberts, Bahn, Ciuti, Boyce, Elith,
Guillera-Arroita, Hauenstein, Lahoz-Monfort, Schröder, Thuiller, Warton,
Wintle, Hartig, and Dormann}}?><label>Roberts et al.(2017)Roberts, Bahn, Ciuti, Boyce, Elith,
Guillera-Arroita, Hauenstein, Lahoz-Monfort, Schröder, Thuiller, Warton,
Wintle, Hartig, and Dormann</label><?label roberts2017?><mixed-citation>Roberts, D. R., Bahn, V., Ciuti, S., Boyce, M. S., Elith, J., Guillera-Arroita,
G., Hauenstein, S., Lahoz-Monfort, J. J., Schröder, B., Thuiller, W.,
Warton, D. I., Wintle, B. A., Hartig, F., and Dormann, C. F.:
Cross-validation strategies for data with temporal, spatial, hierarchical, or
phylogenetic structure, Ecography, 40, 913–929,
<ext-link xlink:href="https://doi.org/10.1111/ecog.02881" ext-link-type="DOI">10.1111/ecog.02881</ext-link>, 2017.</mixed-citation></ref>
      <ref id="bib1.bibx62"><?xmltex \def\ref@label{{Sauter et~al.(2020)Sauter, Arndt, and Schneider}}?><label>Sauter et al.(2020)Sauter, Arndt, and Schneider</label><?label sauter2020cosipy?><mixed-citation>Sauter, T., Arndt, A., and Schneider, C.: COSIPY v1.3 – an open-source coupled snowpack and ice surface energy and mass balance model, Geosci. Model Dev., 13, 5645–5662, <ext-link xlink:href="https://doi.org/10.5194/gmd-13-5645-2020" ext-link-type="DOI">10.5194/gmd-13-5645-2020</ext-link>, 2020.</mixed-citation></ref>
      <ref id="bib1.bibx63"><?xmltex \def\ref@label{{Schultz et~al.(2021)Schultz, Betancourt, Gong, Kleinert, Langguth,
Leufen, Mozaffari, and Stadtler}}?><label>Schultz et al.(2021)Schultz, Betancourt, Gong, Kleinert, Langguth,
Leufen, Mozaffari, and Stadtler</label><?label schultz2021numDL?><mixed-citation>Schultz, M. G., Betancourt, C., Gong, B., Kleinert, F., Langguth, M., Leufen,
L. H., Mozaffari, A., and Stadtler, S.: Can deep learning beat numerical
weather prediction?, Philosophical Transactions of the Royal Society A:
Mathematical, Physical and Engineering Sciences, 379, 20200097,
<ext-link xlink:href="https://doi.org/10.1098/rsta.2020.0097" ext-link-type="DOI">10.1098/rsta.2020.0097</ext-link>, 2021.</mixed-citation></ref>
      <ref id="bib1.bibx64"><?xmltex \def\ref@label{{Seidou et~al.(2006)Seidou, Ouarda, Bilodeau, Hessami, St-Hilaire, and
Bruneau}}?><label>Seidou et al.(2006)Seidou, Ouarda, Bilodeau, Hessami, St-Hilaire, and
Bruneau</label><?label seidou2006ice?><mixed-citation>Seidou, O., Ouarda, T. B. M. J., Bilodeau, L., Hessami, M., St-Hilaire, A., and
Bruneau, P.: Modeling ice growth on Canadian lakes using artificial neural
networks, Water Resour. Res., 42, W11407,
<ext-link xlink:href="https://doi.org/10.1029/2005WR004622" ext-link-type="DOI">10.1029/2005WR004622</ext-link>, 2006.</mixed-citation></ref>
      <ref id="bib1.bibx65"><?xmltex \def\ref@label{{Shean et~al.(2020)Shean, Bhushan, Montesano, Rounce, Arendt, and
Osmanoglu}}?><label>Shean et al.(2020)Shean, Bhushan, Montesano, Rounce, Arendt, and
Osmanoglu</label><?label shean2020systematic?><mixed-citation>Shean, D. E., Bhushan, S., Montesano, P., Rounce, D. R., Arendt, A., and
Osmanoglu, B.: A Systematic, Regional Assessment of High Mountain Asia
Glacier Mass Balance, Frontiers in Earth Science, 7, 363,
<ext-link xlink:href="https://doi.org/10.3389/feart.2019.00363" ext-link-type="DOI">10.3389/feart.2019.00363</ext-link>, 2020.</mixed-citation></ref>
      <ref id="bib1.bibx66"><?xmltex \def\ref@label{{Shwartz{-}Ziv and Armon(2021)}}?><label>Shwartz-Ziv and Armon(2021)</label><?label shwartz2021?><mixed-citation>Shwartz-Ziv, R. and Armon, A.: Tabular Data: Deep Learning is Not All You
Need, arXiv [preprint], <ext-link xlink:href="https://doi.org/10.48550/arXiv.2106.03253" ext-link-type="DOI">10.48550/arXiv.2106.03253</ext-link>,
6 June 2021.</mixed-citation></ref>
      <ref id="bib1.bibx67"><?xmltex \def\ref@label{{Steiner et~al.(2005)Steiner, Walter, and Zumbühl}}?><label>Steiner et al.(2005)Steiner, Walter, and Zumbühl</label><?label steiner200519th?><mixed-citation>Steiner, D., Walter, A., and Zumbühl, H.: The application of a non-linear
back-propagation neural network to study the mass balance of Grosse
Aletschgletscher, Switzerland, J. Glaciol., 51, 313–323,
<ext-link xlink:href="https://doi.org/10.3189/172756505781829421" ext-link-type="DOI">10.3189/172756505781829421</ext-link>, 2005.</mixed-citation></ref>
      <ref id="bib1.bibx68"><?xmltex \def\ref@label{{Thibert et~al.(2008)Thibert, Blanc, Vincent, and
Eckert}}?><label>Thibert et al.(2008)Thibert, Blanc, Vincent, and
Eckert</label><?label thibert2008glaciological?><mixed-citation>Thibert, E., Blanc, R., Vincent, C., and Eckert, N.: Glaciological and
volumetric mass-balance measurements: error analysis over 51 years for
Glacier de Sarennes, French Alps, J. Glaciol., 54, 522–532,
<ext-link xlink:href="https://doi.org/10.3189/002214308785837093" ext-link-type="DOI">10.3189/002214308785837093</ext-link>, 2008.</mixed-citation></ref>
      <ref id="bib1.bibx69"><?xmltex \def\ref@label{{Trantow and Herzfeld(2016)}}?><label>Trantow and Herzfeld(2016)</label><?label trantow2016spatiotemporal?><mixed-citation>Trantow, T. and Herzfeld, U. C.: Spatiotemporal mapping of a large mountain
glacier from CryoSat-2 altimeter data: surface elevation and elevation change
of Bering Glacier during surge (2011–2014), Int. J. Remote
Sens., 37, 2962–2989, <ext-link xlink:href="https://doi.org/10.1080/01431161.2016.1187318" ext-link-type="DOI">10.1080/01431161.2016.1187318</ext-link>, 2016.</mixed-citation></ref>
      <ref id="bib1.bibx70"><?xmltex \def\ref@label{{Tshering and Fujita(2016)}}?><label>Tshering and Fujita(2016)</label><?label tshering2016first?><mixed-citation>Tshering, P. and Fujita, K.: First in situ record of decadal glacier mass
balance (2003–2014) from the Bhutan Himalaya, Ann. Glaciol., 57,
289–294, <ext-link xlink:href="https://doi.org/10.3189/2016AoG71A036" ext-link-type="DOI">10.3189/2016AoG71A036</ext-link>, 2016.</mixed-citation></ref>
      <ref id="bib1.bibx71"><?xmltex \def\ref@label{{Van~Tricht et~al.(2021)Van~Tricht, Huybrechts, Van~Breedam, Vanhulle,
Van~Oost, and Zekollari}}?><label>Van Tricht et al.(2021)Van Tricht, Huybrechts, Van Breedam, Vanhulle,
Van Oost, and Zekollari</label><?label vantricht2021?><mixed-citation>Van Tricht, L., Huybrechts, P., Van Breedam, J., Vanhulle, A., Van Oost, K., and Zekollari, H.: Estimating surface mass balance patterns from unoccupied aerial vehicle measurements in the ablation area of the Morteratsch–Pers glacier complex (Switzerland), The Cryosphere, 15, 4445–4464, <ext-link xlink:href="https://doi.org/10.5194/tc-15-4445-2021" ext-link-type="DOI">10.5194/tc-15-4445-2021</ext-link>, 2021.</mixed-citation></ref>
      <ref id="bib1.bibx72"><?xmltex \def\ref@label{{Vapnik(1999)}}?><label>Vapnik(1999)</label><?label vapnik1999nature?><mixed-citation>Vapnik, V.: The Nature of Statistical Learning Theory, Springer Science &amp;
Business Media, <ext-link xlink:href="https://doi.org/10.1007/978-1-4757-2440-0" ext-link-type="DOI">10.1007/978-1-4757-2440-0</ext-link>, 1999.</mixed-citation></ref>
      <ref id="bib1.bibx73"><?xmltex \def\ref@label{{Vincent et~al.(2018)Vincent, Soruco, Azam, Basantes-Serrano, Jackson,
Kjøllmoen, Thibert, Wagnon, Six, Rabatel, Ramanathan, Berthier, Cusicanqui,
Vincent, and Mandal}}?><label>Vincent et al.(2018)Vincent, Soruco, Azam, Basantes-Serrano, Jackson,
Kjøllmoen, Thibert, Wagnon, Six, Rabatel, Ramanathan, Berthier, Cusicanqui,
Vincent, and Mandal</label><?label vincent2018?><mixed-citation>Vincent, C., Soruco, A., Azam, M. F., Basantes-Serrano, R., Jackson, M.,
Kjøllmoen, B., Thibert, E., Wagnon, P., Six, D., Rabatel, A., Ramanathan,
A., Berthier, E., Cusicanqui, D., Vincent, P., and Mandal, A.: A Nonlinear
Statistical Model for Extracting a Climatic Signal From Glacier Mass Balance
Measurements, J. Geophys. Res.-Earth, 123,
2228–2242, <ext-link xlink:href="https://doi.org/10.1029/2018JF004702" ext-link-type="DOI">10.1029/2018JF004702</ext-link>, 2018.</mixed-citation></ref>
      <ref id="bib1.bibx74"><?xmltex \def\ref@label{{Wang et~al.(2020)Wang, Yuan, Shen, Liu, Li, Yue, Shi, and
Zhang}}?><label>Wang et al.(2020)Wang, Yuan, Shen, Liu, Li, Yue, Shi, and
Zhang</label><?label wang2020estimating?><mixed-citation>Wang, J., Yuan, Q., Shen, H., Liu, T., Li, T., Yue, L., Shi, X., and Zhang, L.:
Estimating snow depth by combining satellite data and ground-based
observations over Alaska: A deep learning approach, J. Hydrol.,
585, 124828, <ext-link xlink:href="https://doi.org/10.1016/j.jhydrol.2020.124828" ext-link-type="DOI">10.1016/j.jhydrol.2020.124828</ext-link>, 2020.</mixed-citation></ref>
      <ref id="bib1.bibx75"><?xmltex \def\ref@label{{Werder et~al.(2020)Werder, Huss, Paul, Dehecq, and
Farinotti}}?><label>Werder et al.(2020)Werder, Huss, Paul, Dehecq, and
Farinotti</label><?label werder2020bayesian?><mixed-citation>Werder, M. A., Huss, M., Paul, F., Dehecq, A., and Farinotti, D.: A Bayesian
ice thickness estimation model for large-scale applications, J.
Glaciol., 66, 137–152, <ext-link xlink:href="https://doi.org/10.1017/jog.2019.93" ext-link-type="DOI">10.1017/jog.2019.93</ext-link>, 2020.</mixed-citation></ref>
      <ref id="bib1.bibx76"><?xmltex \def\ref@label{{White et~al.(1998)White, Harrison, and
Mottershead}}?><label>White et al.(1998)White, Harrison, and
Mottershead</label><?label white1998environmental?><mixed-citation>
White, I. D., Harrison, S. J., and Mottershead, D. N.: Environmental systems:
an introductory text, Psychology Press, ISBN 9780748740819, 1998.</mixed-citation></ref>
      <ref id="bib1.bibx77"><?xmltex \def\ref@label{{WGMS(2021)}}?><label>WGMS(2021)</label><?label fogDB?><mixed-citation>World Glacier Monitoring Service (WGMS): Fluctuations of Glaciers Database, World Glacier Monitoring Service (WGMS) [data set], <ext-link xlink:href="https://doi.org/10.5904/wgms-fog-2021-05" ext-link-type="DOI">10.5904/wgms-fog-2021-05</ext-link>, 2021.</mixed-citation></ref>
      <ref id="bib1.bibx78"><?xmltex \def\ref@label{{Wu et~al.(2018)Wu, Liu, Jiang, Xu, Wei, and Guo}}?><label>Wu et al.(2018)Wu, Liu, Jiang, Xu, Wei, and Guo</label><?label wu2018recent?><mixed-citation>Wu, K., Liu, S., Jiang, Z., Xu, J., Wei, J., and Guo, W.: Recent glacier mass balance and area changes in the Kangri Karpo Mountains from DEMs and glacier inventories, The Cryosphere, 12, 103–121, <ext-link xlink:href="https://doi.org/10.5194/tc-12-103-2018" ext-link-type="DOI">10.5194/tc-12-103-2018</ext-link>, 2018.</mixed-citation></ref>
      <ref id="bib1.bibx79"><?xmltex \def\ref@label{{Wu et~al.(2023)Wu, Su, Ren, Lü, and Yuan}}?><label>Wu et al.(2023)Wu, Su, Ren, Lü, and Yuan</label><?label WU2023129384?><mixed-citation>Wu, X., Su, J., Ren, W., Lü, H., and Yuan, F.: Statistical comparison and
hydrological utility evaluation of ERA5-Land and IMERG precipitation products
on the Tibetan Plateau, J. Hydrol., 620, 129384,
<ext-link xlink:href="https://doi.org/10.1016/j.jhydrol.2023.129384" ext-link-type="DOI">10.1016/j.jhydrol.2023.129384</ext-link>, 2023.</mixed-citation></ref>
      <ref id="bib1.bibx80"><?xmltex \def\ref@label{{Xie et~al.(2021)Xie, Asari, and Haritashya}}?><label>Xie et al.(2021)Xie, Asari, and Haritashya</label><?label xie2021evaluating?><mixed-citation>Xie, Z., Asari, V. K., and Haritashya, U. K.: Evaluating deep-learning models
for debris-covered glacier mapping, Applied Computing and Geosciences, 12,
100071, <ext-link xlink:href="https://doi.org/10.1016/j.acags.2021.100071" ext-link-type="DOI">10.1016/j.acags.2021.100071</ext-link>, 2021.</mixed-citation></ref>
      <ref id="bib1.bibx81"><?xmltex \def\ref@label{{Zandler et~al.(2019)Zandler, Haag, and
Samimi}}?><label>Zandler et al.(2019)Zandler, Haag, and
Samimi</label><?label zandler2019evaluation?><mixed-citation>Zandler, H., Haag, I., and Samimi, C.: Evaluation needs and temporal
performance differences of gridded precipitation products in peripheral
mountain regions, Scientific Reports, 9, 15118, <ext-link xlink:href="https://doi.org/10.1038/s41598-019-51666-z" ext-link-type="DOI">10.1038/s41598-019-51666-z</ext-link>, 2019.</mixed-citation></ref>
      <ref id="bib1.bibx82"><?xmltex \def\ref@label{{Zemp et~al.(2013)Zemp, Thibert, Huss, Stumm, Rolstad~Denby, Nuth,
Nussbaumer, Moholdt, Mercer, Mayer, Joerg, Jansson, Hynek, Fischer,
Escher-Vetter, Elveh{\o}y, and Andreassen}}?><label>Zemp et al.(2013)Zemp, Thibert, Huss, Stumm, Rolstad Denby, Nuth,
Nussbaumer, Moholdt, Mercer, Mayer, Joerg, Jansson, Hynek, Fischer,
Escher-Vetter, Elvehøy, and Andreassen</label><?label zemp2013?><mixed-citation>Zemp, M., Thibert, E., Huss, M., Stumm, D., Rolstad Denby, C., Nuth, C., Nussbaumer, S. U., Moholdt, G., Mercer, A., Mayer, C., Joerg, P. C., Jansson, P., Hynek, B., Fischer, A., Escher-Vetter, H., Elvehøy, H., and Andreassen, L. M.: Reanalysing glacier mass balance measurement series, The Cryosphere, 7, 1227–1245, <ext-link xlink:href="https://doi.org/10.5194/tc-7-1227-2013" ext-link-type="DOI">10.5194/tc-7-1227-2013</ext-link>, 2013.</mixed-citation></ref>
      <ref id="bib1.bibx83"><?xmltex \def\ref@label{{Zemp et~al.(2021)Zemp, Nussbaumer, G{\"{a}}rtner-Roer, Bannwart, Paul,
and Hoelzle}}?><label>Zemp et al.(2021)Zemp, Nussbaumer, Gärtner-Roer, Bannwart, Paul,
and Hoelzle</label><?label zora209777?><mixed-citation>Zemp, M., Nussbaumer, S. U., Gärtner-Roer, I., Bannwart, J., Paul, F., and
Hoelzle, M.: Global Glacier Change Bulletin Nr. 4 (2018-2019), Tech. Rep.,
World Glacier Monitoring Service, Zürich,
<ext-link xlink:href="https://doi.org/10.5167/uzh-209777" ext-link-type="DOI">10.5167/uzh-209777</ext-link>, 2021.</mixed-citation></ref>
      <ref id="bib1.bibx84"><?xmltex \def\ref@label{{Zhang et~al.(2019)Zhang, Liu, and Huang}}?><label>Zhang et al.(2019)Zhang, Liu, and Huang</label><?label zhang2019automatically?><mixed-citation>Zhang, E., Liu, L., and Huang, L.: Automatically delineating the calving front of Jakobshavn Isbræ from multitemporal TerraSAR-X images: a deep learning approach, The Cryosphere, 13, 1729–1741, <ext-link xlink:href="https://doi.org/10.5194/tc-13-1729-2019" ext-link-type="DOI">10.5194/tc-13-1729-2019</ext-link>, 2019.</mixed-citation></ref>
      <ref id="bib1.bibx85"><?xmltex \def\ref@label{{Zhao and He(2022)}}?><label>Zhao and He(2022)</label><?label zhao2022?><mixed-citation>Zhao, P. and He, Z.: A First Evaluation of ERA5-Land Rean<?pagebreak page2828?>alysis Temperature
Product Over the Chinese Qilian Mountains, Frontiers in Earth Science, 10, 907730,
<ext-link xlink:href="https://doi.org/10.3389/feart.2022.907730" ext-link-type="DOI">10.3389/feart.2022.907730</ext-link>, 2022.</mixed-citation></ref>
      <ref id="bib1.bibx86"><?xmltex \def\ref@label{{Zhu et~al.(2021)Zhu, Zhang, Wang, Tian, Liu, Ma, Kan, and
Chu}}?><label>Zhu et al.(2021)Zhu, Zhang, Wang, Tian, Liu, Ma, Kan, and
Chu</label><?label zhu2021downscaling?><mixed-citation>Zhu, L., Zhang, Y., Wang, J., Tian, W., Liu, Q., Ma, G., Kan, X., and Chu, Y.:
Downscaling Snow Depth Mapping by Fusion of Microwave and Optical
Remote-Sensing Data Based on Deep Learning, Remote Sensing, 13, 584,
<ext-link xlink:href="https://doi.org/10.3390/rs13040584" ext-link-type="DOI">10.3390/rs13040584</ext-link>, 2021.</mixed-citation></ref>

  </ref-list></back>
    <!--<article-title-html>Modelling point mass balance for the glaciers of the Central European Alps using machine learning techniques</article-title-html>
<abstract-html/>
<ref-html id="bib1.bib1"><label>Altmann et al.(2010)Altmann, Toloşi, Sander, and
Lengauer</label><mixed-citation>
      
Altmann, A., Toloşi, L., Sander, O., and Lengauer, T.: Permutation
importance: a corrected feature importance measure, Bioinformatics, 26,
1340–1347, <a href="https://doi.org/10.1093/bioinformatics/btq134" target="_blank">https://doi.org/10.1093/bioinformatics/btq134</a>, 2010.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib2"><label>Anilkumar et al.(2022)Anilkumar, Bharti, and
Chutia</label><mixed-citation>
      
Anilkumar, R., Bharti, R., and Chutia, D.: Point Mass Balance Regression using Deep Neural Networks: A Transfer Learning Approach, EGU General Assembly 2022, Vienna, Austria, 23–27 May 2022, EGU22-5317, <a href="https://doi.org/10.5194/egusphere-egu22-5317" target="_blank">https://doi.org/10.5194/egusphere-egu22-5317</a>, 2022.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib3"><label>Bair et al.(2018)Bair, Abreu Calfa, Rittger, and Dozier</label><mixed-citation>
      
Bair, E. H., Abreu Calfa, A., Rittger, K., and Dozier, J.: Using machine learning for real-time estimates of snow water equivalent in the watersheds of Afghanistan, The Cryosphere, 12, 1579–1594, <a href="https://doi.org/10.5194/tc-12-1579-2018" target="_blank">https://doi.org/10.5194/tc-12-1579-2018</a>, 2018.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib4"><label>Bash et al.(2018)Bash, Moorman, and Gunther</label><mixed-citation>
      
Bash, E. A., Moorman, B. J., and Gunther, A.: Detecting Short-Term Surface Melt
on an Arctic Glacier Using UAV Surveys, Remote Sensing, 10, 1547,
<a href="https://doi.org/10.3390/rs10101547" target="_blank">https://doi.org/10.3390/rs10101547</a>, 2018.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib5"><label>Bolibar et al.(2020)Bolibar, Rabatel, Gouttevin, Galiez, Condom, and
Sauquet</label><mixed-citation>
      
Bolibar, J., Rabatel, A., Gouttevin, I., Galiez, C., Condom, T., and Sauquet, E.: Deep learning applied to glacier evolution modelling, The Cryosphere, 14, 565–584, <a href="https://doi.org/10.5194/tc-14-565-2020" target="_blank">https://doi.org/10.5194/tc-14-565-2020</a>, 2020.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib6"><label>Bolibar et al.(2022)Bolibar, Rabatel, Gouttevin, Zekollari, and
Galiez</label><mixed-citation>
      
Bolibar, J., Rabatel, A., Gouttevin, I., Zekollari, H., and Galiez, C.:
Nonlinear sensitivity of glacier mass balance to future climate change
unveiled by deep learning, Nat. Commun., 13, 409,
<a href="https://doi.org/10.1038/s41467-022-28033-0" target="_blank">https://doi.org/10.1038/s41467-022-28033-0</a>, 2022.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib7"><label>Braithwaite(1984)</label><mixed-citation>
      
Braithwaite, R. J.: Can the Mass Balance of a Glacier be Estimated from its
Equilibrium-Line Altitude?, J. Glaciol., 30, 364–368,
<a href="https://doi.org/10.3189/S0022143000006237" target="_blank">https://doi.org/10.3189/S0022143000006237</a>, 1984.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib8"><label>Breiman(1996)</label><mixed-citation>
      
Breiman, L.: Bagging predictors, Machine Learning, 24, 123–140,
<a href="https://doi.org/10.1007/BF00058655" target="_blank">https://doi.org/10.1007/BF00058655</a>, 1996.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib9"><label>Breiman(2001)</label><mixed-citation>
      
Breiman, L.: Random forests, Machine Learning, 45, 5–32,
<a href="https://doi.org/10.1023/A:1010933404324" target="_blank">https://doi.org/10.1023/A:1010933404324</a>, 2001.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib10"><label>Carturan et al.(2009)Carturan, Cazorzi, and
Dalla Fontana</label><mixed-citation>
      
Carturan, L., Cazorzi, F., and Dalla Fontana, G.: Enhanced estimation of
glacier mass balance in unsampled areas by means of topographic data, Ann. Glaciol., 50, 37–46, <a href="https://doi.org/10.3189/172756409787769519" target="_blank">https://doi.org/10.3189/172756409787769519</a>, 2009.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib11"><label>Chen et al.(2021)Chen, Sharma, Zhou, Yang, Li, Niu, Hu, and
Khadka</label><mixed-citation>
      
Chen, Y., Sharma, S., Zhou, X., Yang, K., Li, X., Niu, X., Hu, X., and Khadka,
N.: Spatial performance of multiple reanalysis precipitation datasets on the
southern slope of central Himalaya, Atmos. Res., 250, 105365,
<a href="https://doi.org/10.1016/j.atmosres.2020.105365" target="_blank">https://doi.org/10.1016/j.atmosres.2020.105365</a>, 2021.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib12"><label>Chi and Kim(2017)</label><mixed-citation>
      
Chi, J. and Kim, H.-C.: Prediction of Arctic Sea Ice Concentration Using a
Fully Data Driven Deep Neural Network, Remote Sensing, 9, 1305,
<a href="https://doi.org/10.3390/rs9121305" target="_blank">https://doi.org/10.3390/rs9121305</a>, 2017.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib13"><label>Cortes and Vapnik(1995)</label><mixed-citation>
      
Cortes, C. and Vapnik, V.: Support-vector networks, Machine Learning, 20,
273–297, <a href="https://doi.org/10.1007/BF00994018" target="_blank">https://doi.org/10.1007/BF00994018</a>, 1995.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib14"><label>Cucchi et al.(2020)Cucchi, Weedon, Amici, Bellouin, Lange,
Müller Schmied, Hersbach, and Buontempo</label><mixed-citation>
      
Cucchi, M., Weedon, G. P., Amici, A., Bellouin, N., Lange, S., Müller Schmied, H., Hersbach, H., and Buontempo, C.: WFDE5: bias-adjusted ERA5 reanalysis data for impact studies, Earth Syst. Sci. Data, 12, 2097–2120, <a href="https://doi.org/10.5194/essd-12-2097-2020" target="_blank">https://doi.org/10.5194/essd-12-2097-2020</a>, 2020.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib15"><label>Dietterich(2000)</label><mixed-citation>
      
Dietterich, T. G.: An experimental comparison of three methods for constructing
ensembles of decision trees: Bagging, boosting, and randomization, Machine
Learning, 40, 139–157, <a href="https://doi.org/10.1023/A:1007607513941" target="_blank">https://doi.org/10.1023/A:1007607513941</a>, 2000.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib16"><label>Dobhal et al.(2021)Dobhal, Pratap, Bhambri, and
Mehta</label><mixed-citation>
      
Dobhal, D., Pratap, B., Bhambri, R., and Mehta, M.: Mass balance and
morphological changes of Dokriani Glacier (1992–2013), Garhwal Himalaya,
India, Quaternary Science Advances, 4, 100033,
<a href="https://doi.org/10.1016/j.qsa.2021.100033" target="_blank">https://doi.org/10.1016/j.qsa.2021.100033</a>, 2021.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib17"><label>Fernández-Delgado et al.(2014)Fernández-Delgado, Cernadas,
Barro, and Amorim</label><mixed-citation>
      
Fernández-Delgado, M., Cernadas, E., Barro, S., and Amorim, D.: Do We Need
Hundreds of Classifiers to Solve Real World Classification Problems?, J.
Mach. Learn. Res., 15, 3133–3181, 2014.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib18"><label>Friedman(2001)</label><mixed-citation>
      
Friedman, J. H.: Greedy function approximation: a gradient boosting machine,
Ann. Stat., 29, 1189–1232, <a href="https://doi.org/10.1214/aos/1013203451" target="_blank">https://doi.org/10.1214/aos/1013203451</a>, 2001.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib19"><label>Gabbi et al.(2014)Gabbi, Carenzo, Pellicciotti, Bauder, and
Funk</label><mixed-citation>
      
Gabbi, J., Carenzo, M., Pellicciotti, F., Bauder, A., and Funk, M.: A
comparison of empirical and physically based glacier surface melt models for
long-term simulations of glacier response, J. Glaciol., 60,
1140–1154, <a href="https://doi.org/10.3189/2014JoG14J011" target="_blank">https://doi.org/10.3189/2014JoG14J011</a>, 2014.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib20"><label>Gerbaux et al.(2005)Gerbaux, Genthon, Etchevers, Vincent, and
Dedieu</label><mixed-citation>
      
Gerbaux, M., Genthon, C., Etchevers, P., Vincent, C., and Dedieu, J.: Surface
mass balance of glaciers in the French Alps: distributed modeling and
sensitivity to climate change, J. Glaciol., 51, 561–572,
<a href="https://doi.org/10.3189/172756505781829133" target="_blank">https://doi.org/10.3189/172756505781829133</a>, 2005.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib21"><label>Grinsztajn et al.(2022)Grinsztajn, Oyallon, and
Varoquaux</label><mixed-citation>
      
Grinsztajn, L., Oyallon, E., and Varoquaux, G.: Why do tree-based models still
outperform deep learning on tabular data?, in: Advances in Neural Information Processing Systems, edited by:  Koyejo, S.,  Mohamed, S.,  Agarwal, A., Belgrave, D.,  Cho, K., and  Oh, A., Curran Associates, Inc., vol. 35, 507–520, <a href="https://proceedings.neurips.cc/paper_files/paper/2022/file/0378c7692da36807bdec87ab043cdadc-Paper-Datasets_and_Benchmarks.pdf" target="_blank"/> (last access: 7 July 2023), 2022.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib22"><label>Guidicelli et al.(2023)Guidicelli, Huss, Gabella, and
Salzmann</label><mixed-citation>
      
Guidicelli, M., Huss, M., Gabella, M., and Salzmann, N.: Spatio-temporal reconstruction of winter glacier mass balance in the Alps, Scandinavia, Central Asia and western Canada (1981–2019) using climate reanalyses and machine learning, The Cryosphere, 17, 977–1002, <a href="https://doi.org/10.5194/tc-17-977-2023" target="_blank">https://doi.org/10.5194/tc-17-977-2023</a>, 2023.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib23"><label>Guo et al.(2020)Guo, Chen, Liu, and Zhao</label><mixed-citation>
      
Guo, X., Chen, Y., Liu, X., and Zhao, Y.: Extraction of snow cover from
high-resolution remote sensing imagery using deep learning on a small
dataset, Remote Sens. Lett., 11, 66–75,
<a href="https://doi.org/10.1080/2150704X.2019.1686548" target="_blank">https://doi.org/10.1080/2150704X.2019.1686548</a>, 2020.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib24"><label>Haq et al.(2014)Haq, Jain, and Menon</label><mixed-citation>
      
Haq, M. A., Jain, K., and Menon, K.: Modelling of Gangotri glacier thickness
and volume using an artificial neural network, Int. J.
Remote Sens., 35, 6035–6042, <a href="https://doi.org/10.1080/01431161.2014.943322" target="_blank">https://doi.org/10.1080/01431161.2014.943322</a>, 2014.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib25"><label>Haq et al.(2021)Haq, Azam, and Vincent</label><mixed-citation>
      
Haq, M. A., Azam, M. F., and Vincent, C.: Efficiency of artificial neural
networks for glacier ice-thickness estimation: a case study in western
Himalaya, India, J. Glaciol., 67, 671–684,
<a href="https://doi.org/10.1017/jog.2021.19" target="_blank">https://doi.org/10.1017/jog.2021.19</a>, 2021.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib26"><label>Hersbach et al.(2020)Hersbach, Bell, Berrisford, Hirahara, Horányi,
Muñoz-Sabater, Nicolas, Peubey, Radu, Schepers, Simmons, Soci, Abdalla,
Abellan, Balsamo, Bechtold, Biavati, Bidlot, Bonavita, De Chiara, Dahlgren,
Dee, Diamantakis, Dragani, Flemming, Forbes, Fuentes, Geer, Haimberger,
Healy, Hogan, Hólm, Janisková, Keeley, Laloyaux, Lopez, Lupu, Radnoti,
de Rosnay, Rozum, Vamborg, Villaume, and Thépaut</label><mixed-citation>
      
Hersbach, H., Bell, B., Berrisford, P., Hirahara, S., Horányi, A.,
Muñoz-Sabater, J., Nicolas, J., Peubey, C., Radu, R., Schepers, D., Simmons,
A., Soci, C., Abdalla, S., Abellan, X., Balsamo, G., Bechtold, P., Biavati,
G., Bidlot, J., Bonavita, M., De Chiara, G., Dahlgren, P., Dee, D.,
Diamantakis, M., Dragani, R., Flemming, J., Forbes, R., Fuentes, M., Geer,
A., Haimberger, L., Healy, S., Hogan, R. J., Hólm, E., Janisková, M.,
Keeley, S., Laloyaux, P., Lopez, P., Lupu, C., Radnoti, G., de Rosnay, P.,
Rozum, I., Vamborg, F., Villaume, S., and Thépaut, J.-N.: The ERA5 global
reanalysis, Q. J. Roy. Meteor. Soc., 146,
1999–2049, <a href="https://doi.org/10.1002/qj.3803" target="_blank">https://doi.org/10.1002/qj.3803</a>, 2020.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib27"><label>Hoinkes(1968)</label><mixed-citation>
      
Hoinkes, H. C.: Glacier Variation and Weather, J. Glaciol., 7,
3–18, <a href="https://doi.org/10.3189/S0022143000020384" target="_blank">https://doi.org/10.3189/S0022143000020384</a>, 1968.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib28"><label>Hornik(1991)</label><mixed-citation>
      
Hornik, K.: Approximation capabilities of multilayer feedforward networks,
Neural Networks, 4, 251–257,
<a href="https://doi.org/10.1016/0893-6080(91)90009-T" target="_blank">https://doi.org/10.1016/0893-6080(91)90009-T</a>, 1991.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib29"><label>Huss et al.(2008)Huss, Farinotti, Bauder, and
Funk</label><mixed-citation>
      
Huss, M., Farinotti, D., Bauder, A., and Funk, M.: Modelling runoff from highly
glacierized alpine drainage basins in a changing climate, Hydrol.
Process., 22, 3888–3902, <a href="https://doi.org/10.1002/hyp.7055" target="_blank">https://doi.org/10.1002/hyp.7055</a>, 2008.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib30"><label>Ismail et al.(2023)Ismail, Bogacki, Disse, Schäfer, and
Kirschbauer</label><mixed-citation>
      
Ismail, M. F., Bogacki, W., Disse, M., Schäfer, M., and Kirschbauer, L.: Estimating degree-day factors of snow based on energy flux components, The Cryosphere, 17, 211–231, <a href="https://doi.org/10.5194/tc-17-211-2023" target="_blank">https://doi.org/10.5194/tc-17-211-2023</a>, 2023.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib31"><label>Jouvet et al.(2021)Jouvet, Cordonnier, Kim, Lüthi, Vieli, and
Aschwanden</label><mixed-citation>
      
Jouvet, G., Cordonnier, G., Kim, B., Lüthi, M., Vieli, A., and Aschwanden, A.:
Deep learning speeds up ice flow modelling by several orders of magnitude,
J. Glaciol., 68, 651–664, <a href="https://doi.org/10.1017/jog.2021.120" target="_blank">https://doi.org/10.1017/jog.2021.120</a>, 2021.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib32"><label>Kan et al.(2018)Kan, Zhang, Zhu, Xiao, Wang, Tian, and
Tan</label><mixed-citation>
      
Kan, X., Zhang, Y., Zhu, L., Xiao, L., Wang, J., Tian, W., and Tan, H.: Snow
cover mapping for mountainous areas by fusion of MODIS L1B and geographic
data based on stacked denoising auto-encoders, Computers, Materials &amp; Continua, 57, 49–68, <a href="https://doi.org/10.32604/cmc.2018.02376" target="_blank">https://doi.org/10.32604/cmc.2018.02376</a>, 2018.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib33"><label>King et al.(2020)King, Erler, Frey, and Fletcher</label><mixed-citation>
      
King, F., Erler, A. R., Frey, S. K., and Fletcher, C. G.: Application of machine learning techniques for regional bias correction of snow water equivalent estimates in Ontario, Canada, Hydrol. Earth Syst. Sci., 24, 4887–4902, <a href="https://doi.org/10.5194/hess-24-4887-2020" target="_blank">https://doi.org/10.5194/hess-24-4887-2020</a>, 2020.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib34"><label>Kingma and Ba(2014)</label><mixed-citation>
      
Kingma, D. P. and Ba, J.: Adam: A method for stochastic optimization, arXiv
[preprint], <a href="https://doi.org/10.48550/arxiv.1412.6980" target="_blank">https://doi.org/10.48550/arxiv.1412.6980</a>, 22 December 2014.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib35"><label>Kuhn et al.(1999)Kuhn, Dreiseitl, Hofinger, Markl, Span, and
Kaser</label><mixed-citation>
      
Kuhn, M., Dreiseitl, E., Hofinger, S., Markl, G., Span, N., and Kaser, G.:
Measurements and models of the mass balance of hintereisferner, Geogr.
Ann. A, 81, 659–670,
<a href="https://onlinelibrary.wiley.com/doi/abs/10.1111/1468-0459.00094" target="_blank"/> (last access: 7 July 2023), 1999.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib36"><label>Lefauconnier and Hagen(1990)</label><mixed-citation>
      
Lefauconnier, B. and Hagen, J.: Glaciers and Climate in Svalbard: Statistical
Analysis and Reconstruction of the Brøggerbreen Mass Balance for the Last 77
Years, Ann. Glaciol., 14, 148–152, <a href="https://doi.org/10.3189/S0260305500008466" target="_blank">https://doi.org/10.3189/S0260305500008466</a>,
1990.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib37"><label>Le Meur et al.(2007)Le Meur, Gerbaux, Schäfer, and
Vincent</label><mixed-citation>
      
Le Meur, E., Gerbaux, M., Schäfer, M., and Vincent, C.: Disappearance of an
Alpine glacier over the 21st Century simulated from modeling its future
surface mass balance, Earth Planet. Sc. Lett., 261, 367–374,
<a href="https://doi.org/10.1016/j.epsl.2007.07.022" target="_blank">https://doi.org/10.1016/j.epsl.2007.07.022</a>, 2007.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib38"><label>Li et al.(2017)Li, Wang, Wang, Zhang, Fu, and Wang</label><mixed-citation>
      
Li, J., Wang, C., Wang, S., Zhang, H., Fu, Q., and Wang, Y.: Gaofen-3 sea ice
detection based on deep learning, in: 2017 Progress in Electromagnetics
Research Symposium – Fall (PIERS – FALL), Nanyang Technological Univeiisty, Singapore, 19–22 November 2017, 933–939,
<a href="https://doi.org/10.1109/PIERS-FALL.2017.8293267" target="_blank">https://doi.org/10.1109/PIERS-FALL.2017.8293267</a>, 2017.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib39"><label>Lin et al.(2018)Lin, Chen, Yang, and Ou</label><mixed-citation>
      
Lin, C., Chen, D., Yang, K., and Ou, T.: Impact of model resolution on
simulating the water vapor transport through the central Himalayas:
implication for models’ wet bias over the Tibetan Plateau, Clim.
Dynam., 51, 3195–3207, 2018.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib40"><label>Liu(2021)</label><mixed-citation>
      
Liu, L.: A Review of Deep Learning for Cryospheric Studies, chap. 17,
John Wiley and Sons, Ltd, 258–268,
<a href="https://doi.org/10.1002/9781119646181.ch17" target="_blank">https://doi.org/10.1002/9781119646181.ch17</a>, 2021.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib41"><label>Lliboutry(1974)</label><mixed-citation>
      
Lliboutry, L.: Multivariate Statistical Analysis of Glacier Annual Balances,
J. Glaciol., 13, 371–392, <a href="https://doi.org/10.3189/S0022143000023169" target="_blank">https://doi.org/10.3189/S0022143000023169</a>, 1974.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib42"><label>Lu et al.(2021)Lu, Zhang, Shangguan, and Yang</label><mixed-citation>
      
Lu, Y., Zhang, Z., Shangguan, D., and Yang, J.: Novel Machine Learning Method
Integrating Ensemble Learning and Deep Learning for Mapping Debris-Covered
Glaciers, Remote Sensing, 13, 2595, <a href="https://doi.org/10.3390/rs13132595" target="_blank">https://doi.org/10.3390/rs13132595</a>, 2021.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib43"><label>Manciati et al.(2014)Manciati, Villacís, Taupin, Cadier,
Galárraga-Sánchez, and Cáceres</label><mixed-citation>
      
Manciati, C., Villacís, M., Taupin, J.-D., Cadier, E., Galárraga-Sánchez,
R., and Cáceres, B.: Empirical mass balance modelling of South American
tropical glaciers: case study of Antisana volcano, Ecuador, Hydrolog.
Sci. J., 59, 1519–1535, <a href="https://doi.org/10.1080/02626667.2014.888490" target="_blank">https://doi.org/10.1080/02626667.2014.888490</a>, 2014.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib44"><label>Masiokas et al.(2016)</label><mixed-citation>
      
Masiokas, M. H., Christie, D. A., Le Quesne, C., Pitte, P., Ruiz, L., Villalba, R., Luckman, B. H., Berthier, E., Nussbaumer, S. U., González-Reyes, Á., McPhee, J., and Barcaza, G.: Reconstructing the annual mass balance of the Echaurren Norte glacier (Central Andes, 33.5°&thinsp;S) using local and regional hydroclimatic data, The Cryosphere, 10, 927–940, <a href="https://doi.org/10.5194/tc-10-927-2016" target="_blank">https://doi.org/10.5194/tc-10-927-2016</a>, 2016.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib45"><label>Mattews and Hodgkins(2016)</label><mixed-citation>
      
Mattews, T. and Hodgkins, R.: Interdecadal variability of degree-day factors on
Vestari Hagafellsjökull (Langjökull, Iceland) and the importance of
threshold air temperatures, J. Glaciol., 62, 310–322,
<a href="https://doi.org/10.1017/jog.2016.21" target="_blank">https://doi.org/10.1017/jog.2016.21</a>, 2016.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib46"><label>Maussion et al.(2019)Maussion, Butenko, Champollion, Dusch, Eis,
Fourteau, Gregor, Jarosch, Landmann, Oesterle, Recinos, Rothenpieler, Vlug,
Wild, and Marzeion</label><mixed-citation>
      
Maussion, F., Butenko, A., Champollion, N., Dusch, M., Eis, J., Fourteau, K., Gregor, P., Jarosch, A. H., Landmann, J., Oesterle, F., Recinos, B., Rothenpieler, T., Vlug, A., Wild, C. T., and Marzeion, B.: The Open Global Glacier Model (OGGM) v1.1, Geosci. Model Dev., 12, 909–931, <a href="https://doi.org/10.5194/gmd-12-909-2019" target="_blank">https://doi.org/10.5194/gmd-12-909-2019</a>, 2019.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib47"><label>McCulloch and Pitts(1943)</label><mixed-citation>
      
McCulloch, W. S. and Pitts, W.: A logical calculus of the ideas immanent in
nervous activity, B. Math. Biophys., 5, 115–133,
<a href="https://doi.org/10.1007/BF02478259" target="_blank">https://doi.org/10.1007/BF02478259</a>, 1943.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib48"><label>Mohajerani et al.(2021)Mohajerani, Jeong, Scheuchl, Velicogna,
Rignot, and Milillo</label><mixed-citation>
      
Mohajerani, Y., Jeong, S., Scheuchl, B., Velicogna, I., Rignot, E., and
Milillo, P.: Automatic delineation of glacier grounding lines in differential
interferometric synthetic-aperture radar data using deep learning, Scientific
Reports, 11, 4992, <a href="https://doi.org/10.1038/s41598-021-84309-3" target="_blank">https://doi.org/10.1038/s41598-021-84309-3</a>, 2021.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib49"><label>Moya Quiroga et al.(2013)Moya Quiroga, Mano, Asaoka, Kure, Udo, and
Mendoza</label><mixed-citation>
      
Moya Quiroga, V., Mano, A., Asaoka, Y., Kure, S., Udo, K., and Mendoza, J.: Snow glacier melt estimation in tropical Andean glaciers using artificial neural networks, Hydrol. Earth Syst. Sci., 17, 1265–1280, <a href="https://doi.org/10.5194/hess-17-1265-2013" target="_blank">https://doi.org/10.5194/hess-17-1265-2013</a>, 2013.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib50"><label>Muñoz Sabater(2019)</label><mixed-citation>
      
Muñoz Sabater, J.: ERA5-Land hourly data from 1950 to present, Copernicus Climate Change Service (C3S) Climate Data Store (CDS) [data set], <a href="https://doi.org/10.24381/cds.e2161bac" target="_blank">https://doi.org/10.24381/cds.e2161bac</a>, 2019.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib51"><label>Muñoz Sabater et al.(2021)Muñoz Sabater, Dutra,
Agustí-Panareda, Albergel, Arduini, Balsamo, Boussetta, Choulga,
Harrigan, Hersbach, Martens, Miralles, Piles, Rodríguez-Fernández,
Zsoter, Buontempo, and Thépaut</label><mixed-citation>
      
Muñoz-Sabater, J., Dutra, E., Agustí-Panareda, A., Albergel, C., Arduini, G., Balsamo, G., Boussetta, S., Choulga, M., Harrigan, S., Hersbach, H., Martens, B., Miralles, D. G., Piles, M., Rodríguez-Fernández, N. J., Zsoter, E., Buontempo, C., and Thépaut, J.-N.: ERA5-Land: a state-of-the-art global reanalysis dataset for land applications, Earth Syst. Sci. Data, 13, 4349–4383, <a href="https://doi.org/10.5194/essd-13-4349-2021" target="_blank">https://doi.org/10.5194/essd-13-4349-2021</a>, 2021.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib52"><label>Natekin and Knoll(2013)</label><mixed-citation>
      
Natekin, A. and Knoll, A.: Gradient boosting machines, a tutorial, Front.
Neurorobotics, 7, 21, <a href="https://doi.org/10.3389/fnbot.2013.00021" target="_blank">https://doi.org/10.3389/fnbot.2013.00021</a>, 2013.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib53"><label>Nijhawan et al.(2019)Nijhawan, Das, and Raman</label><mixed-citation>
      
Nijhawan, R., Das, J., and Raman, B.: A hybrid of deep learning and
hand-crafted features based approach for snow cover mapping, Int.
J. Remote Sens., 40, 759–773, <a href="https://doi.org/10.1080/01431161.2018.1519277" target="_blank">https://doi.org/10.1080/01431161.2018.1519277</a>,
2019.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib54"><label>Pratap et al.(2016)Pratap, Dobhal, Bhambri, Mehta, and
Tewari</label><mixed-citation>
      
Pratap, B., Dobhal, D. P., Bhambri, R., Mehta, M., and Tewari, V. C.: Four
decades of glacier mass balance observations in the Indian Himalaya, Reg.
Environ. Change, 16, 643–658,
<a href="https://doi.org/10.1007/s10113-015-0791-4" target="_blank">https://doi.org/10.1007/s10113-015-0791-4</a>, 2016.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib55"><label>Rabatel et al.(2016)Rabatel, Dedieu, and Vincent</label><mixed-citation>
      
Rabatel, A., Dedieu, J. P., and Vincent, C.: Spatio-temporal changes in
glacier-wide mass balance quantified by optical remote sensing on 30 glaciers
in the French Alps for the period 1983–2014, J. Glaciol., 62,
1153–1166, <a href="https://doi.org/10.1017/jog.2016.113" target="_blank">https://doi.org/10.1017/jog.2016.113</a>, 2016.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib56"><label>Radić and Hock(2011)</label><mixed-citation>
      
Radić, V. and Hock, R.: Regionally differentiated contribution of mountain
glaciers and ice caps to future sea-level rise, Nat. Geosci,, 4, 91–94,
<a href="https://doi.org/10.1038/ngeo1052" target="_blank">https://doi.org/10.1038/ngeo1052</a>, 2011.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib57"><label>Radić et al.(2014)Radić, Bliss, Beedlow, Hock, Miles, and
Cogley</label><mixed-citation>
      
Radić, V., Bliss, A., Beedlow, A. C., Hock, R., Miles, E., and Cogley,
J. G.: Regional and global projections of twenty-first century glacier mass
changes in response to climate scenarios from global climate models, Clim.
Dynam., 42, 37–58, 2014.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib58"><label>Rasouli et al.(2012)Rasouli, Hsieh, and Cannon</label><mixed-citation>
      
Rasouli, K., Hsieh, W. W., and Cannon, A. J.: Daily streamflow forecasting by
machine learning methods with weather and climate inputs, J.
Hydrol., 414-415, 284–293,
<a href="https://doi.org/10.1016/j.jhydrol.2011.10.039" target="_blank">https://doi.org/10.1016/j.jhydrol.2011.10.039</a>, 2012.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib59"><label>Rasp et al.(2018)Rasp, Pritchard, and Gentine</label><mixed-citation>
      
Rasp, S., Pritchard, M. S., and Gentine, P.: Deep learning to represent subgrid
processes in climate models, P. Natl. Acad. Sci. USA,
115, 9684–9689, <a href="https://doi.org/10.1073/pnas.1810286115" target="_blank">https://doi.org/10.1073/pnas.1810286115</a>, 2018.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib60"><label>RGI(2017)</label><mixed-citation>
      
RGI: Randolph Glacier Inventory (RGI) – A Dataset of Global Glacier Outlines:
Version 6.0. Technical Report, <a href="https://doi.org/10.7265/N5-RGI-60" target="_blank">https://doi.org/10.7265/N5-RGI-60</a>, 2017.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib61"><label>Roberts et al.(2017)Roberts, Bahn, Ciuti, Boyce, Elith,
Guillera-Arroita, Hauenstein, Lahoz-Monfort, Schröder, Thuiller, Warton,
Wintle, Hartig, and Dormann</label><mixed-citation>
      
Roberts, D. R., Bahn, V., Ciuti, S., Boyce, M. S., Elith, J., Guillera-Arroita,
G., Hauenstein, S., Lahoz-Monfort, J. J., Schröder, B., Thuiller, W.,
Warton, D. I., Wintle, B. A., Hartig, F., and Dormann, C. F.:
Cross-validation strategies for data with temporal, spatial, hierarchical, or
phylogenetic structure, Ecography, 40, 913–929,
<a href="https://doi.org/10.1111/ecog.02881" target="_blank">https://doi.org/10.1111/ecog.02881</a>, 2017.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib62"><label>Sauter et al.(2020)Sauter, Arndt, and Schneider</label><mixed-citation>
      
Sauter, T., Arndt, A., and Schneider, C.: COSIPY v1.3 – an open-source coupled snowpack and ice surface energy and mass balance model, Geosci. Model Dev., 13, 5645–5662, <a href="https://doi.org/10.5194/gmd-13-5645-2020" target="_blank">https://doi.org/10.5194/gmd-13-5645-2020</a>, 2020.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib63"><label>Schultz et al.(2021)Schultz, Betancourt, Gong, Kleinert, Langguth,
Leufen, Mozaffari, and Stadtler</label><mixed-citation>
      
Schultz, M. G., Betancourt, C., Gong, B., Kleinert, F., Langguth, M., Leufen,
L. H., Mozaffari, A., and Stadtler, S.: Can deep learning beat numerical
weather prediction?, Philosophical Transactions of the Royal Society A:
Mathematical, Physical and Engineering Sciences, 379, 20200097,
<a href="https://doi.org/10.1098/rsta.2020.0097" target="_blank">https://doi.org/10.1098/rsta.2020.0097</a>, 2021.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib64"><label>Seidou et al.(2006)Seidou, Ouarda, Bilodeau, Hessami, St-Hilaire, and
Bruneau</label><mixed-citation>
      
Seidou, O., Ouarda, T. B. M. J., Bilodeau, L., Hessami, M., St-Hilaire, A., and
Bruneau, P.: Modeling ice growth on Canadian lakes using artificial neural
networks, Water Resour. Res., 42, W11407,
<a href="https://doi.org/10.1029/2005WR004622" target="_blank">https://doi.org/10.1029/2005WR004622</a>, 2006.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib65"><label>Shean et al.(2020)Shean, Bhushan, Montesano, Rounce, Arendt, and
Osmanoglu</label><mixed-citation>
      
Shean, D. E., Bhushan, S., Montesano, P., Rounce, D. R., Arendt, A., and
Osmanoglu, B.: A Systematic, Regional Assessment of High Mountain Asia
Glacier Mass Balance, Frontiers in Earth Science, 7, 363,
<a href="https://doi.org/10.3389/feart.2019.00363" target="_blank">https://doi.org/10.3389/feart.2019.00363</a>, 2020.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib66"><label>Shwartz-Ziv and Armon(2021)</label><mixed-citation>
      
Shwartz-Ziv, R. and Armon, A.: Tabular Data: Deep Learning is Not All You
Need, arXiv [preprint], <a href="https://doi.org/10.48550/arXiv.2106.03253" target="_blank">https://doi.org/10.48550/arXiv.2106.03253</a>,
6 June 2021.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib67"><label>Steiner et al.(2005)Steiner, Walter, and Zumbühl</label><mixed-citation>
      
Steiner, D., Walter, A., and Zumbühl, H.: The application of a non-linear
back-propagation neural network to study the mass balance of Grosse
Aletschgletscher, Switzerland, J. Glaciol., 51, 313–323,
<a href="https://doi.org/10.3189/172756505781829421" target="_blank">https://doi.org/10.3189/172756505781829421</a>, 2005.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib68"><label>Thibert et al.(2008)Thibert, Blanc, Vincent, and
Eckert</label><mixed-citation>
      
Thibert, E., Blanc, R., Vincent, C., and Eckert, N.: Glaciological and
volumetric mass-balance measurements: error analysis over 51 years for
Glacier de Sarennes, French Alps, J. Glaciol., 54, 522–532,
<a href="https://doi.org/10.3189/002214308785837093" target="_blank">https://doi.org/10.3189/002214308785837093</a>, 2008.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib69"><label>Trantow and Herzfeld(2016)</label><mixed-citation>
      
Trantow, T. and Herzfeld, U. C.: Spatiotemporal mapping of a large mountain
glacier from CryoSat-2 altimeter data: surface elevation and elevation change
of Bering Glacier during surge (2011–2014), Int. J. Remote
Sens., 37, 2962–2989, <a href="https://doi.org/10.1080/01431161.2016.1187318" target="_blank">https://doi.org/10.1080/01431161.2016.1187318</a>, 2016.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib70"><label>Tshering and Fujita(2016)</label><mixed-citation>
      
Tshering, P. and Fujita, K.: First in situ record of decadal glacier mass
balance (2003–2014) from the Bhutan Himalaya, Ann. Glaciol., 57,
289–294, <a href="https://doi.org/10.3189/2016AoG71A036" target="_blank">https://doi.org/10.3189/2016AoG71A036</a>, 2016.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib71"><label>Van Tricht et al.(2021)Van Tricht, Huybrechts, Van Breedam, Vanhulle,
Van Oost, and Zekollari</label><mixed-citation>
      
Van Tricht, L., Huybrechts, P., Van Breedam, J., Vanhulle, A., Van Oost, K., and Zekollari, H.: Estimating surface mass balance patterns from unoccupied aerial vehicle measurements in the ablation area of the Morteratsch–Pers glacier complex (Switzerland), The Cryosphere, 15, 4445–4464, <a href="https://doi.org/10.5194/tc-15-4445-2021" target="_blank">https://doi.org/10.5194/tc-15-4445-2021</a>, 2021.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib72"><label>Vapnik(1999)</label><mixed-citation>
      
Vapnik, V.: The Nature of Statistical Learning Theory, Springer Science &amp;
Business Media, <a href="https://doi.org/10.1007/978-1-4757-2440-0" target="_blank">https://doi.org/10.1007/978-1-4757-2440-0</a>, 1999.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib73"><label>Vincent et al.(2018)Vincent, Soruco, Azam, Basantes-Serrano, Jackson,
Kjøllmoen, Thibert, Wagnon, Six, Rabatel, Ramanathan, Berthier, Cusicanqui,
Vincent, and Mandal</label><mixed-citation>
      
Vincent, C., Soruco, A., Azam, M. F., Basantes-Serrano, R., Jackson, M.,
Kjøllmoen, B., Thibert, E., Wagnon, P., Six, D., Rabatel, A., Ramanathan,
A., Berthier, E., Cusicanqui, D., Vincent, P., and Mandal, A.: A Nonlinear
Statistical Model for Extracting a Climatic Signal From Glacier Mass Balance
Measurements, J. Geophys. Res.-Earth, 123,
2228–2242, <a href="https://doi.org/10.1029/2018JF004702" target="_blank">https://doi.org/10.1029/2018JF004702</a>, 2018.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib74"><label>Wang et al.(2020)Wang, Yuan, Shen, Liu, Li, Yue, Shi, and
Zhang</label><mixed-citation>
      
Wang, J., Yuan, Q., Shen, H., Liu, T., Li, T., Yue, L., Shi, X., and Zhang, L.:
Estimating snow depth by combining satellite data and ground-based
observations over Alaska: A deep learning approach, J. Hydrol.,
585, 124828, <a href="https://doi.org/10.1016/j.jhydrol.2020.124828" target="_blank">https://doi.org/10.1016/j.jhydrol.2020.124828</a>, 2020.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib75"><label>Werder et al.(2020)Werder, Huss, Paul, Dehecq, and
Farinotti</label><mixed-citation>
      
Werder, M. A., Huss, M., Paul, F., Dehecq, A., and Farinotti, D.: A Bayesian
ice thickness estimation model for large-scale applications, J.
Glaciol., 66, 137–152, <a href="https://doi.org/10.1017/jog.2019.93" target="_blank">https://doi.org/10.1017/jog.2019.93</a>, 2020.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib76"><label>White et al.(1998)White, Harrison, and
Mottershead</label><mixed-citation>
      
White, I. D., Harrison, S. J., and Mottershead, D. N.: Environmental systems:
an introductory text, Psychology Press, ISBN 9780748740819, 1998.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib77"><label>WGMS(2021)</label><mixed-citation>
      
World Glacier Monitoring Service (WGMS): Fluctuations of Glaciers Database, World Glacier Monitoring Service (WGMS) [data set], <a href="https://doi.org/10.5904/wgms-fog-2021-05" target="_blank">https://doi.org/10.5904/wgms-fog-2021-05</a>, 2021.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib78"><label>Wu et al.(2018)Wu, Liu, Jiang, Xu, Wei, and Guo</label><mixed-citation>
      
Wu, K., Liu, S., Jiang, Z., Xu, J., Wei, J., and Guo, W.: Recent glacier mass balance and area changes in the Kangri Karpo Mountains from DEMs and glacier inventories, The Cryosphere, 12, 103–121, <a href="https://doi.org/10.5194/tc-12-103-2018" target="_blank">https://doi.org/10.5194/tc-12-103-2018</a>, 2018.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib79"><label>Wu et al.(2023)Wu, Su, Ren, Lü, and Yuan</label><mixed-citation>
      
Wu, X., Su, J., Ren, W., Lü, H., and Yuan, F.: Statistical comparison and
hydrological utility evaluation of ERA5-Land and IMERG precipitation products
on the Tibetan Plateau, J. Hydrol., 620, 129384,
<a href="https://doi.org/10.1016/j.jhydrol.2023.129384" target="_blank">https://doi.org/10.1016/j.jhydrol.2023.129384</a>, 2023.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib80"><label>Xie et al.(2021)Xie, Asari, and Haritashya</label><mixed-citation>
      
Xie, Z., Asari, V. K., and Haritashya, U. K.: Evaluating deep-learning models
for debris-covered glacier mapping, Applied Computing and Geosciences, 12,
100071, <a href="https://doi.org/10.1016/j.acags.2021.100071" target="_blank">https://doi.org/10.1016/j.acags.2021.100071</a>, 2021.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib81"><label>Zandler et al.(2019)Zandler, Haag, and
Samimi</label><mixed-citation>
      
Zandler, H., Haag, I., and Samimi, C.: Evaluation needs and temporal
performance differences of gridded precipitation products in peripheral
mountain regions, Scientific Reports, 9, 15118, <a href="https://doi.org/10.1038/s41598-019-51666-z" target="_blank">https://doi.org/10.1038/s41598-019-51666-z</a>, 2019.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib82"><label>Zemp et al.(2013)Zemp, Thibert, Huss, Stumm, Rolstad Denby, Nuth,
Nussbaumer, Moholdt, Mercer, Mayer, Joerg, Jansson, Hynek, Fischer,
Escher-Vetter, Elvehøy, and Andreassen</label><mixed-citation>
      
Zemp, M., Thibert, E., Huss, M., Stumm, D., Rolstad Denby, C., Nuth, C., Nussbaumer, S. U., Moholdt, G., Mercer, A., Mayer, C., Joerg, P. C., Jansson, P., Hynek, B., Fischer, A., Escher-Vetter, H., Elvehøy, H., and Andreassen, L. M.: Reanalysing glacier mass balance measurement series, The Cryosphere, 7, 1227–1245, <a href="https://doi.org/10.5194/tc-7-1227-2013" target="_blank">https://doi.org/10.5194/tc-7-1227-2013</a>, 2013.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib83"><label>Zemp et al.(2021)Zemp, Nussbaumer, Gärtner-Roer, Bannwart, Paul,
and Hoelzle</label><mixed-citation>
      
Zemp, M., Nussbaumer, S. U., Gärtner-Roer, I., Bannwart, J., Paul, F., and
Hoelzle, M.: Global Glacier Change Bulletin Nr. 4 (2018-2019), Tech. Rep.,
World Glacier Monitoring Service, Zürich,
<a href="https://doi.org/10.5167/uzh-209777" target="_blank">https://doi.org/10.5167/uzh-209777</a>, 2021.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib84"><label>Zhang et al.(2019)Zhang, Liu, and Huang</label><mixed-citation>
      
Zhang, E., Liu, L., and Huang, L.: Automatically delineating the calving front of Jakobshavn Isbræ from multitemporal TerraSAR-X images: a deep learning approach, The Cryosphere, 13, 1729–1741, <a href="https://doi.org/10.5194/tc-13-1729-2019" target="_blank">https://doi.org/10.5194/tc-13-1729-2019</a>, 2019.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib85"><label>Zhao and He(2022)</label><mixed-citation>
      
Zhao, P. and He, Z.: A First Evaluation of ERA5-Land Reanalysis Temperature
Product Over the Chinese Qilian Mountains, Frontiers in Earth Science, 10, 907730,
<a href="https://doi.org/10.3389/feart.2022.907730" target="_blank">https://doi.org/10.3389/feart.2022.907730</a>, 2022.

    </mixed-citation></ref-html>
<ref-html id="bib1.bib86"><label>Zhu et al.(2021)Zhu, Zhang, Wang, Tian, Liu, Ma, Kan, and
Chu</label><mixed-citation>
      
Zhu, L., Zhang, Y., Wang, J., Tian, W., Liu, Q., Ma, G., Kan, X., and Chu, Y.:
Downscaling Snow Depth Mapping by Fusion of Microwave and Optical
Remote-Sensing Data Based on Deep Learning, Remote Sensing, 13, 584,
<a href="https://doi.org/10.3390/rs13040584" target="_blank">https://doi.org/10.3390/rs13040584</a>, 2021.

    </mixed-citation></ref-html>--></article>
