the Creative Commons Attribution 4.0 License.
the Creative Commons Attribution 4.0 License.

Physics-aware machine learning for glacier ice thickness estimation: a case study for Svalbard
Jonathan Louis Bamber
Xiao Xiang Zhu
The ice thickness of the world's glaciers is mostly unmeasured, and physics-based models to reconstruct ice thickness cannot always deliver accurate estimates. In this study, we use deep learning paired with physical knowledge to generate ice thickness estimates for all glaciers of Spitsbergen, Barentsøya, and Edgeøya in Svalbard. We incorporate mass conservation and other physically derived conditions into a neural network to predict plausible ice thicknesses even for glaciers without any in situ ice thickness measurements. With a glacier-wise cross-validation scheme, we evaluate the performance of the physics-informed neural network. The results of these proof-of-concept experiments let us identify several challenges and opportunities that affect the model's performance in a real-world setting.
- Article
(8515 KB) - Full-text XML
- BibTeX
- EndNote
Glacier ice thickness is a fundamental variable required for modelling the evolution of a glacier. Ice thickness is the most important input for modelling the dynamics of an ice mass because surface velocity is proportional to the fourth power of thickness (Cuffey and Paterson, 2010). Combined with surface elevation, it provides bed topography, also key for modelling flow (van der Veen, 2013). However, direct measurements of ice thickness are scarce. In situ ice thickness measurements exist for only a fraction of the 215 000 glaciers in the world (Welty et al., 2020).
There are physics-based and process-based approaches that aim to reconstruct glacier ice thicknesses from in situ data and ice dynamical considerations. Farinotti et al. (2017) compared 17 models and found that their ice thickness estimates differ considerably for the test glaciers. Following these results, Farinotti et al. (2019) created an ensemble of five models to develop a consensus estimate of ice thicknesses for the world's glaciers in 2019. Later, Millan et al. (2022) derived ice thickness estimates for the world's glaciers using ice motion as the primary constraint. However, these results still differ from the Farinotti et al. (2019) consensus estimate. It is evident, therefore, that significant uncertainty remains in ice thickness estimates.
Machine learning approaches are flexible and adapt well to complex structures and non-linear behaviour. They have already been employed to model glacier quantities like surface mass balance (Bolibar et al., 2020; Anilkumar et al., 2023) or ice thickness (Haq et al., 2021; Leong and Horgan, 2020), classify surge-type glaciers (Bouchayer et al., 2022), or model glacier flow (Jouvet, 2023; Min et al., 2019). One advantage of machine learning approaches is their efficient optimization and evaluation compared to process-based models (Jouvet et al., 2022). The disadvantages of purely data-driven machine learning models are that they do not guarantee the physical correctness of the predicted quantities and that they often need huge quantities of training data to fully represent the system's complexity.
Recently, a new framework of data-driven but physically constrained models was described as physics-informed neural networks (PINNs) by Raissi et al. (2018). They exploit the fact that neural networks can represent solutions to partial differential equations (PDEs) if the squared residual of the governing PDE acts as the loss function of the neural network (Lagaris et al., 1998). Partial derivatives with respect to the model inputs are easy to calculate with the automatic differentiation algorithm that is used to train neural networks. PINNs do not require a discrete grid to be evaluated. Therefore, the physics-based loss can be evaluated at any point within the training domain (Xu et al., 2023). Additional ground truth data can be used to compute a data loss that acts as an internal condition for constraining solutions to the PDE.
PINNs and variations thereof have also already been used for predicting ice flow (Jouvet and Cordonnier, 2023), inferring the basal drag of ice streams (Riel et al., 2021), or ice shelf rheology (Wang et al., 2022; Iwasaki and Lai, 2023) for example.
Cheng et al. (2024) built a unified framework involving a PINN to model ice sheet flow by enforcing momentum conservation derived from the shelfy-stream approximation. They apply their framework to a single glacier in Greenland to showcase the ability of the PINN to reconstruct ice thickness and basal friction simultaneously.
Instead of using momentum conservation, Teisberg et al. (2021) created a mass-conserving PINN to produce realistic ice thickness and depth-averaged ice flow maps for a single glacier in Antarctica. They showed that, for their case, solving for mass conservation and additional constraints produces a valid ice thickness estimate.
This work extends the mass-conserving approach to predict ice thickness for multiple glaciers. As a proof of concept, we include all non-surging glaciers in Spitsbergen, Barentsøya, and Edgeøya in Svalbard to show that it is possible to use a PINN architecture for a heterogeneous region. These regions include glaciers with various morphologies, from valley glaciers to ice caps. To better account for the variety of glacier geometries, sizes, and flow velocities, we include additional input features, e.g. slope and elevation. The challenge is to find a configuration of inputs and physical constraints that is general enough to describe the variety of glaciers in the study region. At the same time, the constraints and inputs should be strict enough to force the model to produce physically correct outputs.
Ice thickness measurements exist only for a fraction of the glaciers in the dataset, and there is no benchmark dataset to measure the model's performance. Therefore, we need a validation method that assesses the performance, although there is no ground truth. To this end, we estimate the expected drop in performance for glaciers without ice thickness measurements using glacier-wise cross-validation. In addition, we compare our ice thickness estimates to those of Millan et al. (2022) and the consensus estimate of Farinotti et al. (2019) for our study region. These estimates are not benchmark datasets but are widely accepted in the community. Additionally, we compare our results to the recently published ice thickness estimate of van Pelt and Frank (2025) tailored to the region of Svalbard.
Finally, we discuss the challenges and opportunities of the approach and pathways to improve the ice thickness estimate.
2.1 Mass conservation
Assuming ice to be incompressible and integrating vertically along the depth of a glacier, we retrieve the two-dimensional form of the mass conservation:
with H being the ice thickness and denoting the mass balance of the glacier. is the velocity caused by the deformation of ice, averaged along the vertical axis. We will refer to as the depth-averaged velocity in the following. Equation (1) can be reformulated as
with known as the apparent mass balance. In other words, the flux divergence on a glacier equals its apparent mass balance.
2.2 Depth-averaged velocity and basal sliding
Glacier flow is the result of gravity-induced stresses on the ice. Friction between the ice and the glacier bed or sidewalls, friction between more slowly moving and faster-moving ice within the glacier, and gradients in longitudinal tension or compression balance the gravitational stress (van der Veen, 2013). The resulting ice movements depend on many factors, such as the physical properties of the ice like temperature, impurities, or density, and also conditions at the glacier bed (Jiskoot, 2011).
From space, we can observe the surface velocity of glaciers. To infer thickness from mass conservation, we would need to know the depth-averaged velocity.
There are models with different degrees of approximations of the full Navier–Stokes equations to describe ice flow. The simplest one, the shallow-ice approximation (SIA) assumes lamellar flow, so the driving forces are entirely opposed by basal drag. It neglects lateral shear and longitudinal stresses. The rate factor A from Glen's flow law is taken to be constant with depth (van der Veen, 2013).
From this model, we can derive the outcome that the depth-averaged velocity relates to the surface velocity as assuming the flow velocity at the base of the glacier is 0 (see Appendix A for derivation). However, basal velocity is unlikely to be 0.
The basal sliding velocity tightly relates to the properties of the glacier bed and complex interactions between water, sediment, and ice at the glacier bed (Cuffey and Paterson, 2010). Millan et al. (2022) introduced an empirical factor β with vb=βvs to account for contributions from basal sliding. They derive the factor from the ratio between the surface slope and surface velocity.
If the ice velocity is entirely by slip along the glacier bed, then . Accordingly, we estimate the depth-averaged velocity to be within the bounds of
where llower acts as a parameterization for the vertical integration of the velocity and can be set between 0 and 1. Depending on the factor β, which lies between 0.1 and 1, the lower boundary is close to the defined llower or closer to 1. For β=1, the lower boundary for the depth-averaged velocity equals the surface velocity.
Depth-averaged velocities are estimated for the x and y direction and the velocity magnitude. Therefore, we calculate three separate β values.
2.3 PINN model
As already mentioned, a PINN consists of a neural network that is able to approximate the solution to a PDE (Karniadakis et al., 2021). A neural network, sometimes also called a multi-layered perceptron, consists of layers of connected nodes, also called neurons, where the connections each have an associated weight. At each node, the weighted outputs from each node of the previous layer are passed through a non-linear activation function (Goodfellow et al., 2016). By minimizing a loss, the weights of the network are updated to make accurate predictions.
In a PINN model, the loss is given by the residual of the PDE we want to solve. In theory, PINNs only require input features that are needed to calculate the terms in the PDE (Raissi et al., 2018). In our work, we also provide the neural network with auxiliary data that are related to glacier ice thickness but are not needed to solve the PDE. Therefore, we can exploit information from observable data as we would do with a non-physics-aware neural network.
Additionally, we use a Fourier feature encoding layer as described by Tancik et al. (2020) preceding the neural network. A Fourier feature encoding layer maps the input vector x to a higher-dimensional feature space using
The embedding of spatial coordinates was originally developed to overcome spectral bias in neural networks and speed up convergence in the reconstruction of images. It enables the network to learn high-frequency functions in low-dimensional problem domains.
Figure 1 shows a schematic of the PINN model with its input features, outputs, and loss components. The exact architecture of the PINN is described in Appendix B. The inputs to the model are vectors for each grid cell in the study region. They contain the spatial coordinates and surface velocities in the x and y directions and three β values to correct for basal sliding in the x and y directions and in the magnitude. Additionally, the vectors contain auxiliary data like elevation, slope, the grid cell's distance to the border of its glacier, and the area of the glacier it belongs to.

Figure 1Physics-informed model with inputs, outputs, and components adding to the loss function. The physics-aware losses are in purple boxes. The data loss in the blue box is the only loss depending on ice thickness measurement data. Surface velocity and β values also add to the physics-aware losses. The connection is not shown in order to increase readability. “Apparent MB” stands for apparent mass balance.
Only the spatial coordinates get mapped to higher-dimensional Fourier features.
The model outputs three quantities at each point of query: the ice thickness H and depth-averaged velocity in the x and y directions. The predicted quantities must fulfil the mass conservation described in Eq. (2). The squared deviation from this equation is the first component of the loss function:
The second component of the loss function is the amount by which the depth-averaged velocity estimates in the horizontal plane exceed the boundaries given in Eq. (3).
with and . ℒvel is calculated separately for the x and y component and the magnitude of the depth-averaged velocity. As basal drag is most likely not the only drag the ice experiences, we decided to fix the lower bound as llower=0.7 in order to give more flexibility in the estimate.
We include two more physics-aware constraints to improve the model performance. First, the ice thickness is assumed to be smooth, so the model will penalize large spatial derivatives within the ice thickness prediction:
Secondly, ice thickness cannot be smaller than 0. Therefore, we add a loss component that punishes negative ice thicknesses.
The final loss component is the data loss. It penalizes the deviation from the in situ ice thickness measurements and acts as the internal condition to solve the mass-conserving PDE. Each loss component will have a different scale, so we balance them with individual weights λi. Summing up all the loss components, we obtain the complete loss function as
All the physics-aware losses are evaluated at any point in the study region. In contrast, the data loss is only evaluated wherever ice thickness measurements are available. We refer to the points with ice thickness measurements as labelled, and points without ice thickness measurements are referred to as unlabelled.
The training data are scaled to have a mean of 0 and a variance of 1. Before computing the physics-aware loss components, we scale the quantities back to their original units for physical consistency.
We tested a slim version of the PINN model on a one-dimensional dataset of a single glacier. The results are given in Appendix C. The experiment shows the added value of introducing physics-aware loss components.
2.4 Validation
We evaluate the performance of the PINN model by calculating the root mean squared difference (RMSD) and the mean absolute percentage difference (MAPD) between the model prediction and the in situ ice thickness measurements. However, in situ measurements within a glacier are highly correlated due to their proximity. Therefore, a simple random split of the data into training and test datasets will not yield a realistic view of the model performance.
We employ a glacier-wise cross-validation (CV) approach as done by Bolibar et al. (2020) to better judge the model performance. This also allows us to make assumptions about how well the model will perform on a glacier without any measured ice thicknesses.
For the leave-one-glacier-out (LOGO) CV, we choose seven glaciers that serve as test glaciers. In an alternating way, we train the model without one of those glaciers' ice thickness measurements.
It is important to note that only data labelled with ice thickness measurements of the test glacier are left out of the dataset. All the data needed to enforce the physical consistency for the test glacier stay in the training dataset. The mass-conserving PDE of Eq. (1) will still be solved at the test glacier but without enforcing internal conditions with ice thickness measurements.
Upon validation, the RMSD and the MAPD are calculated for the test glacier. All the test glaciers are thoroughly mapped with ice thickness measurements and differ in size, mean measured ice thickness, and location in Svalbard.
In this study, we focus on the glaciers on the islands of Spitsbergen, Barentsøya, and Edgeøya. Glaciers in an active surge phase during the data acquisition period for the surface velocity are not considered. The information on active surge phases is collected from Koch et al. (2023).
3.1 Data management
We processed all the data needed for the training of the PINN using the Open Global Glacier Model (OGGM) framework developed by Maussion et al. (2019). OGGM is an open-source framework to simulate glacier evolution. It provides models for mass balance, distributed ice thickness, and ice flow, as well as downloading tools for glacier outlines, digital elevation models (DEMs), and climate data. The mass balance model is a temperature index melt model relying on climate data.
OGGM saves all the information for each glacier separately in glacier directories. The Randolph Glacier Inventory (RGI) Version 6.0 contains the outlines for the glaciers (RGI Consortium, 2017).
Using the outlines, OGGM defines a spatial grid for each glacier. The grid resolution is adapted individually according to the size of the glacier. In our study region, the grid resolution ranges from 12 to 200 m. OGGM reprojects and scales the data for each glacier to the glacier grids. We collect these data and transform the coordinates from the individual grids into a common projection.
3.2 Surface velocity data
Millan et al. (2022) derived the surface flow velocity of the world's glaciers using image pairs acquired between 2017 and 2018 by Landsat 8, Sentinel-2, and Sentinel-1. They tracked glacier motion using a cross-correlation approach. The resolution of the velocity product is 50 m with an estimated accuracy of about 10 yr m−1. Using OGGM, velocity in the x and y directions and velocity magnitude are projected onto the individual glacier grids and then smoothed with a two-dimensional Gaussian filter.
We introduce the aforementioned β value (see Sect. 2.2) to incorporate the influence of basal sliding on the measured surface velocity vb=βvs. Following Millan et al. (2022), we set β equal to 0.1 in areas where the ratio between slope and observed surface velocity is greater than 0.001 yr m−1 and modulate up to 0.9 for areas where the ratio is less than 0.001 yr m−1. For each point, we compute three β values from the surface velocities in the x and y directions and the magnitude of the surface velocity.
3.3 Apparent mass balance
The apparent mass balance is the difference between the point-wise mass balance and the thickness change rate at each grid point. The mass balance at each point of a glacier grid is estimated using the ConstantMassBalance model from OGGM. It calculates the average mass balance during a chosen period from given climate data, calibrated with geodetic mass balance data from Hugonnet et al. (2021). To match the acquisition period of the surface velocity, we set the climate period for the mass balance model to 2016–2018.
The rate of thickness change is retrieved from ASTER DEM differences between 2015 and 2019 (Hugonnet et al., 2021). The data are projected onto the glacier grids using OGGM and are then smoothed with a two-dimensional Gaussian filter.
3.4 Thickness data
The data-driven machine learning model needs ice thickness data as ground truth for its supervised training. The Glacier Thickness Database (GlaThiDa) is a comprehensive public database of in situ glacier thickness measurements collected from various studies (GlaThiDa Consortium, 2020). Version 3.1.0 was released in 2020 with nearly 1 million measurements from ice-penetrating radar (IPR) on 207 glaciers or ice caps in Svalbard.
In situ ice thickness measurements are not error-free. GlaThiDa lists reported uncertainties of almost 80 % of the measurements in Svalbard. The mean and standard deviation of the thickness uncertainty are 6.2 and 4.4 m, with a maximum uncertainty of 21 m.

Figure 2Glaciers in the training dataset. (a) The locations of in situ measurements are marked in red. (b) The locations of the test glaciers. The coastline is retrieved from Moholdt et al. (2021).
During the preprocessing, the measurements are projected onto the OGGM glacier grids by aggregating and averaging them at their nearest point on the glacier grid. We only consider aggregated ice thicknesses where the average acquisition year is after 2000. That leaves us with 27 554 points labelled with ice thickness on 65 glaciers.
3.5 Auxiliary data
Adding to the data that we need to impose the physics-aware losses, we also feed the network with extra information from auxiliary data as input features. We chose the features because they were easily available through OGGM and are related to the glacier's ice thicknesses. In Appendix F we analyse how each of the features impacts the model output. The elevation of each point comes from the global DEM from Copernicus DEM GLO-90, which was acquired from 2010 to 2015 (Copernicus, 2019). The slope is then computed by OGGM based on the glacier's smoothed topography and over the length of a grid cell on the glacier. The distance to the border of the glacier outline is computed for each point within a glacier. The glacier area is also retrieved from the RGI.
The full dataset of points with and without ice thickness labels consists of over 3 million data points from the grids of 1465 glaciers. Figure 2a displays the glaciers considered in light blue and the acquisition lines of the IPR measurements in red.
3.6 Test glaciers
The seven glaciers with the most in situ measurements are chosen as test glaciers for the LOGO CV. They differ in size and mean thickness and are located in different areas of Spitsbergen. No glaciers on Barentsøya and Edgeøya are mapped well enough to use them as test glaciers.
Figure 2b shows the location of the test glaciers. Table 1 lists the area, mean, and standard deviation of the measured ice thicknesses and the mean of the survey year for those glaciers. Measurements on glaciers RGI60-07.00496 and RGI60-07.00497 are all from one survey, while the others are from multiple surveys carried out in different years.
The LOGO CV produces seven models with the same architecture but different model weights. Each model was trained on all the unlabelled data to enforce the physical constraints at every point. After putting aside the test glacier's labelled data, the remaining glaciers' labelled data were randomly split into 60 % training and 40 % validation data.
The in-sample performance is measured based on the validation data that the model did not see during the training. Table 2 lists the RMSD and MAPD for the in-sample validation data. The PINNs predict glacier ice thickness with a mean in-sample RMSD of 30 m corresponding to an MAPD of 36 %.
The training and validation data are spatially correlated. Therefore, the in-sample evaluation of the model probably overestimates its performance.
Figure 3 shows the mean ice thickness prediction and the coefficient of variation over all seven LOGO models for the study region. The coefficient of variation measures the variability in relation to the mean ice thickness at each point of the grid. Of all the points, 90 % have a variability below 0.16. As the in-sample validation scores of each model are also similar, we are confident that the method is robust to varyingly labelled data. The PINN models agree with their predictions, although they were trained with different sets of ice thickness measurements as target data.

Figure 3Mean (a) and coefficient of variation (b) of the ice thickness predictions from all seven models from CV.
Table 2Results of the LOGO CV. The in-sample validation scores are measured from the validation set that contains in situ data from every glacier except the test glacier that was left out during training. The LOGO validation scores are measured from the in situ data for the left-out test glacier.

4.1 LOGO results
The model performance for the test glaciers delivers insights into the performance that we can expect for glaciers where we do not have any in situ measurements. Table 2 shows the results of the LOGO validation for each of the test glaciers. As expected, the RMSD and MAPD are significantly higher than for the in-sample validation data. Figure 4 shows the difference between the model's ice thickness estimate and the in situ measurements along the IPR acquisition lines for the seven LOGO test glaciers that were excluded from the dataset during training. Overall, the models underestimated the ice thickness. The ice thickness estimates for all the grids of the glaciers are displayed in Fig. D1.

Figure 4Difference between the predicted ice thickness and the IPR ice thickness measurements for the seven LOGO test glaciers.
The test glaciers differ significantly in mean ice thicknesses (see Table 1). For thinner glaciers like RGI60-07.00496 and RGI60-07.00497, the MAPD is very high, although their RMSD is comparable to the in-sample scores. The RMSD of glacier RGI60-07.01482 is 4 times as high as the mean RMSD score of the in-sample glaciers, but as, on average, in situ measurements are very thick, the MAPD is closer to the in-sample MAPD than the MAPD for a test glacier that is thinner on average. This makes it clear that considering both validation scores is necessary to view the model performance accurately. Another example of this would be the comparison of performances on glaciers RGI60-07.00240 and RGI60-07.01481. They have similar measured ice thicknesses and RMSD scores, but their MAPDs differ greatly.
Over all seven test glaciers, the mean RMSD and the mean MAPD are about 66 m and 0.58 %, respectively, i.e. significantly worse than the in-sample metric. This indicates that the PINN overfits on glaciers with thickness measurements.
4.2 Comparison to other estimates
As we do not have full coverage with in situ measurements, the model scores only represent the model's performance at the acquisition lines of the IPR measurements. Therefore, we compare the ice thickness predictions to the estimates of Millan et al. (2022), Farinotti et al. (2019), and van Pelt and Frank (2025) to see how much the estimates differ. All of those ice thickness products are derived from physics-based models. Farinotti et al. (2019) estimated ice thickness using an ensemble of up to five models, hence the name consensus estimate. Millan et al. (2022) rely on a single model that uses the shallow-ice approximation and surface motion to compute ice thickness. Two inverse methods are used by van Pelt and Frank (2025), one for small and one for larger and surging glaciers, to create their ice thickness product.
Figure 5 shows the scatterplot of the ice thicknesses of the PINN ensemble estimates versus the three other estimates for each point in the study region. The solid red line shows the linear fit between the two ice thickness products, while the dashed black line is the 1:1 line. The values of slope and intersect for the linear fits indicate that the PINN estimates agree slightly less than the three physics-based models on the ice thickness at each grid point. Comparing the mean ice thickness estimate from the ensemble of PINNs to the estimates of physics-based models shows that the deviations from the other estimates are within the range of the differences between the physics-based models. Mean absolute differences (MADs) between the PINN and physics-based predictions are all close, with a mean of 34 m. The MAD between the physics-based models is on the same order of magnitude, and their mean MAD is 34.3 m.
Since physics-based models also work with simplifications of ice dynamics, their ice thickness products cannot be taken as the definitive truth. Comparing the predictions of the PINN ensemble to those products only serves to estimate the qualitative validity of the ice thickness predicted by the PINN ensemble. The overall correlation between the ice thickness estimates leads us to believe that the PINN ensemble produces valid ice thickness estimates.

Figure 5Comparisons of ice thickness estimates: (a, b, c) ice thickness prediction from the ensemble of PINNs versus ice thickness estimates of Millan et al. (2022), Farinotti et al. (2019), and van Pelt and Frank (2025). (d, e, f) Ice thickness estimates from physics-based models against each other. The colour indicates the point density (the brighter, the denser). The solid red line shows the linear fit between the two ice thickness products. The dashed black line is the 1:1 line.
4.3 Depth-averaged velocity
The models estimate the depth-averaged velocities in the x and y directions. There are no ground truth data for the depth-averaged velocities, so we cannot evaluate the models's accuracy. However, during training, the loss of the predicted velocities is reduced significantly, showing that the constraints of Eq. (3) are enforced.
Evaluating the PINN performance with the in-sample validation set and comparing predicted ice thicknesses to other products suggest that the PINN produces reasonable ice thickness estimates. However, testing the model with a glacier-wise CV scheme unveils the lack of generalizability to glaciers without any measurement data. The differences between the predicted ice thicknesses and the measurements are much higher for the test glaciers of the LOGO CV than for the in-sample validation. We identified several factors that may improve generalizability but are also challenging to address. The schematic of Fig. 6 shows an overview of the domains and the particular issues we judge to be the most pressing to address.

Figure 6Challenges for PINNs in a real-world setting like the prediction of glacier ice thickness. The separate realms interfere with each other, complicating the optimization of the model. Weighting of the physical constraints could have the biggest positive benefit.
5.1 Data
Although PINNs are generally less dependent than purely data-driven methods on training data, their performance relies on the quality of input data (Iwasaki and Lai, 2023). This study collected the thickness measurements from 65 glaciers. The individual measurements lie close together along the acquisition lines. As a result, most of the measurements have high spatial correlations with each other. The number of independent training data to learn the physics of glaciers is, therefore, far less than the actual number of measurements. On the other hand, redundant data introduce a bias. To reduce the overfitting, we could reduce the correlations in the training dataset by averaging or sub-sampling the observations for example. This should improve the performance on glaciers without any labelled training data.
Secondly, the training data are not aligned temporally. In situ ice thickness measurements were collected between 2000 and 2017, while the surface velocity was derived from satellite data acquired between 2017 and 2018. Using surface velocity for the same year as ice thickness was measured would result in a better estimate of the ice flux for the labelled data and therefore improve physical consistency within the model. However, this is only important if either ice thickness or velocity changes significantly with time.
Lastly, we do not consider measurement errors in our dataset. Ice thickness measurements from ice-penetrating radar, for example, are subject to errors not only due to the varying density of glacier ice but also due to the unknown thickness of snow and firn layers (Lindbäck et al., 2018). Future work should account for measurement errors with standard uncertainty quantification methods or even introduce error margins into the loss components as Morlighem et al. (2011) did.
5.2 Model training
Training PINNs is difficult (Xu et al., 2023). One major challenge, which is also reported in other studies using multiple loss components in their PINNs (Iwasaki and Lai, 2023; Cheng et al., 2024), is to find the optimal balance between the multi-scale contributions of each loss component. We empirically set the loss weights to a fixed value for the entire training process.
In an experiment to test the importance of the loss components, we found that the relative importance is not very pronounced (see Appendix E). Therefore, we assume that the individual loss components are not optimally weighted in the reported model.
A more sophisticated approach would be to automatically update the weights of the loss components during the training. This introduces some computational overhead but prevents the model from minimizing certain loss components faster than others (Wang et al., 2023).
The multiple loss components create another challenge: the loss landscape is highly complex, and finding its global minimum is difficult. Recently, Rathore et al. (2024) investigated different optimizers for PINNs and showed that combined first- and second-order optimizers lead to faster convergence. Implementing their newly introduced second-order optimizer could improve the PINN convergence. Development of new optimization strategies is a rapidly evolving area of research, and we expect significant advances to be achieved soon.
5.3 Physical constraints
Our model uses physics-aware loss components to enforce physical consistency. However, the physics-aware losses are based on simplifications to make the problem tractable. We identified several challenges that are tied to the incorporation of physical constraints.
Firstly, we use a simple model from OGGM out of the box without further calibration to derive the mass balance at every point in the study region. The mass balance data appear in the computation of the mass conservation loss, but if they are erroneous, the model will never be able achieve perfect mass conservation. Therefore, using a more sophisticated mass balance model and calibrating it for our purpose could enhance physical consistency.
Another option could be to use in situ mass balance data. This way, we circumvent the need for a mass balance model. The mass conservation loss would only be evaluated where data are available. This, however, would come with two restrictions. First, we would not be able to train the model in the entire study region. Secondly, in situ mass balance data are also not error-free. We would have to make a careful selection of the data so as not to introduce even more uncertainty.
A different way to improve the physical consistency is through a better estimate of the depth-averaged velocity. In the current model, the estimate of depth-averaged velocity is coupled to estimating the amount of basal sliding and using a parameter as a lower bound for the vertical integration of the velocity. For now, the estimate of the sliding velocity is based on a simplified approach using a threshold calibrated by Millan et al. (2022). We could, in principle, circumvent the need for this parameter using surface velocity data acquired during the winter months, when basal sliding is inhibited. The measurable surface velocity will have less contribution from basal sliding, and we could avoid estimating the β parameters. This would eliminate one source of uncertainty.
We also want to mention that there are several processes affecting ice dynamics, especially in Svalbard, that are not very simplified or are neglected in the model. One example is that our model assumes ice to be incompressible, when Svalbard glaciers actually have thick firn layers (Pälli et al., 2003). The varying density could introduce a non-negligible densification term in the mass balance equation (Eq. 1).
Another example is the assumption of a temperature-independent creep coefficient A. Many glaciers in Svalbard are believed to be polythermal (Glasser, 2011), so the creep coefficient may vary within the ice, affecting the validity of our lower boundary for the estimate of depth-averaged velocity. However, the influence of these effects should carefully be weighed against the possibility of introducing errors if we decide to include better representations of these processes.
Lastly, one major challenge is that we deal with a highly under-constrained problem. We only provide the model with the ice thickness measurements as a sort of internal condition, but we do not provide boundary conditions. Also, the depth-averaged velocity is only loosely constrained by a set of inequalities. Therefore, it would be beneficial to incorporate additional constraints like momentum conservation to actually derive depth-averaged velocity instead of estimating it. While this is technically easy to do, it comes at the cost of introducing uncertainties from approximating required parameters. To enforce momentum conservation, we would need to make assumptions about ice viscosity and resistance from the bedrock for example. Depending on the data quality, we risk introducing more uncertainty instead of improving the physical consistency.
In our view, the two elements that are most promising to improve the model performance, if revised, are the modelling of mass balance for Svalbard and the choice of surface velocity data for the estimation of depth-averaged velocities.
5.4 Evaluation
In geospatial machine learning, evaluation is generally challenging (Rolf, 2023). As mentioned in Sect. 2.4, the in situ measurements are heavily correlated since they are clustered on only a fraction of the glaciers. Therefore, we employ a spatially aware evaluation method to estimate the true model performance. However, the CV procedure only includes 7 of 65 glaciers with measurements. Moreover, we have varying numbers of IPR measurements for the evaluation of each of the test glaciers, as already mentioned in Sect. 4.1. There is no guarantee that we fully capture the model performance with our approach.
Additionally, in situ ice thicknesses are subject to measurement errors, and some measurements might have higher errors than others. To be as precise as possible when evaluating the model performance, we should consider the trustworthiness of every ice thickness label.
Ultimately, our problem has no benchmark dataset, so it is impossible to know the model performance exactly. Although we compare our ice thickness estimates with others, these also have errors that are not well constrained and are in no respect benchmarks that can be used for uncertainty quantification. It is, therefore, difficult to state which method considered here produces the most reliable ice thickness estimate.
Despite the above-mentioned limitations, we show that a relatively simple PINN can produce reasonable ice thickness estimates while treating an entire area and not only a single glacier at once. Although the lack of high-quality data is an overarching challenge that can hardly be overcome, we expect that by implementing the proposed adjustments in data curation, model training, and physical constraints, the physical consistency and accuracy of the model will be improved. Without changing the dataset, we believe that optimizing the loss weights λ would have the biggest positive benefit, as the optimal configuration depends on the noise in the data (Iwasaki and Lai, 2023). This may especially be the case for glaciers without measured ice thicknesses.
We have demonstrated that it is possible to train a physics-aware machine learning model to produce ice thickness estimates for multiple glaciers, including glaciers without any ground truth ice thickness – in other words, out-of-sample targets. We deploy a relatively simple physical constraint by imposing mass conservation in the loss function of the PINN. This serves as a proof of concept that physics-informed models not only can be applied to one single closed system but also, together with auxiliary data, can make meaningful predictions for entire regions.
More complex approaches and physical constraints could be employed (Karniadakis et al., 2021) and would, we anticipate, improve the results further. Nonetheless, we demonstrate that physics-aware machine learning is a promising approach for tackling this geophysical problem where a physical law and multiple conditions provide constraints for the solution. There are many other geophysical problems where, for example, including conservation of mass, energy, or momentum would provide a similarly effective constraint and would lead to a more scientifically meaningful result, as breaching such constraints is non-physical.
To derive a relation between the surface velocity and the depth-averaged velocity, we follow the analysis in Cuffey and Paterson (2010). Let u be the x component of velocity and H be the ice thickness. Assuming parallel flow, the glacier deforms in simple shear, so the only nonzero deviatoric stress is τxz, and the z component of the velocity is also zero. Therefore, the creep relation derived by Glen (1955) simplifies to
where A is the creep parameter and n the creep exponent. A is, in general, dependent on the temperature of the ice, so A=A(T). We assume a linear increase in shear stress along the glacier depth:
with τb being the shear stress at the bed of the glacier. We further assume constant temperature within the ice, so A does not depend on z. If we integrate Eq. (A1) along the vertical direction up to z, we get
Accordingly, the velocity at the surface is given by
and integrating Eq. (A3) along the vertical axis to derive the depth-averaged velocity, we get
A is assumed to be constant although it depends on temperature and other variables that change within a glacier profile. Temperate glaciers are nearly isothermal, whereas in cold-based glaciers, the temperature increases with a smaller distance to the bed. The highest values of A are found near the glacier bed. Therefore, the shear deformation is concentrated closer to the base than in a temperate glacier. The velocities within the bottom half of the glacier are sensitive to the value of A as it multiplies stress to the power of n. As stress decreases in the upper half of the glacier, the velocity is insensitive to the values of A there.
From Eqs. (A4) and (A5), we can derive the relation between the surface velocity and the depth-averaged velocity in the case of parallel flow and if there is no basal sliding:
for n=3. However, parallel flow in a glacier is a strong assumption, and n=3 is not always the case. Therefore, we set llower to 0.7 to allow more flexibility in estimating the depth-averaged velocity.
The PINN employed in this work consists of a fully connected neural network with 8 layers and 256 neurons each. We chose Softplus as an activation function after each layer as it is infinitely differentiable.
The Fourier feature encoding layer maps each of the spatial coordinates to a 32-dimensional vector using a matrix B with entries drawn from a Gaussian distribution with a variance of 10.0. The loss weights λi are set to keep all the loss components roughly at the same order of magnitude. We chose the Adam optimizer with default settings from PyTorch and a learning rate of 0.0001 at a batch size of 8192. In the LOGO cross-validation, each model is trained for 100 epochs.
As an illustrative example to better explore the validity and performance of the PINN model, we created a synthetic dataset of a one-dimensional glacier. We generated the synthetic data using an artificial bedrock and ice thickness from where we calculated basal stress: (van der Veen, 2013), with ρ being the density of ice, g the gravitational acceleration, and α the surface slope.
The velocity is given by Eq. (A4), where we choose , n=3, and ub=0. The apparent mass balance is calculated from mass conservation (Eq. 1). During the training, we provided the model with the x values, surface velocity, and the β value as described in Sect. 3.2. The model only received 10 points labelled with ice thickness unevenly spread over a domain of 5 km length. We did not include the auxiliary data like the point's distance to the glacier border or the area of the glacier a point lies on, as this does not make sense in the one-dimensional case. For this test case, we switched off the Fourier feature encoding layer in the model, as well as the loss on the magnitude of the depth-averaged velocity prediction.
Figure C1 compares the results of the ice thickness prediction from the model with physics-aware losses and without. Clearly, in the regime where we did not provide data for the ice thickness, the model profits from the physical constraints that keep the prediction closer to the true ice thickness than the prediction from a model without physics-aware losses.
The ice thicknesses of the seven test glaciers were estimated by the model that was trained without the in situ measurements of the respective glaciers as ground truth data. The results are displayed in Fig. D1.
The importance of the individual loss components is tricky to evaluate, as we are dealing with unevenly distributed, correlated, noisy in situ measurements as labels to evaluate the PINN performance. Despite this, we ran the LOGO experiments in which we set the weight of each of the loss components to zero one after another. We then compared the performance to the scores of the models reported in Sect. 4.1 by calculating a relative RMSD applied to the validation set as . The relative RMSD will be positive if the score improves and negative if the score gets worse by setting the weight of a loss component to 0.
The relative differences are below 5 % on average. It is interesting to see that, while the scores using the in-sample validation data improve on average when switching off the physics-aware loss components (Table E1), the scores using the out-of-sample test glaciers get worse on average (see Table E2). This fits with our intuition that the model is overfitting on the in situ ice thickness data that we provide it with during training. The physics-aware loss components act like a regularization while demanding physical consistency.
Nevertheless, we want to emphasize again that the configuration of the loss weights is certainly not optimal, as we discussed in Sect. 5.2, so there might be another distribution of importance if all the loss components are better balanced. In addition, as already mentioned, this experiment depends a lot on the dataset, so the importance of loss components also only applies to this specific study.
The physics-aware model takes not only features that it would need to evaluate the physics-aware losses but also auxiliary data. To gain insights into the model's inner workings and evaluate how it handles the auxiliary data, we estimated the feature importance for the ice thickness predictions.
One way to approximate feature importance is by calculating Shapley values. This concept is rooted in game theory and estimates a player's contribution to a cooperative game (Shapley, 1953). Shapley values represent the contribution of each feature to the model prediction.
However, analytically deriving Shapley values for deep neural networks is very costly (Höhl et al., 2024). Therefore, Shapley values are approximated using techniques like the SHapley Additive exPlanations (SHAP) framework introduced by Lundberg and Lee (2017).
Within the framework, they describe a method with improved computational performance to estimate SHAP values for deep networks: Deep SHAP.
We used the implementation in the Captum library (DeepLiftShap) (Kokhlikyan et al., 2020) to calculate SHAP values for our network. The validation data served as a representative subset of the entire dataset to save computational resources. We calculated the SHAP values for each of the seven models from the LOGO CV.
The framework explains feature contributions to the model prediction, usually for purely data-driven machine learning. In Fig. F1a high values signify a high impact on the output of the model and low values signify a low impact. For our PINN, the spatial coordinates are by far the most important features. This is expected as they define the domain in which we want to find a solution for the mass conservation PDE. Figure F1a shows the mean absolute SHAP values for the features over all seven models from the LOGO CV.
Besides the spatial coordinates, the slope has the biggest impact on the prediction. Figure F1b shows the impact of the features on the output of the model for each data point separately. For better readability, the plot shows the result of the SHAP analysis for only one of the models from the LOGO CV, as they are all similar.

Figure F1SHAP analysis: (a) mean and standard deviation of the absolute SHAP values over all seven LOGO CV models. The values were first averaged for each model separately and then averaged for all seven models. (b) SHAP values for each data point by feature. The colour shows the relative value the feature takes for each data point. The SHAP values are calculated for the model trained without data from glacier RGI60-07.00240.
The colour indicates the feature values: red signifies a relatively high value (within the range of the feature), and blue signifies a relatively low value. For example, the plot in Fig. F1b shows that high slope values lead to rather small values for the predicted ice thickness, while low values increase the predicted ice thickness. This is what we would expect given that ice thickness and slope are indirectly proportionally related in the SIA: steep slopes lead to thinner ice.
The SHAP values for the distance-to-border feature tell us that the model thinks that at the border the ice thickness should be smaller than within the glacier. For the surface velocity values, the interpretation is less clear, also because we only see the component-wise features. High surface velocities do not seem to have much impact on the ice thickness prediction, although, following glacier physics, they should have a strong influence on ice thickness.
We want to emphasize that the SHAP analysis has several limitations. First of all, it expects features to be independent of each other, which clearly is not the case here. The three β values are derived from slope and velocity values for example. Also, the analysis depends very much on the dataset. SHAP tries to replicate the model behaviour, and the model is trained with our specific dataset. Therefore, the results can only show the impact of the features on the ice thickness prediction for our specific dataset and model setup. Additionally, machine learning models can only learn correlations from the data. Causal relationships cannot be extracted. Hence, we cannot derive universal feature importance from the analysis.
However, the results from the analysis are what we would expect from physical considerations. Therefore, it serves as a sanity check if the model is retrieving sensible correlations.
The code and data that were used to train and evaluate the model as well as generate the figures are available at https://doi.org/10.5281/zenodo.13834016 (Steidl, 2024). Additionally, the code can be viewed at https://github.com/viola1593/glacier_pinn (last access: 31 January 2025).
VS conceived the study, conducted the analysis, developed the machine learning framework, and wrote the manuscript with contributions from all authors. JLB provided advice on how to incorporate the physics into the model and model training setup. XZ conceived the study and contributed to the interpretation of the results.
The contact author has declared that none of the authors has any competing interests.
Publisher’s note: Copernicus Publications remains neutral with regard to jurisdictional claims made in the text, published maps, institutional affiliations, or any other geographical representation in this paper. While Copernicus Publications makes every effort to include appropriate place names, the final responsibility lies with the authors.
This work is jointly supported by the German Federal Ministry for Economic Affairs and Climate Action in the framework of the national centre of excellence “ML4Earth” (grant number 50EE2201C); the German Federal Ministry of Education and Research (BMBF) in the framework of the international future AI lab “AI4EO – Artificial Intelligence for Earth Observation: Reasoning, Uncertainties, Ethics and Beyond” (grant number 01DD20001); the Helmholtz Association under the joint research school “Munich School for Data Science – MUDS”; and the Munich Center for Machine Learning (MCML). Jonathan Bamber also received funding from the European Union's Horizon 2020 research and innovation programme through the project Arctic PASSION (grant number 101003472). We thank Ward J. J. van Pelt and Thomas Frank for providing us with their ice thickness dataset that was used for comparison in this study. We also want to thank Thomas O. Teisberg for the insightful suggestion to use Fourier feature embedding to improve the model performance. Lastly, we want to thank Adrian Höhl for sharing his expertise in explainable AI methods with us.
The work of Viola Steidl is jointly supported by the German Federal Ministry for Economic Affairs and Climate Action in the framework of the national centre of excellence “ML4Earth” (grant no. 50EE2201C) and by the Excellence Strategy of the German federal government and the states through the TUM Innovation Network EarthCare. Jonathan Louis Bamber was supported by the German Federal Ministry of Education and Research (BMBF) in the framework of the international future AI lab “AI4EO – Artificial Intelligence for Earth Observation: Reasoning, Uncertainties, Ethics and Beyond” (grant no. 01DD20001) and the European Union's Horizon 2020 research and innovation programme through the project Arctic PASSION (grant no. 101003472). The work of Xiao Xiang Zhu was supported by the Munich Center for Machine Learning and by the Excellence Strategy of the German federal government and the states through the TUM Innovation Network EarthCare.
This paper was edited by Ben Marzeion and reviewed by Guillaume Jouvet, Thomas Teisberg, and two anonymous referees.
Anilkumar, R., Bharti, R., Chutia, D., and Aggarwal, S. P.: Modelling point mass balance for the glaciers of the Central European Alps using machine learning techniques, The Cryosphere, 17, 2811–2828, https://doi.org/10.5194/tc-17-2811-2023, 2023. a
Bolibar, J., Rabatel, A., Gouttevin, I., Galiez, C., Condom, T., and Sauquet, E.: Deep learning applied to glacier evolution modelling, The Cryosphere, 14, 565–584, https://doi.org/10.5194/tc-14-565-2020, 2020. a, b
Bouchayer, C., Aiken, J. M., Thøgersen, K., Renard, F., and Schuler, T. V.: A Machine Learning Framework to Automate the Classification of Surge-Type Glaciers in Svalbard, J. Geophys. Res.-Earth, 127, e2022JF006597, https://doi.org/10.1029/2022JF006597, 2022. a
Cheng, G., Morlighem, M., and Francis, S.: Forward and Inverse Modeling of Ice Sheet Flow Using Physics-Informed Neural Networks: Application to Helheim Glacier, Greenland, Journal of Geophysical Research: Machine Learning and Computation, 1, e2024JH000169, https://doi.org/10.1029/2024JH000169, 2024. a, b
Copernicus: Copernicus DEM GLO-90, Copernicus [data set], https://doi.org/10.5270/ESA-c5d3d65, 2019. a
Cuffey, K. and Paterson, W.: The Physics of Glaciers, Elsevier Science, 4th Edn., ISBN 978-0-08-091912-6, 2010. a, b, c
Farinotti, D., Brinkerhoff, D. J., Clarke, G. K. C., Fürst, J. J., Frey, H., Gantayat, P., Gillet-Chaulet, F., Girard, C., Huss, M., Leclercq, P. W., Linsbauer, A., Machguth, H., Martin, C., Maussion, F., Morlighem, M., Mosbeux, C., Pandit, A., Portmann, A., Rabatel, A., Ramsankaran, R., Reerink, T. J., Sanchez, O., Stentoft, P. A., Singh Kumari, S., van Pelt, W. J. J., Anderson, B., Benham, T., Binder, D., Dowdeswell, J. A., Fischer, A., Helfricht, K., Kutuzov, S., Lavrentiev, I., McNabb, R., Gudmundsson, G. H., Li, H., and Andreassen, L. M.: How accurate are estimates of glacier ice thickness? Results from ITMIX, the Ice Thickness Models Intercomparison eXperiment, The Cryosphere, 11, 949–970, https://doi.org/10.5194/tc-11-949-2017, 2017. a
Farinotti, D., Huss, M., Fürst, J. J., Landmann, J., Machguth, H., Maussion, F., and Pandit, A.: A consensus estimate for the ice thickness distribution of all glaciers on Earth, Nat. Geosci., 12, 168–173, https://doi.org/10.1038/s41561-019-0300-3, 2019. a, b, c, d, e, f
Glasser, N. F.: Polythermal Glaciers, in: Encyclopedia of Snow, Ice and Glaciers, edited by: Singh, V. P., Singh, P., and Haritashya, U. K., pp. 865–867, Springer Netherlands, Dordrecht, ISBN 978-90-481-2642-2, https://doi.org/10.1007/978-90-481-2642-2_417, 2011. a
GlaThiDa Consortium: Glacier Thickness Database 3.1.0, World Glacier Monitoring Service [data set], https://doi.org/10.5904/wgms-glathida-2020-10, 2020. a
Glen, J. W.: The Creep of Polycrystalline Ice, P. Roy. Soc. Lond. Ser. A-Math., 228, 519–538, https://doi.org/10.1098/rspa.1955.0066, 1955. a
Goodfellow, I., Bengio, Y., and Courville, A.: Deep Learning, MIT Press, https://www.deeplearningbook.org/ (last access: 31 January 2025), 2016. a
Haq, M. A., Azam, M. F., and Vincent, C.: Efficiency of artificial neural networks for glacier ice-thickness estimation: A case study in western Himalaya, India, J. Glaciol., 67, 671–684, https://doi.org/10.1017/jog.2021.19 2021. a
Höhl, A., Obadic, I., Fernández-Torres, M.-Á., Najjar, H., Oliveira, D. A. B., Akata, Z., Dengel, A., and Zhu, X. X.: Opening the Black Box: A systematic review on explainable artificial intelligence in remote sensing, IEEE Geosci. Remote Sens. Mag., 12, 261–304, https://doi.org/10.1109/MGRS.2024.3467001, 2024 a
Hugonnet, R., McNabb, R., Berthier, E., Menounos, B., Nuth, C., Girod, L., Farinotti, D., Huss, M., Dussaillant, I., Brun, F., and Kääb, A.: Accelerated global glacier mass loss in the early twenty-first century, Nature, 592, 726–731, https://doi.org/10.1038/s41586-021-03436-z, 2021. a, b
Iwasaki, Y. and Lai, C.-Y.: One-dimensional ice shelf hardness inversion: Clustering behavior and collocation resampling in physics-informed neural networks, J. Comput. Phys., 492, 112435, https://doi.org/10.1016/j.jcp.2023.112435, 2023. a, b, c, d
Jiskoot, H.: Dynamics of Glaciers, in: Encyclopedia of Snow, Ice and Glaciers, edited by: Singh, V. P., Singh, P., and Haritashya, U. K., Springer Netherlands, Dordrecht, 245–256, ISBN 978-90-481-2642-2, https://doi.org/10.1007/978-90-481-2642-2_127, 2011. a
Jouvet, G.: Inversion of a Stokes glacier flow model emulated by deep learning, J. Glaciol., 69, 13–26, https://doi.org/10.1017/jog.2022.41, 2023. a
Jouvet, G. and Cordonnier, G.: Ice-flow model emulator based on physics-informed deep learning, J. Glaciol., 69, 1941–1955, https://doi.org/10.1017/jog.2023.73, 2023. a
Jouvet, G., Cordonnier, G., Kim, B., Lüthi, M., Vieli, A., and Aschwanden, A.: Deep learning speeds up ice flow modelling by several orders of magnitude, J. Glaciol., 68, 651–664, https://doi.org/10.1017/jog.2021.120, 2022. a
Karniadakis, G. E., Kevrekidis, I. G., Lu, L., Perdikaris, P., Wang, S., and Yang, L.: Physics-informed machine learning, Nat. Rev. Phys., 3, 422–440, https://doi.org/10.1038/s42254-021-00314-5, 2021. a, b
Koch, M., Seehaus, T., Friedl, P., and Braun, M.: Automated Detection of Glacier Surges from Sentinel-1 Surface Velocity Time Series—An Example from Svalbard, Remote Sens., 15, 1545, https://doi.org/10.3390/rs15061545, 2023. a
Kokhlikyan, N., Miglani, V., Martin, M., Wang, E., Alsallakh, B., Reynolds, J., Melnikov, A., Kliushkina, N., Araya, C., Yan, S., and Reblitz-Richardson, O.: Captum: A unified and generic model interpretability library for PyTorch, arXiv [preprint], https://doi.org/10.48550/arXiv.2009.07896, 2020. a
Lagaris, I., Likas, A., and Fotiadis, D.: Artificial neural networks for solving ordinary and partial differential equations, IEEE T. Neur. Netw., 9, 987–1000, https://doi.org/10.1109/72.712178, 1998. a
Leong, W. J. and Horgan, H. J.: DeepBedMap: a deep neural network for resolving the bed topography of Antarctica, The Cryosphere, 14, 3687–3705, https://doi.org/10.5194/tc-14-3687-2020, 2020. a
Lindbäck, K., Kohler, J., Pettersson, R., Nuth, C., Langley, K., Messerli, A., Vallot, D., Matsuoka, K., and Brandt, O.: Subglacial topography, ice thickness, and bathymetry of Kongsfjorden, northwestern Svalbard, Earth Syst. Sci. Data, 10, 1769–1781, https://doi.org/10.5194/essd-10-1769-2018, 2018. a
Lundberg, S. M. and Lee, S.-I.: A Unified Approach to Interpreting Model Predictions, in: Advances in Neural Information Processing Systems 30 (NIPS 2017), edited by: Guyon, I., Luxburg, U. V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., and Garnett, R., Curran Associates, Inc., 4765–4774, ISBN 9781510860964, 2017. a
Maussion, F., Butenko, A., Champollion, N., Dusch, M., Eis, J., Fourteau, K., Gregor, P., Jarosch, A. H., Landmann, J., Oesterle, F., Recinos, B., Rothenpieler, T., Vlug, A., Wild, C. T., and Marzeion, B.: The Open Global Glacier Model (OGGM) v1.1, Geosci. Model Dev., 12, 909–931, https://doi.org/10.5194/gmd-12-909-2019, 2019. a
Millan, R., Mouginot, J., Rabatel, A., and Morlighem, M.: Ice velocity and thickness of the world’s glaciers, Nat. Geosci., 15, 124–129, https://doi.org/10.1038/s41561-021-00885-z, 2022. a, b, c, d, e, f, g, h, i
Min, Y., Mukkavilli, S. K., and Bengio, Y.: Predicting ice flow using machine learning, arXiv [preprint], https://doi.org/10.48550/arXiv.1910.08922, 2019. a
Moholdt, G., Maton, J., Majerska, M., and Kohler, J.: Annual coastlines for Svalbard, Norwegian Polar Institute [data set], https://doi.org/10.21334/npolar.2021.21565514, 2021. a
Morlighem, M., Rignot, E., Seroussi, H., Larour, E., Ben Dhia, H., and Aubry, D.: A mass conservation approach for mapping glacier ice thickness, Geophys. Res. Lett., 38, L19503, https://doi.org/10.1029/2011GL048659, 2011. a
Pälli, A., Moore, J. C., and Rolstad, C.: Firn–ice transition-zone features of four polythermal glaciers in Svalbard seen by ground-penetrating radar, Ann. Glaciol., 37, 298–304, https://doi.org/10.3189/172756403781816059, 2003. a
Raissi, M., Perdikaris, P., and Karniadakis, G.: Physics-informed neural networks: A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations, J. Comput. Phys., 378, 686–707, https://doi.org/10.1016/j.jcp.2018.10.045, 2018. a, b
Rathore, P., Lei, W., Frangella, Z., Lu, L., and Udell, M.: Challenges in Training PINNs: A Loss Landscape Perspective, arXiv [preprint], https://doi.org/10.48550/arXiv.2402.01868, 2024. a
RGI Consortium: Randolph Glacier Inventory – A Dataset of Global Glacier Outlines, Version 6, Region 7, National Snow and Ice Data Center [data set], https://doi.org/10.7265/4M1F-GD79, 2017. a
Riel, B., Minchew, B., and Bischoff, T.: Data-Driven Inference of the Mechanics of Slip Along Glacier Beds Using Physics-Informed Neural Networks: Case Study on Rutford Ice Stream, Antarctica, J. Adv. Model. Earth Sy., 13, e2021MS002621, https://doi.org/10.1029/2021MS002621, 2021. a
Rolf, E.: Evaluation Challenges for Geospatial ML, arXiv [preprint], https://doi.org/10.48550/arXiv.2303.18087, 2023. a
Shapley, L. S.: A Value for n-Person Games, in: Contributions to the Theory of Games (AM-28), Volume II, edited by: Kuhn, H. W. and Tucker, A. W., Princeton University Press, 307–318, ISBN 978-1-4008-8197-0, https://doi.org/10.1515/9781400881970-018, 1953. a
Steidl, V.: GlacierPINN: Case Study Svalbard, Zenodo [code and data set], https://doi.org/10.5281/zenodo.13834016, 2024. a
Tancik, M., Srinivasan, P. P., Mildenhall, B., Fridovich-Keil, S., Raghavan, N., Singhal, U., Ramamoorthi, R., Barron, J. T., and Ng, R.: Fourier Features Let Networks Learn High Frequency Functions in Low Dimensional Domains, arXiv [preprint], https://doi.org/10.48550/arXiv.2006.10739, 2020. a
Teisberg, T. O., Schroeder, D. M., and MacKie, E. J.: A Machine Learning Approach to Mass-Conserving Ice Thickness Interpolation, in: 2021 IEEE International Geoscience and Remote Sensing Symposium IGARSS, pp. 8664–8667, IEEE, Brussels, Belgium, https://doi.org/10.1109/IGARSS47720.2021.9555002, 2021. a
van der Veen, C. J.: Fundamentals of Glacier Dynamics, CRC Press, 2nd Edn., ISBN 978-0-429-06380-0, 2013. a, b, c, d
van Pelt, W. and Frank, T.: New glacier thickness and bed topography maps for Svalbard, The Cryosphere, 19, 1–17, https://doi.org/10.5194/tc-19-1-2025, 2025. a, b, c, d
Wang, S., Sankaran, S., Wang, H., and Perdikaris, P.: An Expert's Guide to Training Physics-informed Neural Networks, arXiv [preprint], https://doi.org/10.48550/arXiv.2308.08468, 2023. a
Wang, Y., Lai, C.-Y., and Cowen-Breen, C.: Discovering the rheology of Antarctic Ice Shelves via physics-informed deep learning, Research Square [preprint], https://doi.org/10.21203/rs.3.rs-2135795/v1, 2022. a
Welty, E., Zemp, M., Navarro, F., Huss, M., Fürst, J. J., Gärtner-Roer, I., Landmann, J., Machguth, H., Naegeli, K., Andreassen, L. M., Farinotti, D., Li, H., and GlaThiDa Contributors: Worldwide version-controlled database of glacier thickness observations, Earth Syst. Sci. Data, 12, 3039–3055, https://doi.org/10.5194/essd-12-3039-2020, 2020. a
Xu, Q., Shi, Y., Bamber, J., Tuo, Y., Ludwig, R., and Zhu, X. X.: Physics-aware Machine Learning Revolutionizes Scientific Paradigm for Machine Learning and Process-based Hydrology, arXiv [preprint], https://doi.org/10.48550/arXiv.2310.05227, 2023. a, b
- Abstract
- Introduction
- Physics-aware machine learning
- Data
- Results
- Discussion
- Conclusions
- Appendix A: Relation between surface and depth-averaged velocity to set llower
- Appendix B: PINN architecture and training
- Appendix C: Experiment on synthetic 1D data
- Appendix D: PINN ice thickness prediction on LOGO test glaciers
- Appendix E: Importance of physics-aware loss components
- Appendix F: Importance of input features
- Code and data availability
- Author contributions
- Competing interests
- Disclaimer
- Acknowledgements
- Financial support
- Review statement
- References
- Abstract
- Introduction
- Physics-aware machine learning
- Data
- Results
- Discussion
- Conclusions
- Appendix A: Relation between surface and depth-averaged velocity to set llower
- Appendix B: PINN architecture and training
- Appendix C: Experiment on synthetic 1D data
- Appendix D: PINN ice thickness prediction on LOGO test glaciers
- Appendix E: Importance of physics-aware loss components
- Appendix F: Importance of input features
- Code and data availability
- Author contributions
- Competing interests
- Disclaimer
- Acknowledgements
- Financial support
- Review statement
- References