Articles | Volume 19, issue 10
https://doi.org/10.5194/tc-19-4759-2025
© Author(s) 2025. This work is distributed under
the Creative Commons Attribution 4.0 License.Learning to filter: snow data assimilation using a Long Short-Term Memory network
Download
- Final revised paper (published on 21 Oct 2025)
- Preprint (discussion started on 12 Feb 2025)
Interactive discussion
Status: closed
Comment types: AC – author | RC – referee | CC – community | EC – editor | CEC – chief editor
| : Report abuse
-
RC1: 'Comment on egusphere-2025-423', Anonymous Referee #1, 13 Mar 2025
- AC1: 'Reply on RC1', Giulia Blandini, 09 May 2025
-
RC2: 'Comment on egusphere-2025-423', Anonymous Referee #2, 31 Mar 2025
- AC2: 'Reply on RC2', Giulia Blandini, 09 May 2025
Peer review completion
AR: Author's response | RR: Referee report | ED: Editor decision | EF: Editorial file upload
ED: Publish subject to revisions (further review by editor and referees) (11 May 2025) by Nora Helbig

AR by Giulia Blandini on behalf of the Authors (29 May 2025)
Author's response
Author's tracked changes
Manuscript
ED: Referee Nomination & Report Request started (02 Jun 2025) by Nora Helbig
RR by Anonymous Referee #2 (11 Jun 2025)

RR by Anonymous Referee #1 (12 Jun 2025)

RR by Anonymous Referee #3 (02 Jul 2025)

ED: Publish subject to revisions (further review by editor and referees) (03 Jul 2025) by Nora Helbig

AR by Giulia Blandini on behalf of the Authors (20 Aug 2025)
Author's response
Author's tracked changes
Manuscript
ED: Referee Nomination & Report Request started (21 Aug 2025) by Nora Helbig
RR by Anonymous Referee #3 (04 Sep 2025)

ED: Publish as is (04 Sep 2025) by Nora Helbig

AR by Giulia Blandini on behalf of the Authors (11 Sep 2025)
This work developed a surrogate for EnKF-DA using an LSTM network. The introduction and methods sections are well written and structured. However, there are several errors in the results that are inconsistent with the plots. More importantly, the results lack sufficient explanation and analysis regarding why the LSTM performs differently from EnKF at different sites or scenarios. The discussion could benefit from additional comparisons with previous studies and a deeper analysis of the results. Currently, it leans more toward reinforcing the need for LSTM in data assimilation, which somewhat repeats points already made in the introduction. Therefore, I recommend a major revision before publication.
Line 103-105: What is the source of the meteorological forcing data? Are they derived from gridded datasets?
Table 2: The data time span for each site needs to be mentioned.
Line 171: Forecasted model state is x_k^f
Line 254: Double “the”
Line 271: “predictions”
Line 277-278: Please clarify how the data were split: by individual data points or by continuous time spans?
Line 276: Please clarify what are site-specific limits here
Line 280: The inline formula here should not include 'star,' as 'star' was previously used to represent the LSTM output, not the input from S3M. Please keep consistent.
Line 288-290: Please use a formula to clarify this configuration. Do you mean that x^f and forcing at both time steps k and k-1 are used as LSTM inputs in the second test? Please refer to Figure 2 for clarity.
Line 292-294: This part is confusing. What is the difference between Configuration 1 and Configuration 3? Was a single LSTM selected from Configuration 1 and then applied to other sites? Please clarify.
Line 299-300: Is there a specific reason to randomly sample water years for data splitting rather than using a continuous historical time span to train the model and a continuous future time span to test it? Random sampling can create artificially easier test conditions by allowing test data (time period) to fall between training water years, which may provide the model with indirect information about future conditions.
LSTM structure and hyperparameters were not mentioned in this work.
Line 309-311 (Figure 3): Is this result from testing or operational testing? Please clarify
Line 313-314: It is somewhat difficult to distinguish the EnKF-DA and LSTM boxes in the plots. If the last box in each panel represents LSTM-DA, it suggests that the RMSE values of LSTM-DA for KHT, RME, and FMI-ARC increased compared to EnKF-DA, with KHT showing the largest increase. This appears inconsistent with the narrative presented here. Please check.
Figure 3&4: The Nash-Sutcliffe coefficient can be used as a score to evaluate the accuracy of the time series in (a)–(d).
Line 321-324: Why is the LSTM trained with outputs (states) from EnKF-DA more sensitive to the sparsity of observation data? Could you explain this here? Including observation data as an input may introduce artificial errors when filling in missing data in the input.
Line 336-337: Only Figure 5b shows improvement with memory component, rather than c and d
Line 342-348: Cite Figure 6 here.
Line 344: 0.5 m? The reduction shown in figure 6f is not that large.
Line 346: These strategies were not mentioned and explained in the method.
Section 3.3: This result does not seem meaningful, as the spatial transferability of all models appears to be poor. Please consider removing it.
Line 370-371: Any explanation for this result?
Section 3.4: Instead of presenting the spatial transferability of a single model, it might be more meaningful to compare and discuss the site-specific LSTM and the multi-site LSTM.
Please refer (this is not my work and no need to cite it.): Kratzert, Frederik, Martin Gauch, Daniel Klotz, and Grey Nearing. "HESS Opinions: Never train a Long Short-Term Memory (LSTM) network on a single basin." Hydrology and Earth System Sciences 28, no. 17 (2024): 4187-4201.
Line 410: No results were shown to support this.
Line 415: 7 sites?