Attention-Based Deep Learning for Runoff Forecasting: Evaluating the Temporal Fusion Transformer Against Traditional Machine Learning Models
Discuss this preprint
Start a discussion What are Sciety discussions?Listed in
This article is not in any list yet, why not save it to one of your lists.Abstract
Reliable runoff forecasting is critical for water management and flood preparedness in Nepal’s steep, data-scarce catchments. Traditional models such as SWAT provide process insights but demand extensive calibration and detailed inputs often unavailable in such regions. Recent advances in attentionbased deep learning offer new opportunities to capture temporal dependencies with improved interpretability. This study evaluates the Temporal Fusion Transformer (TFT) for monthly runoff prediction using 40 years (1980–2020) of hydrometeorological data from Nepal, benchmarked against Random Forest (RF) and Long Short-Term Memory (LSTM) networks. Results show that RF underestimates peaks, LSTM captures seasonality but falters under monsoon extremes, while TFT consistently achieves superior accuracy (RMSE = 22.5, R2 = 0.88). Attention weights further reveal precipitation and antecedent runoff as dominant drivers, reinforcing hydrological understanding. These findings highlight attention-based architectures as accurate and interpretable tools for operational flood forecasting and climate-resilient water management.