This is a Preprint and has not been peer reviewed. This is version 1 of this Preprint.
Unified Cross-Modal Learning for Hydrological Processes Using Multi-Task Transformer Framework
Downloads
Authors
Abstract
Most deep learning studies in hydrology adopt single-task frameworks that address individual variables such as rainfall or streamflow independently, limiting opportunities for shared learning across related environmental processes. This study introduces a unified multi-task, multi-modal deep learning framework capable of jointly performing 24-hour horizon streamflow forecasting and rainfall temporal super-resolution. The model employs a shared Transformer encoder with task-specific decoders to integrate temporal and spatial hydrological information within a single architecture. To assess the influence of joint optimization, the same model is also trained individually for each task, enabling direct comparison between single-task and multi-task configurations and performance. Results show that multi-task training maintains or modestly improves predictive accuracy relative to individually trained counterparts while preserving performance comparable to established baselines. The framework demonstrates stable streamflow forecasts and hydrologically consistent rainfall reconstructions, highlighting the potential of unified, process-aware architectures for representing multiple components of the hydrological cycle within one coherent learning system.
DOI
https://doi.org/10.31223/X5FQ8Q
Subjects
Environmental Engineering, Environmental Monitoring, Hydraulic Engineering, Hydrology, Water Resource Management
Keywords
multi-modal, Deep learning, Multi-Task Learning (MTL), Transformers, hydrology, streamflow, rainfall
Dates
Published: 2025-10-31 23:51
Last Updated: 2025-10-31 23:51
License
CC BY Attribution 4.0 International
Additional Metadata
Data Availability (Reason not available):
Used data is cited in the article.
There are no comments or no comments have been made public for this article.