*Result*: Implementation and Evaluation of Parallel Computing Approaches for Large‐Domain, Process‐Based Hydrologic Simulations.

Title:
Implementation and Evaluation of Parallel Computing Approaches for Large‐Domain, Process‐Based Hydrologic Simulations.
Authors:
Guo, Junwei1 (AUTHOR) junwei.guo@ucalgary.ca, Clark, Martyn P.1 (AUTHOR), Knoben, Wouter J. M.1 (AUTHOR), Keshavarz, Kasra1 (AUTHOR), Klenk, Kyle2 (AUTHOR), Van Beusekom, Ashley2 (AUTHOR), Guenter, Victoria2 (AUTHOR), Spiteri, Raymond J.2 (AUTHOR)
Source:
Journal of Advances in Modeling Earth Systems. Jan2026, Vol. 18 Issue 1, p1-18. 18p.
Geographic Terms:
Database:
GreenFILE

*Further Information*

*Process‐based hydrologic simulations in large domains generally require intensive computing resources. In this study, we implement various parallelization approaches within a process‐based hydrologic solver, SUMMA, including the Message Passing Interface (MPI), Open Multi‐Processing (OMP), and the Actor Model, to enable high‐performance computing for large‐domain hydrologic simulations. We provide detailed guidelines on these implementations to assist hydrologists in parallelizing their models effectively. Using a hydrologic simulation over North America as a case study, we compare the scalability, computational cost, input/output performance, and coupling capabilities of these parallel approaches with the original sequential approach. Our results show that the SUMMA‐MPI exhibits linear scaling up to 1,024 cores, whereas SUMMA‐OMP is only recommended for smaller numbers of cores. The MPI approach exhibited a straggler effect, resulting in core utilization of only 80%. To address this, we introduced a load‐balancing calibration based on historical runs, which increases SUMMA‐MPI core usage to 95% and thereby mitigates the straggler effect. With regard to coupling capabilities, MPI is the most effective for large‐scale simulations involving multiple nodes and extensive core counts, supporting strong coupling and synchronization. The Actor Model reveals its excellent fault tolerance that enables automatic modification and recommencement of specific Grouped Response Units (GRUs) rather than restarting the entire simulation in the event of a failure within the simulation. Through this study, the implementation details of multiple parallelization schemes are documented and their advantages and limitations are discussed, which provides parallel computing insights for advancing computational hydrology in the Earth System Science community. Plain Language Summary: We provide detailed guidelines for Earth System Science developers on the implementation and evaluation of multiple parallel computing frameworks for large‐domain process‐based hydrologic simulations, including the Message Passing Interface (MPI), Open Multi‐Processing (OMP), and the Actor Model. As a case study, we conduct a North American hydrologic simulation consisting of 517,315 land units. Among these approaches, SUMMA‐MPI exhibits linear scaling up to 1,024 cores and offers the best synchronization and coupling capabilities. In contrast, SUMMA‐OMP is only recommended for smaller core counts due to significant overhead when scaling to larger allocations. The SUMMA‐Actors application encounters a bottleneck in input/output operations caused by its sequential I/O, a limitation stemming from SUMMA's reliance on the netCDF library. However, SUMMA‐Actors offers excellent fault tolerance by enabling automatic modification and recommencement of specific land units, eliminating the need to restart entire simulations in case of failure. Key Points: We implemented and evaluated Message Passing Interface (MPI), OpenMP, and Actor Model for a process‐based hydrologic solverThe MPI framework demonstrates linear scaling up to 1,024 cores and has excellent coupling capabilities [ABSTRACT FROM AUTHOR]

Copyright of Journal of Advances in Modeling Earth Systems is the property of Wiley-Blackwell and its content may not be copied or emailed to multiple sites without the copyright holder's express written permission. Additionally, content may not be used with any artificial intelligence tools or machine learning technologies. However, users may print, download, or email articles for individual use. This abstract may be abridged. No warranty is given about the accuracy of the copy. Users should refer to the original published version of the material for the full abstract. (Copyright applies to all Abstracts.)*