GPU Optimization of Lattice Boltzmann Method with Local Ensemble Transform Kalman Filter
August 07, 2023 ยท Declared Dead ยท ๐ 2022 IEEE/ACM Workshop on Latest Advances in Scalable Algorithms for Large-Scale Heterogeneous Systems (ScalAH)
"No code URL or promise found in abstract"
Evidence collected by the PWNC Scanner
Authors
Yuta Hasegawa, Toshiyuki Imamura, Takuya Ina, Naoyuki Onodera, Yuuichi Asahi, Yasuhiro Idomura
arXiv ID
2308.03310
Category
physics.flu-dyn
Cross-listed
cs.DC,
physics.comp-ph
Citations
2
Venue
2022 IEEE/ACM Workshop on Latest Advances in Scalable Algorithms for Large-Scale Heterogeneous Systems (ScalAH)
Last Checked
1 month ago
Abstract
The ensemble data assimilation of computational fluid dynamics simulations based on the lattice Boltzmann method (LBM) and the local ensemble transform Kalman filter (LETKF) is implemented and optimized on a GPU supercomputer based on NVIDIA A100 GPUs. To connect the LBM and LETKF parts, data transpose communication is optimized by overlapping computation, file I/O, and communication based on data dependency in each LETKF kernel. In two dimensional forced isotropic turbulence simulations with the ensemble size of $M=64$ and the number of grid points of $N_x=128^2$, the optimized implementation achieved $\times3.80$ speedup from the naive implementation, in which the LETKF part is not parallelized. The main computing kernel of the local problem is the eigenvalue decomposition (EVD) of $M\times M$ real symmetric dense matrices, which is computed by a newly developed batched EVD in $\verb|EigenG|$. The batched EVD in $\verb|EigenG|$ outperforms that in $\verb|cuSOLVER|$, and $\times65.3$ speedup was achieved.
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
๐ Similar Papers
In the same crypt โ physics.flu-dyn
R.I.P.
๐ป
Ghosted
R.I.P.
๐ป
Ghosted
Efficient collective swimming by harnessing vortices through deep reinforcement learning
R.I.P.
๐ป
Ghosted
NVIDIA SimNet^{TM}: an AI-accelerated multi-physics simulation framework
R.I.P.
๐ป
Ghosted
Teaching the Incompressible Navier-Stokes Equations to Fast Neural Surrogate Models in 3D
R.I.P.
๐ป
Ghosted
Prediction of Reynolds Stresses in High-Mach-Number Turbulent Boundary Layers using Physics-Informed Machine Learning
R.I.P.
๐ป
Ghosted
From Deep to Physics-Informed Learning of Turbulence: Diagnostics
Died the same way โ ๐ป Ghosted
R.I.P.
๐ป
Ghosted
Language Models are Few-Shot Learners
R.I.P.
๐ป
Ghosted
PyTorch: An Imperative Style, High-Performance Deep Learning Library
R.I.P.
๐ป
Ghosted
XGBoost: A Scalable Tree Boosting System
R.I.P.
๐ป
Ghosted