Parareal with a physics-informed neural network as coarse propagator
March 07, 2023 Β· Declared Dead Β· π European Conference on Parallel Processing
"No code URL or promise found in abstract"
Evidence collected by the PWNC Scanner
Authors
Abdul Qadir Ibrahim, Sebastian GΓΆtschel, Daniel Ruprecht
arXiv ID
2303.03848
Category
math.NA: Numerical Analysis
Cross-listed
cs.CE,
cs.DC
Citations
13
Venue
European Conference on Parallel Processing
Last Checked
1 month ago
Abstract
Parallel-in-time algorithms provide an additional layer of concurrency for the numerical integration of models based on time-dependent differential equations. Methods like Parareal, which parallelize across multiple time steps, rely on a computationally cheap and coarse integrator to propagate information forward in time, while a parallelizable expensive fine propagator provides accuracy. Typically, the coarse method is a numerical integrator using lower resolution, reduced order or a simplified model. Our paper proposes to use a physics-informed neural network (PINN) instead. We demonstrate for the Black-Scholes equation, a partial differential equation from computational finance, that Parareal with a PINN coarse propagator provides better speedup than a numerical coarse propagator. Training and evaluating a neural network are both tasks whose computing patterns are well suited for GPUs. By contrast, mesh-based algorithms with their low computational intensity struggle to perform well. We show that moving the coarse propagator PINN to a GPU while running the numerical fine propagator on the CPU further improves Parareal's single-node performance. This suggests that integrating machine learning techniques into parallel-in-time integration methods and exploiting their differences in computing patterns might offer a way to better utilize heterogeneous architectures.
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
π Similar Papers
In the same crypt β Numerical Analysis
R.I.P.
π»
Ghosted
R.I.P.
π»
Ghosted
Deep learning-based numerical methods for high-dimensional parabolic partial differential equations and backward stochastic differential equations
R.I.P.
π»
Ghosted
PDE-Net: Learning PDEs from Data
R.I.P.
π»
Ghosted
Efficient tensor completion for color image and video recovery: Low-rank tensor train
R.I.P.
π»
Ghosted
Tensor Ring Decomposition
R.I.P.
π»
Ghosted
Machine learning approximation algorithms for high-dimensional fully nonlinear partial differential equations and second-order backward stochastic differential equations
Died the same way β π» Ghosted
R.I.P.
π»
Ghosted
Language Models are Few-Shot Learners
R.I.P.
π»
Ghosted
PyTorch: An Imperative Style, High-Performance Deep Learning Library
R.I.P.
π»
Ghosted
XGBoost: A Scalable Tree Boosting System
R.I.P.
π»
Ghosted