h-Analysis and data-parallel physics-informed neural networks
Abstract
We explore the data-parallel acceleration of physics-informed machine learning (PIML) schemes, with a focus on physics-informed neural networks (PINNs) for multiple graphics processing units (GPUs) architectures. In order to develop scale-robust and high-throughput PIML models for sophisticated applications which may require a large number of training points (e.g., involving complex and high-dimensional domains, non-linear operators or multi-physics), we detail a novel protocol based on h-analysis and data-parallel acceleration through the Horovod training framework. The protocol is backed by new convergence bounds for the generalization error and the train-test gap. We show that the acceleration is straightforward to implement, does not compromise training, and proves to be highly efficient and controllable, paving the way towards generic scale-robust PIML. Extensive numerical experiments with increasing complexity illustrate its robustness and consistency, offering a wide range of possibilities for real-world simulations.
Más información
Título según WOS: | h-Analysis and data-parallel physics-informed neural networks |
Título según SCOPUS: | ID SCOPUS_ID:85174241886 Not found in local SCOPUS DB |
Título de la Revista: | SCIENTIFIC REPORTS |
Volumen: | 13 |
Editorial: | NATURE PORTFOLIO |
Fecha de publicación: | 2023 |
DOI: |
10.1038/S41598-023-44541-5 |
Notas: | ISI, SCOPUS |