AI Accelerates OpenFOAM with Data-Driven Models

  • Data-driven physics models are improving efficiency in simulations, reducing runtime from hours to minutes.
  • Intel Labs showcased General Physics-Informed Neural Networks (PINNs) at ISC 2024, integrating physical equations into neural networks for enhanced accuracy.
  • Intel’s open-source framework demonstrated a 1.3× speedup in training and a 3.6× speedup in inference compared to NVIDIA H100, and a 6.8× improvement over PyTorch implementations.
  • PINNs effectively address complex, high-dimensional problems, offering superior performance in uncertain scenarios.
  • The framework uses SmartSim and SmartRedis to integrate machine learning models with OpenFOAM, providing benefits such as Bayesian optimization, reduced basis calculations, and online training.

Main AI News:

Data-driven physics models are rapidly advancing within the Modeling and Simulation sector due to their efficiency compared to traditional simulations, achieving results in minutes rather than hours, and their capability to model complex, coupled physical phenomena. A prime example is Intel Labs’ demonstration at ISC 2024 in Hamburg, Germany, where they showcased the power of General Physics-Informed Neural Networks (PINNs). These frameworks integrate the fundamental physical equations of a system into a neural network, allowing the network to adapt and learn from data while adhering to these governing equations.

PINNs excel in providing accurate solutions to partial differential equations (PDEs) by learning from data rather than requiring explicit boundary condition definitions, which are often a major source of uncertainty and error. This approach has proven effective, with reported error rates deemed “good enough” for many systems, making it possible to develop simulations that are both accurate and efficient for further analysis.

Intel Labs’ open-source framework demonstrated a substantial performance boost, achieving a 1.3× speedup in Computational Fluid Dynamics (CFD) workloads on an Intel Data Center GPU Max compared to an NVIDIA H100 for training, and a 3.6× speedup for inference. The framework also outperformed the PyTorch implementation by 6.8× in a physics-informed machine learning task using SmartSim. These initial results aim to introduce physics-informed machine learning to a broader high-performance computing (HPC) audience and set the stage for Intel’s next-generation GPUs, codenamed Falcon Shores, which promise even greater performance enhancements.

Advancing Uncertain and Complex Problem Solving

PINNs offer a promising solution for integrating data with mathematical physics models, especially in scenarios where problems are partially understood, uncertain, or high-dimensional. Research indicates that PINNs are particularly effective for ill-posed and inverse problems and can be scaled to large issues through domain decomposition. Despite their potential, there is a need for new frameworks, standardized benchmarks, and advanced mathematical techniques to ensure the robustness and scalability of next-generation physics-informed learning systems.

PINNs can even surpass the accuracy of traditional methods, as highlighted by Yuan, who noted improved results in mesh motion problems using physics-informed machine learning compared to classical methods.

Framework Overview and Computational Benefits

Intel Labs’ framework, presented at ISC 2024, introduces a gateway to PINNs optimized for Intel GPU Accelerators. It utilized two OpenFOAM CFD case studies—airfoil and a scalable motorcycle benchmark—for performance evaluation. The framework, built on the SmartSim solver, integrates machine learning models with OpenFOAM via the SmartRedis library, which facilitates data exchange and model execution.

The framework’s benefits include:

  1. Bayesian optimization for tuning turbulence model parameters to align low-resolution model results with high-fidelity references.
  2. Streaming CFD data to calculate reduced basis via partitioned singular value decomposition (SVD).
  3. Online training and inference for approximating mesh-point displacements in mesh-motion scenarios.

Solving PDEs with Machine Learning

PINNs represent a breakthrough in scientific machine learning, addressing problems involving Partial Differential Equations (PDEs) and Ordinary Differential Equations (ODEs) by training neural networks to minimize a loss function. This function incorporates initial and boundary conditions and PDE residuals at specific points in the domain. By embedding physical equations into the network, PINNs enable the model to learn from data and improve accuracy in physical system simulations.

PINNs offer a structured approach to solving high-dimensional, complex interactions in physical systems, using empirical data combined with physical knowledge to enhance performance. Methods such as Runga-Kutta integration within neural networks facilitate approximation of ODEs and PDEs from experimental data. During inference, these models mimic physical system behaviors, with alternative approaches incorporating various solvers and techniques into a backpropagation framework.

Conclusion:

The advancements presented by Intel Labs signify a transformative shift in simulation and modeling processes. The integration of data-driven PINNs with high-performance computing frameworks like OpenFOAM and Intel’s GPUs represents a significant leap forward in computational efficiency and accuracy. This evolution is poised to enhance simulation capabilities across various industries, offering faster and more precise results. As companies adopt these technologies, the competitive edge in scientific and industrial applications will likely expand, driving greater innovation and optimization in modeling and simulation practices.

Source