Hostname: page-component-89b8bd64d-7zcd7 Total loading time: 0 Render date: 2026-05-09T04:27:22.218Z Has data issue: false hasContentIssue false

Routes towards an effective AI in CFD: an epistemological and technical perspective

Published online by Cambridge University Press:  19 March 2026

Michaël Bauerheim*
Affiliation:
ISAE-SUPAERO, 10 Av. Marc Pélegrin, Toulouse, France
*
Corresponding author: Michaël Bauerheim, michael.bauerheim@isae-supaero.fr

Abstract

The integration of Artificial Intelligence (AI) into computational science (CS) and computational fluid dynamics (CFD) has raised profound epistemological debates concerning the nature of knowledge and its effectiveness in science. A central question in this discourse is whether AI can rival, or potentially surpass, the effectiveness of traditional mathematical methods in addressing the intricate challenges of CFD. In this work, I examine the concept of effectiveness within this context, highlighting the fundamental epistemological distinctions between AI-driven approaches and classical mathematical techniques. First, this analysis identifies four foundational pillars of effectiveness (PoEs) in scientific methods: (i) symmetries, which impose internal structure and coherence; (ii) scale separation, allowing specific treatments for the different scales and their interactions; (iii) sparsity, which simplifies complexity and enhances explicability; and (iv) semantic significance, which fosters abstraction, reasoning and interpretability. Yet, unlike mathematics where rigour ensures credibility by default, AI methods raise additional concerns of robustness and trust. Therefore, beyond the four PoEs, I also discuss credibility as a complementary pillar essential for the adoption of AI in the CFD community. The next critical step is to assess whether, and to what extent, AI can emulate or even outperform the roles and functions traditionally fulfilled by mathematical models. I therefore systematically review if, and how, these four pillar of effectiveness can be applied to AI-based algorithms. I show that those pillars are actually declined in a succession of technical advances that have shown promising results when using AI in CFD.

Information

Type
JFM Perspectives
Copyright
© The Author(s), 2026. Published by Cambridge University Press
Figure 0

Table 1. Various well-known neural architectures, and their respective input format $x_n$ and associated aggregation function $h_\theta$ parametrised by $\theta = W$ (bias is omitted for brevity). Architectures considered include the multilayer perceptron (MLP), convolutional neural network (CNN), Lipschitz-constrained neural network (LCNN), recurrent neural network (RNN) or graph neural network (GNN).

Figure 1

Table 2. Overview of various neural network tasks with their associated learned function and training objectives.

Figure 2

Figure 1. (a) Classical network, where training data $(x, y_T)$ are sampled and fed to a parametrised function $f_\theta$. Training is achieved by measuring a loss function $\mathcal{L}$ between the prediction ($y$) and the target ($y_T$). Network’s weights $\theta$ are updated by a back-propagation algorithm and gradient descent methods. (b) PINN employs an equation $\mathcal{R}(y, x) = 0$. Training is achieved by computing the residual $\mathcal{R}$ of the physical equation from the predictions $y(x)$. To do so, partial derivatives $ {\partial ^n y}/{\partial x^n}$ are computed by automatic differentiation. Network’s weights are updated using the residual $\mathcal{R}$ as physical-based loss function. Since $f_\theta$ is a continuous function of $x$, the output $y(x)$ is theoretically continuous in space and time (infinite resolution).

Figure 3

Figure 2. (a) CNN auto-encoder (AE) architecture to predict a dynamical system (Usandivaras et al.2024b), where a network $\phi _e$ encodes the snapshot $y_t$, producing a latent code $z_t$. This code is propagated in time by a network $\zeta$ such that $z_{t+\Delta t} = \zeta (z_t)$. This new code is then decoded by the network $\phi _d$, providing the high-dimensional snapshot $t_{t+\Delta t}$ of the system. (b) The AE algorithm shows excellent results compared with CFD, while classical POD fails to effectively reproduce the turbulent dynamics. (c) Sketch of the fluid–structure problem where a flexible drone flies over a gust (transverse velocity $U_{{gust}}$) (Colombo, Morlier & Bauerheim 2023). This problem is formulated through a graph, where elements of the structures are connected together. Thus, graph networks can be trained to replicate the forced fluid–structure dynamics. (d) Results show that a simple graph network (GN, blue) produces slight inaccuracy on node displacements, yielding large errors on the global displacements. This issue can be limited by hybridising graph networks and neural ODE (GDEC, red).

Figure 4

Table 3. Various well-known neural architectures, and their respective domain and underlying symmetry. The first part of the table is extracted from Bronstein et al. (2021). Note that table 1 details how aggregation functions must be adapted for each type of network presented here.

Figure 5

Figure 3. (a) Sketch of the experiment to evaluate empirically the receptive field (RF) of a neural network (NN). The network is randomly initialised and the central pixel of a constant image is perturbed, here in white. The size of the non-uniform zone of the output highlights the RF value. (b) This test is done on three models: (1) a network with one convolutional layer with kernel size $3 \times 3$; (2) a network with five sequential convolutional layers with kernel size $3 \times 3$; and (3) a UNet architecture containing also five convolutional layers with kernel size $3 \times 3$, but dispatch over three branches, acting at different resolutions (N, N/2 and N/4). Convolutional layers are marked with a blue rectangle, D refers to down-sampling, U to up-sampling and + to concatenation. Output of the three models are given in panel (c), showing clearly the advantage of UNet to provide larger RF. This empirical RF value is given for the three models: 3, 11 and 26.

Figure 6

Figure 4. (a) Results obtain by a deep neural network to solve the Poisson equation embedded in an incompressible solver (Illarramendi et al.2020b). Training is performed on non-viscous flows. Test results on viscous flows at various Reynolds numbers ($Re = 100$, $300$ and $1000$) are presented for a feedforward CNN (MONO), UNet architecture (MULTI) and the reference CFD solutions. Note that MONO and MULTI have the same number of weights, they only differ by their architecture, revealing the excellent performance of multi-scale networks like UNet in CFD. (b) Experiment on a test consisting in copying a value of a source node ($\star$) in all other nodes of an unstructured mesh. Results are provided for a GNN with 3 layers (GNN 3) and 12 layers (GNN 12), as well as with a graph network with global attention (GNN ATT). This illustrates the locality of shallow network (GNN 3), the difficulty of training large graph networks (GNN 12) and the excellent behaviours of graph when equipped to capture long-distance relationships (GNN ATT).

Figure 7

Figure 5. (a) Illustration of a FNO layer, which combines a learned linear operator mixing channels (black arrows, matrix $R$) in the Fourier space with a local linear map in physical space (green arrows, matrix $W$) acting through a skip connection. This allows to incorporate long-distance relationships between pixels, since spectral modes are specifically designed to separate and transform features at each scale. (b) INR implicitly learns the long-distance relationship by outputing directly the desired quantity $F_\theta (x, y)$ at a query point $(x, y)$. However, training can be limited by spectral bias, which makes high frequencies more difficult to learn. To avoid this, spectral embeddings are incorporated (Sitzmann et al.2020). A drawback of INR is its difficulty to input field, which requires specific compression and representation techniques (latent variable $z$), hyper-networks ($\phi$) that modulate the weight $\theta = \phi (z)$ of the main network, and meta-learning strategies.

Figure 8

Table 4. Mean $L^2$ relative error between the true and predicted results using various operator learning approaches, among which Graph Neural Operator (GNO), DeepONet, Fourier Neural Operator (FNO) and Wavelet Neural Operator (WNO). Best results are highlighted in bold. Table is extracted from Tripura & Chakraborty (2023).

Figure 9

Figure 6. (a) Illustration of a multi-scale INR architecture with various embedding vectors $\gamma _\sigma$. (b) Experimental results for fitting a single signal using a 3-layer INR with different Fourier encodings. This experiment can be reproduced from our Git repository. (c) Results obtained by CFD, graph network (MGN) and INR on 3-D wing pressure predictions, with (d) a quantitative comparison on the pressure coefficient $C_p$ predictions.

Figure 10

Figure 7. (a) Multi-fidelity configuration where fine (HF, top) and coarse (LF, bottom) LES have been performed by Usandivaras et al. (2024a). (b) A 2-D portion of the dataset, highlighting the LF and HF datasets when the recess length ($l_r$) and chamber diameter ($d_c$) are varied. (c) Results reveal that the LF AI model (green) is inaccurate compared with the true HF LES data (CFD, blue), while the multi-fidelity model trained on LF and fine-tuned on HF is in good agreement with HF LES data.

Figure 11

Figure 8. Algorithm SINDy proposed by Brunton et al. (2016a) for sparse identification of dynamical system. Here, illustration on the 63 Lorentz system (a) with $\sigma = 10$, $\rho = 28$ and $\beta = 8/3$. SINDy search for a linear model $\varXi$, where features are given by a function library $\varTheta (A)$. (c) Training is achieved by minimising the error $\| \dot A - \varTheta (A) \xi \|_2$, in addition to a regularisation term $\lambda \| \xi \|_0$ to enforce sparsity. It yields identified systems (panel d), which are both accurate and sparse, fostering interpretability. Here, true coefficients are well approximated: $\sigma ^* \approx 9.9998$, $\rho ^* \approx 27.998$ and $\beta ^* \approx 2.6665$. Figure is modified from Brunton et al. (2016a).

Figure 12

Figure 9. Illustration of the active learning approach introduced by Corban, Bauerheim & Jardin (2023). (a) Sketch of the framework where data from CFD (i) are collected in an iterative dataset (ii) to train a deep neural network (iii) predicting the lift ($C_L$), power ($C_P$) and efficiency ($\eta = C_L/C_P$) of a 3-D flapping wing depending on its kinematics. NSGA II reuses this network to optimise the kinematics (iv), obtaining a predicted Pareto front (magenta). To guarantee its accuracy, new data are collected along this Pareto front ($\circ$) by an acquisition function $\mathcal{F}$ (v), and evaluated by CFD, providing new training data ($\triangle$). The Pareto front obtained at the first (top) and third (bottom) steps of this iterative active learning are displayed (panel b), showing a good convergence. At each step, suboptimal test data are also displayed (right), showing the mismatch between the CFD ($\bullet$) and network prediction (line).

Figure 13

Figure 10. (a) Illustration of classical approaches for uncertainty estimations through a variance $\mathbb{V}(y)$: deep ensemble (DE, left) and dropout (DO, right). (b) Sketch of the zigzag (ZZ) methodology introduced by Durasov et al. (2024). It consists in a network $f_\theta (x, y)$ which is fed inputs $x$, as well as by its own prediction $y$. A two step approach, starting from $y = 0$, provides a first answer $y^1$, which is then fed for a second step, providing a new output $y^2$. The distance $\| y^2 - y^1 \|$ measures the uncertainty of the model, and thus allows to discriminate between ID or OOD scenario. (c) Uncertainty estimations by deep ensemble (DE), Monte Carlo dropout (DO), zigzag (ZZ) and Gaussian process (GP) applied to the trailing edge noise dataset (Brooks, Pope & Marcolini 1989; Markelle, Longjohn & Nottingham 2014). It shows the mean prediction $\mu _\theta$, as well as the aleatoric and total uncertainty (confidence intervals, CI).

Figure 14

Figure 11. (a) Results obtained by Wei et al. (2024) with a conditioned DM (CLSDM) to generate possible aerofoil geometries. (b) Visualisation of the results using a t-sne map revealed that CLSDM can generate more diverse blade better than CGAN, especially when only limited training data are available (here, 100 training aerofoils). (c) Example of 3-D turbomachinery blades and their 2-D projection generated by CLSDM, when constrained on thickness (top) and twist (bottom).