Hostname: page-component-77f85d65b8-hzqq2 Total loading time: 0 Render date: 2026-03-28T09:47:33.896Z Has data issue: false hasContentIssue false

Prediction and control of two-dimensional decaying turbulence using generative adversarial networks

Published online by Cambridge University Press:  21 February 2024

Jiyeon Kim
Affiliation:
School of Mathematics and Computing, Yonsei University, Seoul 03722, Korea
Junhyuk Kim
Affiliation:
Korea Atomic Energy Research Institute, Daejeon 34057, Korea
Changhoon Lee*
Affiliation:
School of Mathematics and Computing, Yonsei University, Seoul 03722, Korea Department of Mechanical Engineering, Yonsei University, Seoul 03722, Korea
*
Email address for correspondence: clee@yonsei.ac.kr

Abstract

An accurate prediction of turbulence has been very costly since it requires an infinitesimally small time step for advancing the governing equations to resolve the fast-evolving small-scale motions. With the recent development of various machine learning (ML) algorithms, the finite-time prediction of turbulence became one of promising options to relieve the computational burden. Yet, a reliable prediction of the small-scale motions is challenging. In this study, PredictionNet, a data-driven ML framework based on generative adversarial networks (GANs), was developed for fast prediction of turbulence with high accuracy down to the smallest scale using a relatively small number of parameters. In particular, we conducted learning of two-dimensional (2-D) decaying turbulence at finite lead times using direct numerical simulation data. The developed prediction model accurately predicted turbulent fields at a finite lead time of up to half the Eulerian integral time scale over which the large-scale motions remain fairly correlated. Scale decomposition was used to interpret the predictability depending on the spatial scale, and the role of latent variables in the discriminator network was investigated. The good performance of the GAN in predicting small-scale turbulence is attributed to the scale-selection and scale-interaction capability of the latent variable. Furthermore, by utilising PredictionNet as a surrogate model, a control model named ControlNet was developed to identify disturbance fields that drive the time evolution of the flow field in the direction that optimises the specified objective function.

Information

Type
JFM Papers
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2024. Published by Cambridge University Press.
Figure 0

Figure 1. Range of selected data for training with a time interval of $100\delta t/t^*$ shown in the yellow region in (a) the vorticity r.m.s. and dissipation rate and in (b) Reynolds number and the Taylor length scale. Example vorticity fields at (c) $t_0$ and (d) $t_{100}$ from a normalised test data.

Figure 1

Figure 2. Distribution of the temporal autocorrelation function of (a) the whole vorticity field and (b) the scale-decomposed vorticity fields.

Figure 2

Figure 3. Ensemble-averaged two-point correlation functions of vorticity extracted from 500 training data.

Figure 3

Figure 4. Simplified network schematics of (a) the baseline CNN, (b) cGAN-based PredictionNet and (c) ControlNet.

Figure 4

Figure 5. Network architecture of (a) PredictionNet (the generator of cGAN) and ControlNet and (b) discriminator of cGAN.

Figure 5

Table 1. Number of feature maps used at each convolutional layer of Convblks and Disblks.

Figure 6

Figure 6. Convergence of data loss depending on the lead time of (a) baseline CNN and (b) cGAN. The order of the converged value increases as the lead time increases. In addition, relatively large overfitting was observed in the case of $T=2T_L$.

Figure 7

Figure 7. Visualised examples of the performance of the cGAN and CNN for one test dataset. (a) Input field at $t_0$, prediction results at the lead time (b) $0.25T_L$, (c) $0.5 T_L$, (d) $T_L$ and (e) $2T_L$. Panels (b i,c i,d i,e i), (b ii,c ii,d ii,e ii) and (b iii,c iii,d iii,e iii) show the target DNS, cGAN predictions and CNN predictions, respectively.

Figure 8

Table 2. Quantitative comparison of the performance of the cGAN and CNN against the target in the prediction of the CC, MSE and r.m.s., and the $n$th standardised moments, Kolmogorov length scale ($\eta$) and dissipation rate ($\varepsilon '$) depending on the lead time for the input at $t_0$. All the dimensional data such as MSE, r.m.s. and $\varepsilon '$ are provided just for the relative comparison.

Figure 9

Figure 8. Comparison of statistics such as the p.d.f., two-point correlation and enstrophy spectrum of the target and prediction results at different lead times (a) $0.25T_L$, (b) $0.5 T_L$, (c) $T_L$ and (d) $2T_L$ for the same input time point $t_0$.

Figure 10

Figure 9. Prediction results of the scale-decomposed field for lead time $T_L$: (a) original fields, (b) large scales ($k\leq 4$), (c) intermediate scales ($4< k \leq 20$) and (d) small scales ($k>20$). Panels (a i,b i,c i,d i), (a ii,b ii,c ii,d ii) and (a iii,b iii,c iii,d iii) display the target DNS fields, cGAN predictions and CNN predictions, respectively.

Figure 11

Table 4. Correlation coefficient between the target and prediction of the scale-decomposed fields depending on the lead time.

Figure 12

Figure 10. Prediction results of the phase error of each model depending on $T$.

Figure 13

Figure 11. Evolution over training iteration of (a) $L_{true}$ and $L_{false}$ and (b) their difference evaluated every 2000 iterations.

Figure 14

Figure 12. Distributions of discriminator output (latent variable) for various fields at several iterations: (a) $T=0.25T_L$, (b) $T=0.5T_L$, (c) $T=T_L$ and (d) $T=2T_L$. All distributions are shifted by the target mean to fix the target distribution at the zero mean. The vertical black solid line indicates the target mean (zero) and the red line is the prediction mean. When the mean difference between the target and prediction is smaller than 0.5, the vertical lines are omitted for visibility.

Figure 15

Figure 13. Visualised prediction results for $T=2T_L$ of (a) target DNS field and single predictions using cGAN and CNN, (b) cGAN recursive predictions and (c) CNN recursive predictions.

Figure 16

Figure 14. Plots of the CC and MSE of the recursive predictions in terms of the lead time of the base recursive model. Only $T_L$ and $2T_L$ cases are included for CNN for visibility.

Figure 17

Figure 15. Enstrophy spectra of the best case of recursive prediction compared with the target and single prediction for $T=2T_L$ using (a) cGAN and (b) CNN.

Figure 18

Table 3. Statistical results of recursive predictions for various values of $T$ including the single prediction at the last entries. Best prediction for each lead time by cGAN is bold-faced.

Figure 19

Figure 16. Convergence of the data loss of ControlNet for $0.5T_L$. Here $\Delta X_{rms}=0.1X_{rms}$.

Figure 20

Figure 17. Visualised example of disturbance fields: (a) input and the undisturbed prediction, (b) the optimum disturbance ($\textrm {Control}(X)$), disturbance-added prediction ($\textrm {Pred}(X+\Delta X_C)$) and the difference. Comparison cases of (c) $\Delta X_{SI}$, (d) $\Delta X_{TV}$ and (e) $\Delta X_{TG}$.

Figure 21

Figure 18. (a) Comparison of MSEs between various $\tilde {Y}$ and $Y^*$ using different disturbance fields. Black dots denote the sorted MSE of phase-shifted $\Delta X_C$. The first node of black dots (red dashed line) shows the result of the optimum $\Delta X_{C}$. (b) Enstrophy spectra of the input $X$, $\Delta X_{C}$, $\Delta X_{SI}$ and $\Delta X_{TV}$.

Figure 22

Figure 19. Simulated propagations of the control effect along with the time horizon presented by differences between the disturbance-added simulations and the original simulation: (a) $\Delta X_{C}$, (b) $\Delta X_{SI}$, (c) $\Delta X_{TV}$ and (d) $\Delta X_{TG}$. The RMSE result of the propagation of each disturbance is shown in (e) after being normalised by the input r.m.s.

Figure 23

Figure 20. Distribution of the velocity vector field with the input vorticity contours for (a) $\Delta X_{C}$, (b) $\Delta X_{TG}$ and (c) $\Delta X_{TG,min}$. (d) Conditional p.d.f. of the angle between the velocity vectors of the input and disturbance.

Figure 24

Figure 21. Scale decompositions of other values of $T$. (a) 0.25TL, (b) 0.5TL and (c) 2TL.

Figure 25

Figure 22. Comparison between the velocity statistics of the target and prediction results depending on $T$: (a) $T=0.25T_L$, (b) $T=0.5 T_L$, (c) $T=T_L$ and (d) $T=2T_L$ for the same input time point of $t_0$. Panels (a i,b i,c i,d i), (a ii,b ii,c ii,d ii), and (a iii,b iii,c iii,d iii) display velocity p.d.f.s, longitudinal and transverse correlation functions and velocity gradient p.d.f.s, respectively.

Figure 26

Figure 23. Visualised examples of prediction result and enstrophy spectra for $0.5T_L$ prediction using (a) cGAN and Spect 0 (baseline CNN) and (b) spectrum loss added CNNs.

Figure 27

Figure 24. Prediction results of the phase error for $0.5T_L$ comparing cGAN, Spect 0 and spectrum loss added CNNs.

Figure 28

Figure 25. Example of visualising the effect of the disturbance strength on a pre-trained surrogate model using one test dataset with cGAN-$0.5T_L$ and $\Delta X_{rms}$=0.2$X_{rms}$: (a) disturbance-added simulation ($\mathscr {N}(X+\Delta X)$), (b) disturbance-added prediction ($\textrm {Pred}(X+\Delta X)=\tilde {Y}$) and (c) squared difference of (a) and (b).