Hostname: page-component-6766d58669-kn6lq Total loading time: 0 Render date: 2026-05-16T01:47:26.217Z Has data issue: false hasContentIssue false

Reinforcement-learning-based control of turbulent channel flows at high Reynolds numbers

Published online by Cambridge University Press:  07 March 2025

Zisong Zhou
Affiliation:
Max Planck Institute for Solar System Research, Göttingen 37077, Germany
Mengqi Zhang
Affiliation:
Department of Mechanical Engineering, National University of Singapore, 9 Engineering Drive 1, Singapore 117575, Republic of Singapore
Xiaojue Zhu*
Affiliation:
Max Planck Institute for Solar System Research, Göttingen 37077, Germany
*
Corresponding author: Xiaojue Zhu, zhux@mps.mpg.de

Abstract

Deep reinforcement learning (DRL) is employed to develop control strategies for drag reduction in direct numerical simulations of turbulent channel flows at high Reynolds numbers. The DRL agent uses near-wall streamwise velocity fluctuations as input to modulate wall blowing and suction velocities. These DRL-based strategies achieve significant drag reduction, with maximum rates $35.6\,\%$ at $Re_{\tau }\thickapprox 180$, $30.4\,\%$ at $Re_{\tau }\thickapprox 550$, and $27.7\,\%$ at $Re_{\tau }\thickapprox 1000$, outperforming traditional opposition control methods. An expanded range of wall actions further enhances drag reduction, although effectiveness decreases at higher Reynolds numbers. The DRL models elevate the virtual wall through blowing and suction, aiding in drag reduction. However, at higher Reynolds numbers, the amplitude modulation of large-scale structures significantly increases the residual Reynolds stress on the virtual wall, diminishing the drag reduction. Analysis of budget equations provides a systematic understanding of the underlying drag reduction dynamics. The DRL models reduce skin friction by inhibiting the redistribution of wall-normal turbulent kinetic energy. This further suppresses the wall-normal velocity fluctuations, reducing the production of Reynolds stress, thereby decreasing skin friction. This study showcases the successful application of DRL in turbulence control at high Reynolds numbers, and elucidates the nonlinear control mechanisms underlying the observed drag reduction.

Information

Type
JFM Papers
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2025. Published by Cambridge University Press
Figure 0

Table 1. Computational parameters. Here, $\Delta _x$, $\Delta _y$ and $\Delta _z$ are the resolutions in the streamwise, wall-normal and spanwise directions, respectively.

Figure 1

Figure 1. The flow chart of reinforcement-learning-driven control in turbulent channel flows.

Figure 2

Table 2. Parameters of state steps and episodes. Here, $\Delta t$ and $\Delta T$ are the time lengths of each state step and episode, respectively, while $N_{st}$ is the number of state steps in one episode.

Figure 3

Table 3. Comparison of computational details in DRL-based turbulent channel control studies. Here, DDPG denotes the deep deterministic policy gradient algorithm. In all the studies, the output actions selected are the wall blowing and suction velocities $v_{w}^{\prime }$.

Figure 4

Table 4. The DNS cases and drag reduction results. Here, $DR$ represents the drag reduction rate, $P_{S}/P_{I}$ denotes the power saving ratio (where $P_{S}$ is the power saving, and $P_{I}$ is the power input), $\Delta U_{s}^{+}$ denotes the shift of the mean velocity profile in the logarithmic region, $y_{vw}$ indicates the height of the virtual wall, and $-\langle u^{\prime }v^{\prime }\rangle _{vw}$ is the averaged residual Reynolds stress on the virtual wall.

Figure 5

Figure 2. The evolution of the normalized reward over episodes during the training process: (a) C180, (b) C550, (c) C1000. denotes cases with suffix 1; denotes cases with suffix 2; denotes cases with suffix 3.

Figure 6

Figure 3. Mean velocity profile under different control strategies: (a) C180, (b) C550, (c) C1000. denotes cases with suffix 0; denotes cases with suffix 1; denotes cases with suffix 2; denotes cases with suffix 3.

Figure 7

Figure 4. Wall-normal distributions of the velocity fluctuations under different control strategies. Solid, dashed and dash-dotted lines represent $u_{rms}^{\prime }$, $v_{rms}^{\prime }$ and $w_{rms}^{\prime }$, respectively: (a) C180, (b) C550, (c) C1000. denotes cases with suffix 0; denotes cases with suffix 1; denotes cases with suffix 2; denotes cases with suffix 3.

Figure 8

Figure 5. Wall-normal distributions of the averaged Reynolds shear stress under different control strategies: (a) C180, (b) C550, (c) C1000. denotes cases with suffix 0; denotes cases with suffix 1; denotes cases with suffix 2; denotes cases with suffix 3.

Figure 9

Figure 6. Joint p.d.f. of the wall blowing and suction $v_{w}^{\prime }$ with (a) $v^{\prime }$ and (b) $u^{\prime }$ at $y^+=15$ in case C1000-3. The white diagonals denote (a) $v^{\prime }=-v_{w}^{\prime }$ and (b) $u^{\prime }=v_{w}^{\prime }$, respectively. Contour levels are $0.1 (0.1) 0.8$ of the maximum probability density.

Figure 10

Figure 7. Premultiplied spanwise energy spectra $k_{z}E_{uu}$ of streamwise velocity fluctuations $u^{\prime }$ under different control strategies. For $Re_{\tau }^{0}\thickapprox 180$, (a) C180-0, (b) C180-opp, (c) C180-3. For $Re_{\tau }^{0}\thickapprox 550$, (d) C550-0, (e) C550-opp, (f) C550-3. For $Re_{\tau }^{0}\thickapprox 1000$, (g) C1000-0, (h) C1000-opp, (i) C1000-3.

Figure 11

Figure 8. Instantaneous distributions of $u^{\prime }$ on the $(x,z)$ plane at (a,c,e,g) $y^{+}=y_{vw}^{+}$ and (b,d,f,h) $y^{+}=150$, for cases (a,b) C1000-opp, (c,d) C1000-1, (e,f) C1000-2, (g,h) C1000-3. The black rectangles represent some sample areas on the virtual wall where velocity fluctuations are stronger.

Figure 12

Figure 9. Joint p.d.f. of the streamwise velocity fluctuations $u^{\prime }_{O}(\Delta x_{m})$ at the centre of the logarithmic region and the envelope of Reynolds stress $|\mathcal {H}(\langle u^{\prime }v^{\prime }\rangle _{vw})|$ at the virtual wall, for cases (a) C1000-1, (b) C1000-2, (c) C1000-3. Contour levels are $0.1 (0.1) 0.8$ of the maximum probability density.

Figure 13

Figure 10. Wall-normal distributions of the budget terms of Reynolds shear stress $\langle -u^{\prime }v^{\prime }\rangle$ in (3.3): $P_{12}$, $VP_{12}$, $\varepsilon _{12}$, $D_{12,t}$, $D_{12,\nu }$, for (a,b) C180, (c,d) C550, (e,f) C1000. Lines without markers indicate cases with suffix 0; plus signs indicate cases with suffix 1; circles indicate cases with suffix 2; triangles indicate cases with suffix 3.

Figure 14

Figure 11. Wall-normal distributions of the budget terms of wall-normal kinetic energy $\langle v^{\prime }v^{\prime }\rangle$ in (3.4): $\Phi _{22}$, $\varepsilon _{22}$, $D_{22,p}$, $D_{22,t}$, $D_{22,\nu }$, for (a,b) C180, (c,d) C550, (e,f) C1000. Lines without markers indicate cases with suffix 0; plus signs indicate cases with suffix 1; circles indicate cases with suffix 2; triangles indicate cases with suffix 3.

Figure 15

Figure 12. Instantaneous distributions of $u^{\prime }$ on the $(x,z)$ plane at $y^{+}=20$, for cases (a) C180-0, (b) C180-3, (c) C1000-0, (d) C1000-3.

Figure 16

Figure 13. Wall-normal distributions of the budget terms of streamwise kinetic energy $\langle u^{\prime }u^{\prime }\rangle$ in (3.5): $P_{11}$, $\Phi _{11}$, $\varepsilon _{11}$, $D_{11,t}$, $D_{11,\nu }$, for (a,b) C180, (c,d) C550, (e,f) C1000. Lines without markers indicate cases with suffix 0; plus signs indicate cases with suffix 1; circles indicate cases with suffix 2; triangles indicate cases with suffix 3.

Figure 17

Figure 14. Schematic diagram of the dynamic mechanism through which DRL-based control strategies influence skin friction.

Figure 18

Table 5. Drag reduction results under different input states and rewards.