Hostname: page-component-89b8bd64d-x2lbr Total loading time: 0 Render date: 2026-05-08T22:45:09.445Z Has data issue: false hasContentIssue false

A deep operator network for Bayesian parameter identification of self-oscillators

Published online by Cambridge University Press:  25 November 2024

Tobias Sugandi
Affiliation:
CAPS Laboratory, Department of Mechanical and Process Engineering, ETH Zürich, Zürich, Switzerland
Bayu Dharmaputra*
Affiliation:
CAPS Laboratory, Department of Mechanical and Process Engineering, ETH Zürich, Zürich, Switzerland
Nicolas Noiray*
Affiliation:
CAPS Laboratory, Department of Mechanical and Process Engineering, ETH Zürich, Zürich, Switzerland
*
Corresponding authors: Bayu Dharmaputra and Nicolas Noiray; Emails: bayud@ethz.ch; noirayn@ethz.ch
Corresponding authors: Bayu Dharmaputra and Nicolas Noiray; Emails: bayud@ethz.ch; noirayn@ethz.ch

Abstract

Many physical systems exhibit limit-cycle oscillations that can typically be modeled as stochastically driven self-oscillators. In this work, we focus on a self-oscillator model where the nonlinearity is on the damping term. In various applications, it is crucial to determine the nonlinear damping term and the noise intensity of the driving force. This article presents a novel approach that employs a deep operator network (DeepONet) for parameter identification of self-oscillators. We build our work upon a system identification methodology based on the adjoint Fokker–Planck formulation, which is robust to the finite sampling interval effects. We employ DeepONet as a surrogate model for the operator that maps the first Kramers–Moyal (KM) coefficient to the first and second finite-time KM coefficients. The proposed approach can directly predict the finite-time KM coefficients, eliminating the intermediate computation of the solution field of the adjoint Fokker–Planck equation. Additionally, the differentiability of the neural network readily facilitates the use of gradient-based optimizers, further accelerating the identification process. The numerical experiments demonstrate that the proposed methodology can recover desired parameters with a significant reduction in time while maintaining an accuracy comparable to that of the classical finite-difference approach. The low computational time of the forward path enables Bayesian inference of the parameters. Metropolis-adjusted Langevin algorithm is employed to obtain the posterior distribution of the parameters. The proposed method is validated against numerical simulations and experimental data obtained from a linearly unstable turbulent combustor.

Information

Type
Research Article
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2024. Published by Cambridge University Press
Figure 0

Figure 1. A realization of the stochastic Van der Pol oscillator with parameters $ \left(\nu = 2.5,\kappa = 1.6,{D}^{(2)}= 2.5\right) $ that models the local acoustic pressure fluctuation $ \eta (t) $ in a combustion chamber (red), and the corresponding envelope $ A(t) $ (black). The probability distributions for both the oscillation ($ \eta (t) $ taking value $ \eta $) and the envelope ($ A(t) $ taking value $ a $) are displayed on the left and right sides of the time signal, respectively, illustrating the statistical properties of the entire time signal.

Figure 1

Figure 2. Illustration of the calculation of the first (left) and second (right) finite-time KM coefficients from the solution of the AFP equation for $ a= 5.5,\nu = 6.5,\kappa = 1.6,{D}^{(2)}= 2.5 $. The finite-time KM coefficients are obtained by extracting the solution $ {P}_{n,a}^{\dagger}\left({a}^{\prime },\tau \right) $ at $ {a}^{\prime }=a $ for several values of $ \tau $.

Figure 2

Figure 3. A DeepONet architecture designed for the parameter identification based on the AFP formalism. It approximates the operator that maps the first Kramers–Moyal (KM) coefficient $ {D}^{(1)}(a) $ to both the first $ {D}_{\tau}^{(1)}\left(a,\tau \right) $ and second $ {D}_{\tau}^{(2)}\left(a,\tau \right) $ finite-time KM coefficients. Although only the first KM coefficient $ {D}^{(1)}(a) $ appears as the input to the branch network, the second KM coefficient $ {D}^{(2)} $ implicitly influences the forward calculation process through $ {D}^{(1)}(a) $ according to Eq. (2.15).

Figure 3

Figure 4. This set of diagrams visualizes the method for extracting finite-time KM coefficients by processing an acoustic pressure time series. (a) The envelope $ A(t) $ and the time-shifted envelope $ A\left(t+\tau \right) $. (b) The stationary PDF of the envelope $ P\left(A=a\right) $ obtained by histogram binning. (c) The joint PDF of the envelope and the time-shifted envelope also obtained by histogram binning. (d) The conditional (transition) PDF $ P\left({a}^{\prime },t+\tau |a,t\right) $ obtained by normalizing the joint PDF with $ P(A) $. (e) Transition moments calculated by numerically integrating $ {\left({a}^{\prime }-a\right)}^nP\left({a}^{\prime },t+\tau |a,t\right) $. Finite-time KM coefficients can then be obtained by normalizing the moments with $ n!\tau $.

Figure 4

Figure 5. Visualization on the parameter identification process by matching the first (left) and second (right) finite-time KM coefficients at several values of $ a $ and $ \tau $, simultaneously. Estimated finite-time KM coefficients from the time signal are depicted as colored dots, with darker shades indicating higher statistical significance. The DeepONet initial guess is presented as the light wireframe, while the final prediction is presented as the dark wireframe. The thick black curves at $ \tau =0 $ depict the true KM coefficients. The true parameters are $ \nu = 6.5,\kappa = 1.6,{D}^{(2)}= 2.5 $.

Figure 5

Figure 6. Violin plots representing the relative errors of the identified parameters $ \nu, \kappa $, and $ {D}^{(2)} $(a) across three different approaches (100-Hz bandwidth) and (b) across three different bandwidths (DeepONet—LBFGS). The inverse problem is solved for synthetic signals with 100 different combinations of $ \left\{\nu, \kappa, {D}^{(2)}\right\} $. The white dots indicate the median of the relative errors, while the encompassing thick black lines denote the interquartile range, extending from the first to the third quartile. The shaded regions of the violins represent the overall distribution of the error.

Figure 6

Figure 7. The time required to solve the inverse problem (left) and to perform the forward pass (right) for 100 synthetic test data with different values of $ \left\{\nu, \kappa, {D}^{(2)}\right\} $.

Figure 7

Table 1. Comparison of the computational efficiency of three different approaches in solving forward and inverse problems

Figure 8

Figure 8. Evolution of the parameters $ \nu $, $ \kappa $, and $ {D}^{(2)} $ from the initial guess to the final prediction for three different parameter identification approaches. DeepONet can leverage automatic differentiation to use LBFGS to significantly reduce the number of iterations required to achieve convergence.

Figure 9

Figure 9. The relative error for out-of-distribution test data, where the parameters $ \nu $, $ \kappa $, and $ {D}^{(2)} $ extend beyond the range covered by the training data. Specifically, $ \nu $ and $ {D}^{(2)} $ vary between 22.5 and 34.5, and $ \kappa $ ranges from 5.6 to 8.6. This contrasts with the training dataset, where the maximum values for $ \nu $, $ \kappa $, and $ {D}^{(2)} $ are capped at 20, 5.25, and 20, respectively.

Figure 10

Table 2. The identified parameters $ \left\{\nu, \kappa, {D}^{(2)}\right\} $ corresponding to the experimental measurements in Noiray and Denisov (2017)

Figure 11

Figure 10. Probability density function of the acoustic pressure envelope of the experimental time signal data corresponding to operating conditions c2 (left) and c3 (right). Parameters $ \nu, \kappa, {D}^{(2)} $ are extracted using both DeepONet and finite-difference approaches and then used to plot the analytical stationary PDFs.

Figure 12

Figure 11. Posterior distribution of each parameter $ \nu, \kappa, {D}^{(2)} $ corresponding to the operating condition c2 (marginally stable).

Figure 13

Figure 12. Joint distributions over each two of the three parameters $ \nu, \kappa, {D}^{(2)} $ for the operating condition c2 (marginally stable).

Figure 14

Figure 13. Posterior distribution of each parameter $ \nu, \kappa, {D}^{(2)} $ corresponding to the operating condition c3 (linearly unstable).

Figure 15

Figure 14. Joint distributions over each two of the three parameters $ \nu, \kappa, {D}^{(2)} $ for the operating condition c3 (linearly unstable).

Figure 16

Figure 15. Trace and autocorrelation plots of the MCMC algorithm for the operating condition c2 demonstrating good mixing.

Figure 17

Figure 16. Trace and autocorrelation plots of the MCMC algorithm for the operating condition c3 demonstrating good mixing.

Figure 18

Figure A1. Median time to solve the inverse problem using the finite-difference method across 100 test datasets. (Left) The spatial resolution is set to remain constant at $ da= 0.05 $, with variations in the time step achieved by modifying the multiplier $ k $, defined as $ dt=k\times \frac{1}{2}\frac{D^{(2)}}{da^2} $. Note that the y-axis is broken into parts to enhance the visibility of the optimal time-step parameter. (Right) The time step multiplier is set at $ k= 300 $, while the spatial resolution $ da $ is varied.

Submit a response

Comments

No Comments have been published for this article.