Hostname: page-component-77f85d65b8-7lfxl Total loading time: 0 Render date: 2026-03-28T03:05:36.445Z Has data issue: false hasContentIssue false

Adversarial disentanglement by backpropagation with physics-informed variational autoencoder

Published online by Cambridge University Press:  13 November 2025

Ioannis Christoforos Koune*
Affiliation:
Faculty of Civil Engineering and Geosciences, Delft University of Technology, Delft, the Netherlands
Alice Cicirello
Affiliation:
Department of Engineering, University of Cambridge , Cambridge, UK
*
Corresponding author: Ioannis Christoforos Koune; Email: i.c.koune@tudelft.nl

Abstract

Inference and prediction under partial knowledge of a physical system is challenging, particularly when multiple confounding sources influence the measured response. Explicitly accounting for these influences in physics-based models is often infeasible due to epistemic uncertainty, cost, or time constraints, resulting in models that fail to accurately describe the behavior of the system. On the other hand, data-driven machine learning models such as variational autoencoders are not guaranteed to identify a parsimonious representation. As a result, they can suffer from poor generalization performance and reconstruction accuracy in the regime of limited and noisy data. We propose a physics-informed variational autoencoder architecture that combines the interpretability of physics-based models with the flexibility of data-driven models. To promote disentanglement of the known physics and confounding influences, the latent space is partitioned into physically meaningful variables that parametrize a physics-based model, and data-driven variables that capture variability in the domain and class of the physical system. The encoder is coupled with a decoder that integrates physics-based and data-driven components, and constrained by an adversarial training objective that prevents the data-driven components from overriding the known physics, ensuring that the physics-grounded latent variables remain interpretable. We demonstrate that the model is able to disentangle features of the input signal and separate the known physics from confounding influences using supervision in the form of class and domain observables. The model is evaluated on a series of synthetic case studies relevant to engineering structures, demonstrating the feasibility of the proposed approach.

Information

Type
Research Article
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Open Practices
Open materials
Copyright
© The Author(s), 2025. Published by Cambridge University Press
Figure 0

Figure 1. Illustrative examples of the problem setting: a) Beam, b) Oscillator, and c) One member of a population of bridges. The objective is to learn components of the measured response (bottom row) that are not explicitly included in the nominal physics-based model (top row) using observations of related quantities.

Figure 1

Figure 2. Demonstration of the data-driven component of the decoder $ {g}_{\boldsymbol{\theta}}\left({\mathbf{z}}_{\mathrm{c}},{\mathbf{z}}_{\mathrm{y}}\right) $ overriding the physics-based model $ f\left({\mathbf{z}}_{\mathrm{x}}\right) $. The effect of varying the position of the load $ {x}_{\mathrm{F}} $ should be described by the known physics, but is instead captured by the data-driven components.

Figure 2

Figure 3. a) Schematic diagram illustrating the components of the model and the encoder-decoder architecture, and b) Detailed structure of the dependencies in the generative and inference models.

Figure 3

Figure 4. Illustration of the procedure used to obtain the full and nominal physics-based models (left), and to generate the datasets used in the case studies (right).

Figure 4

Table 1. Summary of generative factors and the corresponding ground truth and prior distributions

Figure 5

Figure 5. Mean prediction and $ \pm 2\sigma $ uncertainty bounds for the physics-based $ {\hat{\mathbf{x}}}_{\mathrm{p}} $ and data-driven $ {\hat{\mathbf{x}}}_{\mathrm{d}} $ components, and combined prediction $ \hat{\mathbf{x}} $ while traversing the generative factors. The input response measurements are denoted as dots in the bottom row.

Figure 6

Figure 6. Visualizations of the VAE latent space during traversal of the generative factors $ {x}_{\mathrm{F}} $ and $ \log {k}_{\mathrm{v}} $. Each column corresponds to variation of a single generative factor, and each row shows the marginal approximate posterior distribution of a single latent variable.

Figure 7

Table 2. Summary of generative factors and the corresponding ground truth and prior distributions

Figure 8

Figure 7. Physics-based model prediction $ {\hat{\mathbf{x}}}_{\mathrm{p}} $, data-driven model prediction $ {\hat{\mathbf{x}}}_{\mathrm{d}} $, and combined prediction $ \hat{\mathbf{x}} $ for varying initial displacement $ {x}_0 $. With $ \lambda =-1.0 $ (top) the data-driven components in the VAE are free to account for the variability in the initial position. For $ \lambda =1/128 $ (bottom) the model does not learn this component of the response.

Figure 9

Figure 8. Physics-based model prediction $ {\hat{\mathbf{x}}}_{\mathrm{p}} $, data-driven model prediction $ {\hat{\mathbf{x}}}_{\mathrm{d}} $, and combined prediction $ \hat{\mathbf{x}} $ for varying viscous damping coefficient $ {c}_{\mathrm{d}} $. The data-driven decoder components are prevented from fully accounting for the discrepancies between the physics-based model and measurements, resulting in wider uncertainty bounds for the proposed model.

Figure 10

Figure 9. $ {R}^2 $ value per subset of the latent variables and generative factor as a function of $ \lambda $, averaged over $ 6 $ runs. The shaded intervals correspond to two standard deviations.

Figure 11

Table 3. Summary of physics-based, class and domain variables for the two-span bridge case study

Figure 12

Figure 10. Mean prediction and $ \pm 2\sigma $ uncertainty bounds for the physics-based $ {\hat{\mathbf{x}}}_{\mathrm{p}} $ and data-driven $ {\hat{\mathbf{x}}}_{\mathrm{d}} $ components, and combined prediction $ \hat{\mathbf{x}} $ while traversing the generative factors. The input response measurements are denoted as dots in the bottom row.

Figure 13

Figure 11. Visualization of the VAE latent space during traversal of the generative factors. Each column corresponds to variation of a single generative factor, and each row shows the marginal approximate posterior distribution of a single latent variable.

Figure 14

Figure 12. Samples of physics-grounded generative factors used for creating the synthetic training set (blue) and test set (orange). Two cases are constructed in order to evaluate performance in interpolation (top) and extrapolation (bottom).

Figure 15

Table 4. Mean and standard deviation of $ {R}^2 $ and $ \mathrm{MSE} $ for the task of predicting $ \mathbf{y} $, averaged over $ 6 $ runs

Submit a response

Comments

No Comments have been published for this article.