Hostname: page-component-77f85d65b8-8wtlm Total loading time: 0 Render date: 2026-03-28T08:59:35.449Z Has data issue: false hasContentIssue false

Meta-models for transfer learning in source localization

Published online by Cambridge University Press:  27 December 2024

Lawrence A. Bull*
Affiliation:
School of Mathematics and Statistics, University of Glasgow, Glasgow, G12 8TA, UK Department of Engineering, University of Cambridge, Cambridge, CB3 0FA, UK
Matthew R. Jones
Affiliation:
Department of Engineering, University of Sheffield, Sheffield, S1 3JD, UK
Elizabeth J. Cross
Affiliation:
Department of Engineering, University of Sheffield, Sheffield, S1 3JD, UK
Andrew Duncan
Affiliation:
Department of Mathematics, Imperial College London, London, SW7 2AZ, UK The Alan Turing Institute, The British Library, London, NW1 2DB, UK
Mark Girolami
Affiliation:
Department of Engineering, University of Cambridge, Cambridge, CB3 0FA, UK The Alan Turing Institute, The British Library, London, NW1 2DB, UK
*
Corresponding author: Lawrence A. Bull; Email: lawrence.bull@glasgow.ac.uk

Abstract

In practice, nondestructive testing (NDT) procedures tend to consider experiments (and their respective models) as distinct, conducted in isolation, and associated with independent data. In contrast, this work looks to capture the interdependencies between acoustic emission (AE) experiments (as meta-models) and then use the resulting functions to predict the model hyperparameters for previously unobserved systems. We utilize a Bayesian multilevel approach (similar to deep Gaussian Processes) where a higher-level meta-model captures the inter-task relationships. Our key contribution is how knowledge of the experimental campaign can be encoded between tasks as well as within tasks. We present an example of AE time-of-arrival mapping for source localization, to illustrate how multilevel models naturally lend themselves to representing aggregate systems in engineering. We constrain the meta-model based on domain knowledge, then use the inter-task functions for transfer learning, predicting hyperparameters for models of previously unobserved experiments (for a specific design).

Information

Type
Research Article
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2024. Published by Cambridge University Press
Figure 0

Figure 1. A visual example of multilevel models for multitask learning: predictive tasks $ {f}_k $ are parametrised by $ {\theta}_k=\left\{{\theta}_k^{\prime },{\theta}_k^{{\prime\prime}}\right\} $ or $ {\theta}_k=\left\{{\theta}_k^{\prime },{\theta}_k^{{\prime\prime\prime}}\right\} $; in turn, $ {\theta}_k $ is generated by shared intertask functions $ \left\{{g}^{\prime },{g}^{{\prime\prime}}\right\} $ or $ \left\{{g}^{\prime },{g}^{{\prime\prime\prime}}\right\} $. The notation $ g\to \theta $ shows a function g that predicts a parameter $ \theta $.

Figure 1

Figure 2. Image of plate used in the AE experiments.

Figure 2

Figure 3. Left: all source locations (blue $ \bullet $) and sensor locations (black $ \bullet $). Right: heatmap of the $ \Delta $ToA (response $ {y}_i $) given locations (inputs $ {\mathbf{x}}_i=\left\{{x}_i^{(1)},{x}_i^{(2)}\right\} $) for the experiment $ k=15 $, sensor pair (3, 5).

Figure 3

Figure 4. Heatmaps of the measured $ \Delta $ToA ($ {y}_i $) with respect to locations ($ {\mathbf{x}}_i $) for all sensor combinations. Large black markers plot sensor pairs, while small black markers plot training observations. The remaining data are used to test out-of-sample performance. The number in the bottom right is the experiment index $ k\in \left\{1,2,\dots \mathrm{28}\right\} $.

Figure 4

Figure 5. The inferred sensor pair (3,5) for the experiment k =15 (left) and a length-wise slice to visualize the heteroscedastic noise (right).

Figure 5

Figure 6. Posterior distribution of hyperparameters for independent GPs of each experiment (STL, green), compared to sharing hyperparameters (the prior) between all experiments (MTL-A, orange). Top row: $ f $-process hyperparameters (AE map). Bottom row: $ r $-process hyperparameters (heteroscedastic noise).

Figure 6

Figure 7. Ordering the hyperparameters (conditional posterior distribution) from the STL experiments $ {\theta}_k $ with respect to sensor separation $ \delta {S}_k $. Black markers correspond to models of the hold-out tests $ k\in \left\{\mathrm{4,7,16,22}\right\} $ which do not contribute training data for MTL.

Figure 7

Table 1. Comparison of methods based on the hierarchy of latent variables

Figure 8

Figure 8. Posterior distributions $ p\left({\boldsymbol{\theta}}_k|\mathbf{y}\right) $ following each method of learning from the series of AE experiments. Independent learners (STL, green); Shared GP prior (MTL-A, orange); Hyperparameter modeling (MTL-B, purple). Black markers correspond to STL models of the hold-out tests $ k\in \left\{\mathrm{4,7,16,22}\right\} $ which do not contribute training data for MTL.

Figure 9

Figure 9. Predictive log-likelihood for the out-of-sample test data. Excluding hold-out experiments $ k\in \left\{\mathrm{4,7,16,22}\right\} $ which are modeled using transfer learning in s:transfer. Lines plot the average across all tasks, the combined log-likelihood (summed) for each method is: (STL, 84217), (MTL A, 85149), and (MTL B, 85204).

Figure 10

Figure 10. Predictive log-likelihood for an increasing training budget for the hold-out experiments $ k\in \left\{\mathrm{4,7,16,22}\right\} $. Conventional single-task learning (STL) compared to multitask learning (MTL) which predicts hyperparameter values from similar experiments. Averaged for 100 repeats.

Figure 11

Table 2. Experiment indices $ k $, sensor pairs, and their separation $ \delta S $ (Euclidean distance in normalized space)

Submit a response

Comments

No Comments have been published for this article.