Hostname: page-component-89b8bd64d-mmrw7 Total loading time: 0 Render date: 2026-05-07T02:30:01.988Z Has data issue: false hasContentIssue false

Distributed digital twins for health monitoring: resource constrained aero-engine fleet management

Part of: ISABE 2024

Published online by Cambridge University Press:  15 April 2024

A. Hartwell
Affiliation:
Department of Automatic Control and Systems Engineering, University of Sheffield, Sheffield, UK
F. Montana
Affiliation:
Department of Automatic Control and Systems Engineering, University of Sheffield, Sheffield, UK
W. Jacobs
Affiliation:
Department of Automatic Control and Systems Engineering, University of Sheffield, Sheffield, UK
V. Kadirkamanathan
Affiliation:
Department of Automatic Control and Systems Engineering, University of Sheffield, Sheffield, UK
N. Ameri
Affiliation:
Rolls-Royce Plc, Bristol, UK
A. R. Mills*
Affiliation:
Department of Automatic Control and Systems Engineering, University of Sheffield, Sheffield, UK
*
Corresponding author: Andrew R. Mills; Email: a.r.mills@sheffield.ac.uk
Rights & Permissions [Opens in a new window]

Abstract

Sensed data from high-value engineering systems is being increasingly exploited to optimise their operation and maintenance. In aerospace, returning all measured data to a central repository is prohibitively expensive, often causing useful, high-value data to be discarded. The ability to detect, prioritise and return useful data on asset and in real-time is vital to move toward more sustainable maintenance activities.

We present a data-driven solution for on-line detection and prioritisation of anomalous data that is centrally processed and used to update individualised digital twins (DT) distributed onto remote machines. The DT is embodied as a convolutional neural network (CNN) optimised for real-time execution on a resource constrained gas turbine monitoring computer. The CNN generates a state prediction with uncertainty, which is used as a metric to select informative data for transfer to a remote fleet monitoring system. The received data is screened for faults before updating the weights on the CNN, which are synchronised between real and virtual asset.

Results show the successful detection of a known in-flight engine fault and the collection of data related to high novelty pre-cursor events that were previously unrecognised. We demonstrate that data related to novel operation are also identified for transfer to the fleet monitoring system, allowing model improvement by retraining. In addition to these industrial dataset results, reproducible examples are provided for a public domain NASA dataset.

The data prioritisation solution is capable of running in real-time on production-standard low-power embedded hardware and is deployed on the Rolls-Royce Pearl 15 engines.

Information

Type
Research Article
Creative Commons
Creative Common License - CCCreative Common License - BYCreative Common License - NCCreative Common License - ND
This is an Open Access article, distributed under the terms of the Creative Commons Attribution-NonCommercial-NoDerivatives licence (http://creativecommons.org/licenses/by-nc-nd/4.0/), which permits non-commercial re-use, distribution, and reproduction in any medium, provided that no alterations are made and the original article is properly cited. The written permission of Cambridge University Press must be obtained prior to any commercial use and/or adaptation of the article.
Copyright
© The Author(s), 2024. Published by Cambridge University Press on behalf of Royal Aeronautical Society
Figure 0

Figure 1. On-board architecture. Data from the gas turbine is streamed into a temporary buffer before being run through the currently loaded models. Data with a higher novelty or lower confidence score replaces lower ranked items in the data store.

Figure 1

Figure 2. Overview of the network architecture. Number of hidden units and stride size given in brackets, output size given in yellow. The design is flexible such that ‘T’ matches the number of incoming timesteps in each window and ‘S’ is the number of signals. The ‘Dense Blocks’ are three dense layers each followed by a dropout layer with rate = 0.5. The number of hidden units (N_X) in different layers may be adjusted based on the complexity of the input-output set and available computation.

Figure 2

Figure 3. Solution architecture. Data is continuously assessed for novelty by an on-board twin synchronised with a ground-based digital twin. Novel data is returned and triaged before re-training the on-ground twin.

Figure 3

Figure 4. (a) Distribution of the novelty score $\tilde d$ based on position within a test run. Larger means, variances and maximum values are observed towards the end of runs where faults occur. (b) Maximum values of the novelty score $\tilde d$ across all model outputs for a single engine run. The operating condition is indicated by colour. (c) Prediction error, y_n-μ_n, for Sensor 13 over a single engine run. (d) Standard deviation, σ_n, for model predictions on Sensor 13.

Figure 4

Table 1. Model prediction performance.

Figure 5

Figure 5. Normalised signals for the flight with the confirmed fault. The highlighted timespan in the top graph is enlarged on the bottom graph. The spike in the internal pressure signal (P30) in the bottom graph is a result of a known fault as is all behaviour afterwards. The temperature T30 is an internal temperature detected in the same locale as P30, N1 is the low-speed shaft and TGT is the turbine gas temperature.

Figure 6

Figure 6. Top four graphs: prediction around the fault, in descending order of the novelty score $\tilde d$. Highlighted regions indicate a 95% confidence interval (predicted). Bottom graph: $\tilde d$ for each model around the fault.

Figure 7

Figure 7. Potential precursor fault, three flights before known fault.

Figure 8

Figure 8. Distribution of $\tilde d$ across test flights. The known fault has a large amount of mass at a much higher novelty scores than other flights, and potential precursors show smaller mass at similar novelty scores ($\tilde d \gt 100$). Flight 111 also exhibits high novelty score density due to irregularity in the shaft speed signal not found in other flights. Red lines represent a threshold for data collection, set by communications capacity. Anomalous data with high novelty score is prioritised.

Figure 9

Figure 9. P30 prediction confidence for different inputs. The test data is coloured based on confidence. Large values indicate a wide confidence interval. Black points show the distribution of the training data.

Figure 10

Figure 10. 95th quantile of run durations for the NASA turbofan degradation data. Smaller peak likely caused by cache hits.

Figure 11

Figure 11. 95th quantile of run durations for the fault case study data. Smaller number of models and overall smaller model size relative to NASA data case study lead to the lower average run duration and likely the reversal of shape due to more cache hits.

Figure 12

Figure 12. Neural network architecture with seven input signals and a 10 second window.