Hostname: page-component-89b8bd64d-mmrw7 Total loading time: 0 Render date: 2026-05-06T11:59:20.208Z Has data issue: false hasContentIssue false

Robust design using multiobjective optimisation and artificial neural networks with application to a heat pump radial compressor

Published online by Cambridge University Press:  06 January 2022

Soheyl Massoudi*
Affiliation:
Ecole Polytechnique Fédérale de Lausanne (EPFL), Laboratory for Applied Mechanical Design, CH-1015 Lausanne, Switzerland
Cyril Picard
Affiliation:
Ecole Polytechnique Fédérale de Lausanne (EPFL), Laboratory for Applied Mechanical Design, CH-1015 Lausanne, Switzerland
Jürg Schiffmann
Affiliation:
Ecole Polytechnique Fédérale de Lausanne (EPFL), Laboratory for Applied Mechanical Design, CH-1015 Lausanne, Switzerland
*
Corresponding author S. Massoudi soheyl.massoudi@epfl.ch
Rights & Permissions [Opens in a new window]

Abstract

Although robustness is an important consideration to guarantee the performance of designs under deviation, systems are often engineered by evaluating their performance exclusively at nominal conditions. Robustness is sometimes evaluated a posteriori through a sensitivity analysis, which does not guarantee optimality in terms of robustness. This article introduces an automated design framework based on multiobjective optimisation to evaluate robustness as an additional competing objective. Robustness is computed as a sampled hypervolume of imposed geometrical and operational deviations from the nominal point. In order to address the high number of additional evaluations needed to compute robustness, artificial neutral networks are used to generate fast and accurate surrogates of high-fidelity models. The identification of their hyperparameters is formulated as an optimisation problem. In the frame of a case study, the developed methodology was applied to the design of a small-scale turbocompressor. Robustness was included as an objective to be maximised alongside nominal efficiency and mass-flow range between surge and choke. An experimentally validated 1D radial turbocompressor meanline model was used to generate the training data. The optimisation results suggest a clear competition between efficiency, range and robustness, while the use of neural networks led to a speed-up by four orders of magnitude compared to the 1D code.

Information

Type
Research Article
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2022. Published by Cambridge University Press
Figure 0

Figure 1. Graphical representation of robustness evaluated as a sensitivity analysis from the nominal point, where the value of the metric of interest is computed at each sampled point. Here, the sensitivity analysis is done with respect to three variables $ {g}_1 $, $ {g}_2 $ and $ {g}_3 $. The hypervolume therefore degenerates to a volume, and its projected faces are represented with dashed lines. The red dot in the middle of the hypervolume represents the nominal point, whereas the grey scale of the dot represents the evolution of the considered performance metric.

Figure 1

Figure 2. Robustness evaluation of function $ f $ leads to a continuous domain defined by $ {S}_F $. Outside of $ {S}_F $, the constraints of the problem are no longer respected and the value of $ f $ for these samples is set to zero.

Figure 2

Figure 3. Visualisation of a feed-forward artificial neural network (FWANN) made of several perceptrons and layers. The artificial neuron, or perceptron, takes entries from the previous layers, multiplies them with weights, sums them, inputs the sum into an activation function and returns the result to the perceptrons of the next layer. Perceptrons assembled into hidden layers form the bulk of the FWANN.

Figure 3

Figure 4. Nested optimisation of a feed-forward artificial neural network. Backpropagation is used in gradient descent to find the best weights and biases minimising the loss function. The hyperparamaters of the network, such as learning rate, number of perceptrons, hidden layers, and so forth are found via genetic algorithm optimisation (hyperparameter tuning).

Figure 4

Table 1. Description of the inputs of the centrifugal compressor 1D model indicating the ranges considered for the generation of the dataset of the surrogate model

Figure 5

Figure 5. A cut view of the compressor. The shroud-tip radius ratio $ \overline{r_{2s}}={r}_{2s}/{r}_4 $ and blade height ratio $ \overline{b_4}={b}_4/{r}_4 $, as well as the tip radius $ {r}_4 $ and the rotational speed $ \Omega $ are independent variables. All other variables are either dependent of the aforementioned or fixed.

Figure 6

Table 2. Set of dimensionless variables used to train the FWANNs to supplant the 1D compressor code

Figure 7

Figure 6. Architecture of the compressor surrogate model. In the compressor evaluation process, design variables $ X $ are made dimensionless and standardised. The ANN classifier discriminates functioning designs from the rest and their isentropic efficiency $ \eta $ and pressure ratio $ \Pi $ is evaluated. These two outputs are inverse standardised. In the surrogate model architecture, this allows for the computation of nominal performance as well as a sweep of values for the computation of robustness $ \mathrm{HV} $ and mass-flow modulation $ \Delta {\dot{m}}_{\Pi =2} $. Internal constraints are applied to define the feasible regions in each case. The designs and their respective performance are returned.

Figure 8

Table 3. Description of the hyperparameters searched for the optimisation of the classifier FWANN

Figure 9

Table 4. Description of the hyperparameters searched for the optimisation of the predictor FWANNs for pressure ratio and isentropic efficiency

Figure 10

Table 5. Description of the parameters for the surrogate compressor model MOORD

Figure 11

Figure 7. Visualisation of the compressor optimisation. The surrogate model is used to compute the nominal isentropic efficiency $ {\eta}_{\mathrm{is}} $, robustness $ \mathrm{HV} $ and mass-flow modulation $ \Delta {\dot{m}}_{\Pi =2} $. It is called only once per generation of the evolutionary algorithm, as a matrix of candidate designs $ {G}_{imp} $ – alongside operating conditions OP, fluid Fld and deviations $ \Delta {g}_i $ – is inputted and expanded within the model.

Figure 12

Table 6. Hyperparameters of the ANNs

Figure 13

Figure 8. Log–log graph of computation time versus number of design evaluations for the 1D code executed on CPU, the ANNs executed on CPU (subscript C) and the ANNs executed on GPU (subscript G). The computation time is averaged over 10 runs for the ANNs, while single evaluations are summed for the 1D code.

Figure 14

Figure 9. Pareto front for $ 1\times {10}^{-3} $ deviation on $ {r}_{2s} $ and $ {b}_4 $. Robustness $ \mathrm{HV} $ is represented on the x-axis, efficiency-weighted mass-flow modulation $ \Delta {\dot{m}}_{\Pi =2} $ on the y-axis, and the isentropic efficiency $ {\eta}_{\mathrm{is}} $ is represented by a colour gradient. A clear competition between the three objectives is suggested. Aiming at a robust design with high efficiency, the candidate is selected on the Pareto front and represented by a square.

Figure 15

Figure 10. Relative error, on isentropic efficiency and pressure ratio respectively, for the Pareto optima between the ANN and the 1D code. Pressure ratio is predicted with a better accuracy than isentropic efficiency. The largest relative error observed for efficiency is of −1.2 and 0.4%. The ANN are accurate and conservative with respect to the 1D code.

Figure 16

Figure 11. Compressor map of the selected compressor design. Typical of a high efficiency design, the compressor characteristic is centred on the operating point of $ \dot{m}=0.024 $ kg s−1 and $ \Pi =2 $. The space on the left is delimited by the surge line, while the one on the right is marked by the choke line. On the edges of the compressor map, the efficiency can drop below 0.71.

Figure 17

Table 7. Design point and objectives of the selected design

Figure 18

Figure 12. The evaluation of robustness for the selected solution as a hypervolume of deviations from the nominal point is limited to a volume in this case study. Isentropic efficiency at the sampled points is represented by a colour gradient. A slice at the nominal rotational speed is represented on the right. On the edges of the feasible domain, the value of efficiency can drop below 0.71.

Figure 19

Figure 13. Evolution of each of the four decision variables against the three objectives. Decision variables of the Pareto optima are graphed in pair-plots and coloured by objective value. The left column represents the robustness $ \mathrm{HV} $, while the middle one represents the efficiency-weighted mass-flow range $ \Delta {\dot{m}}_{\Pi =2} $ and the right one the isentropic efficiency $ {\eta}_{\mathrm{is}} $. Specific speed $ Ns $ and specific diameter $ Ds $ instead of $ N $ and $ D $ are picked to favour comparison.

Figure 20

Figure A1. The distribution of the data for the multiclass classifier is represented with normalised boxplots. A total of 1,841,648 designs are errors, 2,564,775 are functioning, 330,741 are in surge and 3,554,436 are in choke.

Figure 21

Figure A2. The distribution of the data is represented with normalised boxplots, after selecting only functioning designs. It is used to train the ANNs predicting efficiency $ \eta $ and pressure ratio $ \Pi $. Their distributions have also been added.

Figure 22

Figure A3. Confusion matrix for the FWANN classifying the states of the compressor for the test set: numerical error (0), functioning (1), surge (2) and choke (3). Given in thousands and rounded.

Figure 23

Figure A4. Evolution of RMSE in function of number of epochs for the FWANN predicting efficiency. Overfitting is avoided by interrupting the network training when the validation error starts increasing again. Training and validation converge towards a 4% asymptote after 500 epochs.

Figure 24

Figure A5. Evolution of RMSE in function of number of epochs for the FWANN predicting pressure ratio. Overfitting is avoided by interrupting the network training when the validation error starts increasing again. Training and validation converge towards a 1% asymptote after 500 epochs.