Hostname: page-component-77f85d65b8-grvzd Total loading time: 0 Render date: 2026-03-29T01:34:25.778Z Has data issue: false hasContentIssue false

A sparsity preserving genetic algorithm for extracting diverse functional 3D designs from deep generative neural networks

Published online by Cambridge University Press:  12 May 2020

James D. Cunningham
Affiliation:
Mechanical Engineering, Carnegie Mellon University, Pittsburgh, PA, 15213, USA
Dule Shu
Affiliation:
Mechanical Engineering, Carnegie Mellon University, Pittsburgh, PA, 15213, USA
Timothy W. Simpson
Affiliation:
Mechanical Engineering, Industrial and Manufacturing Engineering, Penn State University, University Park, PA, 16802, USA
Conrad S. Tucker*
Affiliation:
Mechanical Engineering, Carnegie Mellon University, Pittsburgh, PA, 15213, USA Machine Learning, Carnegie Mellon University, Pittsburgh, PA, 15213, USA
*
Email address for correspondence: conradt@andrew.cmu.edu
Rights & Permissions [Opens in a new window]

Abstract

Generative neural networks (GNNs) have successfully used human-created designs to generate novel 3D models that combine concepts from disparate known solutions, which is an important aspect of design exploration. GNNs automatically learn a parameterization (or latent space) of a design space, as opposed to alternative methods that manually define a parameterization. However, GNNs are typically not evaluated using an explicit notion of physical performance, which is a critical capability needed for design. This work bridges this gap by proposing a method to extract a set of functional designs from the latent space of a point cloud generating GNN, without sacrificing the aforementioned aspects of a GNN that are appealing for design exploration. We introduce a sparsity preserving cost function and initialization strategy for a genetic algorithm (GA) to optimize over the latent space of a point cloud generating autoencoder GNN. We examine two test cases, an example of generating ellipsoid point clouds subject to a simple performance criterion and a more complex example of extracting 3D designs with a low coefficient of drag. Our experiments show that the modified GA results in a diverse set of functionally superior designs while maintaining similarity to human-generated designs in the training data set.

Information

Type
Research Article
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted re-use, distribution, and reproduction in any medium, provided the original work is properly cited.
Copyright
Copyright © The Author(s) 2020
Figure 0

Figure 1. Example of GNN output from Goodfellow et al. (2014). The rightmost column shows the nearest training sample to the neighboring generated sample column. These results demonstrate the two aspects of GNNs desirable for conceptual CAD, diversity of output while maintaining similarity to the training set overall to capture its embedded information.

Figure 1

Figure 2. Simplified illustration of the architecture of a deep autoencoder. In reality, there are additional hidden layers, and the hidden layers are themselves often constructed as blocks containing more complex neural network architectures such as ResNet (He et al.2016).

Figure 2

Table 1. Comparison of features of this proposed method to most closely related literature

Figure 3

Figure 3. Flow diagram of how the loss function is calculated from a latent variable $\mathbf{z}$, as detailed in Equation (2).

Figure 4

Figure 4. Performance function of the ellipsoid design space.

Figure 5

Figure 5. Distribution of training data.

Figure 6

Figure 6. (a) Sample ellipsoid point cloud from the training data set. (b) AtlasNet’s reconstruction of this ellipsoid after 120 epochs of training on the ellipsoid data set.

Figure 7

Figure 7. Sample of mesh (left) and corresponding point cloud (right) models from the watercraft data set.

Figure 8

Figure 8. Fluid simulation environment for calculating drag coefficient.

Figure 9

Figure 9. Flow diagram of the mesh reconstruction process.

Figure 10

Figure 10. View of 3D model prior to projection onto 2D plane. Fluid flow is from $z$ axis. Projection of 3D model onto 2D plane perpendicular to fluid flow.

Figure 11

Figure 11. Performance evolution for each method averaged across all trials with shaded 0.95 confidence interval.

Figure 12

Figure 12. Sparsity evolution averaged across all trials for each TDI method with shaded 0.95 confidence interval. The naïve method is omitted due to it having a sparsity ratio of 0 at all times.

Figure 13

Table 2. (H1) Average difference between average scores of initial and final populations for each method across all trials. Percentage improvement and $p$-values are calculated between the initial and final generations for each method

Figure 14

Table 3. (H2 and H3) Average mean Hausdorff distance across the five trials for different point clouds within the final set of designs $\mathbf{P}$ for each method. The TDI percentage improvement and $p$-value are calculated with respect to the naïve method, and the TDI-SP percentage improvement and $p$-values are calculated with respect to both the naïve method and the TDI method.

Figure 15

Table 4. (H4) Average mean Hausdorff distance across the five trials between point clouds in training data set $\mathbf{T}$ and point clouds of final generation of each method $\mathbf{P}$. Both percentage improvement and $p$-values are calculated with respect to the naïve method.

Figure 16

Figure 13. Distribution of the 120 ellipsoids in the final generation for each method. The top row shows the initial population for each trial, which was held constant for each method.

Figure 17

Figure 14. (Top) Three randomly sampled designs from naïve method. (Middle) Three randomly sampled designs from TDI-LSO method. (Bottom) Three randomly sampled designs from the proposed TDI-SP-LSO method ($\unicode[STIX]{x1D706}=0.2$). All samples are taken from the first trial.

Figure 18

Figure 15. Evolution of the drag ratio, $R$, for each method averaged across all trials with shaded 0.95 confidence interval.

Figure 19

Table 5. (H1) Average difference between average scores of initial and final generations for each method across the three trials. Percentage improvement and $p$-values are calculated between the initial and final generations for each method.

Figure 20

Figure 16. Sparsity evolution averaged across all trials for each TDI method with shaded 0.95 confidence interval. Naïve is omitted due to it having a sparsity ratio of 0 at all times.

Figure 21

Table 6. (H2 and H3) Average mean Hausdorff distance across the three trials for different point clouds within the final set of designs $\mathbf{P}$ for each method. The TDI percentage improvement and $p$-value are calculated with respect to the naïve method, and the TDI-SP percentage improvement and $p$-values are calculated with respect to both the naïve method and the TDI method.

Figure 22

Figure 17. (Top) Three randomly sampled designs from naïve method. (Middle) Three randomly sampled designs from TDI-LSO method. (Bottom) Three randomly sampled designs from the proposed TDI-SP-LSO method ($\unicode[STIX]{x1D706}=1\times 10^{-4}$).

Figure 23

Table 7. (H4) Average mean Hausdorff distance across the three trials between point clouds in training data set $\mathbf{T}$ and point clouds of final generation of each method $\mathbf{P}$. Both percentage improvement and $p$-values are calculated with respect to the naïve method.