Hostname: page-component-77f85d65b8-8wtlm Total loading time: 0 Render date: 2026-03-28T06:15:58.239Z Has data issue: false hasContentIssue false

Generation of geometric interpolations of building types with deep variational autoencoders

Published online by Cambridge University Press:  28 December 2020

Jaime de Miguel Rodríguez*
Affiliation:
University of Seville, Sevilla, Spain
Maria Eugenia Villafañe
Affiliation:
Imperial College London, London, UK
Luka Piškorec
Affiliation:
Aalto University, Espoo, Finland
Fernando Sancho Caparrini
Affiliation:
University of Seville, Sevilla, Spain
*
Corresponding author J. de Miguel Rodríguez jdemiguel@us.es
Rights & Permissions [Opens in a new window]

Abstract

This work presents a methodology for the generation of novel 3D objects resembling wireframes of building types. These result from the reconstruction of interpolated locations within the learnt distribution of variational autoencoders (VAEs), a deep generative machine learning model based on neural networks. The data set used features a scheme for geometry representation based on a ‘connectivity map’ that is especially suited to express the wireframe objects that compose it. Additionally, the input samples are generated through ‘parametric augmentation’, a strategy proposed in this study that creates coherent variations among data by enabling a set of parameters to alter representative features on a given building type. In the experiments that are described in this paper, more than 150 k input samples belonging to two building types have been processed during the training of a VAE model. The main contribution of this paper has been to explore parametric augmentation for the generation of large data sets of 3D geometries, showcasing its problems and limitations in the context of neural networks and VAEs. Results show that the generation of interpolated hybrid geometries is a challenging task. Despite the difficulty of the endeavour, promising advances are presented.

Information

Type
Research Article
Creative Commons
Creative Common License - CCCreative Common License - BYCreative Common License - NCCreative Common License - SA
This is an Open Access article, distributed under the terms of the Creative Commons Attribution-NonCommercial-ShareAlike licence (http://creativecommons.org/licenses/by-nc-sa/4.0/), which permits non-commercial re-use, distribution, and reproduction in any medium, provided the same Creative Commons licence is included and the original work is properly cited. The written permission of Cambridge University Press must be obtained for commercial re-use.
Copyright
© The Author(s), 2020. Published by Cambridge University Press
Figure 0

Figure 1. Diagram of the connectivity vector scheme for a 2D geometry.

Figure 1

Figure 2. Diagram of the connectivity vector scheme for a 3D geometry.

Figure 2

Figure 3. Sequential construction of a 3D building type with the proposed connectivity vector scheme (vector detail).

Figure 3

Figure 4. Sequential construction of a 3D building type with the proposed connectivity vector scheme (geometry detail).

Figure 4

Figure 5. CCTV headquarters, its representative geometry as a snapped set of lines within the 3D-canvas and a part of its corresponding connectivity map.

Figure 5

Figure 6. Parameter scheme for Hejduk-inspired samples.

Figure 6

Figure 7. Parameter scheme for CCTV samples.

Figure 7

Figure 8. Parametric generation of samples (Hejduk-inspired castle).

Figure 8

Figure 9. Parametric generation of samples (CCTV).

Figure 9

Figure 10. Generic diagram of the network architecture of a standard VAE.

Figure 10

Figure 11. Workflow diagram of the proposed methodology.

Figure 11

Table 1. Network architecture of preliminary experiments

Figure 12

Table 2. Hyperparameters and results of preliminary experiments

Figure 13

Table 3. Training results for preliminary experiments

Figure 14

Figure 12. Best training results for both the parametric augmentation data set and the previous random displacements + noise data set.

Figure 15

Figure 13a. Latent space. Encoded samples from the parametric augmentation data set (yellow and purple dots represent each of the two training categories).

Figure 16

Figure 13b. Highlight of observed structural features in the latent space from the parametric augmentation data set.

Figure 17

Figure 14. Latent space. Encoded samples from the random displacements + noise data set.

Figure 18

Figure 15. Model complexity versus training and validation errors (overfitting problem).

Figure 19

Table 4a. C21-C7-D512 network architecture (showing only encoder for simplicity)

Figure 20

Table 4b. 2xC21-2 × C7-3 × D512 network architecture (showing only encoder for simplicity)

Figure 21

Table 4c. 3 × C21-3 × C7-4 × D512 network architecture (showing only encoder for simplicity)

Figure 22

Figure 16. Best training results of A-Conv scheme.

Figure 23

Table 5a. D2048-D512 network architecture (showing only encoder for simplicity)

Figure 24

Table 5b. 4 × D512 network architecture (showing only encoder for simplicity)

Figure 25

Table 5c. 6 × D112 network architecture (showing only encoder for simplicity)

Figure 26

Figure 17. Best training results of B-Dense scheme.

Figure 27

Table 6. Training results for A-Conv and B-Dense schemes

Figure 28

Figure 18. Latent space of 2 × C-2 × C-3 × D512 model.

Figure 29

Figure 19. Latent space of 6 × D112 model.

Figure 30

Figure 20. Geometry reconstructed from latent space. Model 2 × C-2 × C-3 × D512.

Figure 31

Figure 21. Geometry reconstructed from latent space (blow-up). Model 2 × C-2 × C-3 × D512.

Figure 32

Figure 22. Geometry reconstructed from latent space. Model 6 × D112.

Figure 33

Figure 23. Geometry reconstructed from latent space (blow-up). Model 6 × D112.

Figure 34

Table 7. Training results for 2 × C21-2 × C7-3 × 512 upon variations of the data set

Figure 35

Figure 24. Latent space of 2 × C21-2 × C7-3 × D512 model with the ‘increased data set’.

Figure 36

Figure 25. Latent space of 2 × C21-2 × C7-3 × D512 model with the ‘modified and increased data set’.

Figure 37

Figure 26. Comparison of training results of 2 × C21-2 × C7-3 × D512 model from the previous experiment with both the ‘increased data set’ and the ‘increased and modified data set’.

Figure 38

Figure 27. Geometry reconstructed from latent space. Model 2 × C-2 × C-3 × D512 using the increased and modified data set.

Figure 39

Figure 28. Geometry reconstructed from latent space (blow-up). Model 2 × C-2 × C-3 × D512 using the increased and modified data set.

Figure 40

Figure 29. Rendered geometry reconstructed from latent space. Model 2 × C-2 × C-3 × D512 using the increased and modified data set.

Figure 41

Figure 30. 3D printed geometry reconstructed from latent space. Model 2 × C-2 × C-3 × D512 using the increased and modified data set.