Hostname: page-component-89b8bd64d-4ws75 Total loading time: 0 Render date: 2026-05-08T01:21:53.989Z Has data issue: false hasContentIssue false

Closing the domain gap: blended synthetic imagery for climate object detection

Published online by Cambridge University Press:  28 November 2023

Caleb Kornfein
Affiliation:
Department of Computer Science, Duke University, Durham, NC, USA
Frank Willard
Affiliation:
Department of Computer Science, Duke University, Durham, NC, USA
Caroline Tang
Affiliation:
Department of Computer Science, Duke University, Durham, NC, USA
Yuxi Long
Affiliation:
Department of Computer Science, Duke University, Durham, NC, USA
Saksham Jain
Affiliation:
Department of Electrical & Computer Engineering, Duke University, Durham, NC, USA
Jordan Malof
Affiliation:
Department of Computer Science, University of Montana, Missoula, MT, USA
Simiao Ren
Affiliation:
Department of Electrical & Computer Engineering, Duke University, Durham, NC, USA
Kyle Bradbury*
Affiliation:
Department of Electrical & Computer Engineering, Duke University, Durham, NC, USA Nicholas Institute for Energy, Environment & Sustainability, Duke University, Durham, NC, USA
*
Corresponding author: Kyle Bradbury; Email: kyle.bradbury@duke.edu.

Abstract

Accurate geospatial information about the causes and consequences of climate change, including energy systems infrastructure, is critical to planning climate change mitigation and adaptation strategies. When up-to-date spatial data on infrastructure is lacking, one approach to fill this gap is to learn from overhead imagery using deep-learning-based object detection algorithms. However, the performance of these algorithms can suffer when applied to diverse geographies, which is a common case. We propose a technique to generate realistic synthetic overhead images of an object (e.g., a generator) to enhance the ability of these techniques to transfer across diverse geographic domains. Our technique blends example objects into unlabeled images from the target domain using generative adversarial networks. This requires minimal labeled examples of the target object and is computationally efficient such that it can be used to generate a large corpus of synthetic imagery. We show that including these synthetic images in the training of an object detection model improves its ability to generalize to new domains (measured in terms of average precision) when compared to a baseline model and other relevant domain adaptation techniques.

Information

Type
Methods Paper
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2023. Published by Cambridge University Press
Figure 0

Figure 1. Example images from selected domain adaptation methods. For each domain mapping, the original image is shown in the first column, while each image to the right shows that same image transformed by the technique to look like it came from the target domain. A detailed mapping of the domains can be found in Figure 2.

Figure 1

Figure 2. Sample images and corresponding locations from our chosen geographic domains: the Northwest, Southwest, and Eastern Midwest United States.

Figure 2

Figure 3. Diagram depicting our process to generate GP-GAN synthetic images. In brief, sampling target objects, randomizing their locations, selecting a background image, and blending via GP-GAN.

Figure 3

Figure 4. Example GP-GAN blended synthetic images. The upper original images without turbines have been transformed into the lower GP-GAN synthetic images by blending in turbine examples.

Figure 4

Figure 5. The experimental setup. First, a pairing of a source and a target domain is selected. For each pairing, a baseline experiment is run as a benchmark by using labeled source domain images and labeled test images from the target domain. Then, a series of domain adaptation experiments are run that augment the baseline training set with additional supplemental images produced from a domain adaptation method.

Figure 5

Figure 6. Baseline experimental results demonstrating evidence of a domain gap. In this plot, all domain pairings with the same test domain are grouped together and divided into the within- and cross-domain settings. The gap in performance between within-domain and cross-domain settings is shown in this figure as the distance between the red and blue points.

Figure 6

Table 1. Synthetic experiment results compared to the baseline experiment using average precision

Figure 7

Figure 7. Experimental results displaying the averaged 95% confidence intervals among cross-domain pairs. See Appendix C for more on how the intervals constructed.

Figure 8

Table A1. Comparison of synthetic imagery to alternative domain adaptation techniques

Figure 9

Figure B1. Average model performance (AP) over all domain pairs as a function of the number of GP-GAN synthetic images included in each mini-batch of 8. The case of 0 represents the baseline experiment as there are no synthetic images per mini-batch, while 8 represents training entirely using synthetic images. The red line represents the baseline experiment within-domain average, which the synthetic experiment hopes to approach. The blue line represents the baseline case of training entirely using non-synthetic imagery.

Figure 10

Table D1. Synthetic experiment results varying the number of labeled turbine instances accessible by the synthetic image generator

Figure 11

Table D2. Density experiment results varying the number of turbines contained in each synthetic image

Figure 12

Table D3. Synthetic image experiment varying the number of synthetic images included

Figure 13

Table E1. Total number of images and labels used for training, synthetic image generation, and validation by domain