Hostname: page-component-89b8bd64d-dvtzq Total loading time: 0 Render date: 2026-05-08T08:06:11.379Z Has data issue: false hasContentIssue false

Informing synthetic passive microwave predictions through Bayesian deep learning with uncertainty decomposition

Published online by Cambridge University Press:  16 September 2024

Pedro Ortiz
Affiliation:
Department of Computer Science, Naval Postgraduate School, Monterey, CA, USA
Eleanor Casas
Affiliation:
Department of Earth Sciences, Millersville University, Millersville, PA, USA
Marko Orescanin*
Affiliation:
Department of Computer Science, Naval Postgraduate School, Monterey, CA, USA
Scott W. Powell
Affiliation:
Department of Meteorology, Naval Postgraduate School, Monterey, CA, USA
Veljko Petkovic
Affiliation:
Cooperative Institute for Satellite Earth System Studies/Earth System Science Interdisciplinary Center, University of Maryland, College Park, MD, USA
*
Corresponding author: Marko Orescanin; Email: marko.orescanin@nps.edu

Abstract

Space-borne passive microwave (PMW) data provide rich information on atmospheric state, including cloud structure and underlying surface properties. However, PMW data are sparse and limited due to low Earth orbit collection, resulting in coarse Earth system sampling. This study demonstrates that Bayesian deep learning (BDL) is a promising technique for predicting synthetic microwave (MW) data and its uncertainties from more ubiquitously available geostationary infrared observations. Our BDL models decompose predicted uncertainty into aleatoric (irreducible) and epistemic (reducible) components, providing insights into uncertainty origin and guiding model improvement. Low and high aleatoric uncertainty values are characteristic of clear sky and cloudy regions, respectively, suggesting that expanding the input feature vector to allow richer information content could improve model performance. The initially high average epistemic uncertainty metrics quantified by most models indicate that the training process would benefit from a greater data volume, leading to improved performance at most studied MW frequencies. Using quantified epistemic uncertainty to select the most useful additional training data (a training dataset size increase of 3.6%), the study reduced the mean absolute error and root mean squared error by 1.74% and 1.38%, respectively. The broader impact of this study is the demonstration of how predicted epistemic uncertainty can be used to select targeted training data. This allows for the curation of smaller, more optimized training datasets and also allows for future active learning studies.

Information

Type
Application Paper
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2024. Published by Cambridge University Press
Figure 0

Figure 1. Executive summary of this study, where: (a) training dataset features are the brightness temperatures (TBs) measured by GOES-16 ABI Bands 7–16 (near IR and IR wavelengths) and are exemplified by full disk imagery from 17:40 UTC on September 12, 2022 (GOES Image Viewer, 2017); (b) training dataset labels are the microwave TBs from GMI, and are exemplified by the global, full-day GMI coverage on February 01, 2020, where shading denotes data coverage (Negri et al., 1989; EOSDIS Worldview, 2013), black denotes data gaps, and the red box denotes the domain in Figure 10); and (c) a flowchart outlines the methods and scientific impact.

Figure 1

Figure 2. Network architecture. (a) 56-block Residual Network (ResNet) with output modeled as a Normal distribution (highlighted in yellow). (b) Internal structure of each ResNet block depicted in panel (a). The skip connection (highlighted in yellow) inside each block is the defining characteristic of a ResNet architecture. The convolutional 2D Flipout layers (highlighted in yellow) implement variational inference, making this a Bayesian model architecture.

Figure 2

Figure 3. Results using models trained on 26,587,000 samples from January to June 2020 to generate synthetic GMI data using the July test dataset of 1.63 million samples. H = horizontal polarization; V = vertical polarization.

Figure 3

Figure 4. Change in metrics for the August test dataset of 1.53 million samples compared to July metrics in Figure 3. H = horizontal polarization; V = vertical polarization. Table of August metrics is in Appendix Table A2.

Figure 4

Figure 5. Box plots for square root epistemic uncertainty (SRE) and absolute error in Kelvin (K) for July data, where orange lines denote the median values and green triangles denotes the mean value for a given frequency. The whiskers span from 2.5% to 97.5%. (a–c) Each label (T$ {}_b^{mw} $ value) is categorized as “in distribution” if the T$ {}_b^{mw} $ belongs to middle 95% of the label distribution (see Appendix Figure A1). Out of distribution (OOD) labels are further divided into “low” and “high” by T$ {}_b^{mw} $, where “low” means lowest 2.5% of T$ {}_b^{mw} $ values and “highest” means highest 2.5% of T$ {}_b^{mw} $ values. The blue horizontal line denotes the 75th percentile square root epistemic uncertainty (SRE) value for all labels. (d–f) The three leftmost box plots depict the prediction error using the label categorization scheme from panels (a) to (c). The two rightmost box plots depict the prediction error using the 75th percentile SRE value for categorization. The mean absolute error for each category is labeled in Kelvin.

Figure 5

Figure 6. Percentage change in error and uncertainty due to updating the model by training on additional data from July that has uncertainty greater than the 75th percentile ([Appendix Table A3 values – Appendix Table A2 values]/Appendix Table A2 values). The dashed line at −3.6% indicates a decrease proportional to the size of the growth in the training dataset.

Figure 6

Figure 7. GMI 183 ± 3 GHz (vertical) change in metrics due to model update. Row 1: GMI observations and the ABI observations for Bands 8 (6.2 $ \mu $m) and 14 (11.2 $ \mu $m) at 13:41 UTC on August 26, 2020. Rows 2–5: Model predictions, prediction absolute error, prediction aleatoric uncertainty, and prediction epistemic uncertainty. The first column was produced using the model before the update and the second column was produced using the model after the update. The third column contains the difference between the values from column 1 and column 2, where red indicates that a metric increased after the model was updated and blue indicates that a metric decreased after the model was updated.

Figure 7

Figure 8. Change in metrics due to model update as in Figure 7, but for GMI 23 GHz (vertical).

Figure 8

Figure 9. Change in metrics due to model update as in Figure 7, but for GMI 19 GHz (horizontal).

Figure 9

Figure 10. (a) $ \sim $15 min of GMI observations. (b) Synthetic T$ {}_b^{mw} $ generated from 15 min of ABI data (GPM orbit number 33679 at 14:40 UTC on February 1, 2020) corresponding to the box in Figure 1. (c) Standard deviation of each T$ {}_b^{mw} $ in panel (b).

Figure 10

Table A1. Results using models trained on 26,587,000 sample from January to June to generate synthetic GMI data using the July test dataset of 1.63 million samples

Figure 11

Table A2. Results as in Appendix Table A1, but for the August test dataset of 1.53 million samples

Figure 12

Table A3. Results as in Appendix Table A2, but using the updated models

Figure 13

Figure A1. GMI observed microwave temperature brightness (Tbmw) value distributions for the August dataset. Orange indicates horizontal polarization; blue indicates vertical polarization. Orange and blue shaded areas indicate the T$ {}_b^{mw} $ value density for a given GMI frequency. Vertical lines mark the 0th, 25th, 50th, 75th, and 100th percentile. White dots mark the mean T$ {}_b^{mw} $ values. Bold lines indicate T$ {}_b^{mw} $ values in the middle 95% of the distribution; thin lines indicate T$ {}_b^{mw} $ values below or above the middle 95% of the distribution.

Figure 14

Figure A2. Percentage of uncertainty by component for each GMI channel and polarization using the predictions from the July dataset.