Hostname: page-component-89b8bd64d-n8gtw Total loading time: 0 Render date: 2026-05-08T19:28:29.342Z Has data issue: false hasContentIssue false

Radio Galaxy Zoo: Tagging radio subjects using text

Published online by Cambridge University Press:  16 October 2023

Dawei Chen
Affiliation:
School of Computing, Australian National University (ANU), Canberra, Australia
Vinay Kerai
Affiliation:
University of Western Australia (UWA), Perth, Australia
Matthew J. Alger
Affiliation:
Google Australia, Pyrmont, Australia
O. Ivy Wong*
Affiliation:
CSIRO Space & Astronomy, Bentley, WA, Australia ICRAR-M468, University of Western Australia, Crawley, WA, Australia ARC Centre of Excellence for All-Sky Astrophysics in 3 Dimensions (ASTRO 3D), Canberra, Australia
Cheng Soon Ong
Affiliation:
School of Computing, Australian National University (ANU), Canberra, Australia Data61, CSIRO, Black Mountain, Australia
*
Corresponding author: O. Ivy Wong; Email: ivy.wong@csiro.au
Rights & Permissions [Opens in a new window]

Abstract

RadioTalk is a communication platform that enabled members of the Radio Galaxy Zoo (RGZ) citizen science project to engage in discussion threads and provide further descriptions of the radio subjects they were observing in the form of tags and comments. It contains a wealth of auxiliary information which is useful for the morphology identification of complex and extended radio sources. In this paper, we present this new dataset, and for the first time in radio astronomy, we combine text and images to automatically classify radio galaxies using a multi-modal learning approach. We found incorporating text features improved classification performance which demonstrates that text annotations are rare but valuable sources of information for classifying astronomical sources, and suggests the importance of exploiting multi-modal information in future citizen science projects. We also discovered over 10000 new radio sources beyond the RGZ-DR1 catalogue in this dataset.

Information

Type
Research Article
Creative Commons
Creative Common License - CCCreative Common License - BYCreative Common License - NCCreative Common License - SA
This is an Open Access article, distributed under the terms of the Creative Commons Attribution-NonCommercial-ShareAlike licence (http://creativecommons.org/licenses/by-nc-sa/4.0/), which permits non-commercial re-use, distribution, and reproduction in any medium, provided the same Creative Commons licence is used to distribute the re-used or adapted article and the original article is properly cited. The written permission of Cambridge University Press must be obtained prior to any commercial use.
Copyright
© The Author(s), 2023. Published by Cambridge University Press on behalf of the Astronomical Society of Australia
Figure 0

Figure 1. Overview of the machine learning workflow for classifying radio sources in this paper. We apply the same pre-trained Vision Transformer (Dosovitskiy et al. 2020), denoted as ViT$^*$ in the workflow, to both the radio and infrared (IR) images to produce their numerical presentations, and the pre-trained BERT language model (Devlin et al. 2019) is used to create a numerical representation of the corresponding text discussions on RadioTalk forum after removing the tags. These representations are combined before passing it to a multi-label classifier which predicts tags likely applicable to the radio source. The classification performance is obtained by comparing the ground-truth tags in the RadioTalk forum text and the tags predicted by the classifier. The screenshot in this figure shows images and discussions text for subject ARG002XAP.

Figure 1

Figure 2. Summary of tag usage in the RadioTalk dataset. Top: Histogram of the 20 most frequently used tags. Bottom: Proportion of the total tag usage covered by the top n tags (i.e. the n most frequently used tags). We can see that the top 50 tags cover 94% of all tag usage in the RadioTalk dataset.

Figure 2

Figure 3. Hierarchical tree of tag clustering created by an astronomer upon closer examination of the most frequently used tags in the RadioTalk dataset. The merger of tags in Section 2.3 are performed according to this tree.

Figure 3

Table 1. Merging 4 sets of tags to create 4 new tags.

Figure 4

Table 2. Statistics of tags and text discussions of radio subjects.

Figure 5

Figure 4. Architecture of the multi-modal classifier. As an example, this figure shows the predicted probabilities of the 11 tags for the radio subject ARG0002FUD, making use of its discussion text on the RadioTalk forum (after pre-processing as discussed in Section 3.1), and its radio and infrared images.

Figure 6

Table 3. Metadata of the dataset archive on Zenodo.

Figure 7

Figure 5. Over 10000 complex sources that are not in the RGZ-DR1 catalogue are available in the RadioTalk dataset.

Figure 8

Table 4. Classification performance (in terms of F1 score) of each tag, for text, image, and multi-modal information. Classes are sorted in increasing order of the number of subjects tagged with the class. Text: Binary relevance classifier using only text features; Text (CC): Classifier chains using only text data; Image: Classifier using only images; Text+Image: Classifier using multi-modal information. The best performance for each tag is shown in bold italic.

Figure 9

Figure 6. Classification performance (in terms of F1 score) versus the number of occurrences of tags in the dataset for the binary relevance classifiers using text (Text), images (Image) and multi-modal information (Text+Image). Text information can be helpful in predicting tags with few (e.g. xshaped), medium (e.g. noir_ifrs and overedge) or large (e.g. triple) number of observations in the dataset.

Figure 10

Figure A.1. RGZ interface for cross-matching radio components to host galaxies (Banfield et al. 2015). Panel (a) shows an example double-lobed radio source and the slider in the central position where both the radio and infrared images are presented in blue and orange heatmaps, respectively. As the slider is transitioned completely towards IR, the radio image reaches 100 percent transparency and the radio emission is represented by the sets of contours (panels b and c). The associated radio components are highlighted as blue contours in panel (b) and the volunteer-identified cross-matched host galaxy is marked by the yellow circle in panel (c).

Figure 11

Figure B.1. Examples of radio subjects for each of the 11 tags. From left to right: (1) artefact describes an image where there is significant image processing residuals; (2) asymmetric describes radio jets or lobes that are not symmetrical; (3) bent describes jets and lobes that appear to have been swept to one side; (4) compact describes an unresolved single component radio source; (5) double describes two radio components that extend away from the host galaxy; (6) hourglass describes two overlapping radio components; (7) noir_ifrs describes a radio source with no visible galaxy counterpart; (8) overedge describes a source which extends beyond the field-of-view of the image; (9) restarted describes a radio jet that consists of more than 3 components which could be due to restarting radio jet activity; (10) triple describes a source with 3 radio components; (11) xshaped describes a source which appears to be a superposition of 2 orthogonal hourglass sources.

Figure 12

Figure B.2. Precision-Recall curves of the multi-label classifier using multi-modal information for each of the 11 tags.

Figure 13

Table D.1. Classification performance evaluated with additional metrics. Text: Binary relevance classifier using only text features; Text (CC): Classifier chains using only text data; Image: Classifier using only images; Text+Image: Classifier using multi-modal information. The best performance in terms of each metric (i.e. each row) is shown in bold italic.

Figure 14

Table D.2. Classification performance for text, image, and multi-modal information. Text: Binary relevance classifier using only text features; Text (CC): Classifier chains using only text data; Image: Classifier using only images; Text+Image: Classifier using multi-modal information. The best performance in terms of each metric among 4 different multi-label classifiers is shown in bold italic.