Buckets:
Title: RadioGalaxyNET: Dataset and Novel Computer Vision Algorithms for the Detection of Extended Radio Galaxies and Infrared Hosts
URL Source: https://arxiv.org/html/2312.00306
Published Time: Mon, 04 Dec 2023 02:04:00 GMT
Markdown Content: Zeeshan Hayder 2 2{}^{2}start_FLOATSUPERSCRIPT 2 end_FLOATSUPERSCRIPT Ray P. Norris 3,4 3 4{}^{3,4}start_FLOATSUPERSCRIPT 3 , 4 end_FLOATSUPERSCRIPT Minh Huynh 1,5 1 5{}^{1,5}start_FLOATSUPERSCRIPT 1 , 5 end_FLOATSUPERSCRIPT Lars Petersson 2 2{}^{2}start_FLOATSUPERSCRIPT 2 end_FLOATSUPERSCRIPT[1 1{}^{1}start_FLOATSUPERSCRIPT 1 end_FLOATSUPERSCRIPT CSIRO Space & Astronomy, PO Box 1130, Bentley WA 6102, Australia
2 2{}^{2}start_FLOATSUPERSCRIPT 2 end_FLOATSUPERSCRIPT CSIRO Data61, Black Mountain ACT 2601, Australia
3 3{}^{3}start_FLOATSUPERSCRIPT 3 end_FLOATSUPERSCRIPT Western Sydney University, Locked Bag 1797, Penrith, NSW 2751, Australia
4 4{}^{4}start_FLOATSUPERSCRIPT 4 end_FLOATSUPERSCRIPT CSIRO Space & Astronomy, P.O. Box 76, Epping, NSW 1710, Australia
5 5{}^{5}start_FLOATSUPERSCRIPT 5 end_FLOATSUPERSCRIPT International Centre for Radio Astronomy Research (ICRAR), M468, The University of Western Australia, 35 Stirling Highway, Crawley, WA 6009, Australia
(dd Mmm YYYY; dd Mmm YYYY; dd Mmm YYYY)
Abstract
Creating radio galaxy catalogues from next-generation deep surveys requires automated identification of associated components of extended sources and their corresponding infrared hosts. In this paper, we introduce RadioGalaxyNET, a multimodal dataset, and a suite of novel computer vision algorithms designed to automate the detection and localization of multi-component extended radio galaxies and their corresponding infrared hosts. The dataset comprises 4,155 instances of galaxies in 2,800 images with both radio and infrared channels. Each instance provides information about the extended radio galaxy class, its corresponding bounding box encompassing all components, the pixel-level segmentation mask, and the keypoint position of its corresponding infrared host galaxy. RadioGalaxyNET is the first dataset to include images from the highly sensitive Australian Square Kilometre Array Pathfinder (ASKAP) radio telescope, corresponding infrared images, and instance-level annotations for galaxy detection. We benchmark several object detection algorithms on the dataset and propose a novel multimodal approach to simultaneously detect radio galaxies and the positions of infrared hosts.
keywords:
galaxies: active; galaxies: peculiar; radio continuum: galaxies; Galaxy: evolution; methods: data analysis
Nikhel Gupta]Nikhel.Gupta@csiro.au \published dd Mmm YYYY
1 Introduction
Recent advancements in radio astronomy have enabled us to scan large areas of the sky in a short timescale while generating incredibly sensitive continuum images of the Universe. This has created new possibilities for detecting millions of galaxies at radio wavelengths. For example, the ongoing Evolutionary Map of the Universe (EMU; Norris et al., 2021b) survey, conducted using the Australian Square Kilometre Array Pathfinder (ASKAP; Johnston et al., 2007; DeBoer et al., 2009; Hotan et al., 2021) telescope, is projected to discover more than 40 million compact and extended galaxies in the next five years (Norris et al., 2021b). Similarly, the Low-Frequency Array (LOFAR; van Haarlem et al., 2013) survey of the entire northern sky is expected to detect more than 10 million galaxies. Other advanced radio telescopes include Murchison Widefield Array (MWA; Wayth et al., 2018), MeerKAT (Jonas & MeerKAT Team, 2016) and the Karl G. Jansky Very Large Array (JVLA Perley et al., 2011). With the advent of the Square Kilometre Array (SKA 1 1 1 https://www.skatelescope.org/the-ska-project/) radio telescope, which is expected to become operational in the coming years, the number of galaxy detections is expected to increase further, potentially reaching hundreds of millions. Such an enormous dataset will significantly impact our understanding of the physics of galaxy evolution. This is set to significantly impact our understanding of the evolution of the universe. However, to optimize the outcomes of these surveys, there is a need to innovate and develop new technologies for handling large datasets.
Radio galaxies are characterized by giant radio emission regions that extend well beyond their host galaxy structure at visible and infrared wavelengths. While most radio galaxies typically appear as simple, compact circular sources, increasing the sensitivity of radio telescopes results in the detection of more radio galaxies with complex extended structures. These structures typically consist of multiple components with distinct peak radio emissions. Based on the Fanaroff-Riley classifications (Fanaroff & Riley, 1974), radio galaxies exhibit two distinct morphologies, namely, Fanaroff-Riley Class I (FR-I) and Class II (FR-II) Active Galactic Nuclei (AGN) sources. FR-I sources have dark edges, causing the collimated jets from the central black holes at the center of the host galaxy to exhibit lower luminosity at the edges. FR-II sources, on the other hand, have brighter edges compared to the central host galaxies and sometimes lack a visible connection to the host galaxy. Constructing catalogues of radio galaxies necessitates the grouping of associated components within extended radio galaxies and the identification of their corresponding infrared host galaxies. Grouping these components is essential for estimating the real number density and overall integrated flux of radio galaxies. This process is crucial for modelling galaxy evolution and the expansion of the Universe. Failure to properly group associated components can lead to the misestimation of number density and total flux, resulting in inaccurate models.
This highlights the critical need for developing computer vision methods, to reduce the dependency on visual inspections to group associated radio source components and identify their infrared host galaxies. Computer vision tasks are typically dependent on the available data. In supervised learning tasks, the model undergoes training by utilizing pairs of images and labels, where these labels provide precise and comprehensive information necessary for the model to make specific predictions. Recently, these machine learning (ML) methods have been applied in the morphological classification and identification of radio sources, as demonstrated in studies such as (e.g. Lukic et al., 2018; Alger et al., 2018; Wu et al., 2019; Bowles et al., 2020; Maslej-Krešňáková et al., 2021; Becker et al., 2021; Brand et al., 2023). Self-supervised learning utilizes unsupervised techniques to train models based on the inherent data structure, removing the need for explicit annotations. This approach has proven effective in uncovering novel galaxy types in radio surveys, as demonstrated in studies such as (e.g. Galvin et al., 2020; Mostert et al., 2021; Gupta et al., 2022). In contrast, semi-supervised learning integrates both labelled and unlabelled data throughout the training process, as exemplified in Slijepcevic et al. (2022). In the realm of weakly supervised learning, indirect labels are utilized for the entire training dataset, serving as a supervisory signal. This particular approach has found utility in the classification and detection of extended radio galaxies (Gupta et al., 2023).
As supervised learning relies on exact ground truth labels for training, this yields results that are more reliable in contrast to approaches that lack or possess limited supervisory signals during training. However, to train and test such supervised algorithms, a large and diverse dataset of labelled radio galaxy images is necessary. Unfortunately, such a dataset is not currently available for next-generation radio surveys, which poses a significant challenge to developing automated methods for grouping multiple components of radio galaxies and identifying host galaxies. Therefore, there is a dire need for the development of a comprehensive dataset of radio galaxy images, along with accurate and consistent labels, to enable the development of effective machine learning algorithms for the automated grouping and identification of associated components of extended radio galaxies and their hosts.
This paper introduces RadioGalaxyNET, which includes a novel dataset and computer vision algorithms designed to tackle the challenge of associating radio galaxy components and identifying their hosts. Notably, our dataset is the first to be entirely curated by professional astronomers through multiple visual inspections and includes multimodal images of radio and infrared sky, as well as pixel-level labels on associated components of extended radio galaxies and their infrared hosts. RadioGalaxyNET dataset has been structured in the COCO dataset format (Lin et al., 2014a), allowing for straightforward comparison studies of various object detection strategies for the machine learning community. It includes 2,800 images with three channels each, incorporating two radio sky channels, one corresponding to the infrared sky, and a total of 4,155 annotated instances. The annotations encompass galaxy class information, bounding box details for capturing associated components of each radio galaxy, radio galaxy segmentation masks, and the host galaxy positions in infrared images.
To summarize, our work contributes to the following aspects:
- •We introduce the first dataset entirely curated by professional astronomers that includes state-of-the-art images from the highly sensitive ASKAP telescope and instance-level annotations for extended radio galaxies.
- •As a novel addition, our dataset also includes corresponding images of the infrared sky, along with the positional information of the host galaxies.
- •We train and evaluate our dataset on seven cutting-edge object detection algorithms to demonstrate the challenge of detecting and grouping components of radio galaxies. Additionally, we propose a novel method to detect the positions of infrared host galaxies simultaneously.
The paper is organized as follows. In Section2, we provide details on the radio and infrared images, as well as the annotations utilized for training and evaluation. This section also offers a summary of the related work in this domain. Section3 is dedicated to explaining the novel computer vision algorithms employed. Section4 provides comprehensive information about the network training process and the evaluation metric. In Section5, we present the outcomes derived from the trained networks. Our findings are summarized in Section6, where we also outline directions for future research.
Figure 1: Raw radio (left column), processed radio (middle column) and processed infrared (right column) images with the frame size of 450×450 450 450 450\times 450 450 × 450 pixels (0.25∘×0.25∘superscript 0.25 superscript 0.25 0.25^{\circ}\times 0.25^{\circ}0.25 start_POSTSUPERSCRIPT ∘ end_POSTSUPERSCRIPT × 0.25 start_POSTSUPERSCRIPT ∘ end_POSTSUPERSCRIPT). The processed radio images highlight the categories of extended radio galaxies, and the bounding boxes denote their total radio extent encompassing all of its components. The infrared images show host galaxies inside the circles.
2 RadioGalaxyNET: Dataset
2.1 Radio and Infrared Images
The RadioGalaxyNET dataset contains radio images derived from observations with the ASKAP radio telescope. ASKAP is located at Inyarrimanha Ilgari Bundara, the Murchison Radio-astronomy Observatory (MRO). Utilizing advanced technologies like the phased array feed (PAF) (PAF Hay et al., 2006), ASKAP can efficiently scan vast sections of the sky, significantly enhancing survey capabilities. Comprising 36 antennas, ASKAP has baselines reaching up to 6.4 km, with 30 antennas concentrated within a 2.3 km radius (Hotan et al., 2021). ASKAP is currently involved in various surveys, each designed for specific scientific objectives. One notable survey is the EMU survey, which aims to identify approximately 40 million galaxies over its five-year operational period. In 2019, the EMU pilot survey was conducted to develop technologies in preparation for the main survey.
The EMU-PS survey encompasses a sky area of 270 deg 2 2{}^{2}start_FLOATSUPERSCRIPT 2 end_FLOATSUPERSCRIPT, ranging from 301∘superscript 301 301^{\circ}301 start_POSTSUPERSCRIPT ∘ end_POSTSUPERSCRIPT to 336∘superscript 336 336^{\circ}336 start_POSTSUPERSCRIPT ∘ end_POSTSUPERSCRIPT in Right Ascension and from −63∘superscript 63-63^{\circ}- 63 start_POSTSUPERSCRIPT ∘ end_POSTSUPERSCRIPT to −48∘superscript 48-48^{\circ}- 48 start_POSTSUPERSCRIPT ∘ end_POSTSUPERSCRIPT in Declination. Comprising ten overlapping tiles, each tile underwent a total integration time of approximately 10 hours, resulting in an RMS noise of 25−35μ 25 35 𝜇 25-35~{}\mu 25 - 35 italic_μ Jy/beam. The survey operated within a frequency range of 800 to 1088 MHz, centered at 944 MHz (with a wavelength of 0.37 to 0.28m, centered at 0.32m). The telescope’s visibility data, recording electric fields as a function of time and frequency, was processed through the ASKAPsoft pipeline (Whiting et al., 2017) to generate images for the EMU-PS.
The identification of the 2,800 extended radio galaxies in the EMU-PS image involved a meticulous manual process conducted in three distinct stages, as detailed by Yew et al. (in prep.). While a comprehensive overview of the source identification methodology is available in their work, we provide a brief summary here. Initially, a quick assessment of the EMU-PS radio image was conducted to identify clearly visible extended sources. Subsequently, a thorough visual scan covered the entire radio image systematically, involving the classification of sources and marking the extent of diffuse radio emission for each. In the third and final stage, an exhaustive scan aimed to uncover additional sources possibly overlooked in previous stages. Upon completion of the identification process, the centroid position and approximate size of each source were documented. It is important to note that, despite the three rounds of manual identification, locating all extended sources remains challenging, particularly for faint ones, given the substantial scientific resources required for such an exhaustive search. Following this, cutouts of size 0.25∘×0.25∘superscript 0.25 superscript 0.25 0.25^{\circ}\times 0.25^{\circ}0.25 start_POSTSUPERSCRIPT ∘ end_POSTSUPERSCRIPT × 0.25 start_POSTSUPERSCRIPT ∘ end_POSTSUPERSCRIPT were generated at the centroid positions of these galaxies, resulting in 450×450 450 450 450\times 450 450 × 450 pixel images with a pixel size of 2 arcseconds. The left column of Figure1 presents examples of noisy raw images. Following the procedure outlined by Gupta et al. (2023), we preprocess the raw images. Examples of processed radio images are illustrated in the middle column of Figure1.
At the same sky locations of radio images, we obtain AllWISE (Cutri et al., 2021) infrared images from the Wide-field Infrared Survey Explorer’s (WISE; Wright et al., 2010) W1 band that correspond to 3.4 μ 𝜇\mu italic_μ m wavelength. Similar to radio images, infrared images are processed to reduce noise in the same way. However, the noise in the infrared images is estimated as Median +3×+~{}3\times+ 3 ×MAD due to the non-Gaussian nature of the noise and the saturation of sources in the images. The right column of Figure1 presents examples of processed infrared images, while raw infrared images are not displayed for brevity.
Finally, we create 3-channel RGB images by combining the processed radio and infrared images. In this process, the B and G channels represent radio channels. The original 32-bit FITS image is initially converted to 16-bit, and its 8-16 bit and 0-8 bit information are assigned to the B and G channels, respectively. Similarly, the infrared FITS image undergoes a conversion to 16-bit, and its 8-16 bit information is incorporated into the R channel.
Figure 2: The dataset split distributions of the RadioGalaxyNET. Shown are the distributions of extended radio galaxies in one frame (left), their categories (middle) and the occupied area per radio galaxy (right). The tables presented below the figures display the precise counts of galaxy instances within the training, validation, and test sets.
2.2 Annotations
The RadioGalaxyNET dataset has the following four sets of labels
- •the infrared host galaxy positions.
- •the extended radio galaxy classes,
- •the bounding boxes enclosing all components of each radio galaxy, and
- •the segmentation masks for radio galaxies.
The labelling process details are discussed in Yew et al. (in prep.) and Gupta et al. (2023). In summary, we visually identified host galaxies in the infrared images, with the most likely hosts situated near or over the central peaks of radio emission. Only sources with identified host galaxies in the infrared image were included in the dataset. Radio sources were classified based on criteria from Fanaroff & Riley (1974), distinguishing between FR-I and FR-II sources by considering the distance between the peak radio emission of all components and the total extent of the source. Additionally, some sources were classified as FR-x due to uncertainties in determining peak flux positions and total extent, as explained in (Gupta et al., 2023). Another category, denoted as R, includes resolved sources exhibiting extended emissions without clear peaks other than in the central part. The bounding boxes for each radio source, measured with the CARTA visualization package (Comrie et al., 2021), served as the basis for obtaining segmentation masks. These masks were then created for all pixels within the bounding box, where the flux exceeded 3 σ 𝜎\sigma italic_σ. Annotations in the RadioGalaxyNET dataset adhere to the COCO dataset format (Lin et al., 2014a), facilitating straightforward comparisons of object detection methods. Radio annotations for each galaxy are stored as ‘categories’, ‘bbox’, and ‘segmentation’, while the positions of the infrared hosts are stored as ‘keypoints’.
2.3 Data Statistics and Possible Tasks
The RadioGalaxyNET dataset encompasses 2,800 3-channel images featuring two radio sky channels and one corresponding infrared sky channel. Both noisy and processed radio images are included in the dataset. It comprises 2,800 extended radio galaxies, resulting in a total of 4,155 instances of these galaxies due to their proximity in the sky and their appearance in multiple images. The radio galaxies in the dataset are categorized as 13% FR-I, 48% FR-II, 14% FR-x, and 25% R sources. The dataset is divided into three sets: training, validation, and test sets, with a split ratio of 0.7:0.15:0.15:0.7 0.15:0.15 0.7:0.15:0.15 0.7 : 0.15 : 0.15, respectively. This split ratio is widely used in machine learning and represents a balanced approach that achieves a reasonable trade-off between having sufficient data for training and fine-tuning the model, while also ensuring that the model is evaluated on a suitably large, independent test set. The statistics for the number of objects in one frame, categories of extended radio galaxies, and the occupied area of labelled objects are shown in Figure2. The distribution of target galaxies is approximately matched between the training, validation, and test sets. Small extended radio galaxies (area <48 2 absent superscript 48 2<48^{2}< 48 start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT in pixels) make up the largest proportion of the dataset.
Although the primary goal of this study is to detect radio galaxies, the RadioGalaxyNET dataset’s unique multimodality can be leveraged for the simultaneous identification of infrared host galaxies. Additionally, it could be insightful to explore approaches for grouping multiple components of galaxies. Instead of component association using given box boundaries, the direct segmentation of related galaxy components could present a greater challenge for models. This could also be linked to instance segmentation approaches, where different input modalities could provide an interesting challenge. The dataset’s unlabelled extended radio galaxies in a given frame make it suitable for experimenting with semi-supervised and contrastive learning methods. Furthermore, RadioGalaxyNET dataset could be used to validate signal enhancement in previous radio sky surveys. By downweighing detections that resemble extended galaxies in the dataset images, the dataset could be used to discover new galaxies in the radio sky. Finally, the modalities in the dataset could be used to model extended radio galaxy emissions from infrared hosts and vice versa.
In this paper, we only focused on extended radio galaxies for the current dataset and have not included labels for several compact sources in radio images. Few extended radio galaxies, such as Odd Radio Circles (Norris et al., 2021a), some Giant Radio Galaxies and other peculiar galaxies (Gupta et al., 2022), are also not included in the dataset. Future efforts should focus on including a balanced representation of these sources in the dataset.
2.4 Related Work
2.4.1 Existing Radio Galaxy Dataset
The availability of datasets for detecting and classifying radio galaxies is limited. The MiraBest Batched Dataset is a labelled dataset of FR galaxies extracted from Miraghaei & Best (2017). It contains FR-I, FR-II, and hybrid extended radio sources, with any nonstandard morphology reported. The dataset comprises a total of 1,256 images obtained from the VLA FIRST sky survey (Becker et al., 1995), with all images measuring 150×150 150 150 150\times 150 150 × 150 pixels, with one-pixel corresponding to an angular size of 1.8 arcseconds. The dataset is divided into seven training batches and one test batch, with each batch consisting of approximately 157 images, and the number of objects in each class is distributed relatively evenly across batches. The test batch contains at least one example of each class. It is important to note that the MiraBest dataset only provides information about the classes of radio galaxies without annotations for bounding boxes, segmentation masks, or infrared hosts, making it challenging to apply to multi-component association problems.
Another dataset is sourced from the Radio Galaxy Zoo (RGZ) citizen science initiative, as described by (Wu et al., 2019). Over 12,000 volunteers visually identified radio galaxies, resulting in a dataset of 6,536 extended radio galaxies with a user-weighted consensus level (CL) of at least 0.6. The CL reflects the level of agreement among citizens regarding the classification. The images in this dataset were obtained from the FIRST radio survey, and the annotations include the position of galaxies in the sky and the bounding boxes derived from web portal clicks by volunteers. These boxes are saved as the approximate angular size in the RGZ dataset. However, the dataset does not provide information on the segmentation masks and infrared host positions. The RMS noise in the FIRST radio images is approximately ten times larger than that in the EMU-PS, resulting in a lower density of galaxies in the same area. This high galaxy density makes RadioGalaxyNET dataset images derived from EMU-PS challenging for object detection. Table1 shows a comparison between the existing and our new dataset.
It is important to emphasize that the previous dataset possesses radio image noise levels that are an order of magnitude higher than our dataset. This higher noise level renders low-intensity extended structures essentially invisible. As a result, these previous datasets are not optimal for training networks aimed at detecting radio galaxies in the next generation of radio telescopes.
Table 1: Datasets currently available for the machine learning tasks of classification and object detection involving radio galaxies. The annotations C, B, S, and K are categories, bounding boxes, segmentation and keypoint labels, respectively. Section2 provides a detailed description of the annotations for both our dataset and the existing dataset.
2.4.2 Existing Machine Learning Applications
In recent years, radio galaxy detection in images has attracted increasing attention, particularly focusing on deep learning algorithms. These algorithms are commonly divided into two categories: region-proposal-based methods and classification-based methods. Wu et al. (2019) used Faster Region-based Convolutional Neural Network (Faster-RCNN; Ren et al., 2015a) based method to locate and classify radio galaxies in the RGZ dataset. Lao et al. (2021) used a combination of residual neural network (ResNet; He et al., 2016) and Feature Pyramid Network (FPN; Lin et al., 2017) to detect and classify FR galaxies. Scaife & Porter (2021) classified FR galaxies in the MiraBest dataset using Group equivariant Convolutional Neural Networks (G-CNNs; Cohen & Welling, 2016). Slijepcevic et al. (2022) used FixMatch (Sohn et al., 2020) to classify both MiraBest and RGZ radio galaxies. Zhang et al. (2022b) used You Only Look Once method (YOLOv5; Redmon et al., 2016a) to classify radio galaxies.
Figure 3: An overview of the multimodal modelling strategy introduced in this study. In the context of the Gal-DETR model (refer to Section3.1), we introduce a keypoint estimation module within the transformer encoder-decoder framework. This enables the simultaneous detection of categories and bounding boxes for radio galaxies, and the positions of infrared hosts. A similar multimodal strategy is introduced for Gal-Deformable DETR and Gal-DINO (as detailed in Sections3.2 and 3.3).
3 RadioGalaxyNET: Novel Computer Vision Methods
While the machine learning techniques employed in prior studies (refer to Section2.4.2) have their own strengths. Recent advancements in computer vision have introduced several more effective algorithms that outperform the previously utilized methods in real-world scenarios. In addition, none of the previously employed techniques possess the ability to concurrently detect both multiple-component radio galaxies and their corresponding infrared hosts. In this section, we describe these state-of-the-art object detection approaches and explain our novel enhancements to these methods, which allow for the simultaneous detection of both radio galaxies and infrared hosts.
3.1 Gal-DETR
In this paper, we introduce Gal-DETR, multimodal model for computer vision. The Gal-DETR model consists of two primary components: the DEtection TRansformers (DETR), as described in Carion et al. (2020) that is trained for the class and bounding box prediction for radio galaxies, and our novel Keypoint detection module, trained concurrently to detect infrared hosts. Note that the existing multimodal methods are tailored to specific tasks. Here we have radio images where galaxies appear larger due to extended emission, while in infrared images, the host galaxies look like point objects (as depicted in columns 2 and 3 of Figure1). To the best of our knowledge, there are no specific models that deal with objects that look completely different in two image modalities. As a result, we introduce our own approach to modelling, illustrated in Figure3.
3.1.1 Class and Bounding Box Detection for Radio Galaxies
DETR model leverages the Transformer architecture (Vaswani et al., 2017), initially developed for natural language processing, to tackle the complex task of object detection in images following (Dosovitskiy et al., 2020). Unlike traditional methods that rely on region proposal networks (e.g. Faster RCNN; Ren et al., 2015b), DETR introduces end-to-end object detection using Transformers. The model begins with a convolutional neural network (CNN) backbone, such as ResNet, to process the input image and extract relevant feature maps. ResNet-50 is a convolutional neural network architecture designed to address the challenge of vanishing gradients in very deep neural networks. It incorporates 50 layers and utilizes skip connections, also referred to as residual connections, to enhance gradient flow during training (He et al., 2015). To incorporate spatial information into the Transformer architecture, DETR introduces positional encodings. These encodings are added to the feature maps, providing the model with information about the relative positions of objects within the image. The feature maps, enhanced with positional encodings, are then passed through a Transformer encoder. This component enables the model to simultaneously process the entire image, capturing contextual relationships between various regions. DETR introduces learned object queries, similar to class-specific anchor boxes in traditional detectors (e.g. Tian et al., 2019). Note that the anchor boxes are prior boxes defined as a set of pre-determined, fixed-size boxes of different scales and aspect ratios. These anchor boxes are chosen based on prior knowledge of the expected object sizes and shapes within the dataset. However, unlike fixed anchor boxes, the object queries in DETR are learned during training through the transformer decoder and represent the classes of objects the model aims to detect. The final output from the decoder has two distinct output heads: one for class prediction, which estimates the class probabilities for each object query, and another for bounding box prediction, which calculates the coordinates (x, y, width, height) of each object query’s bounding box (see Figure3).
DETR employs a Hungarian loss 2 2 2 https://github.com/mmgalushka/hungarian-loss function to establish associations between predicted bounding boxes and ground-truth boxes. This loss enforces a one-to-one mapping between predicted and ground-truth objects, ensuring the model’s permutation invariance. The overall loss function for DETR combines the classification loss (cross-entropy) for class predictions and the box loss (smooth L1 loss) for bounding box predictions,
ℒ DETR=ℒ class+ℒ box,subscript ℒ DETR subscript ℒ class subscript ℒ box\mathcal{L}{\text{DETR}}=\mathcal{L}{\text{class}}+\mathcal{L}_{\text{box}},caligraphic_L start_POSTSUBSCRIPT DETR end_POSTSUBSCRIPT = caligraphic_L start_POSTSUBSCRIPT class end_POSTSUBSCRIPT + caligraphic_L start_POSTSUBSCRIPT box end_POSTSUBSCRIPT ,(1)
where ℒ class subscript ℒ class\mathcal{L}{\text{class}}caligraphic_L start_POSTSUBSCRIPT class end_POSTSUBSCRIPT is the cross-entropy loss for class predictions and ℒ box subscript ℒ box\mathcal{L}{\text{box}}caligraphic_L start_POSTSUBSCRIPT box end_POSTSUBSCRIPT is the smooth L1 loss for bounding box predictions. In the context of object detection, the L1 loss is often used for bounding box regression. It calculates the absolute difference between the predicted bounding box coordinates (e.g., x, y, width, height) and the ground truth bounding box coordinates. This loss penalizes the model for deviations between the predicted and true bounding box values, with larger differences leading to higher loss values.
3.1.2 Keypoint Detection for Infrared Hosts
In addition to the object detection method that employs bounding boxes to detect extended radio galaxies, the integration of keypoint detection techniques (e.g. Simon et al., 2017) can offer a complementary approach for identifying the corresponding infrared hosts. Keypoints, representing specific landmarks or distinctive features within the radio and infrared images, provide valuable spatial information that aids in precisely determining the location of the host galaxy. Unlike bounding boxes that enclose the entire source, keypoints allow for a more fine-grained localization, providing a more accurate representation of the host’s position within the radio source. This precise localization is especially valuable when the radio emission exhibits complex morphologies.
In the present work, we extended the capabilities of DETR by incorporating keypoint detection. We refer to this enhanced version as Gal-DETR. In Gal-DETR, we integrated keypoint detection to complement the existing object detection capabilities of the original DETR algorithm (see Figure3). By incorporating keypoint detection, Gal-DETR gains the ability to identify salient keypoints within the objects, providing additional spatial information and fine-grained details. The keypoint detection module in Gal-DETR leverages the transformer-based architecture of DETR to capture global context and local details. By attending to the keypoint embeddings and utilizing self-attention mechanisms, Gal-DETR localizes and associates keypoints for the infrared host galaxies. The overall loss function for Gal-DETR combines the DETR loss for class and bounding box predictions, in addition to the keypoint detection loss as
ℒ Gal-DETR=ℒ DETR+ℒ keypoint subscript ℒ Gal-DETR subscript ℒ DETR subscript ℒ keypoint\mathcal{L}{\text{Gal-DETR}}=\mathcal{L}{\text{DETR}}+\mathcal{L}_{\text{% keypoint}}caligraphic_L start_POSTSUBSCRIPT Gal-DETR end_POSTSUBSCRIPT = caligraphic_L start_POSTSUBSCRIPT DETR end_POSTSUBSCRIPT + caligraphic_L start_POSTSUBSCRIPT keypoint end_POSTSUBSCRIPT(2)
where ℒ keypoint subscript ℒ keypoint\mathcal{L}_{\text{keypoint}}caligraphic_L start_POSTSUBSCRIPT keypoint end_POSTSUBSCRIPT is the L1 loss for keypoint detection to detect infrared hosts. In the context of keypoint detection, the L1 loss calculates the absolute difference between the predicted keypoint coordinates (e.g., x, y position of host) and the ground truth keypoint coordinates.
3.2 Gal-Deformable DETR
The second multimodal model we introduce in this paper is Gal-Deformable DETR. Similar to Gal-DETR, the Gal-Deformable DETR model consists of two primary components: the Deformable DEtection TRansformers (Deformable DETR), as described in Zhu et al. (2021) that is trained for the class and bounding box prediction for radio galaxies, and our novel Keypoint detection module, trained concurrently to detect infrared hosts.
Deformable DETR builds upon the foundation of DETR and introduces several key enhancements. Deformable attention mechanisms allow the model to adaptively adjust the spatial sampling locations for better feature extraction, especially for objects with complex shapes or poses. Unlike DETR, which relies solely on learned positional encodings, Deformable DETR adds spatial positional encodings to the feature maps. This addition helps the model better capture the spatial relationships between objects.
In this study, we have expanded the functionalities of Deformable DETR by introducing keypoint detection, resulting in an enhanced version known as Gal-Deformable DETR. Much like the Gal-DETR model, Gal-Deformable DETR incorporates keypoint detection as a complementary component to the existing object detection capabilities inherited from the original Deformable DETR algorithm. Specifically, we have integrated keypoint detection in conjunction with the deformable convolutional layers originally introduced in the Deformable DETR framework. Similar to Equation2, the L1 loss for keypoint detection is combined with the class and bounding box loss during the training of the Gal-Deformable DETR model.
3.3 Gal-DINO
The third multimodal model we introduce in this paper is Gal-DINO. Following Gal-DETR and Gal-Deformable DETR models, Gal-DINO also consists of two primary components: the DETR with Improved deNoising anchOr boxes (DINO), as described in Zhang et al. (2022a) that is trained for the class and bounding box prediction for radio galaxies, and our novel Keypoint detection module, trained simultaneously to detect infrared hosts.
DINO extends the DETR model with improved denoising anchor boxes, introducing several key enhancements. DINO improves anchor boxes, which are pre-defined boxes used for object detection. It introduces better strategies for selecting and placing these anchor boxes, enhancing the model’s ability to detect objects of different sizes and aspect ratios. DINO introduces an improved mechanism for matching anchor boxes to ground truth objects during training, making the model more accurate in localization and classification. DINO employs adaptive convolutional features, allowing the model to focus on informative regions of the image, thus improving both efficiency and accuracy.
Much like our approach with Gal-DETR, we integrated keypoint detection into the DINO algorithm, which already has improved de-noising anchor boxes. This enhanced version, featuring keypoint detection, is denoted as Gal-DINO. By reducing the impact of noise and outliers, Gal-DINO produces more robust and precise bounding box predictions, resulting in better localization of the extended radio sources and their associated infrared hosts within these enhanced bounding boxes. In a manner similar to the approach outlined in Equation2, the Gal-DINO model incorporates the L1 loss for keypoint detection combined with the class and bounding box loss during the training process.
3.4 Gal-SIOD
Detection of objects in situations with imperfect data has recently become a focal point. Weakly supervised object detection (WSOD) encounters notable challenges in localization due to the lack of direct annotations (e.g. Gupta et al., 2023). Li et al. (2022) introduced an approach known as Single Instance Annotated Object Detection (SIOD). We refer the reader to their paper for detailed information about the SIOD method. Briefly, SIOD requires just one instance annotation for each category present in an image. SIOD offers a more robust and informative source of prior knowledge for detecting the remaining unlabelled instances. This approach strikes a balance between annotation cost and performance, providing a valuable solution to object detection challenges under imperfect data conditions.
This holds significance in the context of our current study on radio galaxy detection. As detailed in Section2, even after conducting three rounds of manual search of EMU-PS, not all extended sources could be precisely located. This arises from the fact that visual inspections entail an exhaustive search for faint extended radio galaxies, a process demanding a substantial allocation of scientific resources. This implies that certain extended and peculiar sources within the same image are not annotated in the dataset. Utilizing SIOD for radio galaxy detection offers us the possibility to detect these unlabelled extended radio galaxies within a given image.
Note that we have not incorporated keypoint detection into the radio galaxy detection process using SIOD. This decision stems from our primary objective, which is to assess whether we can detect additional radio galaxies not labelled in the dataset using this method. Future work should integrate keypoint detection into SIOD to enable the simultaneous detection of infrared hosts alongside radio galaxy detections. While our current implementation directly employs SIOD for our task, we have named our implementation as Gal-SIOD for naming consistency.
3.5 Gal-SIOD-DMiner
SIOD-DMiner (Li et al., 2022) improves upon SIOD by introducing the Dual-Mining (DMiner) framework. SIOD, while effective in certain scenarios, faces challenges when it comes to dealing with unlabelled regions. Directly assigning all unlabelled regions as background can adversely affect the training process and detector performance. SIOD-DMiner tackles this challenge and improves object detection efficacy by integrating a Similarity-based Pseudo-Label Generating module (SPLG). This module retrieves instances by assessing feature similarity between labelled reference instances and the remaining unlabelled area in the image. This improves object detection over SIOD, which lacks this mechanism for leveraging unlabelled data effectively.
In addition, the SIOD-DMiner recognizes that relying solely on pseudo-labels generated by SPLG can be problematic. It can lead to confusion due to false pseudo-labels, especially when the model focuses on learning a hyperplane for discriminating each class from the others. To mitigate this issue, SIOD-DMiner introduces a Pixel-level Group Contrastive Learning module (PGCL). PGCL improves the model’s ability to withstand inaccurate pseudo-labels, reducing reliance on potentially flawed annotations. For similar considerations as mentioned earlier, we have not integrated keypoint detection into the radio galaxy detection process using SIOD-DMiner. Future work should integrate keypoint detection into SIOD-DMiner to enable the simultaneous detection of unlabelled infrared hosts. Furthermore, we have employed the SIOD-DMiner from Li et al. (2022) for our task but named our implementation Gal-SIOD-DMiner to maintain naming consistency.
3.6 Gal-Faster RCNN
Faster Region-based Convolutional Neural Network (Faster RCNN; Ren et al., 2015b) significantly improved the speed and accuracy of object detection compared to the proposal-based predecessor methods. Faster RCNN follows a two-stage approach to object detection, which distinguishes it from earlier methods like Fast RCNN (Girshick, 2015). The first stage of Faster R-CNN is a Region Proposal Network (RPN), which efficiently generates region proposals (candidate bounding boxes) that are likely to contain objects of interest. The RPN operates on feature maps extracted from the input image and predicts regions with high scores. In the second phase, the system fine-tunes and categorizes these region proposals. Region of Interest (RoI) pooling is employed to derive standardized feature maps of fixed dimensions from each region proposal, making them suitable for a standard CNN classifier. Faster RCNN uses a CNN classifier to categorize the content within each RoI into predefined object classes (e.g., FR-I, FR-II etc.) and regress the bounding box coordinates for precise localization. The entire network, including both the RPN and the classifier, is trained end-to-end. This means that the model learns to generate region proposals and classify objects simultaneously during training. While Deformable DETR and DINO have demonstrated superior performance over Faster RCNN when applied to real-life images from the COCO dataset (Zhu et al., 2021; Zhang et al., 2022a), we have opted to utilize Faster RCNN on our dataset for the sake of comparison. In line with this primary objective of comparing Faster RCNN with Transformer-based methods, we have refrained from incorporating Keypoint detection into Faster RCNN. We have named our implementation Gal-Faster RCNN to maintain naming consistency.
Figure 4: Object detection results: Shown are the processed radio-radio-infrared images and ground truth annotations (first column), ground truth and Gal-DINO keypoint detections as circles and triangles over infrared images (second column), Gal-DINO (third column) and Gal-SIOD-DMiner (fourth column) class and bounding box predictions over radio images with a confidence threshold of 0.25. These models exhibit the capability to detect additional extended galaxies that lack ground truth annotations.
3.7 Gal-YOLOv8
You Only Look Once version 8 (YOLOv8) is an object detection model latest in the YOLO (Redmon et al., 2016b) series of object detection models. YOLOv8 follows the one-stage detection paradigm, which means it directly predicts bounding boxes and class probabilities for objects in a single pass through the neural network. This makes it faster compared to two-stage detectors like Faster R-CNN. YOLOv8 employs a CNN backbone architecture that is designed to capture features at multiple scales. YOLOv8 uses a detection head consisting of convolutional layers to predict bounding boxes and class probabilities. It predicts class labels and confidence scores for each detected object. Similar to initial YOLO versions, YOLOv8 uses anchor boxes to help predict the shape and location of objects. However, YOLOv8 has the option to automatically calculate anchor box sizes based on the dataset, which can simplify the training process. In this study, we perform fine-tuning on a pre-trained YOLOv8 model using our dataset and designate our customized implementation as Gal-YOLOv8 to maintain naming consistency. It is worth noting that we have not integrated keypoint detection into YOLOv8 in this implementation. Future work should explore methods for detecting infrared hosts alongside radio galaxy detections within the YOLO class of models.
Table 2: Bounding box and keypoint detection results on the test set of RadioGalaxyNET. From left to right, the columns display the multimodal models introduced in this study, the number of model parameters in millions, the number of training epochs, the average precision for IoU (or OKS) thresholds ranging from 0.50 to 0.95 (AP), a specific IoU (or OKS) threshold of 0.5 (AP 50 50{}{50}start_FLOATSUBSCRIPT 50 end_FLOATSUBSCRIPT), IoU (or OKS) threshold of 0.75 (AP 75 75{}{75}start_FLOATSUBSCRIPT 75 end_FLOATSUBSCRIPT), and the average precision for small-sized radio galaxies (AP S S{}{\rm S}start_FLOATSUBSCRIPT roman_S end_FLOATSUBSCRIPT), medium-sized radio galaxies (AP M M{}{\rm M}start_FLOATSUBSCRIPT roman_M end_FLOATSUBSCRIPT), and large-sized radio galaxies (AP L L{}_{\rm L}start_FLOATSUBSCRIPT roman_L end_FLOATSUBSCRIPT), categorized by areas less than 24 2 superscript 24 2 24^{2}24 start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT, between 24 2 superscript 24 2 24^{2}24 start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT and 48 2 superscript 48 2 48^{2}48 start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT, and greater than 48 2 superscript 48 2 48^{2}48 start_POSTSUPERSCRIPT 2 end_POSTSUPERSCRIPT pixels, respectively. Detailed information on the training and development of these models is provided in Section4, while the models themselves are described in Section3.
{NiceTabular} clllllllll Model Params Epochs AP AP 50 50{}{50}start_FLOATSUBSCRIPT 50 end_FLOATSUBSCRIPT AP 75 75{}{75}start_FLOATSUBSCRIPT 75 end_FLOATSUBSCRIPT AP S S{}{\rm S}start_FLOATSUBSCRIPT roman_S end_FLOATSUBSCRIPT AP M M{}{\rm M}start_FLOATSUBSCRIPT roman_M end_FLOATSUBSCRIPT AP L L{}_{\rm L}start_FLOATSUBSCRIPT roman_L end_FLOATSUBSCRIPT
(%) (%) (%) (%) (%) (%)
\Block 3-1\rotate Bbox Gal-DETR 41M 500 22.6 38.1 26.2 16.3 24.8 19.8
Gal-Deformable DETR 40M 100 40.2 52.1 45.9 37.7 39.9 22.2
Gal-DINO 47M 30 53.7 60.2 58.9 41.5 56.9 35.2
\Block 3-1\rotate Keys Gal-DETR 41M 500 35.4 37.5 35.3 9.1 60.0 49.6
Gal-Deformable DETR 40M 100 45.0 49.0 45.3 21.5 79.9 76.1
Gal-DINO 47M 30 48.1 53.4 48.4 17.6 81.4 82.9
4 Training and Evaluation
In this section, we provide details of the network training process. Additionally, we outline the evaluation metrics employed for comparing different computer vision models in this section.
4.1 Training Details
The dataset division into training, validation, and test sets is detailed in Figure2 and Section2.3. The training data is used to train the networks, while the validation and test sets serve as inference datasets during and after training, respectively. The networks described in Section3 are trained for different numbers of epochs, which depend on their convergence speed and the stability of validation results. In this context, one epoch signifies a single pass through the entire training dataset during the model’s training process. During each epoch, the training dataset is subjected to various randomly applied augmentations.
These augmentations involve random flipping, random rotations, random resizing, and random cropping of a randomly selected set of 3-channel images from the training data. Random flipping, performed horizontally (left to right), exposes the model to various object orientations and viewpoints. Random rotations, ranging from -180 to 180 degrees, promote rotational invariance in the network – a critical aspect for handling radio galaxies with random orientations in the sky. Random resizing operations entail scaling images up or down by selecting a target size randomly within predefined bounds. We have set the lower and upper bounds to 400×400 400 400 400\times 400 400 × 400 and 1300×1300 1300 1300 1300\times 1300 1300 × 1300 pixels, respectively. This strategy enables the model to learn from objects at different scales while preventing overfitting to specific sizes or aspect ratios. Additionally, random cropping involves the random selection of a portion (or crop) from an input image, discarding the rest. This cropping operation, always applied after image resizing, introduces spatial variability in object locations within an image. The application of these augmentations enhances the model’s capacity to generalize effectively and perform well on unseen data. These augmentations are applied during the training of all the networks described in this study.
All networks are trained and evaluated on an Nvidia Tesla P100. We retained the original hyperparameters for each algorithm during training, except for specific modifications. These hyperparameters are configuration settings that are not learned from the training data but are set prior to the training process. The most important ones include the learning rate, which dictates the magnitude of weight adjustments during training; the batch size, indicating the number of data samples processed at a time; architecture parameters, specifying neural network architecture details like layer counts and filter sizes; dropout rate for preventing overfitting by deactivating units during training; activation functions applied to the output of neural network layers to introduce non-linearity into the model; and optimizer for weight updates. While we do not present all the hyperparameters for various networks here to keep the discussion concise, we refer the reader to the network architectures available in the provided repositories for these details. Here, we focus on the critical training aspects.
As detailed in Section3.1, Gal-DETR introduces multimodal modelling for radio and infrared detections. Specifically, we implemented keypoint detection to the model architecture and Hungarian loss function. We reduced the learning rate to 5×10−5 5 superscript 10 5 5\times 10^{-5}5 × 10 start_POSTSUPERSCRIPT - 5 end_POSTSUPERSCRIPT and the number of queries to 10 from 100. The number of queries is decreased due to the fact that there are no more than five extended galaxies per image in our dataset, as depicted in the left panel of Figure2. As indicated in Table3.7, Gal-DETR has 41 million parameters and undergoes training over the course of 500 epochs for ∼similar-to\sim∼10 hours on a single P100 GPU.
Similar changes were made for Gal-Deformable DETR model (Section3.2), where keypoint detection was also implemented in the deformable attention mechanism. Training Gal-Deformable DETR, which comprises 40 million parameters, spanned 100 epochs and consumed approximately 8 hours on a single GPU. For Gal-DINO model (Section3.3), we made the same changes as for Gal-DETR and additionally implemented keypoint detection in the de-noising anchor box mechanism. Training Gal-DINO, featuring 47 million parameters, involved a 30-epoch training process that took approximately 6 hours on a single GPU.
As Gal-SIOD (Section3.4) necessitates only one instance annotation per category within an image, we performed a selection process where we retained unique categories of radio galaxies in each image. For example, we ensured that each image contained at most one random annotation for FR-I sources, FR-II sources, FR-x sources, and R radio sources. This selection reduced the overall number of training annotations in all training images from 2,910 to 2,534, resulting in an annotation keeping ratio of 0.87. No additional changes were made for Gal-SIOD-DMiner (Section3.5) except for the reduction in annotations with the same keeping ratio. Both the Gal-SIOD and Gal-SIOD-DMiner networks, each comprising 14.4 million parameters, underwent separate 200-epoch training sessions, with each session taking approximately 8 hours on a single GPU. The Gal-Faster RCNN (Section3.6) model underwent training for 20,000 epochs, which took approximately 5 hours on a single GPU. To train Gal-YOLOv8 (Section3.7), we utilized the pre-trained YOLOv8 model from Ultralytics 3 3 3 https://ultralytics.com for object detection. The training process took approximately 12 hours for 30 epochs and was conducted on a single GPU.
Table 3: Bounding box detection results using Gal-SIOD and Gal-SIOD-DMiner networkd. The AP 50 50{}{50}start_FLOATSUBSCRIPT 50 end_FLOATSUBSCRIPT, AP S S{}{\rm S}start_FLOATSUBSCRIPT roman_S end_FLOATSUBSCRIPT, AP M M{}{\rm M}start_FLOATSUBSCRIPT roman_M end_FLOATSUBSCRIPT, and AP L L{}{\rm L}start_FLOATSUBSCRIPT roman_L end_FLOATSUBSCRIPT reported here correspond to those in Table3.7. The average precision values in this table are provided for various confidence thresholds (S 𝑆 S italic_S), ranging from no limit to 0, 0.3, and 0.5 confidence scores. Comprehensive information regarding the models and their training (or evaluation) can be found in Sections3 and 4, respectively.
cccccccc Model Confidence Params Epochs AP 50 50{}{50}start_FLOATSUBSCRIPT 50 end_FLOATSUBSCRIPT AP S S{}{\rm S}start_FLOATSUBSCRIPT roman_S end_FLOATSUBSCRIPT AP M M{}{\rm M}start_FLOATSUBSCRIPT roman_M end_FLOATSUBSCRIPT AP L L{}{\rm L}start_FLOATSUBSCRIPT roman_L end_FLOATSUBSCRIPT
Constraint (%) (%) (%) (%)
\Block 4-1Gal-SIOD AP@S@𝑆@S@ italic_S 14.4M 200 24.0 14.7 17.8 14.8
AP@S 0@subscript 𝑆 0@S_{0}@ italic_S start_POSTSUBSCRIPT 0 end_POSTSUBSCRIPT 14.4M 200 45.9 28.5 34.5 30.9
AP@S 3@subscript 𝑆 3@S_{3}@ italic_S start_POSTSUBSCRIPT 3 end_POSTSUBSCRIPT 14.4M 200 41.6 26.6 31.4 22.4
AP@S 5@subscript 𝑆 5@S_{5}@ italic_S start_POSTSUBSCRIPT 5 end_POSTSUBSCRIPT 14.4M 200 18.3 10.3 12.7 10.5
\Block 4-1Gal-SIOD-DMiner AP@S@𝑆@S@ italic_S 14.4M 200 26.3 15.1 19.6 16.2
AP@S 0@subscript 𝑆 0@S_{0}@ italic_S start_POSTSUBSCRIPT 0 end_POSTSUBSCRIPT 14.4M 200 46.8 27.3 35.4 30.3
AP@S 3@subscript 𝑆 3@S_{3}@ italic_S start_POSTSUBSCRIPT 3 end_POSTSUBSCRIPT 14.4M 200 43.1 26.0 32.5 24.5
AP@S 5@subscript 𝑆 5@S_{5}@ italic_S start_POSTSUBSCRIPT 5 end_POSTSUBSCRIPT 14.4M 200 24.3 15.6 16.3 12.1
Table 4: Bounding box detection results for Gal-Faster RCNN and Gal-YOLOv8 models. The columns here align with those presented in Table3.7. Additional information regarding the networks can be found in Sections3 and 4.
Xet Storage Details
- Size:
- 69.8 kB
- Xet hash:
- 29c3606169738381faeed2f0fa3d283403f8d60fbc19a0d4aeed294924c0769c
Xet efficiently stores files, intelligently splitting them into unique chunks and accelerating uploads and downloads. More info.










