Surveying Image Segmentation Approaches in Astronomy
Surveying Image Segmentation Approaches in Astronomy
Abstract
Image segmentation plays a critical role in unlocking the mysteries of the universe, providing astronomers with a
clearer perspective on celestial objects within complex astronomical images and data cubes. Manual segmentation,
while traditional, is not only time-consuming but also susceptible to biases introduced by human intervention. As a
result, automated segmentation methods have become essential for achieving robust and consistent results in astro-
nomical studies. This review begins by summarizing traditional and classical segmentation methods widely used in
astronomical tasks. Despite the significant improvements these methods have brought to segmentation outcomes, they
fail to meet astronomers’ expectations, requiring additional human correction, further intensifying the labor-intensive
nature of the segmentation process. The review then focuses on the transformative impact of machine learning, partic-
ularly deep learning, on segmentation tasks in astronomy. It introduces state-of-the-art machine learning approaches,
highlighting their applications and the remarkable advancements they bring to segmentation accuracy in both as-
tronomical images and data cubes. As the field of machine learning continues to evolve rapidly, it is anticipated
that astronomers will increasingly leverage these sophisticated techniques to enhance segmentation tasks in their re-
search projects. In essence, this review serves as a comprehensive guide to the evolution of segmentation methods
in astronomy, emphasizing the transition from classical approaches to cutting-edge machine learning methodologies.
We encourage astronomers to embrace these advancements, fostering a more streamlined and accurate segmentation
process that aligns with the ever-expanding frontiers of astronomical exploration.
Keywords: Segmentation, Artificial Intelligence, Machine Learning, Neural Network, Vision Transformer,
Generative Model, Astronomy image processing(2306)
3
Pixels with a high magnitude and a consistent direction
are considered to be edges. The Sobel and Prewitt oper-
ators are simple and efficient, but they can be sensitive
to noise in the image. To improve robustness to noise,
it is common to pre-smooth the image with a Gaussian
filter before applying the edge detection algorithm.
In contrast to the Sobel and Prewitt operators, the
Canny edge detector distinguishes itself as a multi-stage
edge detection approach celebrated for its robustness
and precision. The Canny method incorporates a se-
quence of stages:
(a) (b) (c) (d)
1. Gaussian smoothing: The image is smoothed with Figure 2: Detected edges using various methods on Cassini Astron-
a Gaussian filter to reduce noise and create a con- omy images, as detailed in Yang et al. (2018). (a) Depicts the Original
Image. (b), (c), and (d) illustrate the edges detected by Canny, Sobel,
tinuous gradient. and Prewitt methods, respectively.
2. Gradient calculation: The gradient magnitude and
direction are calculated at each pixel location.
3. Non-maximum suppression: Only the local max- ods such as Canny, Sobel, and Prewitt (Yang et al.,
ima along the gradient direction are preserved, re- 2018).
sulting in single-pixel-wide edges.
4. Edge tracking by hysteresis: A two-level thresh- 2.3. Watershed
olding strategy is used to link strong edges together Watershed segmentation is a widely used image seg-
and discard weak edges. mentation algorithm that partitions an image into dis-
The Canny edge detector is more computationally ex- tinct regions based on pixel intensity and spatial rela-
pensive than other edge detection algorithms, such as tionships. The algorithm works by treating the image as
the Sobel and Prewitt operators, but it produces more a topographic surface, with pixel intensity values rep-
accurate edge maps. It is also more robust to noise, resenting elevations. Water is flooded onto the surface
making it suitable for a wide range of applications in from markers, which are typically identified by the user
computer vision, image processing, and image analysis. or generated using image processing techniques. The
water flows into valleys and basins, and the watershed
Edge detection is a fundamental image processing
lines (i.e., the boundaries between the basins) define the
technique that is widely used in astronomy to identify
segmented regions. Here is a step-by-step overview of
structures. For example, the Sobel operator is used to
the watershed segmentation process:
detect filamentary structures in molecular clouds, which
can be used to study their alignment with the magnetic 1. Preprocessing: Denoise and enhance the image, if
field direction (Green et al., 2017). The Sobel edge de- needed.
tection technique is also used to measure the magni- 2. Gradient computation: Calculate the image gradi-
tude of the tip of the red giant branch (TRGB), which ent magnitude to emphasize areas of rapid change
is a key parameter for determining distances to nearby in intensity.
galaxies (Mouhcine et al., 2005). Additionally, the
3. Marker selection: Identify seed pixels, either man-
Canny edge detector is used to detect transient events
ually or using image processing techniques.
such as coronal mass ejections from the Sun (Boursier
et al., 2005), and to search for the signatures of cos- 4. Distance transform: Calculate the distance trans-
mic string networks on cosmic microwave background form of the markers to define their potential re-
(CMB) anisotropies (Vafaei Sadr et al., 2018). gions of influence.
The Sobel, Prewitt, and Canny methods are all fun- 5. Watershed flooding: Simulate water rising from
damental edge detection techniques, each with its own the markers and flooding the image, separating
strengths and weaknesses. The choice of which method basins with watershed lines.
to use depends on factors such as the level of image 6. Region merging: Merge adjacent basins as the wa-
noise, the required precision, and the computational ter level rises.
constraints. Figure 2 illustrates the detected edges in 7. Result visualization: Label each pixel in the seg-
Cassini Astronomy images employing different meth- mented image according to its basin.
4
8. Post-processing: Refine the segmentation results 2. Compute the energy functional. The energy func-
as needed, e.g., by removing small regions or tional is computed based on the current position of
merging adjacent regions. the active contour and the image features.
3. Deform the active contour. The active contour is
Watershed segmentation is a powerful image segmen-
deformed to minimize the energy functional. This
tation algorithm, but it is sensitive to the selection of
can be done using a variety of optimization algo-
markers and initial conditions. Careful marker selec-
rithms.
tion and post-processing are essential for accurate and
meaningful results. 4. Repeat steps 2 and 3 until the active contour con-
verges. Convergence occurs when the energy func-
tional cannot be further reduced.
5. The final position of the active contour represents
the segmented object.
1. Data Preparation: Assemble a diverse dataset with Figure 9: Overview of the Mask R-CNN framework, as presented
labeled images for training and validation. Clean in He et al. (2017).
and preprocess the data by resizing, normalizing,
and augmenting to improve model generalization. Mask R-CNN (He et al., 2017), a deep learning ex-
2. Model Selection: Select an appropriate deep learn- tension of Faster R-CNN (Girshick, 2015), has revo-
ing architecture (e.g., U-Net, Mask R-CNN) based lutionized image segmentation, particularly in astron-
on task requirements and available resources. omy. The precise delineation of celestial objects in as-
Choose between building a custom structure or tronomical images is crucial, and Mask R-CNN excels
adopting pretrained models for transfer learning. in this domain by enabling pixel-level object segmen-
3. Model Configuration: Fine-tune parameters like tation. Mask R-CNN is a region-based convolutional
learning rate, batch size, and regularization for op- neural network that is finely tuned for detecting and seg-
timal performance. Define a suitable loss func- menting objects in images. Building on the success of
tion aligned with segmentation objectives (e.g., Faster R-CNN, a leading object detection model, Mask
cross-entropy loss, Dice loss, as discussed in Sec- R-CNN has become instrumental in image segmenta-
tion 3.9). tion tasks.
4. Training: Feed input images into the network and Figure 9, taken from the work by He et al. (2017), il-
optimize the model to minimize the defined loss lustrates the workflow of Mask R-CNN. In addition to
function. Monitor performance on a validation set the general steps introduced at the beginning of this sec-
to prevent overfitting. tion, we elaborate on the essential elements and central
processes of Mask R-CNN as follows:
5. Evaluation: Assess the model’s performance using
evaluation metrics such as IoU, Dice coefficient, or 1. Region Proposal Network (RPN): The RPN takes
pixel-wise accuracy. the extracted features from the backbone network
10
and generates a set of proposal boxes, each asso-
ciated with a score indicating the likelihood of an
object being present within the box.
2. Region of Interest (RoI) Align: The RoI Align op-
eration takes the proposal boxes and extracts cor-
responding feature maps from the backbone net-
work, preserving spatial information and aligning
the feature maps with the regions.
3. Classification and Bounding Box Refinement: For
each proposal box, Mask R-CNN performs two
tasks: classification and bounding box refinement.
It first classifies the object within the box and re-
fines the box’s position if necessary.
4. Mask Generation: Mask R-CNN generates a seg-
mentation mask for each object within the proposal
boxes using a separate neural network that predicts
pixel-wise masks for each object.
5. Final Prediction: The final output of Mask R-CNN
includes:
Figure 10: From Burke et al. (2019), illustration of detection inference
• Classification results indicating the object in an actual DECaLS image of ACO 1689. Galaxy masks are depicted
class for each box in light blue, while star masks are presented in green. The confidence
of the detection, indicating the likelihood that the object belongs to a
• Refined bounding box coordinates specific class, is displayed above each mask.
• Segmentation masks representing the ob-
ject’s precise shape at the pixel level
and galaxies (Burke et al., 2019). Figure 10 provides
6. Post-Processing: The final results can undergo
an illustration of how Mask R-CNN identifies and seg-
post-processing steps to filter out low-confidence
ments stars and galaxies in real observational data.
detections, eliminate duplicate detections, and
As a side note, there were earlier Fully Convolu-
fine-tune the masks.
tional Networks (FCN) based methods that predated
7. Output: The output of the Mask R-CNN algorithm
Mask R-CNN, although Mask R-CNN has largely re-
is a set of bounding boxes, each associated with a
placed these simpler FCN-based methods in image seg-
class label and a high-resolution binary mask that
mentation. These earlier FCN-based approaches em-
accurately delineates the object.
ployed straightforward Convolutional Neural Networks
Mask R-CNN demonstrates resilience to noise and (CNNs) like AlexNet (Krizhevsky et al., 2012), VG-
has maintained a track record of excellence in numerous GNet (Simonyan and Zisserman, 2014), GoogLeNet
image segmentation challenges and competitions. How- (Szegedy et al., 2015), and ResNet (He et al., 2016)
ever, it is important to note that the computational de- as backbones, predicting class labels for each pixel in
mands of training and deploying a Mask R-CNN model an image. For instance, Bhambra et al. (2022) utilized
can be substantial, necessitating the availability of ro- popular image classification model architectures such as
bust GPUs. Additionally, this model often mandates a VGG16 (Simonyan and Zisserman, 2014), ResNet50v2
substantial volume of meticulously annotated data for (He et al., 2016), and Xception (Chollet, 2017) as back-
effective training, a process that can be labor-intensive bones, incorporating saliency maps to highlight signif-
and costly. In certain scenarios, Mask R-CNN has the icant regions for segmentation of various components
potential to yield an abundance of smaller segments, of galaxies. Building upon this eXplainable Artificial
contributing to the challenge of over-segmentation. Intelligence (XAI) technique, Tang et al. (2023) further
In the domain of astronomy, Mask R-CNN has found developed and enhanced it for the classification and seg-
widespread application. It has been utilized for tasks mentation of different types of radio galaxies. Addition-
such as segmenting solar filaments in H-α images of ally, Richards et al. (2023) combined both semantic seg-
the Sun (Ahmadzadeh et al., 2019), distinguishing dif- mentation and instance segmentation, fusing Mask R-
ferent types of galaxies from SDSS data (Farias et al., CNN and FCN, to create a panoptic segmentation model
2020), and efficiently detecting and segmenting stars for segmenting galactic structures.
11
3.2. Self-Organizing Map comprehensive understanding:
1. SOM Initialization: Configure the SOM grid by
determining its size and dimensions. Initialize the
SOM nodes with random weights.
2. Competition: Present each input image to the SOM
and identify the winning node, which possesses
weights most similar to the input. Update the
weights of the winning node and its neighboring
nodes using a learning rate and a neighborhood
function. Repeat this process for multiple itera-
tions, allowing the SOM to adapt to the input data.
3. Quantization: Assign each input image to the node
with the closest weights, effectively mapping the
high-dimensional data onto the lower-dimensional
Figure 11: From Qian et al. (2019), an illustration depicting the SOM grid.
training of SOM with red dots representing training data in high-
dimensional space, and black dots alongside the grid indicating the
4. Clustering and Labeling: Group similar nodes to-
trained SOM. gether to form clusters, representing distinct re-
gions or structures in astronomical images. Ana-
Self-Organizing Map (SOM), also known as Koho- lyze the clustered map and assign labels to differ-
nen maps, is a neural network-based technique rooted in ent regions or clusters based on the characteristics
the principles of unsupervised autonomous learning in- of the astronomical objects they represent.
spired by neurobiological studies (Kohonen, 1990). The
SOM’s inherent strength lies in its adept handling of
SOM algorithm adapts to data through synaptic plastic-
complex, high-dimensional data, such as multi-band or
ity, mirroring the organized mapping of sensory inputs
hyperspectral images prevalent in remote sensing and
in the cerebral cortex. This topographic map maintains
astronomical observations. Its robustness against noise
two critical properties: firstly, it retains incoming infor-
and variations in image intensity further enhances its ef-
mation in its proper context at each processing stage,
ficacy in real-world applications. The SOM technique
and secondly, it ensures that neurons handling closely
offers a unique and adaptive approach to image segmen-
related information are positioned near one another, fos-
tation, proving to be a valuable tool for extracting mean-
tering interaction through short synaptic connections.
ingful structures and patterns from various types of im-
In the realm of image segmentation, SOM stands out
agery. This versatility and effectiveness extend across
as a potent unsupervised learning algorithm, adept at
different domains, including astronomy, medical imag-
discerning intricate patterns and structures within com-
ing, and remote sensing, showcasing its applicability to
plex datasets. Its ability to map high-dimensional input
a wide range of segmentation challenges. As an illus-
data onto a lower-dimensional grid ensures the preser-
tration, Schilliro and Romano (2021) employ SOM in
vation of topology and relationships among data points,
astronomy to analyze a high spatial and spectral resolu-
making it particularly well-suited for tasks such as im-
tion Hα line image of the sun, identifying several fea-
age segmentation, where understanding spatial relation-
tures corresponding to the main structures of the solar
ships is crucial. SOM excels in grouping similar pixels,
photosphere and chromosphere. Figure 12 provides an
facilitating the identification of distinct regions or ob-
example of the segmented output generated by SOM for
jects. Its departure from conventional methods, which
solar images.
often rely on predefined criteria, is notable, as SOM
learns patterns directly from the input data, offering an
adaptive and data-driven approach to segmentation. The 3.3. Encoder-Decoder Architectures
algorithm’s versatility in capturing both global and local Encoder-decoder architectures, a powerful paradigm
features positions it as an effective tool across a spec- in image segmentation, have revolutionized astronomy
trum of image segmentation tasks. by enabling precise delineation of celestial objects and
Figure 11 visually illustrates the training process of regions within astronomical images, a crucial step in
SOM. In addition to the introductory steps provided ear- understanding and analyzing astronomical structures.
lier in this section, we delve into the fundamental com- Encoder-decoder architectures comprise two key com-
ponents and central procedures of SOM to provide a ponents:
12
(A) (B) (C)
Figure 12: From Schilliro and Romano (2021), an illustration of the segmented output using SOM for a solar image. (A): A selection of spectral
images obtained along the Hα line. (B): The feature lattice generated by a 4 × 4 SOM, facilitating the segmentation of 16 distinct regions. (C): The
feature lattice produced by a 3 × 3 SOM, enabling the segmentation of nine different regions.
• Encoder: The encoder extracts informative fea- chitecture provides a robust foundation, while the vari-
tures from the input image using a sequence of ations in implementation cater to specific segmentation
convolutional layers that reduce spatial dimensions needs and challenges. In subsequent sections, we ex-
while capturing high-level features. In astronomy, plore the specific details of these encoder-decoder ar-
the encoder learns to discern and learn from the in- chitectures, unraveling their unique attributes and con-
tricate structures, objects, and phenomena present tributions to image segmentation in astronomy and re-
in astronomical images, even in noisy or complex lated domains.
backgrounds.
• Decoder: The decoder takes the encoded features 3.3.1. U-Net
and generates a segmentation mask using trans-
posed convolutions to upsample the spatial dimen-
sions and produce a high-resolution mask. In as-
tronomy, the decoder translates the learned fea-
tures into precise pixel-level delineation of astro-
nomical objects, providing insights into their size,
shape, and spatial distribution.
13
U-Net (Ronneberger et al., 2015), depicted in Fig- putationally intensive, requiring access to robust hard-
ure 13, is a pivotal encoder-decoder architecture that has ware resources. Achieving optimal performance with
significantly transformed image segmentation, includ- U-Net typically involves meticulous hyperparameter
ing in astronomy. Its symmetrical structure with skip tuning, a process that can be time-consuming and it-
connections enables precise object delineation, even in erative. Additionally, effectively deploying U-Net may
complex and noisy backgrounds. This adaptability has necessitate a deep understanding of deep learning con-
found some applications in astronomy, playing a pivotal cepts. Finally, U-Net can be prone to overfitting on
role in tasks such as segmenting large-scale structures small datasets without the proper application of regu-
in simulations (Aragon-Calvo, 2019), outlining stellar larization techniques, potentially leading to suboptimal
wind-driven bubble structures in simulations (Van Oort generalization when applied to new, unseen data.
et al., 2019), pinpointing stellar feedback structures in
observational data (Xu et al., 2020b,a), precisely cap- 3.3.2. U-Net++
turing the intricate details of spiral arms in disk galax-
ies (Bekki, 2021), and delineating galactic spiral arms
and bars (Walmsley and Spindler, 2023). Moreover, it
has proven effective in segmenting individual galaxies
within cosmological surveys (Boucaud et al., 2020; Bre-
tonnière et al., 2021), as well as in segmenting galaxy-
galaxy strong lensing systems (Ostdiek et al., 2022b)
and locating subhalos from strong lens images (Ostdiek
et al., 2022a). Figure 14 visually depicts the 3D out-
put of the Convolutional Approach to Structure Iden-
tification - 3D (casi-3d) prediction (Xu et al., 2020a).
Employing a U-net architecture, this method identi-
fies the positions of protostellar outflows within a real
3D position-position-velocity data cube of the Perseus
molecular cloud.
blocks equip the model with the ability to access and in-
tegrate a more extensive set of contextual information,
thereby fostering a deeper understanding of complex as-
tronomical structures. This feature enables U-Net++ to
attain more precise and accurate segmentation, which
is of paramount importance in the realm of astronomy,
where objects and structures can be highly complex and
exhibit considerable variation in size and shape. U-
Net++ has demonstrated its dominance by consistently
surpassing U-Net in various medical image segmenta-
tion challenges, thereby cementing its status as the op-
timal choice for these tasks.
3.3.3. TransUNet
Figure 18: From Yang et al. (2023), segmentation results with and
without the Transformer branch. The first column corresponds to the
input image, the second column showcases the segmentation result
without the Transformer branch, and the third column exhibits the
segmentation result obtained using the Transformer branch. The red
boxes highlight examples where the utilization of a dual-branch net-
work enhances feature extraction precision and improves MBP seg-
mentation. The yellow boxes point out instances where the dual-
branch network leads to fewer misidentifications of MBPs.
Figure 17: Overview of the TransUNet framework, as presented in
Chen et al. (2021).
3.3.4. Swin-UNet
Swin-UNet (Cao et al., 2022), depicted in Figure 19,
represents an innovative amalgamation that integrates
the hierarchical architecture of the Swin Transformer
with the encoder-decoder framework of U-Net. Swin
Transformers are renowned for their ability to process
images with diverse structures and scales efficiently.
Figure 19: Overview of the Swin-UNet framework, adapted from Cao
Swin-UNet leverages this efficiency to skillfully capture et al. (2022).
complex structures that may vary significantly in size.
In situations where a single-scale approach may fail to
accurately segment objects within astronomical images,
Swin-UNet provides an effective solution.
Swin-UNet has proven its efficacy in effectively han-
dling complex and extensive medical images, substan-
tially improving the precision of image segmentation
tasks within the medical domain. While the utiliza-
tion of Swin-UNet in astronomical contexts is some-
what limited, there are instances where it has been ap-
plied. For example, it has been employed in segment-
ing clouds from remote sensing images (Gong et al.,
2023) and detecting astronomical targets from multi-
color photometry sky surveys (Jia et al., 2023a). Fig- RGB image Label UNet Swin-Unet Gong23
ure 20 illustrates Swin-UNet’s proficiency in accurately
segmenting clouds from a remote sensing image. Figure 20: From Gong et al. (2023), prediction outcomes of various
models on the AIR-CD remote sensing dataset. The Gong23 model
3.4. Vision Transformers (ViT) integrates both Swin-UNet and traditional CNN in its architecture.
Embedding
* Extra learnable
Attention now applied to images.
[ c l as s ] embedding Linear Projection of Flattened Patches
Norm 3. Positional Encoding: Since ViTs lack the inherent
spatial understanding of CNNs, they incorporate
Embedded
Patches positional encoding to provide spatial information
to the model. This encoding informs the model
Figure 21: Overview of the Vision Transformer (ViT) framework as about the relative locations of patches within the
presented in (Dosovitskiy et al., 2020).
image. Various positional encoding techniques, in-
cluding sinusoidal encoding or learned positional
To understand the capabilities of ViTs in image seg-
encodings, can be used.
mentation, it is essential to explore the foundations of
4. Design Model Architecture: Design a ViT ar-
transformers and attention mechanisms. Transformers
chitecture for image segmentation. ViTs, unlike
are a class of deep learning models that have revolution-
CNNs, use a transformer architecture. A typical
ized various domains, including NLP and computer vi-
ViT model comprises several key components:
sion. These models rely on a mechanism called “atten-
tion,” which enables them to process sequential or spa- • Embedding Layer: This layer converts image
tial data effectively. In the context of computer vision, patches into embedding vectors.
transformers break down images into smaller patches, • Transformer Encoder Blocks: These blocks
treating them as sequences of data. This approach al- process the embeddings and capture spatial
lows them to capture global context and intricate spatial relationships and context information across
relationships, which are vital for tasks like image seg- patches.
mentation.
• Class Token: ViTs add a class token to the
Attention, the key mechanism within transformers,
embeddings to perform classification.
enables the model to assign varying levels of impor-
tance to different parts of the input data. By learn- • Positional Encodings: These encodings help
ing these importance weights, the model can focus on the model understand the spatial position of
relevant information while filtering out noise or irrele- patches.
vant details. This mechanism’s ability to capture long- • Linear Projection: This projection maps the
range dependencies in the data has made transformers, transformer output to the segmentation mask
and subsequently ViTs, exceptionally effective in image space.
17
ViTs have become increasingly prominent in the field have also ventured into various astronomical applica-
of computer vision, demonstrating remarkable profi- tions. For instance, they have been employed in tasks
ciency, particularly in scientific image segmentation. In such as classifying transient astronomical sources (Chen
domains like medical image segmentation, including X- et al., 2023) and estimating strong gravitational lensing
ray, CT, and MRI datasets, ViTs have showcased SOTA parameters (Huang et al., 2022). These initial explo-
performance and remarkable accuracy, as summarized rations into the application of ViTs in astronomy repre-
by (Henry et al., 2022). sent an exciting frontier. ViTs bring the transformative
capabilities of transformers and attention mechanisms
to this domain, potentially reshaping the way we ana-
lyze and comprehend astronomical structures in the vast
expanse of astronomical images in the future.
3.5. Mamba
Figure 22: From Merz et al. (2023): Left: Image showing artifacts
like blooming and optical ghosts around the bright star in the upper
right and large ghosts in the lower middle. Right: Inference results
from an MViTv2 Lupton-scaled network.
Data Latent
Diffusion process
𝑞(𝑥1 |𝑥0 ) 𝑞(𝑥 𝑇 |𝑥 𝑇−1 )
𝑥0 𝑥1 𝑥𝑇−1 𝑥𝑇
𝑝𝜃 (𝑥0 |𝑥1 ) 𝑝𝜃 (𝑥𝑇−1 |𝑥𝑇 )
Denoising process
N classes
Figure 25: Overview of the general frameworks for three generative (a) Discriminative semantic segmentation
models, figure adapted from (Zhu et al., 2022b).
Posterior
Prior learning
learning
Generative models constitute an intriguing subset of
deep learning that transcends the confines of conven-
tional image segmentation. They not only identify and
outline objects within images, but also generate new Latent distribution
data based on discerned patterns and structures. These
models are engineered to grasp the inherent distribution (b) Generative semantic segmentation
of data, enabling them to generate novel, data-coherent
content. In the realm of image segmentation, genera- Figure 26: From Chen et al. (2023a), a schematic comparison illus-
tive models offer a unique proficiency: they can fab- trating (a) conventional discriminative learning and (b) a generative
ricate detailed and lifelike segmentations of objects, a learning-based model for segmentation.
capability that can prove invaluable in various fields, in-
cluding astronomy. Generative models, utilizing tech-
niques like Variational Autoencoders (VAEs)(Kingma 3.6.1. Variational Autoencoders (VAEs)
and Welling, 2014), Generative Adversarial Networks Variational Autoencoders (VAEs) (Kingma and
(GANs) (Goodfellow et al., 2014), and Denoising Dif- Welling, 2014) represent a powerful class of genera-
fusion Probabilistic Models (DDPMs) (Sohl-Dickstein tive models used in various domains, inclding com-
et al., 2015; Ho et al., 2020), are capable of simulta- puter vision and image processing. These models ex-
neously identifying inherent patterns in an image and cel at learning the underlying structure of data and, in
generating segmentations that mirror these patterns ac- the context of image segmentation, play a pivotal role
curately. This vibrant fusion of image interpretation in understanding image patterns and generating coher-
and generation paves the way for new possibilities in ent segmentations. VAEs offer a unique approach to
object identification and precise image outlining. The both data compression and generation, making them in-
schematic representations of these generative models valuable in tasks like image reconstruction and synthe-
are depicted in Figure 25. sis. VAE training usually seeks to minimize the Kull-
In contrast to conventional deep learning models for back–Leibler (KL) divergence (a statistical metric that
segmentation, which rely on the log-likelihood of a con- measures the similarity between two distributions) via
ditional probability (i.e., the classification probability the Gaussian reparametrization in the bottleneck layer
of image pixels), generative segmentation models in- as part of the overall loss function. Additional image
troduce an auxiliary latent variable distribution. This perception losses such as L1 are further introduced to
19
ensure the reconstruction ability of the neural networks. are relatively data-efficient, they still require a certain
The overall learning objective can be formulated as: amount of labeled data for training. Acquiring a suffi-
ciently large and diverse dataset for segmentation tasks
log p(x) = Eq(x|z) [log p(x|z)] − DKL [q(z|x)||q(z)], (1) can be challenging, especially in some domains.
based mask decoder. The image encoder generates an • Intersection over Union (IoU): IoU is a widely em-
image embedding from each input image, while the ployed metric for evaluating segmentation accu-
prompt encoder embeds various input prompts, includ- racy. It is calculated as the area of intersection be-
ing interaction cues like clicks or bounding boxes. The tween the predicted segmentation mask (denoted
mask decoder then predicts object masks using the im- as A) and the ground truth mask (denoted as B), di-
age embedding and prompt embedding. vided by the area of their union. Mathematically,
IoU is defined as:
|A ∩ B|
IoU = .
|A ∪ B|
A variant of the IoU metric is the mean IoU, which
represents the average IoU computed across all
target classes. Another form of generalized in-
tersection over union, proposed by (Rezatofighi
et al., 2019), can serve as the objective function
to optimize in scenarios involving non-overlapping
Figure 32: From Osco et al. (2023), depictions of images segmented
by SAM utilizing point prompts. The initial column displays the RGB
bounding boxes, where traditional IoU methods
image, followed by the second column illustrating the treatment of may be ineffective.
the point prompt. The third column presents the ground-truth mask,
and the fourth column exhibits the prediction result from SAM. The • Dice Coefficient: The Dice Coefficient, a promi-
last column emphasizes the false-positive (FP) pixels identified in the nent metric in image segmentation, particularly in
prediction. medical image analysis, measures the similarity
between two sets. It is commonly used to assess
SAM represents a groundbreaking large-scale model the agreement between a predicted segmentation
with the potential to revolutionize image segmentation, mask and the ground truth mask. The Dice Coeffi-
offering unparalleled flexibility and adaptability that can cient is defined as:
reshape our perception and interaction with the visual
world. Its versatility is demonstrated in several pioneer- 2|A ∩ B|
Dice = .
ing applications in scientific images, such as segment- |A| + |B|
25
The Dice Coefficient ranges from 0 to 1, with 0 in- serve as widely adopted automatic metrics to assess im-
dicating complete dissimilarity and 1 indicating a age quality according to human perceptions. For video
perfect match. This metric is advantageous in tasks generation, Fréchet Video Distance (FVD) (Unterthiner
where the class of interest is a small portion of the et al., 2019), an extension of FID, evaluates video qual-
overall image, effectively addressing class imbal- ity. In audio generation tasks, emphasis is placed on
ance. Similar to IoU in definition, the Dice Coef- beats and rhythms for the generated audio signals (Zhu
ficient is useful for both training and performance et al., 2022a). In applications like image segmentation,
evaluation. evaluations adhere to established norms specific to the
relevant downstream field of research. Despite the di-
• Pixel Accuracy (PA): PA is a metric used in im- versity of automated quantitative assessment methods,
age segmentation tasks to assess the overall accu- human inspection remains a universal benchmark for a
racy of pixel-wise classification. It is calculated as thorough performance evaluation. Given the diversity of
the ratio of correctly classified pixels to the total perspectives, individuals are likely to have varying opin-
number of pixels in the image. Mean Pixel Ac- ions on segmentation results, and even when providing
curacy (mPA) extends this concept by averaging training data, different individuals may produce differ-
PA over each segmentation class. Unlike metrics ent masks, contributing to subjectivity in user opinions
such as IoU and Dice, PA and mPA are primarily and model training. While methods like Mean Opin-
used for evaluation purposes. Pixel Accuracy of- ion Score exist to collect user opinions and rate model
fers a straightforward assessment of the model’s performance, they require substantial participation and
ability to correctly classify individual pixels, re- can be labor-intensive. Only with a large and diverse
gardless of class labels. However, it may not be group of respondents can a more converged opinion be
the most suitable metric for tasks with imbalanced obtained. We acknowledge that both models and indi-
class distributions. This is because PA treats all viduals may exhibit bias in segmentation tasks. Conse-
classes equally and may not provide a comprehen- quently, it is crucial to emphasize that all outputs from
sive assessment of the model’s performance, es- machine learning methods in image segmentation, in-
pecially for minority classes. In such scenarios, cluding those from traditional segmentation methods
metrics like Intersection over Union (IoU) or Dice listed in Section 2, require careful human inspection be-
Coefficient are often preferred for a more nuanced fore proceeding to further analysis.
evaluation.
• In addition to the commonly used metrics and 4. Conclusions
losses discussed earlier, there are other metrics
specifically designed for object segmentation in the In conclusion, our review of segmentation meth-
computer vision field. Metrics like Precision, Re- ods in astronomy has encompassed both classical tech-
call, and F1 scores are frequently used for eval- niques and the state-of-the-art machine learning ap-
uation, while the cross-entropy loss is a common proaches. These methods play a vital role in diverse sci-
choice for training. These metrics are typically de- entific tasks, providing valuable insights into the com-
fined and calculated based on object class annota- plex structures present in astronomical images and data
tions, making them more suitable for multi-object cubes.
semantic segmentation tasks. The emergence of advanced machine learning tech-
niques in the broader computer science community has
It is important to note that differences may exist be- heralded a new era of segmentation methodologies.
tween traditional discriminative machine learning mod- These innovative approaches offer enhanced capabilities
els and generative models in their training and evalua- for precisely delineating objects in astronomical data,
tion processes. Deep generative models aim to estimate marking a significant advancement in the field.
an unknown real-world distribution and establish an ef- As we navigate the evolving landscape of segmenta-
fective mapping between an easy-to-sample prior (e.g., tion methods, we envision a future where the astron-
Gaussian) and the target implicit distribution. While the omy community fully embraces and integrates these
training goals of generative models may align, the eval- advanced techniques into routine research endeavors.
uation protocols are diverse and tailored to specific ap- By harnessing the power of these cutting-edge seg-
plication scenarios. In image generation, where fidelity mentation methods, astronomers can benefit from more
matters, Fréchet Inception Distance (FID) (Heusel et al., precise, reliable, and efficient segmentation outcomes.
2017) and Inception Score (IS) (Salimans et al., 2016) This, in turn, has the potential to alleviate astronomers
26
from laborious manual efforts, allowing them to fo- Krishnarao, D., Lacerna, I., Lan, T.W., Lane, R.R., Law, D.R.,
cus more on interpreting and understanding the intri- Le Goff, J.M., Leung, H.W., Lewis, H., Li, C., Lian, J., Lin, L.,
Long, D., Longa-Peña, P., Lundgren, B., Lyke, B.W., Mackereth,
cate physical processes captured in their datasets. The J.T., MacLeod, C.L., Majewski, S.R., Manchado, A., Maraston,
synergy between advanced segmentation methods and C., Martini, P., Masseron, T., Masters, K.L., Mathur, S., McDer-
astronomical research holds the promise of unlocking mid, R.M., Merloni, A., Merrifield, M., Mészáros, S., Miglio, A.,
deeper insights into the mysteries of the universe. Minniti, D., Minsley, R., Miyaji, T., Mohammad, F.G., Mosser, B.,
Mueller, E.M., Muna, D., Muñoz-Gutiérrez, A., Myers, A.D., Na-
We express our gratitude to the anonymous referees dathur, S., Nair, P., Nandra, K., Correa do Nascimento, J., Nevin,
for their valuable comments and suggestions, particu- R.J., Newman, J.A., Nidever, D.L., Nitschelm, C., Noterdaeme, P.,
larly the references, which have significantly enhanced O’Connell, J.E., Olmstead, M.D., Oravetz, D., Oravetz, A., Oso-
the quality of this review. DX acknowledges support rio, Y., Pace, Z.J., Padilla, N., Palanque-Delabrouille, N., Palicio,
P.A., Pan, H.A., Pan, K., Parker, J., Paviot, R., Peirani, S., Ramŕez,
from the Virginia Initiative on Cosmic Origins (VICO). K.P., Penny, S., Percival, W.J., Perez-Fournon, I., Pérez-Ràfols, I.,
YZ acknowledges support from the VisualAI lab of Petitjean, P., Pieri, M.M., Pinsonneault, M., Poovelil, V.J., Povick,
Princeton University. We recognize the utilization of J.T., Prakash, A., Price-Whelan, A.M., Raddick, M.J., Raichoor,
A., Ray, A., Rembold, S.B., Rezaie, M., Riffel, R.A., Riffel, R.,
ChatGPT, a language model created by OpenAI using
Rix, H.W., Robin, A.C., Roman-Lopes, A., Román-Zúñiga, C.,
the GPT-3.5 architecture, for grammar checking in our Rose, B., Ross, A.J., Rossi, G., Rowlands, K., Rubin, K.H.R.,
review paper. Salvato, M., Sánchez, A.G., Sánchez-Menguiano, L., Sánchez-
Gallego, J.R., Sayres, C., Schaefer, A., Schiavon, R.P., Schimoia,
J.S., Schlafly, E., Schlegel, D., Schneider, D.P., Schultheis, M.,
References Schwope, A., Seo, H.J., Serenelli, A., Shafieloo, A., Shamsi, S.J.,
Shao, Z., Shen, S., Shetrone, M., Shirley, R., Silva Aguirre, V.,
Adithya, H.N., Kariyappa, R., Shinsuke, I., Kanya, K., Zender, J., Simon, J.D., Skrutskie, M.F., Slosar, A., Smethurst, R., Sobeck,
Damé, L., Gabriel, G., DeLuca, E., Weber, M., 2021. Solar Soft J., Sodi, B.C., Souto, D., Stark, D.V., Stassun, K.G., Steinmetz,
X-ray Irradiance Variability, I: Segmentation of Hinode/XRT Full- M., Stello, D., Stermer, J., Storchi-Bergmann, T., Streblyanska,
Disk Images and Comparison with GOES (1 - 8 Å) X-Ray Flux. A., Stringfellow, G.S., Stutz, A., Suárez, G., Sun, J., Taghizadeh-
SoPh 296, 71. doi:10.1007/s11207-021-01785-6. Popp, M., Talbot, M.S., Tayar, J., Thakar, A.R., Theriault, R.,
Ahmadzadeh, A., Mahajan, S.S., Kempton, D.J., Angryk, R.A., Ji, S., Thomas, D., Thomas, Z.C., Tinker, J., Tojeiro, R., Toledo, H.H.,
2019. Toward filament segmentation using deep neural networks, Tremonti, C.A., Troup, N.W., Tuttle, S., Unda-Sanzana, E., Valen-
in: 2019 IEEE International Conference on Big Data (Big Data), tini, M., Vargas-González, J., Vargas-Magaña, M., Vázquez-Mata,
IEEE. pp. 4932–4941. J.A., Vivek, M., Wake, D., Wang, Y., Weaver, B.A., Weijmans,
Ahumada, R., Allende Prieto, C., Almeida, A., Anders, F., Anderson, A.M., Wild, V., Wilson, J.C., Wilson, R.F., Wolthuis, N., Wood-
S.F., Andrews, B.H., Anguiano, B., Arcodia, R., Armengaud, E., Vasey, W.M., Yan, R., Yang, M., Yèche, C., Zamora, O., Zarrouk,
Aubert, M., Avila, S., Avila-Reese, V., Badenes, C., Balland, C., P., Zasowski, G., Zhang, K., Zhao, C., Zhao, G., Zheng, Z., Zheng,
Barger, K., Barrera-Ballesteros, J.K., Basu, S., Bautista, J., Beaton, Z., Zhu, G., Zou, H., 2020. The 16th Data Release of the Sloan
R.L., Beers, T.C., Benavides, B.I.T., Bender, C.F., Bernardi, M., Digital Sky Surveys: First Release from the APOGEE-2 South-
Bershady, M., Beutler, F., Bidin, C.M., Bird, J., Bizyaev, D., ern Survey and Full Release of eBOSS Spectra. ApJS 249, 3.
Blanc, G.A., Blanton, M.R., Boquien, M., Borissova, J., Bovy, doi:10.3847/1538-4365/ab929e, arXiv:1912.02905.
J., Brandt, W.N., Brinkmann, J., Brownstein, J.R., Bundy, K., Bu- Aragon-Calvo, M.A., 2019. Classifying the large-scale structure of
reau, M., Burgasser, A., Burtin, E., Cano-Dı́az, M., Capasso, R., the universe with deep neural networks. MNRAS 484, 5771–5784.
Cappellari, M., Carrera, R., Chabanier, S., Chaplin, W., Chapman, doi:10.1093/mnras/stz393, arXiv:1804.00816.
M., Cherinka, B., Chiappini, C., Doohyun Choi, P., Chojnowski, Austin, J., Johnson, D., Ho, J., Tarlow, D., van den Berg, R., 2021.
S.D., Chung, H., Clerc, N., Coffey, D., Comerford, J.M., Com- Structured denoising diffusion models in discrete state-spaces, in:
parat, J., da Costa, L., Cousinou, M.C., Covey, K., Crane, J.D., Conference on Neural Information Processing Systems (NeurIPS).
Cunha, K., Ilha, G.d.S., Dai, Y.S., Damsted, S.B., Darling, J., Bandyopadhyay, S., Das, S., Datta, A., 2020. Detection of coro-
Davidson, James W., J., Davies, R., Dawson, K., De, N., de la Ma- nal holes using hough simulated parameterized online region-
corra, A., De Lee, N., Queiroz, A.B.d.A., Deconto Machado, A., based active contour method, in: 2020 URSI Regional Confer-
de la Torre, S., Dell’Agli, F., du Mas des Bourboux, H., Diamond- ence on Radio Science ( URSI-RCRS), pp. 1–4. doi:10.23919/
Stanic, A.M., Dillon, S., Donor, J., Drory, N., Duckworth, C., URSIRCRS49211.2020.9113512.
Dwelly, T., Ebelke, G., Eftekharzadeh, S., Davis Eigenbrot, A., Barra, V., Delouille, V., Kretzschmar, M., Hochedez, J.F., 2009. Fast
Elsworth, Y.P., Eracleous, M., Erfanianfar, G., Escoffier, S., Fan, and robust segmentation of solar euv images: algorithm and results
X., Farr, E., Fernández-Trincado, J.G., Feuillet, D., Finoguenov, for solar cycle 23. Astronomy & Astrophysics 505, 361–371.
A., Fofie, P., Fraser-McKelvie, A., Frinchaboy, P.M., Fromenteau, Barra, V., Delouille, V., Kretzschmar, M., Hochedez, J.F., 2009. Fast
S., Fu, H., Galbany, L., Garcia, R.A., Garcı́a-Hernández, D.A., and robust segmentation of solar EUV images: algorithm and re-
Garma Oehmichen, L.A., Ge, J., Geimba Maia, M.A., Geisler, sults for solar cycle 23. A&A 505, 361–371. doi:10.1051/
D., Gelfand, J., Goddy, J., Gonzalez-Perez, V., Grabowski, K., 0004-6361/200811416.
Green, P., Grier, C.J., Guo, H., Guy, J., Harding, P., Hasselquist, Bekki, K., 2021. Quantifying the fine structures of disk galaxies with
S., Hawken, A.J., Hayes, C.R., Hearty, F., Hekker, S., Hogg, deep learning: Segmentation of spiral arms in different Hubble
D.W., Holtzman, J.A., Horta, D., Hou, J., Hsieh, B.C., Huber, types. A&A 647, A120. doi:10.1051/0004-6361/202039797,
D., Hunt, J.A.S., Ider Chitham, J., Imig, J., Jaber, M., Jimenez arXiv:2103.08127.
Angel, C.E., Johnson, J.A., Jones, A.M., Jönsson, H., Jullo, E., Belavin, V., Trofimova, E., Ustyuzhanin, A., 2021. Segmen-
Kim, Y., Kinemuchi, K., Kirkpatrick, Charles C., I., Kite, G.W., tation of em showers for neutrino experiments with deep
Klaene, M., Kneib, J.P., Kollmeier, J.A., Kong, H., Kounkel, M., graph neural networks. Journal of Instrumentation 16,
27
P12035. URL: https://dx.doi.org/10.1088/1748-0221/ Sembroski, G.H., Lin, J.Y.Y., 2019. Deblending and clas-
16/12/P12035, doi:10.1088/1748-0221/16/12/P12035. sifying astronomical sources with Mask R-CNN deep learn-
Bellm, E.C., Kulkarni, S.R., Graham, M.J., Dekany, R., Smith, R.M., ing. MNRAS 490, 3952–3965. doi:10.1093/mnras/stz2845,
Riddle, R., Masci, F.J., Helou, G., Prince, T.A., Adams, S.M., Bar- arXiv:1908.02748.
barino, C., Barlow, T., Bauer, J., Beck, R., Belicki, J., Biswas, R., Cao, H., Wang, Y., Chen, J., Jiang, D., Zhang, X., Tian, Q., Wang,
Blagorodnova, N., Bodewits, D., Bolin, B., Brinnel, V., Brooke, M., 2022. Swin-unet: Unet-like pure transformer for medical im-
T., Bue, B., Bulla, M., Burruss, R., Cenko, S.B., Chang, C.K., age segmentation, in: European conference on computer vision,
Connolly, A., Coughlin, M., Cromer, J., Cunningham, V., De, K., Springer. pp. 205–218.
Delacroix, A., Desai, V., Duev, D.A., Eadie, G., Farnham, T.L., Chen, J., Lu, J., Zhu, X., Zhang, L., 2023a. Generative semantic
Feeney, M., Feindt, U., Flynn, D., Franckowiak, A., Frederick, segmentation, in: Proceedings of the IEEE/CVF Conference on
S., Fremling, C., Gal-Yam, A., Gezari, S., Giomi, M., Goldstein, Computer Vision and Pattern Recognition, pp. 7111–7120.
D.A., Golkhou, V.Z., Goobar, A., Groom, S., Hacopians, E., Hale, Chen, J., Lu, Y., Yu, Q., Luo, X., Adeli, E., Wang, Y., Lu, L.,
D., Henning, J., Ho, A.Y.Q., Hover, D., Howell, J., Hung, T., Hup- Yuille, A.L., Zhou, Y., 2021. Transunet: Transformers make
penkothen, D., Imel, D., Ip, W.H., Ivezić, Ž., Jackson, E., Jones, strong encoders for medical image segmentation. arXiv preprint
L., Juric, M., Kasliwal, M.M., Kaspi, S., Kaye, S., Kelley, M.S.P., arXiv:2102.04306 .
Kowalski, M., Kramer, E., Kupfer, T., Landry, W., Laher, R.R., Chen, Z., Zheng, Y., Li, C., Zhan, Y., 2023b. Fast and robust star
Lee, C.D., Lin, H.W., Lin, Z.Y., Lunnan, R., Giomi, M., Maha- detection algorithm based on the dyadic wavelet transform. IET
bal, A., Mao, P., Miller, A.A., Monkewitz, S., Murphy, P., Ngeow, Image Processing 17, 944–955.
C.C., Nordin, J., Nugent, P., Ofek, E., Patterson, M.T., Penprase, Chen, Z., Zhou, W., Sun, G., Zhang, M., Ruan, J., Zhao, J.,
B., Porter, M., Rauch, L., Rebbapragada, U., Reiley, D., Rigault, 2023. TransientViT: A novel CNN - Vision Transformer hybrid
M., Rodriguez, H., van Roestel, J., Rusholme, B., van Santen, J., real/bogus transient classifier for the Kilodegree Automatic Tran-
Schulze, S., Shupe, D.L., Singer, L.P., Soumagnac, M.T., Stein, sient Survey. arXiv e-prints , arXiv:2309.09937doi:10.48550/
R., Surace, J., Sollerman, J., Szkody, P., Taddia, F., Terek, S., arXiv.2309.09937, arXiv:2309.09937.
Van Sistine, A., van Velzen, S., Vestrand, W.T., Walters, R., Ward, Chollet, F., 2017. Xception: Deep learning with depthwise separable
C., Ye, Q.Z., Yu, P.C., Yan, L., Zolkower, J., 2019. The Zwicky convolutions, in: Proceedings of the IEEE conference on computer
Transient Facility: System Overview, Performance, and First Re- vision and pattern recognition, pp. 1251–1258.
sults. PASP 131, 018002. doi:10.1088/1538-3873/aaecbe, Collet, C., Murtagh, F., 2004. Multiband segmentation based on
arXiv:1902.01932. a hierarchical markov model. Pattern Recognition 37, 2337–
Berry, D.S., 2015. FellWalker-A clump identification algorithm. 2347. URL: https://www.sciencedirect.com/science/
Astronomy and Computing 10, 22–31. doi:10.1016/j.ascom. article/pii/S0031320304001906, doi:https://doi.org/
2014.11.004, arXiv:1411.6267. 10.1016/j.patcog.2004.03.017.
Bertin, E., Arnouts, S., 1996. SExtractor: Software for source extrac- Colombo, D., Rosolowsky, E., Ginsburg, A., Duarte-Cabral, A.,
tion. A&AS 117, 393–404. doi:10.1051/aas:1996164. Hughes, A., 2015. Graph-based interpretation of the molecu-
Bhambra, P., Joachimi, B., Lahav, O., 2022. Explaining lar interstellar medium segmentation. MNRAS 454, 2067–2091.
deep learning of galaxy morphology with saliency mapping. doi:10.1093/mnras/stv2063, arXiv:1510.04253.
MNRAS 511, 5032–5041. doi:10.1093/mnras/stac368, Covas, P.B., Prix, R., 2022. Improved short-segment detec-
arXiv:2110.08288. tion statistic for continuous gravitational waves. Phys. Rev.
Bijaoui, A., Rué, F., 1995. A multiscale vision model D 105, 124007. URL: https://link.aps.org/doi/10.
adapted to the astronomical images. Signal Processing 46, 1103/PhysRevD.105.124007, doi:10.1103/PhysRevD.105.
345–362. URL: https://www.sciencedirect.com/ 124007.
science/article/pii/0165168495000934, doi:https: Dai, Y., Zheng, T., Xue, C., Zhou, L., 2022. Seg-
//doi.org/10.1016/0165-1684(95)00093-4. marsvit: Lightweight mars terrain segmentation network for au-
Boucaud, A., Huertas-Company, M., Heneka, C., Ishida, E.E.O., tonomous driving in planetary exploration. Remote Sensing
Sedaghat, N., de Souza, R.S., Moews, B., Dole, H., Castellano, 14. URL: https://www.mdpi.com/2072-4292/14/24/6297,
M., Merlin, E., Roscani, V., Tramacere, A., Killedar, M., Trindade, doi:10.3390/rs14246297.
A.M.M., Collaboration COIN, 2020. Photometry of high-redshift Dey, A., Schlegel, D.J., Lang, D., Blum, R., Burleigh, K., Fan, X.,
blended galaxies using deep learning. MNRAS 491, 2481–2495. Findlay, J.R., Finkbeiner, D., Herrera, D., Juneau, S., Landriau,
doi:10.1093/mnras/stz3056, arXiv:1905.01324. M., Levi, M., McGreer, I., Meisner, A., Myers, A.D., Moustakas,
Boucheron, L.E., Valluri, M., McAteer, R.T.J., 2016. Segmenta- J., Nugent, P., Patej, A., Schlafly, E.F., Walker, A.R., Valdes, F.,
tion of Coronal Holes Using Active Contours Without Edges. Weaver, B.A., Yèche, C., Zou, H., Zhou, X., Abareshi, B., Ab-
SoPh 291, 2353–2372. doi:10.1007/s11207-016-0985-z, bott, T.M.C., Abolfathi, B., Aguilera, C., Alam, S., Allen, L., Al-
arXiv:1610.01023. varez, A., Annis, J., Ansarinejad, B., Aubert, M., Beechert, J.,
Boursier, Y., Llebaria, A., Goudail, F., Lamy, P., Robelus, S., 2005. Bell, E.F., BenZvi, S.Y., Beutler, F., Bielby, R.M., Bolton, A.S.,
Automatic detection of coronal mass ejections on LASCO-C2 syn- Briceño, C., Buckley-Geer, E.J., Butler, K., Calamida, A., Carl-
optic maps, in: Fineschi, S., Viereck, R.A. (Eds.), Solar Physics berg, R.G., Carter, P., Casas, R., Castander, F.J., Choi, Y., Com-
and Space Weather Instrumentation, pp. 13–24. doi:10.1117/12. parat, J., Cukanovaite, E., Delubac, T., DeVries, K., Dey, S., Dhun-
616011. gana, G., Dickinson, M., Ding, Z., Donaldson, J.B., Duan, Y.,
Bretonnière, H., Boucaud, A., Huertas-Company, M., 2021. Proba- Duckworth, C.J., Eftekharzadeh, S., Eisenstein, D.J., Etourneau,
bilistic segmentation of overlapping galaxies for large cosmolog- T., Fagrelius, P.A., Farihi, J., Fitzpatrick, M., Font-Ribera, A., Ful-
ical surveys. arXiv e-prints , arXiv:2111.15455doi:10.48550/ mer, L., Gänsicke, B.T., Gaztanaga, E., George, K., Gerdes, D.W.,
arXiv.2111.15455, arXiv:2111.15455. Gontcho, S.G.A., Gorgoni, C., Green, G., Guy, J., Harmer, D.,
Buonanno, R., Buscema, G., Corsi, C.E., Ferraro, I., Iannicola, G., Hernandez, M., Honscheid, K., Huang, L.W., James, D.J., Jan-
1983. Automated photographic photometry of stars in globular nuzi, B.T., Jiang, L., Joyce, R., Karcher, A., Karkar, S., Kehoe,
clusters. A&A 126, 278–282. R., Kneib, J.P., Kueter-Young, A., Lan, T.W., Lauer, T.R., Le Guil-
Burke, C.J., Aleo, P.D., Chen, Y.C., Liu, X., Peterson, J.R., lou, L., Le Van Suu, A., Lee, J.H., Lesser, M., Perreault Levasseur,
28
L., Li, T.S., Mann, J.L., Marshall, R., Martı́nez-Vázquez, C.E., Drimmel, R., Katz, D., Lattanzi, M.G., van Leeuwen, F., Bakker,
Martini, P., du Mas des Bourboux, H., McManus, S., Meier, T.G., J., Cacciari, C., Castañeda, J., De Angeli, F., Ducourant, C.,
Ménard, B., Metcalfe, N., Muñoz-Gutiérrez, A., Najita, J., Napier, Fabricius, C., Fouesneau, M., Frémat, Y., Guerra, R., Guerrier,
K., Narayan, G., Newman, J.A., Nie, J., Nord, B., Norman, D.J., A., Guiraud, J., Jean-Antoine Piccolo, A., Masana, E., Messi-
Olsen, K.A.G., Paat, A., Palanque-Delabrouille, N., Peng, X., Pop- neo, R., Mowlavi, N., Nicolas, C., Nienartowicz, K., Pailler, F.,
pett, C.L., Poremba, M.R., Prakash, A., Rabinowitz, D., Raichoor, Panuzzo, P., Riclet, F., Roux, W., Seabroke, G.M., Sordo, R.,
A., Rezaie, M., Robertson, A.N., Roe, N.A., Ross, A.J., Ross, N.P., Tanga, P., Thévenin, F., Gracia-Abril, G., Portell, J., Teyssier, D.,
Rudnick, G., Safonova, S., Saha, A., Sánchez, F.J., Savary, E., Altmann, M., Andrae, R., Bellas-Velidis, I., Benson, K., Berthier,
Schweiker, H., Scott, A., Seo, H.J., Shan, H., Silva, D.R., Slepian, J., Blomme, R., Brugaletta, E., Burgess, P.W., Busso, G., Carry,
Z., Soto, C., Sprayberry, D., Staten, R., Stillman, C.M., Stupak, B., Cellino, A., Cheek, N., Clementini, G., Damerdji, Y., David-
R.J., Summers, D.L., Sien Tie, S., Tirado, H., Vargas-Magaña, son, M., Delchambre, L., Dell’Oro, A., Fernández-Hernández, J.,
M., Vivas, A.K., Wechsler, R.H., Williams, D., Yang, J., Yang, Galluccio, L., Garcı́a-Lario, P., Garcia-Reinaldos, M., González-
Q., Yapici, T., Zaritsky, D., Zenteno, A., Zhang, K., Zhang, T., Núñez, J., Gosset, E., Haigron, R., Halbwachs, J.L., Hambly,
Zhou, R., Zhou, Z., 2019. Overview of the DESI Legacy Imag- N.C., Harrison, D.L., Hatzidimitriou, D., Heiter, U., Hernández,
ing Surveys. AJ 157, 168. doi:10.3847/1538-3881/ab089d, J., Hestroffer, D., Hodgkin, S.T., Holl, B., Janßen, K., Jevardat
arXiv:1804.08657. de Fombelle, G., Jordan, S., Krone-Martins, A., Lanzafame, A.C.,
Domı́nguez Sánchez, H., Huertas-Company, M., Bernardi, M., Kavi- Löffler, W., Lorca, A., Manteiga, M., Marchal, O., Marrese, P.M.,
raj, S., Fischer, J.L., Abbott, T.M.C., Abdalla, F.B., Annis, J., Moitinho, A., Mora, A., Muinonen, K., Osborne, P., Pancino, E.,
Avila, S., Brooks, D., Buckley-Geer, E., Carnero Rosell, A., Car- Pauwels, T., Petit, J.M., Recio-Blanco, A., Richards, P.J., Riello,
rasco Kind, M., Carretero, J., Cunha, C.E., D’Andrea, C.B., da M., Rimoldini, L., Robin, A.C., Roegiers, T., Rybizki, J., Sarro,
Costa, L.N., Davis, C., De Vicente, J., Doel, P., Evrard, A.E., Fos- L.M., Siopis, C., Smith, M., Sozzetti, A., Ulla, A., Utrilla, E.,
alba, P., Frieman, J., Garcı́a-Bellido, J., Gaztanaga, E., Gerdes, van Leeuwen, M., van Reeven, W., Abbas, U., Abreu Aramburu,
D.W., Gruen, D., Gruendl, R.A., Gschwend, J., Gutierrez, G., Hart- A., Accart, S., Aerts, C., Aguado, J.J., Ajaj, M., Altavilla, G.,
ley, W.G., Hollowood, D.L., Honscheid, K., Hoyle, B., James, D.J., Álvarez, M.A., Álvarez Cid-Fuentes, J., Alves, J., Anderson, R.I.,
Kuehn, K., Kuropatkin, N., Lahav, O., Maia, M.A.G., March, M., Anglada Varela, E., Antoja, T., Audard, M., Baines, D., Baker,
Melchior, P., Menanteau, F., Miquel, R., Nord, B., Plazas, A.A., S.G., Balaguer-Núñez, L., Balbinot, E., Balog, Z., Barache, C.,
Sanchez, E., Scarpine, V., Schindler, R., Schubnell, M., Smith, M., Barbato, D., Barros, M., Barstow, M.A., Bartolomé, S., Bassilana,
Smith, R.C., Soares-Santos, M., Sobreira, F., Suchyta, E., Swan- J.L., Bauchet, N., Baudesson-Stella, A., Becciani, U., Bellazzini,
son, M.E.C., Tarle, G., Thomas, D., Walker, A.R., Zuntz, J., 2019. M., Bernet, M., Bertone, S., Bianchi, L., Blanco-Cuaresma, S.,
Transfer learning for galaxy morphology from one survey to an- Boch, T., Bombrun, A., Bossini, D., Bouquillon, S., Bragaglia, A.,
other. MNRAS 484, 93–100. doi:10.1093/mnras/sty3497, Bramante, L., Breedt, E., Bressan, A., Brouillet, N., Bucciarelli,
arXiv:1807.00807. B., Burlacu, A., Busonero, D., Butkevich, A.G., Buzzi, R., Caf-
Dorotovic, I., Shahamatnia, E., Lorenc, M., Rybansky, M., Ribeiro, fau, E., Cancelliere, R., Cánovas, H., Cantat-Gaudin, T., Carballo,
R.A., Fonseca, J.M., 2014. Sunspots and Coronal Bright Points R., Carlucci, T., Carnerero, M.I., Carrasco, J.M., Casamiquela, L.,
Tracking using a Hybrid Algorithm of PSO and Active Contour Castellani, M., Castro-Ginard, A., Castro Sampol, P., Chaoul, L.,
Model. Sun and Geosphere 9, 81–84. Charlot, P., Chemin, L., Chiavassa, A., Cioni, M.R.L., Comoretto,
Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., G., Cooper, W.J., Cornez, T., Cowell, S., Crifo, F., Crosta, M.,
Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, Crowley, C., Dafonte, C., Dapergolas, A., David, M., David, P.,
S., et al., 2020. An image is worth 16x16 words: Transformers for de Laverny, P., De Luise, F., De March, R., De Ridder, J., de
image recognition at scale. arXiv preprint arXiv:2010.11929 . Souza, R., de Teodoro, P., de Torres, A., del Peloso, E.F., del
Ellien, A., Slezak, E., Martinet, N., Durret, F., Adami, C., Gavazzi, R., Pozo, E., Delbo, M., Delgado, A., Delgado, H.E., Delisle, J.B.,
Rabaça, C., Da Rocha, C., Pereira, D.E., 2021. Dawis: a detection Di Matteo, P., Diakite, S., Diener, C., Distefano, E., Dolding,
algorithm with wavelets for intracluster light studies. Astronomy C., Eappachen, D., Edvardsson, B., Enke, H., Esquej, P., Fabre,
& Astrophysics 649, A38. C., Fabrizio, M., Faigler, S., Fedorets, G., Fernique, P., Fienga,
Farias, H., Ortiz, D., Damke, G., Jaque Arancibia, M., Solar, M., A., Figueras, F., Fouron, C., Fragkoudi, F., Fraile, E., Franke,
2020. Mask galaxy: Morphological segmentation of galaxies. As- F., Gai, M., Garabato, D., Garcia-Gutierrez, A., Garcı́a-Torres,
tronomy and Computing 33, 100420. doi:10.1016/j.ascom. M., Garofalo, A., Gavras, P., Gerlach, E., Geyer, R., Giacobbe,
2020.100420. P., Gilmore, G., Girona, S., Giuffrida, G., Gomel, R., Gomez,
Flewelling, H.A., Magnier, E.A., Chambers, K.C., Heasley, J.N., A., Gonzalez-Santamaria, I., González-Vidal, J.J., Granvik, M.,
Holmberg, C., Huber, M.E., Sweeney, W., Waters, C.Z., Calamida, Gutiérrez-Sánchez, R., Guy, L.P., Hauser, M., Haywood, M.,
A., Casertano, S., Chen, X., Farrow, D., Hasinger, G., Hender- Helmi, A., Hidalgo, S.L., Hilger, T., Hładczuk, N., Hobbs, D.,
son, R., Long, K.S., Metcalfe, N., Narayan, G., Nieto-Santisteban, Holland, G., Huckle, H.E., Jasniewicz, G., Jonker, P.G., Juaristi
M.A., Norberg, P., Rest, A., Saglia, R.P., Szalay, A., Thakar, Campillo, J., Julbe, F., Karbevska, L., Kervella, P., Khanna, S., Ko-
A.R., Tonry, J.L., Valenti, J., Werner, S., White, R., Denneau, choska, A., Kontizas, M., Kordopatis, G., Korn, A.J., Kostrzewa-
L., Draper, P.W., Hodapp, K.W., Jedicke, R., Kaiser, N., Ku- Rutkowska, Z., Kruszyńska, K., Lambert, S., Lanza, A.F., Lasne,
dritzki, R.P., Price, P.A., Wainscoat, R.J., Chastel, S., McLean, B., Y., Le Campion, J.F., Le Fustec, Y., Lebreton, Y., Lebzelter, T.,
Postman, M., Shiao, B., 2020. The Pan-STARRS1 Database and Leccia, S., Leclerc, N., Lecoeur-Taibi, I., Liao, S., Licata, E.,
Data Products. ApJS 251, 7. doi:10.3847/1538-4365/abb82d, Lindstrøm, E.P., Lister, T.A., Livanou, E., Lobel, A., Madrero
arXiv:1612.05243. Pardo, P., Managau, S., Mann, R.G., Marchant, J.M., Marconi,
Gaia Collaboration, Brown, A.G.A., Vallenari, A., Prusti, T., de Brui- M., Marcos Santos, M.M.S., Marinoni, S., Marocco, F., Marshall,
jne, J.H.J., Babusiaux, C., Biermann, M., Creevey, O.L., Evans, D.J., Martin Polo, L., Martı́n-Fleitas, J.M., Masip, A., Massari, D.,
D.W., Eyer, L., Hutton, A., Jansen, F., Jordi, C., Klioner, S.A., Mastrobuono-Battisti, A., Mazeh, T., McMillan, P.J., Messina, S.,
Lammers, U., Lindegren, L., Luri, X., Mignard, F., Panem, C., Michalik, D., Millar, N.R., Mints, A., Molina, D., Molinaro, R.,
Pourbaix, D., Randich, S., Sartoretti, P., Soubiran, C., Walton, Molnár, L., Montegriffo, P., Mor, R., Morbidelli, R., Morel, T.,
N.A., Arenou, F., Bailer-Jones, C.A.L., Bastian, U., Cropper, M., Morris, D., Mulone, A.F., Munoz, D., Muraveva, T., Murphy, C.P.,
29
Musella, I., Noval, L., Ordénovic, C., Orrù, G., Osinde, J., Pagani, age Data. ApJS 248, 20. doi:10.3847/1538-4365/ab8868,
C., Pagano, I., Palaversa, L., Palicio, P.A., Panahi, A., Pawlak, M., arXiv:1906.11248.
Peñalosa Esteller, X., Penttilä, A., Piersimoni, A.M., Pineau, F.X., He, K., Gkioxari, G., Dollár, P., Girshick, R., 2017. Mask r-cnn,
Plachy, E., Plum, G., Poggio, E., Poretti, E., Poujoulet, E., Prša, in: Proceedings of the IEEE international conference on computer
A., Pulone, L., Racero, E., Ragaini, S., Rainer, M., Raiteri, C.M., vision, pp. 2961–2969.
Rambaux, N., Ramos, P., Ramos-Lerate, M., Re Fiorentin, P., Reg- He, K., Zhang, X., Ren, S., Sun, J., 2016. Deep residual learning
ibo, S., Reylé, C., Ripepi, V., Riva, A., Rixon, G., Robichon, N., for image recognition, in: Proceedings of the IEEE conference on
Robin, C., Roelens, M., Rohrbasser, L., Romero-Gómez, M., Row- computer vision and pattern recognition, pp. 770–778.
ell, N., Royer, F., Rybicki, K.A., Sadowski, G., Sagristà Sellés, Henry, E.U., Emebob, O., Omonhinmin, C.A., 2022. Vision
A., Sahlmann, J., Salgado, J., Salguero, E., Samaras, N., Sanchez transformers in medical imaging: A review. arXiv preprint
Gimenez, V., Sanna, N., Santoveña, R., Sarasso, M., Schultheis, arXiv:2211.10043 .
M., Sciacca, E., Segol, M., Segovia, J.C., Ségransan, D., Semeux, Herzog, A.D., Illingworth, G., 1977. The Structure of Globular Clus-
D., Shahaf, S., Siddiqui, H.I., Siebert, A., Siltala, L., Slezak, E., ters. I. Direct Plane Automated Reduction Techniques. ApJS 33,
Smart, R.L., Solano, E., Solitro, F., Souami, D., Souchay, J., 55. doi:10.1086/190418.
Spagna, A., Spoto, F., Steele, I.A., Steidelmüller, H., Stephen- Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., Hochreiter,
son, C.A., Süveges, M., Szabados, L., Szegedi-Elek, E., Taris, F., S., 2017. Gans trained by a two time-scale update rule converge
Tauran, G., Taylor, M.B., Teixeira, R., Thuillot, W., Tonello, N., to a local nash equilibrium. Conference on Neural Information
Torra, F., Torra, J., Turon, C., Unger, N., Vaillant, M., van Dillen, Processing Systems (NeurIPS) 30.
E., Vanel, O., Vecchiato, A., Viala, Y., Vicente, D., Voutsinas, S., Ho, J., Jain, A., Abbeel, P., 2020. Denoising diffusion probabilistic
Weiler, M., Wevers, T., Wyrzykowski, Ł., Yoldas, A., Yvard, P., models, in: NeurIPS.
Zhao, H., Zorec, J., Zucker, S., Zurbach, C., Zwitter, T., 2021. Gaia Hopkins, A.M., Miller, C.J., Connolly, A.J., Genovese, C., Nichol,
Early Data Release 3. Summary of the contents and survey prop- R.C., Wasserman, L., 2002. A New Source Detection Algorithm
erties. A&A 649, A1. doi:10.1051/0004-6361/202039657, Using the False-Discovery Rate. AJ 123, 1086–1094. doi:10.
arXiv:2012.01533. 1086/338316, arXiv:astro-ph/0110570.
Gill, C.D., Fletcher, L., Marshall, S., 2010. Using Active Contours for Huang, K.W., Chih-Fan Chen, G., Chang, P.W., Lin, S.C., Hsu,
Semi-Automated Tracking of UV and EUV Solar Flare Ribbons. C.J., Thengane, V., Yao-Yu Lin, J., 2022. Strong Gravitational
SoPh 262, 355–371. doi:10.1007/s11207-010-9508-5. Lensing Parameter Estimation with Vision Transformer. arXiv
Girshick, R., 2015. Fast r-cnn, in: Proceedings of the IEEE interna- e-prints , arXiv:2210.04143doi:10.48550/arXiv.2210.04143,
tional conference on computer vision, pp. 1440–1448. arXiv:2210.04143.
Gong, C., Long, T., Yin, R., Jiao, W., Wang, G., 2023. A hybrid Huertas-Company, M., Lanusse, F., 2023. The Dawes Review 10: The
algorithm with swin transformer and convolution for cloud de- impact of deep learning for the analysis of galaxy surveys. PASA
tection. Remote Sensing 15. URL: https://www.mdpi.com/ 40, e001. doi:10.1017/pasa.2022.55, arXiv:2210.01813.
2072-4292/15/21/5264, doi:10.3390/rs15215264. Irwin, M.J., 1985. Automatic analysis of crowded fields. MNRAS
Gonzalez-Sanchez, E., Saccardo, D., Esteves, P.B., Kuffa, M., We- 214, 575–604. doi:10.1093/mnras/214.4.575.
gener, K., 2024. Automatic characterization of wedm single craters Ivezić, Ž., Kahn, S.M., Tyson, J.A., Abel, B., Acosta, E., Allsman,
through ai based object detection. International Journal of Automa- R., Alonso, D., AlSayyad, Y., Anderson, S.F., Andrew, J., Angel,
tion Technology 18, 265–275. J.R.P., Angeli, G.Z., Ansari, R., Antilogus, P., Araujo, C., Arm-
Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, strong, R., Arndt, K.T., Astier, P., Aubourg, É., Auza, N., Axel-
D., Ozair, S., Courville, A., Bengio, Y., 2014. Generative adver- rod, T.S., Bard, D.J., Barr, J.D., Barrau, A., Bartlett, J.G., Bauer,
sarial nets, in: NeurIPS. A.E., Bauman, B.J., Baumont, S., Bechtol, E., Bechtol, K., Becker,
Grajeda, J.A., Boucheron, L.E., Kirk, M.S., Leisner, A., Arge, C.N., A.C., Becla, J., Beldica, C., Bellavia, S., Bianco, F.B., Biswas,
2023. Quantifying the consistency and characterizing the confi- R., Blanc, G., Blazek, J., Blandford, R.D., Bloom, J.S., Bogart,
dence of coronal holes detected by active contours without edges J., Bond, T.W., Booth, M.T., Borgland, A.W., Borne, K., Bosch,
(acwe). Solar Physics 298, 133. J.F., Boutigny, D., Brackett, C.A., Bradshaw, A., Brandt, W.N.,
Green, C.E., Dawson, J.R., Cunningham, M.R., Jones, P.A., No- Brown, M.E., Bullock, J.S., Burchat, P., Burke, D.L., Cagnoli,
vak, G., Fissel, L.M., 2017. Measuring Filament Orientation: A G., Calabrese, D., Callahan, S., Callen, A.L., Carlin, J.L., Carl-
New Quantitative, Local Approach. ApJS 232, 6. doi:10.3847/ son, E.L., Chandrasekharan, S., Charles-Emerson, G., Chesley,
1538-4365/aa8507, arXiv:1708.01953. S., Cheu, E.C., Chiang, H.F., Chiang, J., Chirino, C., Chow,
Gu, A., Dao, T., 2023. Mamba: Linear-time sequence modeling with D., Ciardi, D.R., Claver, C.F., Cohen-Tanugi, J., Cockrum, J.J.,
selective state spaces. arXiv preprint arXiv:2312.00752 . Coles, R., Connolly, A.J., Cook, K.H., Cooray, A., Covey, K.R.,
Gu, M., Wang, F., Hu, T., Yu, S., 2023. Localization and segmen- Cribbs, C., Cui, W., Cutri, R., Daly, P.N., Daniel, S.F., Daruich, F.,
tation of galaxy morphology based on mask r-cnn, in: 2023 4th Daubard, G., Daues, G., Dawson, W., Delgado, F., Dellapenna,
International Conference on Computer Engineering and Applica- A., de Peyster, R., de Val-Borro, M., Digel, S.W., Doherty, P.,
tion (ICCEA), pp. 512–515. doi:10.1109/ICCEA58433.2023. Dubois, R., Dubois-Felsmann, G.P., Durech, J., Economou, F.,
10135337. Eifler, T., Eracleous, M., Emmons, B.L., Fausti Neto, A., Fergu-
Hale, C.L., Robotham, A.S.G., Davies, L.J.M., Jarvis, M.J., Driver, son, H., Figueroa, E., Fisher-Levine, M., Focke, W., Foss, M.D.,
S.P., Heywood, I., 2019. Radio source extraction with PRO- Frank, J., Freemon, M.D., Gangler, E., Gawiser, E., Geary, J.C.,
FOUND. MNRAS 487, 3971–3989. doi:10.1093/mnras/ Gee, P., Geha, M., Gessner, C.J.B., Gibson, R.R., Gilmore, D.K.,
stz1462, arXiv:1902.01440. Glanzman, T., Glick, W., Goldina, T., Goldstein, D.A., Goodenow,
Hancock, P.J., Murphy, T., Gaensler, B.M., Hopkins, A., Curran, J.R., I., Graham, M.L., Gressler, W.J., Gris, P., Guy, L.P., Guyonnet,
2012. Compact continuum source finding for next generation radio A., Haller, G., Harris, R., Hascall, P.A., Haupt, J., Hernandez, F.,
surveys. MNRAS 422, 1812–1824. doi:10.1111/j.1365-2966. Herrmann, S., Hileman, E., Hoblitt, J., Hodgson, J.A., Hogan, C.,
2012.20768.x, arXiv:1202.4500. Howard, J.D., Huang, D., Huffer, M.E., Ingraham, P., Innes, W.R.,
Hausen, R., Robertson, B.E., 2020. Morpheus: A Deep Learn- Jacoby, S.H., Jain, B., Jammes, F., Jee, M.J., Jenness, T., Jernigan,
ing Framework for the Pixel-level Analysis of Astronomical Im- G., Jevremović, D., Johns, K., Johnson, A.S., Johnson, M.W.G.,
30
Jones, R.L., Juramy-Gilles, C., Jurić, M., Kalirai, J.S., Kallivay- O., 2018. A probabilistic u-net for segmentation of ambiguous
alil, N.J., Kalmbach, B., Kantor, J.P., Karst, P., Kasliwal, M.M., images. Advances in neural information processing systems 31.
Kelly, H., Kessler, R., Kinnison, V., Kirkby, D., Knox, L., Kotov, Kohonen, T., 1990. The self-organizing map. Proceedings of the IEEE
I.V., Krabbendam, V.L., Krughoff, K.S., Kubánek, P., Kuczewski, 78, 1464–1480.
J., Kulkarni, S., Ku, J., Kurita, N.R., Lage, C.S., Lambert, R., Krizhevsky, A., Sutskever, I., Hinton, G.E., 2012. Imagenet classi-
Lange, T., Langton, J.B., Le Guillou, L., Levine, D., Liang, M., fication with deep convolutional neural networks, in: Pereira, F.,
Lim, K.T., Lintott, C.J., Long, K.E., Lopez, M., Lotz, P.J., Lupton, Burges, C., Bottou, L., Weinberger, K. (Eds.), Advances in Neural
R.H., Lust, N.B., MacArthur, L.A., Mahabal, A., Mandelbaum, Information Processing Systems, Curran Associates, Inc.
R., Markiewicz, T.W., Marsh, D.S., Marshall, P.J., Marshall, S., Latorre, F., Spiller, D., Sasidharan, S.T., Basheer, S., Curti, F., 2023.
May, M., McKercher, R., McQueen, M., Meyers, J., Migliore, Transfer learning for real-time crater detection on asteroids using a
M., Miller, M., Mills, D.J., Miraval, C., Moeyens, J., Moolekamp, Fully Convolutional Neural Network. Icarus 394, 115434. doi:10.
F.E., Monet, D.G., Moniez, M., Monkewitz, S., Montgomery, C., 1016/j.icarus.2023.115434, arXiv:2204.00477.
Morrison, C.B., Mueller, F., Muller, G.P., Muñoz Arancibia, F., Lazzati, D., Campana, S., Rosati, P., Panzera, M.R., Tagliaferri, G.,
Neill, D.R., Newbry, S.P., Nief, J.Y., Nomerotski, A., Nordby, M., 1999. The Brera Multiscale Wavelet ROSAT HRI Source Cata-
O’Connor, P., Oliver, J., Olivier, S.S., Olsen, K., O’Mullane, W., log. I. The Algorithm. ApJ 524, 414–422. doi:10.1086/307788,
Ortiz, S., Osier, S., Owen, R.E., Pain, R., Palecek, P.E., Pare- arXiv:astro-ph/9904374.
jko, J.K., Parsons, J.B., Pease, N.M., Peterson, J.M., Peterson, Liu, D., Song, W., Lin, G., Wang, H., 2021. Solar Filament Segmen-
J.R., Petravick, D.L., Libby Petrick, M.E., Petry, C.E., Pierfed- tation Based on Improved U-Nets. SoPh 296, 176. doi:10.1007/
erici, F., Pietrowicz, S., Pike, R., Pinto, P.A., Plante, R., Plate, s11207-021-01920-3.
S., Plutchak, J.P., Price, P.A., Prouza, M., Radeka, V., Rajagopal, Mackovjak, Š., Harman, M., Maslej-Krešňáková, V., Butka, P., 2021.
J., Rasmussen, A.P., Regnault, N., Reil, K.A., Reiss, D.J., Reuter, SCSS-Net: solar corona structures segmentation by deep learn-
M.A., Ridgway, S.T., Riot, V.J., Ritz, S., Robinson, S., Roby, W., ing. MNRAS 508, 3111–3124. doi:10.1093/mnras/stab2536,
Roodman, A., Rosing, W., Roucelle, C., Rumore, M.R., Russo, arXiv:2109.10834.
S., Saha, A., Sassolas, B., Schalk, T.L., Schellart, P., Schindler, Men’shchikov, A., André, P., Didelon, P., Motte, F., Hennemann,
R.H., Schmidt, S., Schneider, D.P., Schneider, M.D., Schoening, M., Schneider, N., 2012. A multi-scale, multi-wavelength source
W., Schumacher, G., Schwamb, M.E., Sebag, J., Selvy, B., Sem- extraction method: getsources. A&A 542, A81. doi:10.1051/
broski, G.H., Seppala, L.G., Serio, A., Serrano, E., Shaw, R.A., 0004-6361/201218797, arXiv:1204.4508.
Shipsey, I., Sick, J., Silvestri, N., Slater, C.T., Smith, J.A., Smith, Merz, G., Liu, Y., Burke, C.J., Aleo, P.D., Liu, X., Carrasco Kind,
R.C., Sobhani, S., Soldahl, C., Storrie-Lombardi, L., Stover, E., M., Kindratenko, V., Liu, Y., 2023. Detection, instance segmenta-
Strauss, M.A., Street, R.A., Stubbs, C.W., Sullivan, I.S., Sweeney, tion, and classification for astronomical surveys with deep learn-
D., Swinbank, J.D., Szalay, A., Takacs, P., Tether, S.A., Thaler, ing (DEEPDISC): DETECTRON2 implementation and demon-
J.J., Thayer, J.G., Thomas, S., Thornton, A.J., Thukral, V., Tice, stration with Hyper Suprime-Cam data. MNRAS 526, 1122–1137.
J., Trilling, D.E., Turri, M., Van Berg, R., Vanden Berk, D., Vet- doi:10.1093/mnras/stad2785, arXiv:2307.05826.
ter, K., Virieux, F., Vucina, T., Wahl, W., Walkowicz, L., Walsh, Mohan, N., Rafferty, D., 2015. PyBDSF: Python Blob Detection
B., Walter, C.W., Wang, D.L., Wang, S.Y., Warner, M., Wiecha, and Source Finder. Astrophysics Source Code Library, record
O., Willman, B., Winters, S.E., Wittman, D., Wolff, S.C., Wood- ascl:1502.007.
Vasey, W.M., Wu, X., Xin, B., Yoachim, P., Zhan, H., 2019. Mouhcine, M., Ferguson, H.C., Rich, R.M., Brown, T.M., Smith, T.E.,
LSST: From Science Drivers to Reference Design and Anticipated 2005. Halos of Spiral Galaxies. I. The Tip of the Red Giant Branch
Data Products. ApJ 873, 111. doi:10.3847/1538-4357/ab042c, as a Distance Indicator. ApJ 633, 810–820. doi:10.1086/468177,
arXiv:0805.2366. arXiv:astro-ph/0510253.
Jia, P., Zheng, Y., Wang, M., Yang, Z., 2023a. A deep learning Newell, B., O’Neil, Earl J., J., 1977. The Reduction of Panoramic
based astronomical target detection framework for multi-colour Photometry 1. Two Search Algorithms. PASP 89, 925. doi:10.
photometry sky survey projects. Astronomy and Computing 42, 1086/130248.
100687. URL: https://www.sciencedirect.com/science/ Núnez, J., Llacer, J., 2003. Astronomical image segmentation by self-
article/pii/S2213133723000021, doi:https://doi.org/ organizing neural networks and wavelets. Neural Networks 16,
10.1016/j.ascom.2023.100687. 411–417.
Jia, Y., Liu, L., Zhang, C., 2021. Moon impact crater detection using Olmedo, O., Zhang, J., Wechsler, H., Poland, A., Borne, K., 2008.
nested attention mechanism based unet++. IEEE Access 9, 44107– Automatic Detection and Tracking of Coronal Mass Ejections in
44116. doi:10.1109/ACCESS.2021.3066445. Coronagraph Time Series. SoPh 248, 485–499. doi:10.1007/
Jia, Y., Su, Z., Wan, G., Liu, L., Liu, J., 2023b. Ae-transunet+: An s11207-007-9104-5.
enhanced hybrid transformer network for detection of lunar south Osco, L.P., Wu, Q., de Lemos, E.L., Gonçalves, W.N., Ramos, A.P.M.,
small craters in lro nac images. IEEE Geoscience and Remote Li, J., Marcato, J., 2023. The segment anything model (sam) for
Sensing Letters 20, 1–5. doi:10.1109/LGRS.2023.3294500. remote sensing applications: From zero to one shot. International
Johnston, K.G., Beuther, H., Linz, H., Schmiedeke, A., Ragan, S.E., Journal of Applied Earth Observation and Geoinformation 124,
Henning, T., 2014. The dynamics and star-forming potential of 103540. URL: https://www.sciencedirect.com/science/
the massive Galactic centre cloud G0.253+0.016. A&A 568, A56. article/pii/S1569843223003643, doi:https://doi.org/
doi:10.1051/0004-6361/201423943, arXiv:1404.1372. 10.1016/j.jag.2023.103540.
Karmakar, A., Mishra, D., Tej, A., 2018. Stellar cluster detection Ostdiek, B., Diaz Rivero, A., Dvorkin, C., 2022a. Extracting the Sub-
using gmm with deep variational autoencoder, in: 2018 IEEE Re- halo Mass Function from Strong Lens Images with Image Segmen-
cent Advances in Intelligent Computational Systems (RAICS), pp. tation. ApJ 927, 83. doi:10.3847/1538-4357/ac2d8d.
122–126. doi:10.1109/RAICS.2018.8634903. Ostdiek, B., Diaz Rivero, A., Dvorkin, C., 2022b. Image
Kingma, D.P., Welling, M., 2014. Auto-encoding variational bayes, segmentation for analyzing galaxy-galaxy strong lensing sys-
in: ICLR. tems. A&A 657, L14. doi:10.1051/0004-6361/202142030,
Kohl, S., Romera-Paredes, B., Meyer, C., De Fauw, J., Ledsam, J.R., arXiv:2009.06663.
Maier-Hein, K., Eslami, S., Jimenez Rezende, D., Ronneberger, Otsu, N., 1979. A threshold selection method from gray-level his-
31
tograms. IEEE transactions on systems, man, and cybernetics 9, doi:10.1093/mnras/sty440, arXiv:1802.00937.
62–66. Rombach, R., Blattmann, A., Lorenz, D., Esser, P., Ommer, B., 2022.
Peracaula, M., Oliver, A., Torrent, A., Lladó, X., Freixenet, J., Martı́, High-resolution image synthesis with latent diffusion models, in:
J., 2011. Segmenting extended structures in radio astronomical im- Proceedings of the IEEE/CVF conference on computer vision and
ages by filtering bright compact sources and using wavelets decom- pattern recognition, pp. 10684–10695.
position, in: 2011 18th IEEE International Conference on Image Ronneberger, O., Fischer, P., Brox, T., 2015. U-net: Convolutional
Processing, pp. 2805–2808. doi:10.1109/ICIP.2011.6116254. networks for biomedical image segmentation, in: Medical Image
Platen, E., van de Weygaert, R., Jones, B.J.T., 2007. A cos- Computing and Computer-Assisted Intervention–MICCAI 2015:
mic watershed: the WVF void detection technique. MNRAS 18th International Conference, Munich, Germany, October 5-9,
380, 551–570. doi:10.1111/j.1365-2966.2007.12125.x, 2015, Proceedings, Part III 18, Springer. pp. 234–241.
arXiv:0706.2788. Rosolowsky, E., Leroy, A., 2006. Bias-free Measurement of Giant
Qian, J., Nguyen, N.P., Oya, Y., Kikugawa, G., Okabe, T., Huang, Y., Molecular Cloud Properties. PASP 118, 590–610. doi:10.1086/
Ohuchi, F.S., 2019. Introducing self-organized maps (som) as a 502982, arXiv:astro-ph/0601706.
visualization tool for materials research and education. Results in Rue, F., Bijaoui, A., 1996. Pyramidal vision model applied to as-
Materials 4, 100020. tronomical images, in: Wavelet Applications in Signal and Image
Rani, R., Moore, T.J.T., Eden, D.J., Rigby, A.J., Duarte-Cabral, A., Processing IV, SPIE. pp. 373–383.
Lee, Y.N., 2023. Identification of molecular clouds in emission Salimans, T., Goodfellow, I., Zaremba, W., Cheung, V., Radford, A.,
maps: a comparison between methods in the 13 CO/C18 O (J = 3-2) Chen, X., 2016. Improved techniques for training gans. Confer-
Heterodyne Inner Milky Way Plane Survey. MNRAS 523, 1832– ence on Neural Information Processing Systems (NeurIPS) 29.
1852. doi:10.1093/mnras/stad1507, arXiv:2305.07874. Sandfort, V., Yan, K., Graffy, P.M., Pickhardt, P.J., Summers, R.M.,
Reiman, D.M., Göhre, B.E., 2019. Deblending galaxy superposi- 2021. Use of variational autoencoders with unsupervised learning
tions with branched generative adversarial networks. MNRAS 485, to detect incorrect organ segmentations at ct. Radiology: Artificial
2617–2627. doi:10.1093/mnras/stz575, arXiv:1810.10098. Intelligence 3, e200218.
Rey Deutsch, T., Bignone, L.A., Pedrosa, S.E., 2023. Galaxy segmen- Schilliro, F., Romano, P., 2021. Segmentation of spectroscopic im-
tation using U-Net deep-learning algorithm. Boletin de la Asocia- ages of the low solar atmosphere by the self-organizing map tech-
cion Argentina de Astronomia La Plata Argentina 64, 253–255. nique. MNRAS 503, 2676–2687. doi:10.1093/mnras/stab507,
Rezatofighi, H., Tsoi, N., Gwak, J., Sadeghian, A., Reid, I., Savarese, arXiv:2102.11595.
S., 2019. Generalized intersection over union: A metric and a loss Shen, J., Zheng, X.W., 2020. The bar and spiral arms in the Milky
for bounding box regression, in: Proceedings of the IEEE/CVF Way: structure and kinematics. Research in Astronomy and
conference on computer vision and pattern recognition, pp. 658– Astrophysics 20, 159. doi:10.1088/1674-4527/20/10/159,
666. arXiv:2012.10130.
Richards, F., Paiement, A., Xie, X., Sola, E., Duc, P.A., 2023. Panop- Simonyan, K., Zisserman, A., 2014. Very deep convolutional
tic segmentation of galactic structures in lsb images, in: 2023 networks for large-scale image recognition. arXiv preprint
18th International Conference on Machine Vision and Applications arXiv:1409.1556 .
(MVA), IEEE. pp. 1–6. Slezak, E., Bijaoui, A., Mars, G., 1988. Galaxy counts in the Coma su-
Ricker, G.R., Winn, J.N., Vanderspek, R., Latham, D.W., Bakos, percluster field. II. Automated image detection and classification.
G.Á., Bean, J.L., Berta-Thompson, Z.K., Brown, T.M., Buchhave, A&A 201, 9–20.
L., Butler, N.R., Butler, R.P., Chaplin, W.J., Charbonneau, D., Slezak, E., Lefèvre, S., Collet, C., Perret, B., 2010. Connected com-
Christensen-Dalsgaard, J., Clampin, M., Deming, D., Doty, J., De ponent trees for multivariate image processing and applications
Lee, N., Dressing, C., Dunham, E.W., Endl, M., Fressin, F., Ge, J., in astronomy, in: Pattern Recognition, International Conference
Henning, T., Holman, M.J., Howard, A.W., Ida, S., Jenkins, J.M., on, IEEE Computer Society, Los Alamitos, CA, USA. pp. 4089–
Jernigan, G., Johnson, J.A., Kaltenegger, L., Kawai, N., Kjeld- 4092. URL: https://doi.ieeecomputersociety.org/10.
sen, H., Laughlin, G., Levine, A.M., Lin, D., Lissauer, J.J., Mac- 1109/ICPR.2010.994, doi:10.1109/ICPR.2010.994.
Queen, P., Marcy, G., McCullough, P.R., Morton, T.D., Narita, N., Sohl-Dickstein, J., Weiss, E., Maheswaranathan, N., Ganguli, S.,
Paegert, M., Palle, E., Pepe, F., Pepper, J., Quirrenbach, A., Rine- 2015. Deep unsupervised learning using nonequilibrium thermo-
hart, S.A., Sasselov, D., Sato, B., Seager, S., Sozzetti, A., Stas- dynamics, in: ICML, PMLR.
sun, K.G., Sullivan, P., Szentgyorgyi, A., Torres, G., Udry, S., Vil- Starck, J.L., Aussel, H., Elbaz, D., Fadda, D., Cesarsky, C., 1999.
lasenor, J., 2015. Transiting Exoplanet Survey Satellite (TESS). Faint source detection in ISOCAM images. A&AS 138, 365–379.
Journal of Astronomical Telescopes, Instruments, and Systems 1, doi:10.1051/aas:1999281.
014003. doi:10.1117/1.JATIS.1.1.014003. Starck, J.L., Murtagh, F.D., Bijaoui, A., 1998. Image processing
Riggi, S., 2018. CAESAR: Compact And Extended Source Auto- and data analysis: the multiscale approach. Cambridge University
mated Recognition. Astrophysics Source Code Library, record Press.
ascl:1807.015. Szalay, A.S., Connolly, A.J., Szokoly, G.P., 1999. Simultaneous Mul-
Robitaille, J.F., Motte, F., Schneider, N., Elia, D., Bontemps, S., ticolor Detection of Faint Galaxies in the Hubble Deep Field. AJ
2019a. Exposing the plural nature of molecular clouds. Extract- 117, 68–74. doi:10.1086/300689, arXiv:astro-ph/9811086.
ing filaments and the cosmic infrared background against the true Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S., Anguelov, D., Er-
scale-free interstellar medium. A&A 628, A33. doi:10.1051/ han, D., Vanhoucke, V., Rabinovich, A., 2015. Going deeper with
0004-6361/201935545, arXiv:1905.11492. convolutions, in: Proceedings of the IEEE conference on computer
Robitaille, T., Rice, T., Beaumont, C., Ginsburg, A., MacDonald, vision and pattern recognition, pp. 1–9.
B., Rosolowsky, E., 2019b. astrodendro: Astronomical data Tajik, H.R., Rahebi, J., 2013. Diffuse objects extraction in coronal
dendrogram creator. Astrophysics Source Code Library, record holes using active contour means model. ACSIJ Adv. Comput. Sci
ascl:1907.016. arXiv:1907.016. 2, 55–61.
Robotham, A.S.G., Davies, L.J.M., Driver, S.P., Koushan, S., Taranu, Tang, H., Yue, S., Wang, Z., Lai, J., Wei, L., Luo, Y., Liang, C.,
D.S., Casura, S., Liske, J., 2018. ProFound: Source Extraction and Chu, J., Xu, D., 2023. A model local interpretation routine for
Application to Modern Survey Data. MNRAS 476, 3137–3159. deep learning based radio galaxy classification, in: 2023 XXXVth
32
General Assembly and Scientific Symposium of the International tion of Convolutional Neural Networks to Identify Stellar Feed-
Union of Radio Science (URSI GASS), IEEE. pp. 1–4. back Bubbles in CO Emission. ApJ 890, 64. doi:10.3847/
Tej, A., Ojha, D.K., Ghosh, S.K., Kulkarni, V.K., Verma, R.P., 1538-4357/ab6607, arXiv:2001.04506.
Vig, S., Prabhu, T.P., 2006. A multiwavelength study of the Yan, Q.Z., Yang, J., Su, Y., Sun, Y., Wang, C., 2020. Distances
massive star-forming region IRAS 06055+2039 (RAFGL 5179). and Statistics of Local Molecular Clouds in the First Galactic
A&A 452, 203–215. doi:10.1051/0004-6361:20054687, Quadrant. ApJ 898, 80. doi:10.3847/1538-4357/ab9f9c,
arXiv:astro-ph/0601535. arXiv:2006.13654.
Unterthiner, T., van Steenkiste, S., Kurach, K., Marinier, R., Michal- Yang, P., Bai, H., Zhao, L., Gong, X., Zhong, L., Yang, Y., Rao, C.,
ski, M., Gelly, S., 2019. Fvd: A new metric for video generation 2023. A deep learning approach for automated segmentation of
. magnetic bright points in the solar photosphere. A&A 677, A121.
Vafaei Sadr, A., Movahed, S.M.S., Farhang, M., Ringeval, C., doi:10.1051/0004-6361/202346914.
Bouchet, F.R., 2018. A Multiscale pipeline for the search of Yang, X., Zhang, Q., Yang, X., Peng, Q., Li, Z., Wang, N., 2018. Edge
string-induced CMB anisotropies. MNRAS 475, 1010–1022. detection in cassini astronomy image using extreme learning ma-
doi:10.1093/mnras/stx3126, arXiv:1710.00173. chine, in: MATEC Web of Conferences, EDP Sciences. p. 06007.
van der Zwaard, R., Bergmann, M., Zender, J., Kariyappa, R., Giono, Zarin Era, I., Ahmed, I., Liu, Z., Das, S., 2023. An unsu-
G., Damé, L., 2021. Segmentation of Coronal Features to Under- pervised approach towards promptable defect segmentation in
stand the Solar EUV and UV Irradiance Variability III. Inclusion laser-based additive manufacturing by Segment Anything. arXiv
and Analysis of Bright Points. SoPh 296, 138. doi:10.1007/ e-prints , arXiv:2312.04063doi:10.48550/arXiv.2312.04063,
s11207-021-01863-9. arXiv:2312.04063.
Van Oort, C.M., Xu, D., Offner, S.S.R., Gutermuth, R.A., 2019. Zavagno, A., Dupé, F.X., Bensaid, S., Schisano, E., Li Causi, G.,
CASI: A Convolutional Neural Network Approach for Shell Iden- Gray, M., Molinari, S., Elia, D., Lambert, J.C., Brescia, M., Arzou-
tification. ApJ 880, 83. doi:10.3847/1538-4357/ab275e, manian, D., Russeil, D., Riccio, G., Cavuoti, S., 2023. Supervised
arXiv:1905.09310. machine learning on Galactic filaments. Revealing the filamentary
Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, structure of the Galactic interstellar medium. A&A 669, A120.
A.N., Kaiser, Ł., Polosukhin, I., 2017. Attention is all you need. doi:10.1051/0004-6361/202244103, arXiv:2212.00463.
Advances in neural information processing systems 30. Zheng, C., Pulido, J., Thorman, P., Hamann, B., 2015. An improved
Verbeeck, C., Delouille, V., Mampaey, B., De Visscher, R., 2014. method for object detection in astronomical images. MNRAS 451,
The SPoCA-suite: Software for extraction, characterization, and 4445–4459. doi:10.1093/mnras/stv1237.
tracking of active regions and coronal holes on EUV images. A&A Zhou, Z., Rahman Siddiquee, M.M., Tajbakhsh, N., Liang, J., 2018.
561, A29. doi:10.1051/0004-6361/201321243. Unet++: A nested u-net architecture for medical image segmen-
Verbeeck, C., Higgins, P.A., Colak, T., Watson, F.T., Delouille, V., tation, in: Deep Learning in Medical Image Analysis and Multi-
Mampaey, B., Qahwaji, R., 2013. A Multi-wavelength Anal- modal Learning for Clinical Decision Support: 4th International
ysis of Active Regions and Sunspots by Comparison of Auto- Workshop, DLMIA 2018, and 8th International Workshop, ML-
matic Detection Algorithms. SoPh 283, 67–95. doi:10.1007/ CDS 2018, Held in Conjunction with MICCAI 2018, Granada,
s11207-011-9859-6. Spain, September 20, 2018, Proceedings 4, Springer. pp. 3–11.
Vikhlinin, A., Forman, W., Jones, C., Murray, S., 1995. Matched Zhu, Y., Olszewski, K., Wu, Y., Achlioptas, P., Chai, M., Yan, Y.,
Filter Source Detection Applied to the ROSAT PSPC and the De- Tulyakov, S., 2022a. Quantized gan for complex music generation
termination of the Number-Flux Relation. ApJ 451, 542. doi:10. from dance videos, in: European Conference on Computer Vision
1086/176242. (ECCV), Springer. pp. 182–199.
Vos, E.E., Francois Luus, P.S., Finlay, C.J., Bassett, B.A., 2019. A Zhu, Y., Wu, Y., Olszewski, K., Ren, J., Tulyakov, S., Yan, Y., 2023.
generative machine learning approach to rfi mitigation for radio Discrete contrastive diffusion for cross-modal music and image
astronomy, in: 2019 IEEE 29th International Workshop on Ma- generation, in: The Eleventh International Conference on Learn-
chine Learning for Signal Processing (MLSP), pp. 1–6. doi:10. ing Representations (ICLR).
1109/MLSP.2019.8918820. Zhu, Y., Wu, Y., Sebe, N., Yan, Y., 2022b. Vision+ x: A sur-
Walmsley, M., Spindler, A., 2023. Deep Learning Segmentation of vey on multimodal learning in the light of data. arXiv preprint
Spiral Arms and Bars. arXiv e-prints , arXiv:2312.02908doi:10. arXiv:2210.02884 .
48550/arXiv.2312.02908, arXiv:2312.02908.
Whiting, M., Humphreys, B., 2012. Source-Finding for the Australian
Square Kilometre Array Pathfinder. PASA 29, 371–381. doi:10.
1071/AS12028, arXiv:1208.2479.
Williams, J.P., de Geus, E.J., Blitz, L., 1994. Determining Structure
in Molecular Clouds. ApJ 428, 693. doi:10.1086/174279.
Xavier, G., Soman, K.P., Tvn, D., Philip, T.E., 2012. An efficient
algorithm for the segmentation of astronomical images. IOSR
Journal of Computer Engineering 6, 21–29. URL: https://api.
semanticscholar.org/CorpusID:54658228.
Xu, D., Kong, S., Kaul, A., Arce, H.G., Ossenkopf-Okada, V.,
2023. CMR Exploration. II. Filament Identification with Machine
Learning. ApJ 955, 113. doi:10.3847/1538-4357/acefce,
arXiv:2308.06641.
Xu, D., Offner, S.S.R., Gutermuth, R., Oort, C.V., 2020a. Application
of Convolutional Neural Networks to Identify Protostellar Out-
flows in CO Emission. ApJ 905, 172. doi:10.3847/1538-4357/
abc7bf, arXiv:2010.12525.
Xu, D., Offner, S.S.R., Gutermuth, R., Oort, C.V., 2020b. Applica-
33