Deep Learning-Based Welding Image Recognition: A Comprehensive Review
Deep Learning-Based Welding Image Recognition: A Comprehensive Review
net/publication/371253385
CITATIONS READS
0 252
3 authors:
Bao Jinsong
Donghua University
161 PUBLICATIONS 1,901 CITATIONS
SEE PROFILE
Some of the authors of this publication are also working on these related projects:
Proactive Human-Robot Collaboration (GRF & AIR@InnoHK & Endowed Young Scholars Scheme, HKSAR) View project
All content following this page was uploaded by Tianyuan Liu on 05 June 2023.
Review
A R T I C L E I N F O A B S T R A C T
Keywords: The reliability and accuracy of welding image recognition (WIR) is critical, which can largely improve domain
Welding system experts’ insight of the welding system. To ensure its performance, deep learning (DL), as the cutting-edge arti
Computer vision ficial intelligence technique, has been prevailingly studied and adopted to empower intelligent WIR in various
Image recognition
industry implementations. However, to date, there still lacks a comprehensive review of the DL-based WIR
Deep learning
(DLBWIR) in literature. Aiming to address this issue, and to better understand its development and application,
Convolutional neural network
Explainable AI this paper undertakes a state-of-the-art survey of the existing DLBWIR research holistically, including the key
technologies, the main applications and tasks, and the public datasets. Moreover, possible research directions are
also highlighted at last, to offer insightful knowledge to both academics and industrial practitioners in their
research and development work in WIR.
* Corresponding authors.
E-mail addresses: [email protected] (P. Zheng), [email protected] (J. Bao).
https://doi.org/10.1016/j.jmsy.2023.05.026
Received 13 September 2022; Received in revised form 4 March 2023; Accepted 29 May 2023
Available online 2 June 2023
0278-6125/© 2023 The Society of Manufacturing Engineers. Published by Elsevier Ltd. All rights reserved.
T. Liu et al. Journal of Manufacturing Systems 68 (2023) 601–625
602
T. Liu et al. Journal of Manufacturing Systems 68 (2023) 601–625
603
T. Liu et al. Journal of Manufacturing Systems 68 (2023) 601–625
effective method for image recognition, it has been studied very little in
the WIR field. Therefore, based on the definition of DL, the paradigm of
supervised DLBWIR can be summarized in Fig. 6. In practice, three
major ways of obtaining welding image labels are: expert experience-
based manual labelling [46], welding process-based automatic label
ling [47] and the high precision inspection-based automatic labelling
[48]. After acquiring the welding image data and the labels corre
sponding to the different tasks, a certain hypothesis function is learned
from the hypothesis space defined by the structure of the model for the
different tasks as the final model according to the training method. The
process is supported by the platform provided by the hardware and
software environment. This well-trained model is an approximation of
the real model.
604
T. Liu et al. Journal of Manufacturing Systems 68 (2023) 601–625
Table 2
Comparison of typical welding image at different stages.
Ref. Year Stage Method Sketch Map Image content Welding information
[15] 2019 PRW CCD Single-stripe laser, a single weld joint and groove. Weld joint position
2020 The back side of the melt pool, and the front laser stripe. Penetration state
[18]
2019 Weld seams are captured in three directions by multiple Penetration state
[29]
specular reflections.
2020 Protective products for manual welding Work safety and welding processes
[19]
2022 Acceleration, displacement, current and voltage. Interpretable transformation to reflect welding failure
[22]
mechanisms
605
T. Liu et al. Journal of Manufacturing Systems 68 (2023) 601–625
Table 2 (continued )
Ref. Year Stage Method Sketch Map Image content Welding information
2021 Simulation Molten pool Internal defects in weld seam
[34]
2020 Multi-stripe laser scan multiple weld bead and groove Surface defects in weld seam
[35]
2021 X-ray Inside the weld seam Internal defects in weld seam
[25]
techniques for welding image data itself. 3.1.2. Welding image dataset
Due to the interference of arc light, spatter, and vibration in the In addition to the problems of the welding image itself, there is also a
welding environment, PreP methods represented by filtering, denoising, widespread problem of small samples at the dataset level, which brings a
and enhancement have also been extensively researched. The PreP re huge obstacle to the in-depth application of the DL methods. From the
sults in high quality welding image that facilitates the learning of feature data point of view, scholars have mainly addressed this problem through
information. DA, TL, and GAN. Fig. 8 plots the number of occurrences of the three
Meanwhile, the purpose of welding image NormL is to compress the methods in the literature.
grey scale range, thus reducing the search space and speeding up the Table 6 summarizes the specific DA, TL, and GAN methods involved
learning and convergence of the DL model. in the existing studies.
The idea of DA is to artificially transform the original image in the
spatial, greyscale and color domains to obtain new samples of welding
606
T. Liu et al. Journal of Manufacturing Systems 68 (2023) 601–625
Fig. 5. Four basic tasks of WIR: (a) classification, (b) detection, (c) segmentation, and (d) tracking.
Table 3
Typical definitions of DL.
Ref. Year Definition Keywords Highlights
607
T. Liu et al. Journal of Manufacturing Systems 68 (2023) 601–625
Table 4
Review articles related to DLBWIR.
Ref. Year Scope Taxonomy Relation to DLBWIR
608
T. Liu et al. Journal of Manufacturing Systems 68 (2023) 601–625
Table 6
The specific technologies for welding image dataset and their comparison.
Ref. Year Key elements Method Pros Cons
609
T. Liu et al. Journal of Manufacturing Systems 68 (2023) 601–625
Table 6 (continued ) summarizes the optimizer and training techniques involved in the
Ref. Year Key elements Method Pros Cons DLBWIR model training process.
The optimizer guides the model parameters to update the appro
[92] 2020 Convolutional GAN The new It is difficult to
layers replace samples are learn a
priate size in the correct direction in the process of DL backpropagation,
perceptrons more diverse reliable data so that the updated parameters make the loss function value continue to
[26] 2020 Add conditional distribution approach the global minimum, which plays an important role in the
information to the while the model training process. It can be seen from the definition that the core of
generator and original
the optimizer includes two aspects, namely the optimization direction
discriminator welding image
[93] 2022 No image pairs data is and the update step size. Different optimizers appearing in the literature
required insufficient. are optimized and improved for these two. It can be seen from Fig. 10
[94] 2022 As an that Adam and SGD are more commonly used at present. This is because
enhancement tool Adam has the advantage of fast convergence speed, SGD has the
to restore the
balance of the
advantage of strong generalization ability, and each DL framework has
dataset better support for these two optimizers. A detailed comparison of DL
[95] 2020 Solve the problem optimizers can be found in the reference [140]. For the problem of weld
of collapse mode appearance detection, Li et al. [141] conducted a comparative study on
the effects of different optimizers. For the TIG welding defect identifi
cation problem, Sekhar et al. [142] compared four optimizers. However,
The current DL is generally regarded as a black box model with poor
there is a lack in DLBWIR to analyze the characteristics of the welding
explainability. The literature [116] argues that a key component of AI
image data and select the optimizer accordingly.
systems is the ability to explain decisions, recommendations, pre
The use of batch training can improve memory utilization and
dictions, actions and the process. DLBWIR mainly adopts the methods of
accelerate the training process of the model. It is the most widely used
feature map visualization and class activation mapping [117,118]
training technique under the condition of rapid development of hard
respectively trying to explain what the model has learned and what is
ware technology. Compared with training based on single sample,
the basis for the model’s decision-making for individuals. Table 7
training based on batch can lock the direction of gradient descent faster
summarizes the commonly used model improvement methods in
and cause smaller training shock. Batch training has applications in
DLBWIR.
multiple tasks such as classification, detection, segmentation, and
tracking. However, the selection of batch size in the current research is
3.3. Training technology
based on experience, and there is no clear basis to select a batch size
suitable for specific DLBWIR tasks.
After acquiring the dataset and determining the hypothesis space
BN first calculates the mean and variance of the current batch, then
from the model structure, the model needs to be properly trained in
normalizes the current batch, and finally introduces a scale factor and a
order to find a hypothesis function that performs well. Fig. 10
Fig. 9. Backbone commonly used in DLBWIR: (a) classification, (b) detection, (c) segmentation, (d) tracking, and (e) others.
610
T. Liu et al. Journal of Manufacturing Systems 68 (2023) 601–625
Table 7
The specific technologies for improved structure and their comparison.
Ref. Year Key elements Method Pros Cons
[119] 2021 Infrared thermal and visual image MML Complementary information, The primary and secondary relationship
[120] 2022 High-speed camera and coherent Limiting the hypothesis space between modalities is not clear.
light imaging
[121] 2022 Image and temperature
[122] 2020 Image and sound
[123] 2022 Eddy current detection and 3D
laser scanning
[124] 2020 Image, sound, voltage and current
[16] 2021 Image and text
[125] 2022 Handcrafted and learned features
[95] 2020 Inception and MobileNet ME Voting decisions with high robustness The training process is difficult and inefficient.
[126] 2021 Lenet5 and MLP
[127] 2019 Alexnet, Densenet, Resnet, VGG,
and Capsnet
[128] 2018 Lenet5 and Lenet5
[20] 2021 CNN, GRU, SVM and KNN
[16] 2021 Label text AM Pay attention to important information and optimize There is a lack of research on task justification
[129] 2019 Spatial the allocation of resources. and where to add it.
[130] 2020 Channel
[56] 2021 Spatial channel mixing
[131] 2022 Spatiotemporal and channel
attention
[132] 2021 Spatial correlation and Senet
[133] 2022 Bilinear attention
[134] 2022 Two modal cross-attention
[135] 2022 Four modal cross-attention
[136] 2022 Self-attention
[137] 2018 Feature map visualisation Explainability Visualization of the model’s learned knowledge and Lack of semantic-level explanations.
[136] 2022 Attention visualization basis for decision making.
[138] 2020 GradCAM
[82] 2020 CAM and Guided GradCAM
[139] 2022 Multi-scale fusion features-based
CAM
translation factor to retain the learned features [143]. This method Wang et al. [153] used a Monte Carlo cross-validation method. The
makes the distribution of each layer’s input data relatively stable, re advantage of this method is that the limited sample space can be utilized
duces the model’s dependence on network initialization, accelerates the as much as possible and data interference can be eliminated to the
model’s convergence speed, and reduces the risk of model overfitting. greatest extent. However, the current literature does not consider the
Dropout refers to making some neurons stop working with a certain original data distribution when dividing the sub-datasets. Therefore, it is
probability when the signal propagates forward [144]. This can make necessary to consider a hierarchical cross-validation method that com
the model not too dependent on some local features, thereby improving bines data distribution in future applications.
the generalization ability. However, dropout is designed for a more The essence of parameter regularization is to limit the parameter set
generalized neural network, ignoring the correlation of parameters in to be optimized, thereby standardizing the training direction of the
the convolution kernel. Therefore, Pan et al. [145] used a more targeted model, and limiting the scale of the hypothesis space to prevent the
Dropblock [146] method. model from overfitting. L1 [154] and L2 [155] parameter regularization
KCV generally divides the original dataset into k sub-datasets, a methods are two commonly used in the literature. However, the pa
single sub-dataset is used for the verification, and other sub-datasets are rameters in the convolution kernel have a local correlation, and the L1
used for training model. KCV is repeated k times, and the result of and L2 regularization methods lack consideration of this correlation.
averaging k times can get a single estimate [147–152]. In addition, Therefore, Liu et al. [156] took into account the characteristics of the
611
T. Liu et al. Journal of Manufacturing Systems 68 (2023) 601–625
612
T. Liu et al. Journal of Manufacturing Systems 68 (2023) 601–625
4. Applications of DLBWIR the problem of small visual differences between different penetration
states by designing a label semantic attention that uses discriminative
This section provides an overview of the current status of DLBWIR inter-label texts to guide the model in learning discriminative visual
applications from a WIR task type perspective. Fig. 13 summarizes the features. In the reference [ 165,351-dimensional features were first
amount of DLBWIR posts in different task applications. The statistics extracted manually, followed by a stacked sparse autoencoder to
were based on 192 papers that had been searched and filtered by the establish a mapping relationship between the weld state and the input
rules described in the Introduction. The different tasks were distin features. Xue et al. [81] used Mobilenet as the skeleton network, which
guished on the basis of the supervisory information and task scenarios as greatly reduces the model size and offers the possibility of practical
described in Section 2.2. As the classification task is the basis and pre deployment of the model. The references [138,166] have improved the
requisite for several other tasks, the development of technology oriented explainability of models through model decision basis visualization and
towards the classification task generally precedes the other tasks. And feature map visualization respectively.
the cost of image-level labeling for classified task is lower than other The classification task for POW focuses on the identification of
tasks. Therefore, the applications oriented to classification task account external or internal defects in the weld seam to enable intelligent quality
for the vast majority, and applications oriented to object detection, inspection task. In the reference [ 167], three resampling methods were
image segmentation, target tracking, and other tasks decrease in turn. designed and compared for the problem of defect sample imbalance in
weld radiographic image recognition. Jiang et al. [168] designed four
4.1. Welding image classification pooling methods for different weld defect types and different regions in
the image. Kumaresan et al. [169] used a TL strategy to address the
Welding image classification is to discover the category of objects problem of a small sample of weld seam radiographic image. Miao et al.
contained in the welding image, which is a coarse-grained WIR task. [170] devised a two-stage identification method based on the fact that
The classification task for PRW is mainly the recognition of the bevel defects in weld appearance do not occur frequently, i.e. identifying the
shape and then the adjustment of the welding process based on the presence or absence of defects and then identifying the specific type of
recognition results. Tian et al. [164] first used different laser sources to defect. In the reference [36], a one-dimensional signal was first acquired
obtain weld bevel profile features and then designed a 10-layer CNN to using the acoustic emission technique, then a two-dimensional image
classify the bevel types. In the reference [15], a 6-layer CNN was was obtained by TFT, and then a CNN was built to identify whether the
designed to identify the bevel type in response to strong noise interfer welded steel pipe was subject to a natural gas leak. Zhang et al. [171]
ence in the field of view, and to narrow the parameter search for sub designed a 7-layer CNN and classified the weld microstructure images
sequent tasks accordingly. Cruz et al. [57] designed a 9-layer CNN to corresponding to the three defects. Feng et al. [132] firstly learned the
identify whether the parts to be welded are aligned. difference between normal and defect samples through comparative
The classification task for IW is mainly to identify the state of the learning, and secondly filtered the same features and extracts the
melting zone, and the purpose is to provide a basis for the traceability different features of the image pairs by means of the spatial correlation
and online control of welding quality. To better predict the melt depth, AM, which provides an effective attention for learning the differences
Jiao et al. [17] synthesised the current image with the previous 1/6 and between images of different armature welding surfaces. Hu et al. [172]
2/6 s images to reflect the dynamic welding phenomenon and built a adopted ELU activation function and grayscale adaptive pooling strat
9-layer CNN to identify the image features. Zhang et al. [29] used egy in the task of X-ray defect recognition. To address the small sample
different data augmentation methods for different welding defects and problem faced by weld radiographic defect detection tasks, Di and Li
designed a 10-layer CNN to achieve classification of the three penetra [173] investigated the effect of freezing different layers on transfer
tion states. Miao et al. [33] designed a two-stage detection strategy to learning methods. Zhang et al. [174] weighted the loss function for the
improve the efficiency of CNN-based defect recognition for welding problem of imbalanced defect categories in weld radiographic images.
image acquired by TFT. The references [124,131] introduced temporal To address the problem of low quality of X-ray images, Liu et al. [175]
information in the identification of weld states and convert disturbing firstly embed defect images into local and global feature spaces, and
information for a single image into useful information to support model secondly maintain the consistency of the feature space based on a new
decisions. Liu et al. [137] found that the molten pool features extracted consistency strategy.
by the convolution kernel were redundant through feature map visual In summary, the main application scenarios of welding image clas
ization, so it was integrated into the LSTM network to adaptively fuse sification include the identification of bevel types, melt zone states and
the redundant features. In order to learn richer molten pool features, Liu weld defects. The current research mainly focuses on issues such as noise
et al. [156] proposed a coarse-grained parameter regularization method, interference, lack of welding feature information, insufficient data vol
taking into account that the weight parameters in the convolution kernel ume, small visual differences, high calculation costs, and poor explain
have local correlations, which facilitated the convolutional kernel to be ability. However, the current research has been modeling DLBWIR as a
trained in the direction of increasing variance. Liu et al. [16] addresses single-label classification, and the actual welding image often contains a
613
T. Liu et al. Journal of Manufacturing Systems 68 (2023) 601–625
variety of label categories. Therefore, the authors believe that the ac in the welding image at the same time, which is a fine-grained WIR task.
curacy and interpretability of the classification model can be improved In view of the fact that the regression task can obtain specific values, this
by modeling DLBWIR as a multi-label classification task. section considers the regression problem as an object detection task and
summarizes it.
Lin et al. [176] first used a CNN to initially localize the laser spot
4.2. Welding object detection area, then increased the optical magnification of the camera and used
template matching for high precision localization. Wang et al. [177] first
Object detection is to determine the category and location of objects
Table 8
Application of DL in welding image object detection.
Ref. Year Stage Issue and challenge Method and result
Engineering Visual task Difficulty and challenge Backbone Key elements Data sources,
requirement volumes and
metrics
[176] 2018 IW Welding quality Laser spot detection Incomplete laser spot Not Two-stage detection of first coarse and Laboratory, 400
online image, mentioned then fine and Acc of 97.96
monitoring large field of view and %
complex calculation.
[177] 2021 Prediction of Back-bead width Weld width is difficult to Lenet5 Simultaneous acquisition of front and Laboratory,
weld forming prediction measure directly back images of the molten pool. Define 28494 and MSE
quality the width detection on the back of the of 0.096
weld as a regression problem.
[178] 2021 Welding quality The monitoring of Global information is Resnet Non-local attention Laboratory, DE of
online weld width and difficult to extract 0.1805 mm
monitoring reinforcement
[34] 2021 Welding quality Weld defect Lack of dataset YOLO Multiple dataset construction methods Simulation, 3892
online detection and MAP of 85 %
monitoring
[179] 2021 Laser beam Top and bottom The laser beam YOLO Dual vision sensing Laboratory, 240
absorbance diameter detection absorptance and GOF of 99.76
monitoring of keyhole inside a keyhole varies %
rapidly
[119] 2021 Welding quality Penetration status Arc interference and Faster Dual vision sensing and TL Laboratory, 2832
online recognition small sample size R-CNN and Acc of 95.58
monitoring %
[19] 2020 Safety monitoring Welding protective The object scale is SSD Multi-scale feature fusion and DSC Production,
of welding work equipment detection different and the model is 10000 and MAP
complex of 87.45 %
[180] 2022 Welding quality Penetration status The penetration state Lenet5- Fusion of timing information Laboratory,
online recognition changes dynamically and LSTM 28560 and DE of
monitoring is difficult to measure 3 mm
directly
[35] 2020 POW Automatic NDT Weld line detection Strong noise and complex SSD Parallel structured light and DSC Laboratory,
model 10000 and DE of
1 mm
[181] 2021 Connector quality Solder joint Small sample size, small Faster DA, TL and the region clustering-based Laboratory, 1654
detection detection object and the model is R-CNN anchor box generation and MAP of 94.1
complex %
[84] 2021 Body welding Solder joint Strong interference, small YOLO Feature pyramid, DA and DSC Production, 400
quality control detection object, small sample size and MAP of 89.32
and the model is complex %
[182] 2022 Automatic NDT Weld defect Low contrast and strong YOLO Contrast-limited adaptive histogram Production, 256
detection noise equalization, non-local mean and MAP of 51.2
denoising and GIOU loss %
[91] 2022 Weld seam Weld seam detection Small sample size Faster DA and TL Laboratory, 5000
polishing R-CNN and Acc 0 f 91.68
%
[78] 2022 Automatic NDT Weld defect Small sample size YOLO Stick-based DA Public dataset,
detection 3000 and MAP of
79.33 %
[183] 2022 Automatic NDT Weld defect Small object Faster Bidirectional feature pyramid Laboratory, MAP
detection R-CNN of 85.5 %
[184] 2022 Quality Weld defect Fine-grained YOLO Label smoothing Laboratory, 1100
inspection of detection and MAP of 90.1
body welded %
joints
[185] 2021 Rail quality Flash butt weld Small object YOLO Multi-scale feature fusion Production, 4529
detection detection and MAP of 87.41
%
[61] 2021 Steel pipe quality Weld defect Small sample size and YOLO Multiple DA methods and single stage Production,
detection detection real-time detection 30672 and MAP
of 98.7 %
[186] 2019 PCB board quality Solder joint Simultaneous multiple Mask Mask R-CNN and TL Laboratory, 88
detection detection recognition tasks with R-CNN and MAP of 97.4
small sample size %
614
T. Liu et al. Journal of Manufacturing Systems 68 (2023) 601–625
used CNN to extract the visual features of the molten pool, and then while obtaining refined recognition results. Therefore, segmentation
established the correspondence between the visual features and the task is often used in quantitative detection of objects to be identified in
width of the back weld to achieve the key feature detection of the molten POW images.
pool object. In the reference [ 178], after CNN extracted the visual Yang et al. [187] designed an attention dense convolutional block
features of the molten pool, a regression method was used to simulta and used pixel segmentation to extract accurate laser stripes in a
neously detect the width and reinforcement of the deposited layer. In the strongly disturbed environment. Wang et al. [153] designed a segmen
reference [34], six auxiliary datasets were created for the small sample tation network with a channel AM and a separable structure for
problem, and both porosity and slag defects were effectively detected extracting laser stripes.
based on YOLOv4. Kim et al. [179] collected the images of the top Mi et al. [188] used dilated convolution to increase the receptive
keyhole and the bottom aperture simultaneously, and then measured the field and designed a semantic segmentation network to segment the
aperture size based on YOLOv4, and finally established the relationship molten pool and the splash. In the literature [52], a segmentation
between the aperture size and the absorption rate of the laser beam. A network Epnet was designed to calculate the molten pool width and then
dual-input Faster R-CNN model with infrared thermal and visual images predict the forming width of the weld. Tan et al. [189] first divided the
as input, including a simultaneous feature extraction module, a con welding image into small blocks, and then designed CNN and threshold
volutional descriptor selection module and a synthetic feature-based neural network to segment each small block.
recognition module, was developed in the reference [119]. Lu et al. Yang et al. [190] proposed a branching structure for encoding spatial
[19] used an idea based on multi-scale feature fusion to improve the SSD and contextual features, respectively, for the small target and cross-scale
structure and apply it to welders’ behaviour detection. Yu et al. [ 180] problems of appearance defects. Ling et al. [191] designed a parallel
developed a CNN-LSTM model and extracted sufficient penetration in network to obtain semantic differences for the problem of small and
formation from the molten pool image sequence. irregular targets of weld defects, and used the Lovasz-softmax loss
Dong et al. [35] combined Mobilenet and SSD structures to achieve function in the training process. Li and Li [192] proposed a
lightweight detection of weld lines. In the reference [181], a high-frequency feature enhancement network for the problem of insig
clustering-based anchor generation method was proposed and combined nificant pixel variation and small area of defects on the welded steel
with Faster R-CNN to achieve connector weld joint defect detection. Dai surface. Pandiyan et al. [193] designed an encoding-decoding type CNN
et al. [84] not only used a lightweight Mobilenet as the backbone of to realize the segmentation of four types of welds, which provides a basis
YOLOv3, but also designed a feature pyramid with cross-scale connec for the optimization of robot grinding parameters. Liu et al. [194] pro
tions and employed a complete IOU loss to speed up model convergence posed a fusion block to extract the most important features in the three
and applied it to quality detection of resistance spot welds. Aiming at the dimensions of the normal map, and then constructed a CNN to segment
non-derivable problem of the IOU loss function, Yun et al. [182] first the key points of the welded stud from the normal map. Oh and Ki [26]
enhanced the X-ray image through contrast-limited adaptive histogram used the first generator to generate the weld segmentation map, and the
equalization, and secondly used GIOU as the loss function for training second generator converted it into an OM image, which realized the
the YOLO model. Liu et al. [91] used DA and TL methods to solve the direct prediction of the cross-sectional welding image from the laser
problem of insufficient weld appearance image data, and compared the welding process parameters. Due to the lack of image data of weld
effects of TL based on different backbone. Mery et al. [78] designed a DA appearance, Yang et al. [24] segmented the weld area based on U-Net,
method called Stick. For the problem of small defect targets in X-ray and then merged with the workpiece image to achieve data augmenta
images, Chen et al. [183] adopted a bidirectional fusion feature pyramid tion. In order to segment laser welding defects of different sizes and
structure. Li et al. [184] adopted the label smoothing technique for the shapes, Zhu et al. [195] proposed a high-precision and high-efficiency
problem of small variation in the features of appearance defective. In the lightweight segmentation algorithm, which mainly includes a feature
reference [185], a minimum scale prediction network was designed to extraction module, an attention module, a localisation module and a
improve the detection accuracy of the model for small-scale objects and boundary anti-aliasing module. Jang et al. [196] regarded Resnet as the
applied to the detection of welding defects in railway tracks. In order to backbone of the segmentation network, and realized the measurement
identify multiple defects in the weld seam of steel pipes, Yang et al. [61] of acicular ferrite content that had a significant impact on the me
used a single-stage inspection structure YOLOv5 to detect 8 types of chanical properties of carbon steel. In the reference [197], a Mask
defects. Wu et al. [186] first pre-trained the skeleton network Resnet101 R-CNN network for defect segmentation of weld X-ray images was
on the COCO dataset, and then detected the location of solder joint constructed based on transfer learning for the problem of small sample
defects based on Mask R-CNN. Table 8 summarizes the key elements of size of weld X-ray images. Zhang et al. [ 198] firstly proposed a com
DL in the welding image object detection tasks. The identification of key pressed U-Net to segment the weld pad and the weld joint, secondly a
elements is based on the targeted approach used by the literature to template based approach to confirm the validity of each weld joint and
address the difficulty and challenge, the innovation point in the litera finally a heuristic algorithm to determine the defect type. In order to
ture, and the associated point with the theme of this paper. obtain more precise details of defects, Yang et al. [25] added cross-scale
In summary, the main application scenarios of welding object skip connections to U-Net. Wang and Shen [199] proposed a
detection include the detection of process quality, forming quality, weld boundary-aware semantic segmentation method, which combined
position and defect location. In addition to the common problems spatial pyramid pooling and spatial AM to improve the segmentation
similar to classification task, the current research also includes small accuracy and robustness of the welding area. For the problem of com
object and cross-scale issues. However, although the current object plex background and low contrast of X-ray images, Yang et al. [200,201]
detection tasks are modeling and positioning the labels and positions of proposed an attention fusion module and a cross-scale connection
multiple objects in the welding image, the relationship between objects method. Dong et al. [90] proposed a novel unsupervised local deep
is not considered. Therefore, the authors believe that the reliability of feature learning method based on image segmentation to build a
DLBWIR can be improved through modeling relationship between network which can extract useful features from an image. Nowroth et al.
multiple objects. [202] proposed an improved method combining dilated convolution
and pyramidal pooling for segmenting OM images. Table 9 summarizes
4.3. Welding image segmentation the key elements of DL in the application of welding image segmentation
task.
Segmentation is the most refined WIR task, and its essence is a pixel- In summary, the main application scenarios of welding object seg
by-pixel classification task. Due to the pixel-by-pixel classification mentation include segmentation of weld key points, melt zone objects,
characteristic of segmentation task, the model inference speed is slow and defect objects. In addition to facing common challenges similar to
615
T. Liu et al. Journal of Manufacturing Systems 68 (2023) 601–625
Table 9
Application of DL in welding image segmentation.
Ref. Year Stage Issue and challenge Method and result
Engineering requirement Visual task Difficulty and challenge Backbone Key elements Data sources, volumes
and metrics
[187] 2022 PRW Weld seam tracking Laser stripe Strong interference, across U-Net Pixel segmentation Laboratory, 192 and
segmentation scales and unbalance task, BiConvLSTM Acc of 99.4 %
module and weighted
loss
[153] 2022 Weld seam tracking Laser stripe Strong interference and real- Erfnet Channel attention and Laboratory, 737 and
segmentation time DSC segmentation success
rate of 96 %
[188] 2021 IW Directed energy Segmentation of Strong interference and real- PSPnet Dilated convolution and Laboratory, 700 and
deposition monitoring molten pool and time pyramid pooling Acc of 94.71 %
spatter
[52] 2021 Width monitoring of Segmentation of Small sample size, real-time Enet DA, DSC and pyramid Laboratory, 600 and
cladding layer in arc molten pool and difficulty in contour pooling Acc of 94.18 %
additive manufacturing extraction
[189] 2020 Powder bed laser melting Segmentation of Strong interference and small VGG Threshold neural Laboratory, 210 and
process monitoring spatter sample size network and processing Rec of 80.48 %
in blocks
[190] 2022 POW Safety valve welding Segmentation of Small sample size and across Proposed Spatial features, Laboratory, 4708 and
quality inspection weld seam scales contextual features and Mean IOU of 86.704
feature fusion %
[191] 2022 PCB solder joint quality Solder joint defect Small sample size, small object Deeplab Parallel network and Laboratory, 340 and
inspection segmentation and irregular shapes Lovasz-softmax Acc of 96.78 %
[192] 2022 Surface defect detection Segmentation of Defects are insignificant and U-Net High-frequency Production, 1580 and
of welded rebar welding seam defect small in size components of low- Mean IOU of 86.2 %
level features
[193] 2019 Robot polishing endpoint Segmentation of Strong interference, unbalance Segnet DA, class- Laboratory, 2000 and
detection weld seam and robustness weighted loss and Acc of 99.35 %
regularization
[194] 2020 Welding stud detection Segmentation of There are noise spots, missing FCN Channel separation and Laboratory, 1427 and
stud key point spots, and the need to measure feature fusion Rec of 0.98 %
multiple studs simultaneously.
[26] 2020 Weld microstructure Segmentation of Small sample size and complex FCN CGAN and two-stage Laboratory, 1638 and
detection weld section OM image content image generation GOF of 93.6 %
[24] 2020 Weld position detection Segmentation of Complex and diverse industrial U-Net DA based on segmented Laboratory, 1940 and
weld seam environment image fusion Acc of 96.2 %
[195] 2021 Welding defect detection Segmentation of Defect size and shape are U-Net Multi-scale AM and DSC Production, 7263 and
of battery cover safety welding defect variable and require real-time Mean IOU of 84.67 %
hole
[196] 2020 Acicular ferrite content Segmentation of Small sample size and FCN DA and category Laboratory, 144 and
measurement weld microstructure imbalance sensitive loss Mean IOU of 86.2 %
[197] 2018 Automatic NDT Segmentation of Simultaneous multiple Mask Mask R-CNN and TL Public dataset, 88 and
welding seam defect recognition tasks with small R-CNN MAP of 85 %
sample size
[198] 2020 Quality inspection of Segmentation of Small sample size and real- U-Net DA and network Laboratory, 37,877
laser solder joints in solder joint and time compression and Acc of 99.85 %
battery production solder pad
[25] 2021 Automatic NDT Segmentation of Small sample size and across U-Net DA and skip Public dataset, 20 and
welding seam defect scales connections Acc of 99.8 %
[199] 2020 Automatic NDT Segmentation of Small sample size, strong U-Net DA, AM and atrous Production, 300 and
welding seam defect interference and variable spatial pyramid pooling Acc of 93.7 %
object shape
[200] 2022 Automatic NDT Segmentation of Complex backgrounds, U-Net Attention fusion block Public dataset, 20 and
welding seam defect poor contrast, weak texture, and cross-scale Acc of 99.8 %
and class imbalance connection
[90] 2021 Automatic NDT Segmentation of Difficulty in labeling data U-Net Unsupervised Production and
welding seam defect simulation, 24,111
and Acc of 89.55 %
[202] 2022 Weld microstructure Segmentation of Small sample size and across Deeplab TL, dilated convolution Laboratory, 282 and
detection weld section scales and pyramid pooling Mean IOU of 76.88 %
classification and detection tasks, this application is particularly affected approach is to first detect the area of the target, and then calculate the
by the difficulty of pixel-level labelling and the difficulty of segmenting object position to achieve the tracking task [203]. The end-to-end DL
pixels near boundaries. Therefore, the authors suggest that weakly su structure has not been applied, such as [204,205]. Furthermore, the
pervised welding image segmentation based on image-level labelling current research focuses on the seam tracking task of PRW.
information could be investigated in the future. In addition, the hard-to- Zou et al. [206] first input the welding image into the convolutional
segment samples can be mined explicitly by modelling the contextual long and short-term memory network, and then detected the welding
relationships of the boundary pixels. feature points through the similarity matching between multiple feature
layers. Zou and Zhou [86] firstly used CNN to learn welding image
4.4. Welding object tracking features, and secondly proposed a weld feature point detection method
based on multi-correlation filter collaboration considering the continu
In the object tracking task of the welding image, the current ity of feature points in adjacent frames. Zou et al. [87] used the SSD
616
T. Liu et al. Journal of Manufacturing Systems 68 (2023) 601–625
model to learn global and local weld features and combined it with RNN generate computed tomography results from eddy current testing re
to learn contextual information of visual features to achieve weld sults. Aiming at the problem of industrial fakes caused by repeated im
tracking under strong noise interference. Xiao et al. [207] used RCNN to aging in the field of NDT, Gao et al. [ 213] proposed an X-ray image
identify the weld type and detect the laser stripe ROI, then designed similarity comparison algorithm based on Siamese neural network.
different centre extraction algorithms for different types of welds and Table 11 summarizes the scenarios and key elements of DL in other
proposed a priori model to ensure the stability of the algorithm. In order subdivided WIR tasks.
to identify laser stripes and track welds in a strong interference envi
ronment, Zhang et al. [208] designed a full convolutional network based
4.6. Brief summary
on the idea of segmentation. Li et al. [209] first used Mask R-CNN to
detect the weld position and then extracted the weld line to assist the
The application of DL in welding image classification, object detec
robot in its inspection. Table 10 summarizes the key elements of DL in
tion, image segmentation and target tracking has realized the instanti
the welding object tracking task.
ation of DLBWIR. However, different task types have different
In summary, the main application scenario for welding object
applicable scenarios. Table 12 summarizes the applicable scenarios for
tracking is the weld seam tracking task by capturing key points/objects
different task types and the pros and cons of the application process. In
in the area to be welded, and the current research focuses on the
general, DL mainly faces challenges at three levels of data (small sample,
problems of strong arc light interference and small samples. As far as
imbalance), feature (strong interference, fine-grained) and model (real-
application scenarios are concerned, the authors believe that the task
time, explainability) in various WIR tasks, and related solutions are
can be applied to the object tracking scenario of the IW stage. In terms of
developed around the key technologies mentioned in Section 3. How
research points, the authors believe that tracking multiple objects in
ever, related research focuses on the application of DL methods in WIR,
welding image will be an interesting direction.
which lacks the guidance of welding domain knowledge.
Due to the difficulty of dataset construction, the DL method is less Welding image datasets are the basis for DLBWIR research. The
used in subdivided WIR tasks. publication of large datasets such as ImageNet [214] for classification
Wang et al. [92] designed a multi-scale feature fusion semantic tasks, COCO [215] or object detection tasks, SBD [216] for segmentation
segmentation network Res-Seg, which detected the edge of the molten tasks, and LaSOT [217] for tracking tasks is an important reason for the
pool with good continuity, and laid the foundation for the calculation of rapid development of DL in the CV field. Table 13 summarizes the public
width for molten pool and weld seam. Wang et al. [130] designed a datasets that exist in the field of DLBWIR. However, the public datasets
generative network to predict the future molten pool image, and then of DLBWIR still have the following shortcomings: (a) small amount of
input it into the regression network to guide weld reinforcement infor data; (b) few label categories; (c) unbalanced data; (d) lack of datasets
mation. In order to avoid extensive analytical calculations, Li et al. [23] for subdivided tasks.
used a LSTM network to establish a mapping mechanism from 2D im
aging points in the imaging plane to 3D reflection points on the molten 6. Future perspectives
pool surface, enabling data-driven 3D reconstruction of the molten pool
surface. In the reference [31], a transfer learning method based on 6.1. Perspectives from the literature
multi-objective instance weighting was designed to detect welding
quality from the perspective of equipment fault diagnosis and was Approximately 50 % of the examined 192 articles gave a clear
validated with robotic spot welding tests. In order to know the crystal research outlook. As shown in Fig. 14, this section summarizes their
lographic orientation during welding, Singh et al. [ 210] established a relevant future perspectives in terms of the DLBWIR paradigm as well as
generative model from ultrasonic wave travel times to crystallographic system integration in the following six areas.
orientations. Mirzapour et al. [ 211] proposed a fast and flexible
denoising CNN for the denoising problem of weld radiographic images. 6.1.1. Sensing
In order to perform NDT with a low-cost method, Rohkohl et al. [ 212] Visual sensing has the advantage of containing a wealth of
Table 10
Application of DL in welding image object tracking.
Ref. Year Stage Issue and challenge Method and result
Engineering Visual task Difficulty and challenge Backbone Key elements Data sources, volumes and
requirement metrics
[206] 2020 PRW Weld seam Welding line Strong interference Faster Fusion of timing information and Laboratory, 46819 and TE
tracking feature point R-CNN periodically initialize network of 0.5 mm
detection
[86] 2019 Weld seam Welding line Strong interference and VGG Fusion of timing information, Laboratory, 0 and TE of
tracking feature point small sample size correlative filtering and TL 0.25 mm
detection
[87] 2021 Weld seam Welding line Strong interference and SSD Combine timing information, Laboratory, 40000 and TE
tracking feature point small sample size multi-scale features, DA and TL of 0.5 mm
detection
[207] 2019 Weld seam Weld line center Strong interference and Faster Two-step extraction and type- Laboratory, 4000 and TE of
tracking point detection robustness R-CNN related weld centre extraction 0.28 mm
method.
[208] 2019 Weld seam Weld line position Strong interference and FCN TL and multi-scale feature fusion Laboratory, 1000 and
tracking detection small sample size PRMSE of 3.87
[209] 2022 POW Robot inspection Weld seam position Complex structure and Mask Multi-scale feature fusion Laboratory and
detection non-salient objects R-CNN production, 1500 and TE of
20 mm
617
T. Liu et al. Journal of Manufacturing Systems 68 (2023) 601–625
Table 11
Application of DL in other WIR tasks.
Ref. Year Stage Issue and challenge Method and result
Engineering Visual task Difficulty and challenge Backbone Key elements Data sources,
requirement volumes and metrics
[92] 2020 IW Weld width Edge detection of Small sample size and arc FCN DA and multi-scale feature fusion Laboratory, 1100
prediction molten pool interference and TE of 0.2 mm
[130] 2020 Weld Generation of Non-salient features Prednet Combine timing information and Laboratory, 25830
reinforcement molten pool image channel attention and SS of 93.27 %
prediction
[23] 2021 3D reconstruction 2D imaging points Strong interference, small LSTM Combine timing information and Laboratory and
of molten pool are mapped to 3D sample size and difficult to train the model based on the simulation, 5000
reflection points measure directly. simulation data and MSE of 0.002
[31] 2021 Welding product Robot spot welding Small sample size and Resnet Case-based TL Laboratory, 2160
fault diagnosis defect diagnosis transferability issue and Acc of 98.33 %
[210] 2022 Prediction of weld Image generation Determining the DNN Build a generative model from Simulation, 10000
tissue structure microstructure of weld ultrasonic wave travel times data and within 3◦
seam by NDT methods to crystallographic orientation.
[211] 2021 POW Weld radiographic Image denoising Image is blurry and details FFDnet Selective image enhancement Public dataset and
image recognition are not noticeable production, DWIQI
of 96 %
[212] 2022 Automatic NDT Image generation Comprehensive and U-Net Generate high-cost test results Laboratory, 294 and
inexpensive online from low-cost test results Acc of 93.7 %
inspection
[213] 2022 Fake film Image similarity Small sample size and VGG Comparisons by block Production, 342 and
identification comparison larger image resolution Acc of 93.458 %
618
T. Liu et al. Journal of Manufacturing Systems 68 (2023) 601–625
Table 13
Public datasets for DLBWIR.
Ref. Year Stage Amount Label Task URL
6.1.6. Integration
The WIR module does not exist independently, but exists in the robot
welding system, the robot welding production cell, and the
manufacturing system. Therefore, Nogay and Akinci [235] consider the
DL model as one of the modules in the robot welding cell, and emphasize
its horizontal integration with other modules such as motion modules.
Wang et al. [236] emphasize the vertical integration between DLBWIR
module and control system. Cruz et al. [57] and Chu et al. [237]
emphasize the end-to-end integration of existing models with produc
tion lines and even information systems through internet of things to
achieve interconnectivity.
619
T. Liu et al. Journal of Manufacturing Systems 68 (2023) 601–625
necessary to establish a public dataset in the field of welding image, but the KG, but also helps developers discover the causal relationship be
also to pre-train a batch of classic models for WIR tasks in subdivision tween the welding image and the welding state. Thus, giving the DL
fields and teams that lack large-scale test conditions. model semantic explainability and causality reasoning capabilities, it
provides a basis for developers to improve the model and optimize the
6.2.2. Semantic explainability for DLBWIR welding process.
The welding process is carried out under clear process guidance. The This part takes the radiographic image detection of welding seams of
welding image contains rich welding semantic information, which is a aerospace structural parts as an example to illustrate the semantic
carrier that reflects the welding state. However, DLBWIR is an end-to- explainability and causal reasoning framework of DLBWIR based on KG.
end way to establish the association relationship between the welding As shown in Fig. 15, it can be divided into three layers: data layer,
image and the welding state. This black box model not only lacks the knowledge layer, and explanation layer. The data layer mainly includes
reflection of the decision-making basis of the model, but also lacks an welding context data and multi-modal data. The multi-modal data in the
explicit expression of the causal relationship between the welding image process design stage mainly includes process cards and text documents.
and the welding state. Although the references [82,137] used feature The multi-modal data in the simulation stage includes temperature field
map visualization and class activation mapping methods to show users simulation data, stress field simulation data, and molten pool flow
the features learned by DL and the basis for making decisions for specific simulation data. The PRW stage mainly includes point cloud and image
examples, this explanation is a coarse-grained explanation oriented by data. The IW stage includes weld pool image data, instantaneous current
human visual perception. This explanation neither include the job se data, welding sound data, and so on. The multi-modal data in the POW
mantics in the welding process, nor the reasoning and revealing of the stage includes weld radiographic image data, weld actual width data,
causal relationship of the welding status information [139]. Therefore, weld metallographic image data, etc. After acquiring welding data, in
people still cannot intuitively capture the decision-making basis of the the knowledge layer, knowledge extraction, knowledge representation,
model through this visual explanation and optimize the welding process semantic alignment, knowledge fusion, and knowledge storage opera
accordingly. tions are performed on each modal data to form knowledge modules of
As a highly readable external knowledge carrier, the KG provides a each welding stage and synthesize to form the WDKG. In the explanation
way for the semantic expression of the welding manufacturing process layer, there are two parts that can be explained by semantics and the
[133]. Through KG, users can infer the associated path as the semantic basis for decision-making. Through causal reasoning technology,
explanation of the model decision. This type of association path not only various path dependencies that produce various defects can be found,
expresses the semantic information of the entities and relationships in thereby forming semantic explainability.
Fig. 15. KG-empowered semantic explanation for DLBWIR. The multimodal data involved in the welding context is integrated in the data layer. In the knowledge
layer, the sub-KG corresponding to each welding stage is constructed. Each sub-KG is fused at the explanation layer, and the basis for explanation reasoning of
welding defects is provided.
620
T. Liu et al. Journal of Manufacturing Systems 68 (2023) 601–625
As shown in Fig. 15, for a weld seam, visual explanation can provide [3] Yan Z, Xu H, Huang P. Multi-scale multi-intensity defect detection in ray image of
weld bead. NDT E Int 2020;116:102342. https://doi.org/10.1016/j.
the basis for decisions made by the DL model. However, this description
ndteint.2020.102342.
is imprecise and does not contain engineering semantics. The fusion of [4] Fan X, Gao X, Liu G, Ma N, Zhang Y. Research and prospect of welding
knowledge modules formed by welding contexts can lead to semantic monitoring technology based on machine vision. Int J Adv Manuf Tech 2021;115:
explanation. In terms of semantic explanation, if the DL-based target 3365–91. https://doi.org/10.1007/s00170-021-07398-4.
[5] Sahoo S, Lo CY. Smart manufacturing powered by recent technological
detection algorithm detects porosity defects, it can correspond to se advancements: a review. J Manuf Syst 2022;64:236–50. https://doi.org/
mantic knowledge such as stress concentration, impure shielding gas, 10.1016/j.jmsy.2022.06.008.
too much welding current, etc. Suppose the DL-based target detection [6] Wang J, Xu C, Zhang J, Zhong R. Big data analytics for intelligent manufacturing
systems: a review. J Manuf Syst 2022;62:738–52. https://doi.org/10.1016/j.
algorithm detects an unperforated defect. In that case, it can correspond jmsy.2021.03.005.
to the semantic knowledge of low molten pool temperature, too small [7] Cai W, Wang J, Jiang P, Cao L, Mi G, Zhou Q. Application of sensing techniques
bevel angle, too small welding current, too large welding speed, etc. The and artificial intelligence-based methods to laser welding real-time monitoring: A
critical review of recent literature. J Manuf Syst 2020;57:1–18. https://doi.org/
explanation formed by semantic association of the causes of welding 10.1016/j.jmsy.2020.07.021.
defects is for front-line operators, and its essence is the semantic [8] Wang B, Hu SJ, Sun L, Freiheit T. Intelligent welding system technologies: State-
knowledge formed by in-depth annotation of multi-stage and multi- of-the-art review and perspectives. J Manuf Syst 2020;56:373–91. https://doi.
org/10.1016/j.jmsy.2020.06.020.
modal information. [9] Hinton GE, Osindero S, Teh YW. A fast learning algorithm for deep belief nets.
Neural Comput 2006;18(7):1527–54. https://doi.org/10.1162/
7. Conclusion neco.2006.18.7.1527.
[10] Krizhevsky A, Sutskever I, Hinton GE. Imagenet classification with deep
convolutional neural networks. Commun Acm 2017;60(6):84–90. https://doi.
DLBWIR is becoming a core technology to provide forecasting and org/10.1145/3065386.
decision making in intelligent welding systems. This paper reviewed a [11] Silver D, Huang A, Maddison CJ, Guez A, Sifre L, Driessche GVD, et al. Mastering
the game of Go with deep neural networks and tree search. Nature 2016;529
total of 192 previous publications to summarize the state-of-the-art of (7587):484–9. https://doi.org/10.1038/nature16961.
the DLBWIR research and application. The main contributions of this [12] Wang J, Ma Y, Zhang L, Gao RX, Wu D. Deep learning for smart manufacturing:
work are summarized as follows: methods and applications. J Manuf Syst, 48(C); 2018. p. 144–56. 〈https://doi.
org/10.1016/j.jmsy.2018.01.003〉.
[13] Günther J, Pilarski PM, Helfrich G, Shen H, Diepold K. Intelligent laser welding
1) The key enabling technologies for the DLBWIR, such as sensing, data, through representation, prediction, and control learning: an architecture with
model, training, and platform were outlined. Moreover, the charac deep neural networks and reinforcement learning. Mechatronics 2016;34:1–11.
teristics and scope of application of the relevant technologies were https://doi.org/10.1016/j.mechatronics.2015.09.004.
[14] Zhang Y, Yang YP, Zhang W, Na SJ. Advanced welding manufacturing: a brief
summarized. analysis and review of challenges and solutions. J Manuf Sci Eng 2020;142(11):
2) The current applications of DL in different WIR tasks are summa 110816. https://doi.org/10.1115/1.4047947.
rized, based on which, the most challenging is the lack of guidance [15] Du R, Xu Y, Hou Z, Shu J, Chen S. Strong noise image processing for vision-based
seam tracking in robotic gas metal arc welding. Int J Adv Manuf Tech 2019;101
from welding domain knowledge. (5):2135–49. https://doi.org/10.1007/s00170-018-3115-2.
3) Two promising research works, namely opening datasets and se [16] Liu T, Bao J, Zheng H, Wang J, Yang C, Gu J. Learning semantic-specific visual
mantic explainability are highlighted, which are the key to promote representation for laser welding penetration status recognition. Sci China Technol
Sc 2021:1–14. https://doi.org/10.1007/s11431-021-1848-7.
the in-depth research and practical application of DLBWIR. [17] Jiao W, Wang Q, Cheng Y, Yu R, Zhang Y. Prediction of weld penetration using
dynamic weld pool arc images. Weld J 2020;99(11):295–302. https://doi.org/
Despite the rapid growth, DLBWIR remains a rapidly evolving field. 10.29391/2020.99.027.
[18] Li C, Wang Q, Jiao W, Johnson, Zhang Y. Deep learning-based detection of
Many pressing issues should be addressed to enhance its viability in penetration from weld pool reflection images. Weld J 2020;99(9):239–45.
practice. In that regard, we hope this paper can provide insightful https://doi.org/10.29391/2020.99.022.
knowledge and guide to attract more researchers in addressing the [19] Lu H, Li C, Chen W, Jiang Z. A single shot multibox detector based on welding
operation method for biometrics recognition in smart cities. Pattern Recogn Lett
future directions of the DLBWIR research and application.
2020;140:295–302. https://doi.org/10.1016/j.patrec.2020.10.016.
[20] Knaak C, von Eßen J, Kröger M, Schulze F, Abels P, Gillner A. A spatio-temporal
ensemble deep learning architecture for real-time defect detection during laser
Declaration of Competing Interest welding on low power embedded computing boards. Sens-Basel 2021;21(12):
4205. https://doi.org/10.3390/s21124205.
The authors declare that they have no known competing financial [21] Jin C, Shin S, Yu J, Rhee S. Prediction model for back-bead monitoring during gas
metal arc welding using supervised deep learning. IEEE Access 2020;8:
interests or personal relationships that could have appeared to influence 224044–58. https://doi.org/10.1109/ACCESS.2020.3041274.
the work reported in this paper. [22] Lee J, Noh I, Lee J, Lee SW. Development of an explainable fault diagnosis
framework based on sensor data imagification: a case study of the robotic spot-
welding process. IEEE Trans Ind Inf 2022;18(10):6895–904. https://doi.org/
Acknowledgement 10.1109/TII.2021.3134250.
[23] Li L, Cheng F, Wu S. An LSTM-based measurement method of 3D weld pool
This research is partially funded by the Mainland-Hong Kong Joint surface in GTAW. Measurement 2021;171:108809. https://doi.org/10.1016/j.
measurement.2020.108809.
Funding Scheme (MHX/001/20), Innovation and Technology Commis [24] Yang L, Fan J, Liu Y, Li E, Peng J, Liang Z. Automatic detection and location of
sion (ITC), Hong Kong Special Administration Region of the People’s weld beads with deep convolutional neural networks. IEEE Trans Instrum Meas
Republic of China, National Key R&D Programs of Cooperation on Sci 2021;70:1–12. https://doi.org/10.1109/TIM.2020.3026514.
[25] Yang L, Wang H, Huo B, Li F, Liu Y. An automatic welding defect location
ence and Technology Innovation with Hong Kong, Macao and Taiwan algorithm based on deep learning. NDT E Int 2021;120:102435. https://doi.org/
(SQ2020YFE020182), Ministry of Science and Technology (MOST) of 10.1016/j.ndteint.2021.102435.
the People’s Republic of China, and the Centrally Funded Postdoctoral [26] Oh S, Ki H. Cross-section bead image prediction in laser keyhole welding of AISI
1020 steel using deep learning architectures. IEEE Access 2020;8:73359–72.
Fellowship Scheme (1-YXBM), The Hong Kong Polytechnic University.
https://doi.org/10.1109/ACCESS.2020.2987858.
[27] Yan Y, Liu D, Gao B, Tian G, Cai Z. A deep learning-based ultrasonic pattern
References recognition method for inspecting girth weld cracking of gas pipeline. IEEE Sens J
2020;20(14):7997–8006. https://doi.org/10.1109/JSEN.2020.2982680.
[28] Wu D, Hu M, Huang Y, Zhang P, Yu Z. In situ monitoring and penetration
[1] Luo Z, Dai JS, Wang C, Wang F, Tian Y, Zhao M. Predictive seam tracking with
prediction of plasma arc welding based on welder intelligence-enhanced deep
iteratively learned feedforward compensation for high-precision robotic laser
random forest fusion. J Manuf Process 2021;66:153–65. https://doi.org/
welding. J Manuf Syst 2012;31(1):2–7. https://doi.org/10.1016/j.
10.1016/j.jmapro.2021.04.007.
jmsy.2011.03.005.
[29] Zhang Z, Wen G, Chen S. Weld image deep learning-based on-line defects
[2] Zhang Y, Zhao Z, Zhang Y, Bai L, Wang K, Han J. Online weld pool contour
detection using convolutional neural networks for Al alloy in robotic arc welding.
extraction and seam width prediction based on mixing spectral vision. Opt Rev
2019;26(1):65–76. https://doi.org/10.1007/s10043-018-0479-3.
621
T. Liu et al. Journal of Manufacturing Systems 68 (2023) 601–625
J Manuf Process 2019;45:208–16. https://doi.org/10.1016/j. digital image processing and deep learning techniques. Sens-Basel 2020;20(16):
jmapro.2019.06.023. 4505. https://doi.org/10.3390/s20164505.
[30] Yu R, He H, Han J, Bai L, Zhao Z, Lu J. Monitoring of back bead penetration based [58] Zhu H, Ge W, Liu Z. Deep Learning-based classification of weld surface defects.
on temperature sensing and deep learning. Measurement 2022;188:110410. Appl Sci-Basel 2019;9(16):3312. https://doi.org/10.3390/app9163312.
https://doi.org/10.1016/j.measurement.2021.110410. [59] Deng H, Cheng Y, Feng Y, Xiang J. Industrial laser welding defect detection and
[31] Lee K, Han S, Pham VH, Cho S, Choi HJ, Lee J, et al. Multi-objective instance image defect recognition based on deep learning model developed. Symmetry-
weighting-based deep transfer learning network for intelligent fault diagnosis. Basel 2021;13(9):1731. https://doi.org/10.3390/sym13091731.
Appl Sci-Basel 2021;11(5):2370. https://doi.org/10.3390/app11052370. [60] Li Y, Hu M, Wang T. Weld image recognition algorithm based on deep learning.
[32] Ren W, Wen G, Xu B, Zhang Z. A novel convolutional neural network based on Int J Pattern Recogn 2020;34(08):2052004. https://doi.org/10.1142/
time–frequency spectrogram of arc sound and its application on GTAW S0218001420520047.
penetration classification. IEEE T Ind Inf 2020;17(2):809–19. https://doi.org/ [61] Yang D, Cui Y, Yu Z, Yuan H. Deep learning based steel pipe weld defect
10.1109/TII.2020.2978114. detection. Appl Artif Intell 2021;35(15):1237–49. https://doi.org/10.1080/
[33] Miao R, Gao Y, Ge L, Jiang Z, Zhang J. Online defect recognition of narrow 08839514.2021.1975391.
overlap weld based on two-stage recognition model combining continuous [62] Cheng Y, Wang Q, Jiao W, Yu R, Chen S, Zhang Y, et al. Detecting dynamic
wavelet transform and convolutional neural network. Comput Ind 2019;112: development of weld pool using machine learning from innovative composite
103115. https://doi.org/10.1016/j.compind.2019.07.005. images for adaptive welding. J Manuf Process 2020;56:908–15. https://doi.org/
[34] Gantala T, Balasubramaniam K. Automated defect recognition for welds using 10.1016/j.jmapro.2020.04.059.
simulation assisted TFM imaging with artificial intelligence. J Nondestruct Eval [63] Wang D, Dong Y, Pan Y, Ma R. Machine vision-based monitoring methodology for
2021;40(1):1–24. https://doi.org/10.1007/s10921-021-00761-1. the fatigue cracks in U-rib-to-deck weld seams. IEEE Access 2020;8:94204–19.
[35] Dong Z, Mai Z, Yin S, Wang J, Yuan J, Fei Y. A weld line detection robot based on https://doi.org/10.1109/ACCESS.2020.2995276.
structure light for automatic NDT. Int J Adv Manuf Tech 2020;111(7):1831–45. [64] Yang L, Fan J, Huo B, Liu Y. Inspection of welding defect based on multi-feature
https://doi.org/10.1007/s00170-020-05964-w. fusion and a convolutional network. J Nondestruct Eval 2021;40(4):1–11.
[36] Song Y, Li S. Gas leak detection in galvanised steel pipe with internal flow noise https://doi.org/10.1007/s10921-021-00823-4.
using convolutional neural network. Process Saf Environ 2021;146:736–44. [65] Tao X, Peng L, Tao Y, Ye C. Inspection of defects in weld using differential array
https://doi.org/10.1016/j.psep.2020.11.053. ECT probe and deep learning algorithm. IEEE Trans Instrum Meas 2021;70:1–9.
[37] Zhi Z, Jiang H, Yang D, Gao J, Wang Q, Wang X, et al. An end-to-end welding https://doi.org/10.1109/TIM.2021.3099566.
defect detection approach based on titanium alloy time-of-flight diffraction [66] Choi SG, Hwang I, Kim YM, Kang B, Kang M. Prediction of the weld qualities
images. J Intell Manuf 2022. https://doi.org/10.1007/s10845-021-01905-w. using surface appearance image in resistance spot welding. Met-Basel 2019;9(8):
[38] Rout A, Deepak BBVL, Biswal BB. Advances in weld seam tracking techniques for 831. https://doi.org/10.3390/met9080831.
robotic welding: a review. Robot Cim-Int Manuf 2019;56:12–37. https://doi.org/ [67] Dong X, Taylor CJ, Cootes TF. Defect detection and classification by training a
10.1016/j.rcim.2018.08.003. generic convolutional neural network encoder. IEEE Trans Signal Proces 2020;68:
[39] Lei T, Rong Y, Wang H, Huang Y, Li M. A review of vision-aided robotic welding. 6055–69. https://doi.org/10.1109/TSP.2020.3031188.
Comput Ind 2020;123:103326. https://doi.org/10.1016/j. [68] Ajmi C, Zapata J, Elferchichi S, Zaafouri A, Laabidi K. Deep learning technology
compind.2020.103326. for weld defects classification based on transfer learning and activation features.
[40] Zhang Y, Wang Q, Liu Y. Adaptive intelligent welding manufacturing. Weld J Adv Mater Sci Eng 2020:1574350. https://doi.org/10.1155/2020/1574350.
2021;100(1):63–83. https://doi.org/10.29391/2021.100.006. [69] Guo S, Wang D, Chen J, Feng Z, Guo W. Predicting nugget size of resistance spot
[41] Bengio Y, Montreal U. Learning deep architectures for AI. Found Trends Mach Le welds using infrared thermal videos with image segmentation and convolutional
2009;2(1):1–87. https://doi.org/10.1561/2200000006. neural network. J Manuf Sci Eng 2022;144(2):021009. https://doi.org/10.1115/
[42] Deng L, Yu D. Deep learning: methods and applications. Found Trends Signal 1.4051829.
2014;7(3–4):197–387. https://doi.org/10.1561/2000000039. [70] Shorten C, Khoshgoftaar TM. A survey on image data augmentation for deep
[43] Schmidhuber J. Deep learning in neural networks: an overview. Neural Netw learning. J Big Data 2019;6(1):1–48. https://doi.org/10.1186/s40537-019-0197-
2015;61:85–117. https://doi.org/10.1016/j.neunet.2014.09.003. 0.
[44] Lecun Y, Bengio Y, Hinton GE. Deep learning. Nature 2015;521(7553):436–44. [71] Pan SJ, Yang Q. A survey on transfer learning. IEEE Trans Knowl Data En 2009;22
https://doi.org/10.1038/nature14539. (10):1345–59. https://doi.org/10.1109/TKDE.2009.191.
[45] Zhang W, Yang G, Lin Y, Ji C, Gupta MM. On definition of deep learning. In: IEEE [72] Goodfellow I, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, et al.
World Automation Congress; 2018. p. 232–6. 〈https://doi.org/10.23919/WAC. Generative adversarial nets. Commun Acm 2020;63(11):139–44. https://doi.org/
2018.8430387〉. 10.1145/3422622.
[46] Oh S, Jung M, Lim C, Shin S. Automatic detection of welding defects using Faster [73] Radford A, Metz L, Chintala S. Unsupervised representation learning with deep
R-CNN. Appl Sci-Basel 2020;10(23):8629. https://doi.org/10.3390/ convolutional generative adversarial networks. arXiv preprint arXiv:1511.06434;
app10238629. 2015.
[47] Zhang Z, Li B, Zhang W, Lu R, Wada S, Zhang Y. Real-time penetration state [74] Mirza M, Osindero S. Conditional generative adversarial nets. arXiv Prepr arXiv
monitoring using convolutional neural network for laser welding of tailor rolled 2014;1411:1784.
blanks. J Manuf Syst 2020;54:348–60. https://doi.org/10.1016/j. [75] Zhu JY, Park T, Isola P, Efros AA. Unpaired image-to-image translation using
jmsy.2020.01.006. cycle-consistent adversarial networks. In: IEEE International Conference on
[48] Shevchik S, Quang T, Meylan B, Farahani F, Olbinado M, Back A, et al. Supervised Computer Vision; 2017. p. 2242–51. 〈https://doi.org/10.1109/ICCV.2017.244〉.
deep learning for real-time quality monitoring of laser welding with X-ray [76] Mariani G, Scheidegger F, Istrate R, Bekas C, Malossi C. BAGAN: data
radiographic guidance. Sci-Rep-UK 2020;10:3389. https://doi.org/10.1038/ augmentation with balancing gan. arXiv preprint arXiv; 2018. p. 1803.09655.
s41598-020-60294-x. [77] Arjovsky M, Chintala S, Bottou L. Wasserstein generative adversarial networks.
[49] Wu D, Zhang P, Yu Z, Gao Y, Zhang H, Chen H, et al. Progress and perspectives of In: Proceedings of the 34th international conference on machine learning; 2017.
in-situ optical monitoring in laser beam welding: sensing, characterization and p. 214–23.
modeling. J Manuf Process 2022;75:767–91. https://doi.org/10.1016/j. [78] Mery D, Kaminetzky A, Golborne L, Figueroa S, Saavedra D. Target detection by
jmapro.2022.01.044. target simulation in X-ray testing. J Nondestruct Eval 2022;41:21. https://doi.
[50] Hou W, Zhang D, Wei Y, Guo J, Zhang X. Review on computer aided weld defect org/10.1007/s10921-022-00851-8.
detection from radiography images. Appl Sci-Basel 2020;10(5):1878. https://doi. [79] Cai W, Jiang P, Shu LS, Geng SN, Zhou Q. Real-time laser keyhole welding
org/10.3390/app10051878. penetration state monitoring based on adaptive fusion images using
[51] Xu Y, Wang Z. Visual sensing technologies in robotic welding: recent research convolutional neural networks. J Intell Manuf 2021. https://doi.org/10.1007/
developments and future interests. Sens Actuat A-Phys 2021;320:112551. s10845-021-01848-2.
https://doi.org/10.1016/j.sna.2021.112551. [80] Cheng YC, Wang QY, Jiao WH, Xiao J, Chen SJ, Zhang YM. Automated
[52] Wang Y, Lu J, Zhao Z, Deng W, Han J, Bai L, et al. Active disturbance rejection recognition of weld pool characteristics from active vision sensing. Weld J 2021;
control of layer width in wire arc additive manufacturing based on deep learning. 100(5):183–92. https://doi.org/10.29391/2021.100.015.
J Manuf Process 2021;67:364–75. https://doi.org/10.1016/j. [81] Xue B, Chang B, Du D. Multi-output monitoring of high-speed laser welding state
jmapro.2021.05.005. based on deep learning. Sens-Basel 2021;21(5):1626. https://doi.org/10.3390/
[53] Park JK, An WH, Kang DJ. Convolutional neural network based surface inspection s21051626.
system for non-patterned welding defects. Int J Precis Eng Man 2019;20(3): [82] Xia C, Pan Z, Fei Z, Zhang S, Li H. Vision based defects detection for keyhole TIG
363–74. https://doi.org/10.1007/s12541-019-00074-4. welding using deep learning with visual explanation. J Manuf Process 2020;56:
[54] Yang Y, Pan L, Ma J, Yang R, Zhu Y, Yang Y, et al. A high-performance deep 845–55. https://doi.org/10.1016/j.jmapro.2020.05.033.
learning algorithm for the automated optical inspection of laser welding. Appl [83] Ajmi C, Zapata J, Martínez-Álvarez JJ, Doménech G, Ruiz R. Using deep learning
Sci-Basel 2020;10(3):933. https://doi.org/10.3390/app10030933. for defect classification on a small weld X-ray image dataset. J Nondestruct Eval
[55] Bacioiu D, Melton G, Papaelias M, Shaw R. Automated defect classification of 2020;39(3):1–13. https://doi.org/10.1007/s10921-020-00719-9.
SS304 TIG welding process using visible spectrum camera and machine learning. [84] Dai W, Li D, Tang D, Jiang Q, Wang D, Wang H, et al. Deep learning assisted
NDT E Int 2019;107:102139. https://doi.org/10.1016/j.ndteint.2019.102139. vision inspection of resistance spot welds. J Manuf Process 2021;62:262–74.
[56] Ma G, Yu L, Yuan H, Xiao W, He Y. A vision-based method for lap weld defects https://doi.org/10.1016/j.jmapro.2020.12.015.
monitoring of galvanized steel sheets using convolutional neural network. [85] Zhou C, Li D, Wang P, Sun J, Huang Y, Li W. ACR-Net: attention integrated and
J Manuf Process 2021;64:130–9. https://doi.org/10.1016/j.jmapro.2020.12.067. cross-spatial feature fused rotation network for tubular solder joint detection.
[57] Cruz YJ, Rivas M, Quiza R, Beruvides G, Haber RE. Computer vision system for IEEE Trans Instrum Meas 2021;70:1–12. https://doi.org/10.1109/
welding inspection of liquefied petroleum gas pressure vessels based on combined TIM.2021.3094837.
622
T. Liu et al. Journal of Manufacturing Systems 68 (2023) 601–625
[86] Zou Y, Zhou W. Automatic seam detection and tracking system for robots based [114] Hu J, Shen L, Albanie S, Sun G, Wu E. Squeeze-and-excitation networks. IEEE
on laser vision. Mechatronics 2019;63:102261. https://doi.org/10.1016/j. Trans Pattern Anal 2020;42(8):2011–23. https://doi.org/10.1109/
mechatronics.2019.102261. TPAMI.2019.2913372.
[87] Zou Y, Zhu M, Chen X. A robust detector for automated welding seam tracking [115] Woo S, Park J, Lee JY, Kweon IS. CBAM: convolutional block attention module.
system. J Dyn Sys Meas Control 2021;143(7):071001. https://doi.org/10.1115/ In: European Conference on Computer Vision; 2018. p. 3–19.
1.4049547. [116] Biran O, Cotton C . Explanation and justification in machine learning: a survey. In:
[88] Gonzalez-Val C, Pallas A, Panadeiro V, Rodriguez A. A convolutional approach to International Joint Conference on Artificial Intelligence; 2017. p. 8–13.
quality monitoring for laser manufacturing. J Intell Manuf 2020;31(3):789–95. [117] Selvaraju RR, Cogswell M, Das A, Vedantam R, Parikh D, Batra D.. Grad-cam:
https://doi.org/10.1007/s10845-019-01495-8. visual explanations from deep networks via gradient-based localization. In: IEEE
[89] Yang L, Wang Z, Gao S, Shi M, Liu B. Magnetic flux leakage image classification International Conference on Computer Vision; 2017. p. 618–626. 〈https://doi.
method for pipeline weld based on optimized convolution kernel. org/10.1109/ICCV.2017.74〉.
Neurocomputing 2019;365:229–38. https://doi.org/10.1016/j. [118] Zhou BL, Khosla A, Lapedriza A, Oliva A, Torralba A. Learning deep features for
neucom.2019.07.083. discriminative localization. In: IEEE Conference on Computer Vision and Pattern
[90] Dong X, Taylor CJ, Cootes TF. Automatic aerospace weld inspection using Recognition; 2016. p. 27–30. 〈https://doi.org/10.1109/CVPR.2016.319〉.
unsupervised local deep feature learning. Knowl-Based Syst 2021;221:106892. [119] Jiang R, Xiao R, Chen S. Prediction of penetration based on infrared thermal and
https://doi.org/10.1016/j.knosys.2021.106892. visual images during pulsed GTAW process. J Manuf Process 2021;69:261–72.
[91] Liu C, Chen S, Huang J. Machine vision-based object detection strategy for weld https://doi.org/10.1016/j.jmapro.2021.07.046.
area. SCI Program-Neth 2022;2022:1188974. https://doi.org/10.1155/2022/ [120] Ma D, Jiang P, Shu L, Geng S. Multi-sensing signals diagnosis and CNN-based
1188974. detection of porosity defect during Al alloys laser welding. J Manuf Syst 2022;62:
[92] Wang Y, Han J, Lu J, Bai L, Zhao Z. TIG stainless steel molten pool contour 334–46. https://doi.org/10.1016/j.jmsy.2021.12.004.
detection and weld width prediction based on res-seg. Met-Basel 2020;10(11): [121] Jamnikar ND, Liu S, Brice C, Zhang X. In-process comprehensive prediction of
1495. https://doi.org/10.3390/met10111495. bead geometry for laser wire-feed DED system using molten pool sensing data and
[93] Ai J, Tian L, Bai L, Zhang J. Data augmentation by a CycleGAN-based extra- multi-modality CNN. Int J Adv Manuf Tech 2022;121:903–17. https://doi.org/
supervised model for nondestructive testing. Meas Sci Technol 2022;33:045017. 10.1007/s00170-022-09248-3.
https://doi.org/10.1088/1361-6501/ac3ec3. [122] Wu D, Huang Y, Zhang P, Yu Z, Chen H, Chen S. Visual-acoustic penetration
[94] Dai W, Li D, Tang D, Wang H, Peng Y. Deep learning approach for defective spot recognition in variable polarity plasma arc welding process using hybrid deep
welds classification using small and class-imbalanced datasets. Neurocomputing learning approach. IEEE Access 2020;8:120417–28. https://doi.org/10.1109/
2022;477:46–60. https://doi.org/10.1016/j.neucom.2022.01.004. ACCESS.2020.3005822.
[95] Le X, Mei J, Zhang H, Zhou B, Xi J. A learning-based approach for surface defect [123] Miao R, Shan Z, Zhou Q, Wu Y, Ge L, Zhang J, et al. Real-time defect identification
detection using small image datasets. Neurocomputing 2020;408:112–20. of narrow overlap welds and application based on convolutional neural networks.
https://doi.org/10.1016/j.neucom.2019.09.107. J Manuf Syst 2022;62:800–10. https://doi.org/10.1016/j.jmsy.2021.01.012.
[96] Lecun Y, Bottou L, Bengio Y, Haffner P. Gradient-based learning applied to [124] Chen C, Xiao R, Chen H, Lv N, Chen S. Prediction of welding quality
document recognition. Prpc IEEE 86(11); 1998. p. 2278–324. 〈https://doi. characteristics during pulsed GTAW process of aluminum alloy by multisensory
org/10.1109/5.726791〉. fusion and hybrid network model. J Manuf Processes, 68(B); 2020. p. 209–24.
[97] He K, Zhang X, Ren SQ, Sun J. Deep residual learning for image recognition. In: 〈https://doi.org/10.1016/j.jmapro.2020.08.028〉.
IEEE conference on computer vision and pattern recognition; 2016. p. 770–778. [125] Hou W, Rao L, Zhu A, Zhang D. Feature fusion for weld defect classification with
〈https://doi.org/10.1109/CVPR.2016.90〉. small dataset. J Sens 2022;2022:8088202. https://doi.org/10.1155/2022/
[98] Simonyan K, Zisserman A. Very deep convolutional networks for large-scale 8088202.
image recognition. arXiv Prepr arXiv 2015;1409:1556. [126] Kershaw J, Yu R, Zhang YM, Wang P. Hybrid machine learning-enabled adaptive
[99] Howard AG Zhu M, Chen B, Kalenichenko D, Wang WJ, Weyand T., et al. welding speed control. J Manuf Process 2021;71:374–83. https://doi.org/
MobileNets: efficient convolutional neural networks for mobile vision 10.1016/j.jmapro.2021.09.023.
applications. arXiv preprint arXiv: 1704.04861; 2017. [127] Feng Y, Chen Z, Wang D, Chen J, Feng Z. DeepWelding: a deep learning enhanced
[100] Girshick R, Donahue J., Darrell T., Malik J. Rich feature hierarchies for accurate approach to GTAW using multisource sensing images. IEEE Trans Ind Inf 2019;16
object detection and semantic segmentation. In: IEEE Conference on Computer (1):465–74. https://doi.org/10.1109/TII.2019.2937563.
Vision and Pattern Recognition; 2014. p. 580–7. 〈https://doi.org/10.1109/CVP [128] Cai N, Cen G, Wu J, Li F, Wang H, Chen X. SMT solder joint inspection via a novel
R.2014.81〉. cascaded convolutional neural network. IEEE Trans Comp Pack Man 2018;8(4):
[101] Girshick, R. Fast R-CNN. In: IEEE International Conference on Computer Vision. 670–7. https://doi.org/10.1109/TCPMT.2018.2789453.
2015: 1440–1448. 〈https://doi.org/10.1109/ICCV.2015.169〉. [129] Yang L, Wang Z, Gao S. Pipeline magnetic flux leakage image detection algorithm
[102] Ren SQ, He K, Girshick R, Sun J. Faster R-CNN: towards real-time object detection based on multiscale SSD network. IEEE Trans Ind Inf 2019;16(1):501–9. https://
with region proposal networks. IEEE Trans Pattern Anal 2017;39(6). https://doi. doi.org/10.1109/TII.2019.2926283.
org/10.1109/TPAMI.2016.2577031. 1137-49. [130] Wang Y, Zhang C, Lu J, Bai L, Zhao Z, Han J. Weld reinforcement analysis based
[103] Redmon J, Divvala S, Girshick R, Farhadi A. You only look once: unified, real- on long-term prediction of molten pool image in additive manufacturing. IEEE
time object detection. In: IEEE Conference on Computer Vision and Pattern Access 2020;8:69908–18. https://doi.org/10.1109/ACCESS.2020.2986130.
Recognition; 2016. p. 779–88. 〈https://doi.org/10.1109/CVPR.2016.91〉. [131] Liu T, Wang J, Huang X, Lu Y, Bao J. 3DSMDA-Net: an improved 3DCNN with
[104] Liu W, Anguelov D, Erhan D, Szegedy C, Reed S, Fu CY, et al. SSD: single shot separable structure and multi-dimensional attention for welding status
multibox detector. In: European Conference on Computer Vision; 2016. p. 21–37. recognition. J Manuf Syst 2022;62:811–22. https://doi.org/10.1016/j.
〈https://doi.org/10.1007/978-3-319-46448-0_2〉. jmsy.2021.01.017.
[105] Shelhamer E, Long J, Darrell T. Fully convolutional networks for semantic [132] Feng T, Huang S, Liu J, Wang J, Fang X. Welding surface inspection of armatures
segmentation. IEEE Trans Pattern Anal 2017;39(4):640–51. https://doi.org/ via CNN and image comparison. IEEE Sens J 2021;21(19):21696–704. https://
10.1109/TPAMI.2016.2572683. doi.org/10.1109/JSEN.2021.3079334.
[106] Ronneberger O, Fischer P. Brox T. U-Net: convolutional networks for biomedical [133] Li Q, Yang B, Wang S, Zhang Z, Tang X, Zhao C. A fine-grained flexible graph
image segmentation. In: International conference on medical image computing convolution network for visual inspection of resistance spot welds using cross-
and computer assisted intervention; 2015. p. 234–41. 〈https://doi.org/10.1007/9 domain features. J Manuf Process 2022;78:319–29. https://doi.org/10.1016/j.
78-3-319-24574-4_28〉. jmapro.2022.04.025.
[107] Yu R, Kershaw J, Wang P, Zhang Y. Real-time recognition of arc weld pool using [134] Li W, Qi J, Sun H. TAF2-Net: triple-branch attentive feature fusion network for
image segmentation network. J Manuf Process 2021;72:159–67. https://doi.org/ ultrasonic flaw detection. IEEE Trans Instrum Meas 2022;71:1–12. https://doi.
10.1016/j.jmapro.2021.10.019. org/10.1109/TIM.2022.3150592.
[108] He K, Gkioxari G, Dollár P, Girshick R. Mask R-CNN. In: IEEE International [135] Li Z, Chen H, Ma X, Chen H, Ma Z. Triple pseudo-siamese network with hybrid
Conference on Computer Vision; 2017. p. 2961–9. 〈https://doi.org/10.1109/ attention mechanism for welding defect detection. Mater Des 2022;217:110645.
ICCV.2017.322〉. https://doi.org/10.1016/j.matdes.2022.110645.
[109] Ma C, Huang J, Yang X, Yang M. Hierarchical convolutional features for visual [136] Wang Z, Chen H, Zhong Q, Lin S, Wu J, Xu M, et al. Recognition of penetration
tracking. In: IEEE International Conference on Computer Vision; 2015. p. state in GTAW based on vision transformer using weld pool image. Int J Adv
3074–3082. 〈https://doi.org/10.1109/ICCV.2015.352〉. Manuf Tech 2022;119:5439–52. https://doi.org/10.1007/s00170-021-08538-6.
[110] Nam H, Han B. Learning multi-domain convolutional neural networks for visual [137] Liu T, Bao J, Wang J, Zhang Y. A hybrid CNN–LSTM algorithm for online defect
tracking. In: IEEE Conference on Computer Vision and Pattern Recognition; 2016. recognition of CO2 welding. Sens-Basel 2018;18(12):4369. https://doi.org/
p. 4293–4302. 〈https://doi.org/10.1109/CVPR.2016.465〉. 10.3390/s18124369.
[111] Danelljan M, Bhat G, Khan F, Felsberg MECO.: efficient convolution operators for [138] Božič A, Kos M, Jezeršek M. Power control during remote laser welding using a
tracking. In: IEEE Conference on Computer Vision and Pattern Recognition; 2017. convolutional neural network. Sens-Basel 2020;20(22):6658. https://doi.org/
p. 6931–6939. 〈https://doi.org/10.1109/CVPR.2017.733〉. 10.3390/s20226658.
[112] Nazir Q, Shao C. Online tool condition monitoring for ultrasonic metal welding [139] Liu T, Zheng H, Bao J, Zheng P, Wang J, Yang C, et al. An explainable laser
via sensor fusion and machine learning. J Manuf Process 2021;62:806–16. welding defect recognition method based on multi-scale class activation mapping.
https://doi.org/10.1016/j.jmapro.2020.12.050. IEEE Trans Instrum Meas 2022;71:1–12. https://doi.org/10.1109/
[113] Jaderberg M, Simonyan K, Zisserman A, Kavukcuoglu K Spatial transformer TIM.2022.3148739.
networks. In: International Conference on Neural Information Processing [140] Ruder S. An overview of gradient descent optimization algorithms. arXiv preprint
Systems; 2015. p. 2017–25. arXiv:1609.04747; 2016.
623
T. Liu et al. Journal of Manufacturing Systems 68 (2023) 601–625
[141] Li Y, Mu H, Polden H, Li H, Wang L, Xia C, Pan Z. Towards intelligent monitoring [169] Kumaresan S, Aultrin KSJ, Kumar SS, Anand MD. Transfer learning with CNN for
system in wire arc additive manufacturing: a surface anomaly detector on a small classification of weld defect. IEEE Access 2021;9:95097–108. https://doi.org/
dataset. Int J Adv Manuf Tech 2022;120:5225–42. https://doi.org/10.1007/ 10.1109/ACCESS.2021.3093487.
s00170-022-09076-5. [170] Miao R, Jiang Z, Zhou Q, Wu Y, Gao Y, Zhang J, et al. Online inspection of narrow
[142] Sekhar R, Sharma D, Shah P. Intelligent classification of tungsten inert gas overlap weld quality using two-stage convolution neural network image
welding defects: a transfer learning approach. Front Mech Eng 2022;8:824038. recognition. Mach Vis Appl 2021;32(1):27. https://doi.org/10.1007/s00138-020-
https://doi.org/10.3389/fmech.2022.824038. 01158-2.
[143] Ioffe S, Szegedy C. Batch normalization: accelerating deep network training by [171] Zhang B, Jaiswal P, Rai R, Guerrier P, Baggs G. Convolutional neural network-
reducing internal covariate shift. In: International Conference on Machine based inspection of metal additive manufacturing parts. Rapid Prototyp J 2019;25
Learning; 2015. p. 448–56. (3):530–40. https://doi.org/10.1108/RPJ-04-2018-0096.
[144] Srivastava N, Hinton G, Krizhevsky A, Sutskever I, Salakhutdinov R. Dropout: a [172] Hu A, Wu L, Huang J, Fan D, Xu Z. Recognition of weld defects from X-ray images
simple way to prevent neural networks from overfitting. J Mach Learn Res 2014; based on improved convolutional neural network. Multimed Tools Appl 2022;81:
15(1):1929–58. 15085–102. https://doi.org/10.1007/s11042-022-12546-3.
[145] Pan H, Pang Z, Wang Y, Wang Y, Chen L. A new image recognition and [173] Di X, Li L. Image defect recognition method based on deep learning network. Mob
classification method combining transfer learning algorithm and mobilenet model Inf Syst 2022;2022:5696334. https://doi.org/10.1155/2022/5696334.
for welding defects. IEEE Access 2020;8:119951–60. https://doi.org/10.1109/ [174] Zhang Z, Liu W, Sun X. Image recognition of limited and imbalanced samples
ACCESS.2020.3005450. based on transfer learning methods for defects in welds. Proc Inst Mech Eng B-J
[146] Ghiasi G, Lin TY, Le Q.V. Dropblock: a regularization method for convolutional Eng 2022. https://doi.org/10.1177/09544054221082779.
networks. arXiv preprint arXiv:1810.12890; 2018. [175] Liu J, Liu X, Qu F, Zhang H, Zhang L. A defect recognition method for low-quality
[147] Yang L, Jiang H. Weld defect classification in radiographic images using unified weld image based on consistent multiscale feature mapping. IEEE Trans Instrum
deep neural network with multi-level features. J Intell Manuf 2020;32(2):459–69. Meas 2022;71:1–11. https://doi.org/10.1109/TIM.2022.3171609.
https://doi.org/10.1007/s10845-020-01581-2. [176] Lin CS, Huang YC, Chen SH, Hsu YL, Lin YC. The application of deep learning and
[148] Chen Y, Su S, Yang H. Convolutional neural network analysis of recurrence plots image processing technology in laser positioning. Appl Sci-Basel 2018;8(9):1542.
for anomaly detection. Int J Bifurcat Chaos 2020;30(01):2050002. https://doi. https://doi.org/10.3390/app8091542.
org/10.1142/S0218127420500029. [177] Wang Q, Jiao W, Wang P, Zhang Y. A tutorial on deep learning-based data
[149] Zhang B, Liu S, Shin YC. In-process monitoring of porosity during laser additive analytics in manufacturing through a welding case study. J Manuf Process 2021;
manufacturing process. Addit Manuf 2019;28:497–505. https://doi.org/10.1016/ 63:2–13. https://doi.org/10.1016/j.jmapro.2020.04.044.
j.addma.2019.05.030. [178] Wang Y, Xu X, Zhao Z, Deng W, Han J, Bai L, et al. Coordinated monitoring and
[150] Liu J, Guo F, Zhang Y, Hou B, Zhou H. Defect classification on limited labeled control method of deposited layer width and reinforcement in WAAM process.
samples with multiscale feature fusion and semi-supervised learning. Appl Intell J Manuf Process 2021;71:306–16. https://doi.org/10.1016/j.
2021:1–16. https://doi.org/10.1007/s10489-021-02917-y. jmapro.2021.09.033.
[151] Li S, Gao J, Zhou E, Pan Q, Wang X. Deep learning-based fusion hole state [179] Kim H, Nam K, Oh S, Ki H. Deep-learning-based real-time monitoring of full-
recognition and width extraction for thin plate TIG welding. Weld World 2022;66: penetration laser keyhole welding by using the synchronized coaxial observation
1329–47. https://doi.org/10.1007/s40194-022-01287-4. method. J Manuf Process 2021;68:1018–30. https://doi.org/10.1016/j.
[152] Xiao M, Yang B, Wang S, Chang Y, Li S, Yi G. Research on recognition methods of jmapro.2021.06.029.
spot-welding surface appearances based on transfer learning and a lightweight [180] Yu R, Kershaw J, Wang P, Zhang Y. How to accurately monitor the weld
high-precision convolutional neural network. J Intell Manuf 2022. https://doi. penetration from dynamic weld pool serial images using CNN-LSTM deep
org/10.1007/s10845-022-01909-0. learning model? IEEE Robot Autom Let, 7(3); 2022. p. 6519–25. 〈https://doi.
[153] Wang B, Li F, Lu R, Ni X, Zhu W. Weld feature extraction based on semantic org/10.1109/LRA.2022.3173659〉.
segmentation network. Sens-Basel 2022;22(11):4130. https://doi.org/10.3390/ [181] Zhang K, Shen H. Solder joint defect detection in the connectors using improved
s22114130. Faster-RCNN algorithm. Appl Sci-Basel 2021;11(2):576. https://doi.org/
[154] Park MY, Hastie T. L1–regularization path algorithm for generalized linear 10.3390/app11020576.
models. J R Stat Soc B 2007;69(4):659–77. https://doi.org/10.1111/j.1467- [182] Yun GH, Oh SJ, Shin SC. Image preprocessing method in radiographic inspection
9868.2007.00607.x. for automatic detection of ship welding defects. Appl Sci-Basel 2022;12(1):123.
[155] Cortes C, Mohri M, Rostamizadeh A. L2 regularization for learning kernels. arXiv https://doi.org/10.3390/app12010123.
Prepr arXiv 2012;1205:2653. [183] Chen Y, Wang J, Wang G. Intelligent welding defect detection model on improved
[156] Liu T, Bao J, Wang J, Zhang Y. A coarse-grained regularization method of R-CNN. IETE J Res 2022:1–10. https://doi.org/10.1080/
convolutional kernel for molten pool defect identification. J Comput Inf Sci Eng 03772063.2022.2040387.
2020;20(2):021005. https://doi.org/10.1115/1.4045294. [184] Li H, Li B, Wen X, Wang H, Wang X, Zhang S, et al. A detection and configuration
[157] Prechelt L. Early stopping-but when? Neural Netw: Tricks Trade 1998:55–69. method for welding completeness in the automotive body‑in‑white panel based
https://doi.org/10.1007/3-540-49430-8_3. on digital twin. Sci Rep-UK 2022;12:7929. https://doi.org/10.1038/s41598-022-
[158] Abadi M, Barham P, Chen J, Chen Z, Davis A, Dean J., et al. Tensorflow: a system 11440-0.
for large-scale machine learning. In: Operating Systems Design and [185] Chen Z, Wang Q, Yang K, Yu T, Yao J, Liu Y, et al. Deep learning for the detection
Implementation; 2016. p. 265–83. and recognition of rail defects in ultrasound B-scan images. Transp Res Rec 2021;
[159] Parvat A., Chavan J., Kadam S., Dev S., Pathak V. A survey of deep-learning 2675(11):888–901. https://doi.org/10.1177/03611981211021547.
frameworks. In: International Conference on Inventive Systems and Control; [186] Wu H, Gao W, Xu X. Solder joint recognition using mask R-CNN method. IEEE
2017. p. 1–7. 〈https://doi.org/10.1109/ICISC.2017.8068684〉. Trans Comp Pack Man 2019;10(3):525–30. https://doi.org/10.1109/
[160] Gulli A, Pal S. Deep learning with Keras. Packt Publishing Ltd; 2017. TCPMT.2019.2952393.
[161] Paszke A, Gross S, Massa F, Lerer A, Bradury J, Chanan G, et al. Pytorch: an [187] Yang L, Fan J, Huo B, Li E, Liu Y. Image denoising of seam images with deep
imperative style, high-performance deep learning library. Adv Neural Inf Process learning for laser vision seam tracking. IEEE Sens J 2022;22(6):6098–107.
Syst 2019;32:8026–37. https://doi.org/10.1109/JSEN.2022.3147489.
[162] Han CM, Kim TW, Choi HW. Assessment of laser joining quality by visual [188] Mi J, Zhang Y, Li H, Shen S, Yang Y, Song C, et al. In-situ monitoring laser based
inspection, computer Simulation, and deep learning. Appl Sci-Basel 2021;11(2): directed energy deposition process with deep convolutional neural network.
642. https://doi.org/10.3390/app11020642. J Intell Manuf 2021:1–11. https://doi.org/10.1007/s10845-021-01820-0.
[163] Mishra A, Patti A. Deep convolutional neural network modeling and laplace [189] Tan Z, Fang Q, Li H, Liu S, Zhu W, Yang D. Neural network based image
transformation algorithm for the analysis of surface quality of friction stir welded segmentation for spatter extraction during laser-based powder bed fusion
joints. ADCAIJ-Adv Distrib C, 10(3); 2021. 3–7-320. 〈http://hdl.handle.net/103 processing. Opt Laser Technol 2020;130:106347. https://doi.org/10.1016/j.
66/147243〉. optlastec.2020.106347.
[164] Tian Y, Liu H, Li L, Yuan G, Feng J, Chen Y, et al. Automatic identification of [190] Yang Y, He Y, Guo H, Chen Z, Zhang L. Semantic segmentation supervised deep-
multi-type weld seam based on vision sensor with silhouette-mapping. IEEE Sens learning algorithm for welding-defect detection of new energy batteries. Neural
J 2020;21(4):5402–12. https://doi.org/10.1109/JSEN.2020.3034382. Comput Appl 2022. https://doi.org/10.1007/s00521-022-07474-0.
[165] Zhang Y, You D, Gao X, Wang C, Li Y, Gao PP. Real-time monitoring of high- [191] Ling Z, Zhang A, Ma D, Shi Y, Wen H. Deep siamese semantic segmentation
power disk laser welding statuses based on deep learning framework. J Intell network for PCB welding defect detection. IEEE Trans Instrum Meas 2022;71:
Manuf 2019;31:799–814. https://doi.org/10.1007/s10845-019-01477-w. 1–11. https://doi.org/10.1109/TIM.2022.3154814.
[166] Jia CB, Liu XF, Zhang GK, Zhang Y, Yu CH, Wu CS. Penetration/keyhole status [192] Li Z, Li D. A high-frequency feature enhancement network for the surface defect
prediction and model visualization based on deep learning algorithm in plasma detection of welded rebar. Struct Control Health 2022;29(8):e2983. https://doi.
arc welding. Int J Adv Manuf Tech 2021;117(11):3577–97. https://doi.org/ org/10.1002/stc.2983.
10.1007/s00170-021-07903-9. [193] Pandiyan V, Murugan P, Tjahjowidodo T, Caesarendra W, Manyar OM, Then DJH.
[167] Hou W, Wei Y, Jin Y, Zhu C. Deep features based on a DCNN model for classifying In-process virtual verification of weld seam removal in robotic abrasive belt
imbalanced weld flaw types. Measurement 2019;131:482–9. https://doi.org/ grinding process using deep learning. Robot Cim-Int Manuf 2019;57:477–87.
10.1016/j.measurement.2018.09.011. https://doi.org/10.1016/j.rcim.2019.01.006.
[168] Jiang H, Hu Q, Zhi Z, Gao J, Gao Z, Wang R, et al. Convolution neural network [194] Liu H, Yan Y, Song K, Chen H, Yu H. Efficient optical measurement of welding
model with improved pooling strategy and feature selection for weld defect studs with normal maps and convolutional neural network. IEEE Trans Instrum
recognition. Weld World 2021;65(4):731–44. https://doi.org/10.1007/s40194- Meas 2020;70:1–14. https://doi.org/10.1109/TIM.2020.3024389.
020-01027-6. [195] Zhu Y, Yang R, He Y, Ma J, Guo H, Yang Y, et al. A lightweight multiscale
attention semantic segmentation algorithm for detecting laser welding defects on
624
T. Liu et al. Journal of Manufacturing Systems 68 (2023) 601–625
safety vent of power battery. IEEE Access 2021;9:39245–54. https://doi.org/ Pattern Recognition; 2019. p. 5374–83. 〈https://doi.org/10.1109/CVPR.201
10.1109/ACCESS.2021.3064180. 9.00552〉.
[196] Jang J, Van D, Jang H, Baik DH, Yoo SD, Park J, et al. Residual neural network- [218] Bacioiu D, Melton G, Papaelias M, Shaw R. Automated defect classification of
based fully convolutional network for microstructure segmentation. Sci Technol aluminium 5083 TIG welding using HDR camera and neural networks. J Manuf
Weld Joi 2020;25(4):282–9. https://doi.org/10.1080/13621718.2019.1687635. Process 2019;45:603–13. https://doi.org/10.1016/j.jmapro.2019.07.020.
[197] Ferguson MK, Ronay A, Lee YTT, Law KH. Detection and segmentation of [219] Martínez RT, Bestard GA, Silva AMA, Alfaro SCA. Analysis of GMAW process with
manufacturing defects with convolutional neural networks and transfer learning. deep learning and machine learning techniques. J Manuf Process 2021;62:
Smart Sustain Manuf 2018;2:1–43. https://doi.org/10.1520/SSMS20180033. 695–703. https://doi.org/10.1016/j.jmapro.2020.12.052.
[198] Zhang H, Di X, Zhang Y. Real-time CU-Net-based welding quality inspection [220] Mery D, Riffo V, Zscherpel U, Mondragón G, Lillo I, Zuccar I, et al. GDXray: the
algorithm in battery production. IEEE Trans Ind Electron 2020;67(12):10942–50. database of X-ray images for nondestructive testing. J Nondestruct Eval 2015;34
https://doi.org/10.1109/TIE.2019.2962421. (4):42. https://doi.org/10.1007/s10921-015-0315-7.
[199] Wang L, Shen Q. Visual inspection of welding zone by boundary-aware semantic [221] Hartl R, Bachmann A, Habedank JB, Semm T, Zaeh MF. Process monitoring in
segmentation algorithm. IEEE T Instrum Meas 2020;70:1–9. https://doi.org/ friction stir welding using convolutional neural networks. Met-Basel 2021;11(4):
10.1109/TIM.2020.3010665. 535. https://doi.org/10.3390/met11040535.
[200] Yang L, Fan J, Huo B, Li E, Liu Y. A nondestructive automatic defect detection [222] Kandala RR, Abhinay A, Aechana S.; 2021. 〈https://www.kaggle.com/ruthka/ma
method with pixelwise segmentation. Knowl-Based Syst 2022;242:108338. skrcnn-mswelddefect〉.
https://doi.org/10.1016/j.knosys.2022.108338. [223] Zhang Y, You D, Gao X, Zhang N, Gap PP. Welding defects detection based on
[201] Yang L, Song S, Fan J, Huo B, Li E, Liu Y. An automatic deep segmentation deep learning with multiple optical sensors during disk laser welding of thick
network for pixel-level welding defect detection. IEEE Trans Instrum Meas 2022; plates. J Manuf Syst 2019;51:87–94. https://doi.org/10.1016/j.
71:1–10. https://doi.org/10.1109/TIM.2021.3127645. jmsy.2019.02.004.
[202] Nowroth C, Gu T, Grajczak J, Nothdruft S, Twiefel J, Hermsdorf J, et al. Deep [224] Cheng Y., Chen S., Xiao J., Zhang Y. Dynamic estimation of joint penetration by
learning-based weld contour and defect detection from micrographs of laser beam deep learning from weld pool image. Sci Technol Weld Joi, 26(4); 2021. p.
welded semi-finished products. Appl Sci-Basel 2022;12(9):4645. https://doi.org/ 279–85. 〈https://doi.org/10.1080/13621718.2021.1896141〉.
10.3390/app12094645. [225] Shang J, An W, Liu Y, Han B, Guo Y. Oil pipeline weld defect identification system
[203] Zou Y, Li J, Chen X, Lan R. Learning siamese networks for laser vision seam based on convolutional neural network. KSII T Internet Inf, 14(3); 2020. p.
tracking. J Opt Soc Am A Opt Image Sci Vis 2018;35(11):1805–13. https://doi. 1086–103. 〈https://doi.org/10.3837/tiis.2020.03.010〉.
org/10.1364/JOSAA.35.001805. [226] Yang Y, Yang R, Pan L, Ma J, Zhu Y, Diao T, et al. A lightweight deep learning
[204] Ma C, Huang JB, Yang X, Yang MH. Hierarchical convolutional features for visual algorithm for inspection of laser welding defects on safety vent of power battery.
tracking. In: IEEE International Conference on Computer Vision; 2015. p. Comput Ind 2020;123:103306. https://doi.org/10.1016/j.
3074–82. 〈https://doi.org/10.1109/ICCV.2015.352〉. compind.2020.103306.
[205] Nam H, Han B. Learning multi-domain convolutional neural networks for visual [227] Zhang B, Zhu J, Su H. Toward the third generation of artificial intelligence. Sci Sin
tracking. In: IEEE Conference on Computer Vision and Pattern Recognition; 2016. Inf 2020;50(9):1281–302. https://doi.org/10.1360/SSI-2020-0204.
p. 4293–302. 〈https://doi.org/10.1109/CVPR.2016.465〉. [228] Kästner L, Ahmadi S, Jonietz F, Jung P, Caire G, Ziegler M, et al. Classification of
[206] Zou Y, Lan R, Wei X, Chen J. Robust seam tracking via a deep learning framework spot-welded joints in laser thermography data using convolutional neural
combining tracking and detection. Appl Opt 2020;59(14):4321–31. https://doi. networks. IEEE Access 2021;9:48303–12. https://doi.org/10.1109/
org/10.1364/AO.389730. ACCESS.2021.3063672.
[207] Xiao R, Xu Y, Hou Z, Chen C, Chen S. An adaptive feature extraction algorithm for [229] Maldonado-Ramirez A, Rios-Cabrera R, Lopez-Juarez I. A visual path-following
multiple typical seam tracking based on vision sensor in robotic arc welding. Sens learning approach for industrial robots using DRL. Robot Cim-Int Manuf, 71;
Actuat A-Phys 2019;297:111533. https://doi.org/10.1016/j.sna.2019.111533. 2021. p. 102130. 〈https://doi.org/10.1016/j.rcim.2021.102130〉.
[208] Zhang Y, Yang G, Wang Q, Ma L, Wang Y. Weld feature extraction based on fully [230] Guo R, Liu H, Xie G, Zhang Y. Weld defect detection from imbalanced
convolutional networks. Chin J Lasers 2019;46(3):0302002. https://doi.org/ radiographic images based on contrast enhancement conditional generative
10.3788/CJL201946.0302002. adversarial network and transfer learning. IEEE Sens J 2021;21(9):10844–53.
[209] Li J, Jin S, Wang C, Xue J, Wang X. Weld line recognition and path planning with https://doi.org/10.1109/JSEN.2021.3059860.
spherical tank inspection robots. J Field Robot 2022;39(2):131–52. https://doi. [231] Nomura K, Fukushima K, Matsumura T, Asai S. Burn-through prediction and weld
org/10.1002/rob.22042. depth estimation by deep learning model monitoring the molten pool in gas metal
[210] Singh J, Tant K, Mulholland A, MacLeod C. Deep learning based inversion of arc welding with gap fluctuation. J Manuf Process 2021;61:590–600. https://doi.
locally anisotropic weld properties from ultrasonic array data. Appl Sci-Basel org/10.1016/j.jmapro.2020.10.019.
2022;12(2):532. https://doi.org/10.3390/app12020532. [232] Fioravanti CCB, Centeno TM, Delgado MRDBDS. A deep artificial immune system
[211] Mirzapour M, Movafeghi A, Yahaghi E. Quantitative weld defect sizing using to detect weld defects in DWDI radiographic images of petroleum pipes. IEEE
convolutional neural network-aided processing of RT images. Insight 2021;63(3): Access 2019;7:180947–64. https://doi.org/10.1109/ACCESS.2019.2959810.
141–5. https://doi.org/10.1784/insi.2021.63.3.141. [233] Dung CV, Sekiya H, Hirano S, Okatani T, Miki C. A vision-based method for crack
[212] Rohkohl E, Kraken M, Schönemann M, Breuer A, Herrmann C. How to detection in gusset plate welded joints of steel bridges using deep convolutional
characterize a NDT method for weld inspection in battery cell manufacturing neural networks. Autom Constr 2019;102:217–29. https://doi.org/10.1016/j.
using deep learning. Int J Adv Manuf Tech 2022;119. https://doi.org/10.1007/ autcon.2019.02.013.
s00170-021-08553-7. 4829-2843. [234] Sassi P, Tripicchio P, Avizzano CA. A smart monitoring system for automatic
[213] Gao W, Li C, Yang T, Li L, Li X. A method for detection of rare industrial fake film welding defect detection. IEEE T Ind Electron 2019;66(12):9641–50. https://doi.
based on deep learning. Secur Commun Netw 2022; 2022. 3253190. 〈https://doi. org/10.1109/TIE.2019.2896165.
org/10.1155/2022/3253190〉. [235] Nogay HS, Akinci TC. Classification of operation cases in electric arc welding
[214] Deng J, Dong W, Socher R, Li LJ, Li K., Fei-Fei L. ImageNet: a large-scale wachine by using deep convolutional neural networks. Neural Comput Appl 2021;
hierarchical image database. In: IEEE Conference on Computer Vision and Pattern 33(12):6657–70. https://doi.org/10.1007/s00521-020-05436-y.
Recognition; 2009. p. 248–55. 〈https://doi.org/10.1109/CVPR.2009.5206848〉. [236] Wang Q, Jiao W, Zhang YM. Deep learning-empowered digital twin for visualized
[215] Lin TY, Maire M, Belongie S, Hays J, Perona P, Ramanan, et al. Microsoft COCO: weld joint growth monitoring and penetration control. J Manuf Syst 2020;57:
common objects in context. In: European Conference on Computer Vision; 2014. 429–39. https://doi.org/10.1016/j.jmsy.2020.10.002.
p. 740–55. 〈https://doi.org/10.1007/978-3-319-10602-1_48〉. [237] Chu Y, Feng D, Liu Z, Zhao Z, Wang Z, Xia X, et al. Hybrid-learning-based
[216] Gould S, Fulton R, Koller D . Decomposing a scene into geometric and operational visual quality inspection for edge-computing-enabled IoT system.
semantically consistent regions. In: International Conference on Computer Vision; IEEE Internet Things 2022;9(7):4958–72. https://doi.org/10.1109/
2009. p. 1–8. 〈https://doi.org/10.1109/ICCV.2009.5459211〉. JIOT.2021.3107902.
[217] Fan H, Lin L, Yang F, Chu P, Deng G, Yu S, et al. Lasot: a high-quality benchmark
for large-scale single object tracking. In: IEEE Conference on Computer Vision and
625