Sensors 19 03987 v3
Sensors 19 03987 v3
Article
Deep Learning for Industrial Computer Vision
Quality Control in the Printing Industry 4.0
Javier Villalba-Diez 1,2,†,∗ , Daniel Schmidt 3,4,† , Roman Gevers 3 , Joaquín Ordieres-Meré 4 ,
Martin Buchwitz 5 and Wanja Wellbrock 1
1 Hochschule Heilbronn, Fakultät Management und Vertrieb, Campus Schwäbisch Hall,
74523 Schwäbisch Hall, Germany; [email protected]
2 Department of Artificial Intelligence, Escuela Técnica Superior de Ingenieros Informáticos,
Universidad Politécnica de Madrid, 28660 Madrid, Spain
3 Matthews International GmbH, Gutenbergstraße 1-3, 48691 Vreden, Germany;
[email protected] (D.S.); [email protected] (R.G.)
4 Departament of Business Intelligence, Escuela Técnica Superior de Ingenieros Industriales,
Universidad Politécnica de Madrid, 28006 Madrid, Spain; [email protected]
5 InspectOnline, Wiley-VCH Verlag GmbH & Co. KGaA, 69469 Weinheim, Germany; [email protected]
* Correspondence: [email protected]
† These authors contributed equally to this work.
Received: 9 July 2019; Accepted: 13 September 2019; Published: 15 September 2019
Abstract: Rapid and accurate industrial inspection to ensure the highest quality standards at a
competitive price is one of the biggest challenges in the manufacturing industry. This paper shows an
application of how a Deep Learning soft sensor application can be combined with a high-resolution
optical quality control camera to increase the accuracy and reduce the cost of an industrial visual
inspection process in the Printing Industry 4.0. During the process of producing gravure cylinders,
mistakes like holes in the printing cylinder are inevitable. In order to improve the defect detection
performance and reduce quality inspection costs by process automation, this paper proposes a deep
neural network (DNN) soft sensor that compares the scanned surface to the used engraving file and
performs an automatic quality control process by learning features through exposure to training
data. The DNN sensor developed achieved a fully automated classification accuracy rate of 98.4%.
Further research aims to use these results to three ends. Firstly, to predict the amount of errors a
cylinder has, to further support the human operation by showing the error probability to the operator,
and finally to decide autonomously about product quality without human involvement.
Keywords: soft sensors; industrial optical quality inspection; deep learning; artificial vision
1. Introduction
Countries aspiring to lead these technological changes and remain in industrial leadership
positions have strategically positioned themselves for the new type of cyber–physical infrastructure
that will emerge from the Industrial Internet of Things (IIoT) and data science. Germany’s Industry 4.0
framework has evolved into a pan-European collaborative effort to perform intelligent automation
at scale [1]. In a similar move, the United States launched the Manufacturing Leadership Coalition
(SMLC) [2] in 2011. Other notable examples include “China Manufacturing 2025” [3] that seeks to
elevate advanced manufacturing technology, or Japanese’s “Society 5.0” [4] with a holistic focus on
the safety and well-being of humans through cyber–physical systems. As a paradigmatic example,
the Japanese manufacturer has consistently gained a competitive edge towards its competition by
providing its value stream elements with the ability not to pass defects to the next step in the
manufacturing process [5].
For the purpose of making it easier for interested readers to recognize the fundamental phases of
this OQC evolutional continuous improvement process that paved the road for a fully automatized
computer vision OQC process have been summarized in Table 1 and is depicted in Figure 2.
Table 1. State of the Art.
(a) Manual Inspection of Printed Product (b) Manual Inspection of Monochrome Printed Product
(c) Expert Evaluation and software cLynx (d) Machine scans and software cLynx
Figure 2. OQC evolutional continuous improvement process.
(a) Detail Cylinder Scan (b) Detail Engraving Layout (c) Absolute Difference
Figure 4. Example 1 of automatic selection of areas around possible errors.
Figure 5. Deep Learning Architecture for Industrial Computer Vision OQC in the Printing Industry 4.0.
Sensors 2019, 19, 3987 8 of 23
3.1. Deep Neural Network Architecture for Computer Vision in Industrial Quality Control in the Printing
Industry 4.0
2. Brightness Adjustment
To get comparable data for all cylinder images, pre-processing is needed and is performed on the
complete scan of a cylinder. From this scan multiple examples are taken. Because there can be
slight deviations due to many influences during the recording of the cylinder surface, this can only
be achieved by having a similar brightness for the cylinder surface and engraved parts. Another
important point is that no essential information gets lost from the images and, that the brightness
between the engraved and not engraved parts are comparable for all cylinder scans. Therefore a
brightness stretch is needed but only few pixels are allowed to become the darkest or brightest
pixels. Notwithstanding, the amount of pixel that become the darkest and brightest pixels ca
not be set to a very low value because noise in the image data would result in big differences.
Sensors 2019, 19, 3987 10 of 23
In conclusion a low percentage of the pixels should be set as darkest and brightest. For example,
the lowest and the highest percentage should each have a maximum of 0.5%. Figure 7 shows a
stretching example for brightness adjustment for one image so that 0.5% of all pixels will have a
value of 0 and 0.5% of all pixels will have the value of 255.
100 × 100 parts. If any one of these is detected as an error, all are marked as an error. As shown
in the Open Access Repository, there are multiple possible ways to handle the bigger data. Every
example also has the actual and target data. There are different ways of using this data as input.
One way is just using the actual data. A different option is to use the difference between the
actual and expected data. The problem in both cases is that information gets lost. Better results
have been achieved by using the differences. These get adjusted, so that the input data is in a
range from [−1,1]. Once this is performed, and because a balanced dataset is important to train
the neural network and the OK-cylinder examples far outnumber the not-OK-cylinder examples,
an OK-cylinder example is only saved if a not-OK-cylinder example has been found previously.
• Classification The first goal of this architecture is not to identify different objects inside of part
of the images but to separate two classes (not-OK and OK images), where the main source of
noise came from the illumination factor from the scanner lectures. Therefore, neither the so deep
architectures nor the identity transference, which was the key for the ResNet [41] is needed in our
case, and just few convolutions shall help identify convenient structural features to rely on.
• Performance. The proposed architecture is even more simplistic than the AlexNet [42] one, as we
do not use five convolution layers but just three. The main reason is to look for a compromise
between the number of parameters and the available dataset of images. Our architecture was
always looking to be frugal in terms of resources, as it is expected to be a soft sensor, running in
real time and having the inherent capability of retrain for reinforced learning, close to such real
time constraint.
After data acquisition and pre-processing, the input data of the DNN are figures represented as
tensors. A type of network that performs well on the classification problem of such data is usually
divided in two main parts: feature extractors and classifiers as shown in Figure 5:
• Feature Extraction. The feature extraction is performed by a deep stack of alternatively fully
connected convolutional and sub-sampling max pooling layers, the even numbered layers are for
convolutions and the odd numbered layers are for max-pooling operations.
– Convolution and ReLu (rectified linear unit) activated convolutional layers. Convolution operations,
by means of activation functions, extract the features from the input information which are
propagated to deeper level layers. A ReLu activation function is a function meant to zero out
negative values. The ReLu activation function was first presented in AlexNet [42] and solves
the vanishing gradient problem for training DNN.
– Max pooling. Consists of extracting windows from the input feature maps and outputting the
max value of each channel. It’s conceptually similar to convolution, except that instead of
transforming local patches via a learned linear transformation (the convolution kernel), they
are transformed via a max tensor operation.
• Classification. The classification is performed by fully connected activation layers [43].
Some examples of such models are LeNet [44], AlexNet [42], Network in Network [45],
GoogLeNet [46–48], DenseNet [49].
– Fully connected activation layers output a probability distribution over the output classes [25].
Because we are facing a binary classification problem and the output of our network is
a probability, it is best to use the binary-crossentropy loss function. Crossentropy is a
Sensors 2019, 19, 3987 12 of 23
quantity from the field of Information Theory that measures the distance between probability
distributions or, in this case, between the ground-truth distribution and the predictions.
It is not the only viable choice: we could use, for instance, mean-squared-error. However,
crossentropy is usually the best choice when dealing with models that output probabilities.
Because we are attacking a binary-classification problem, we end the network with a single
unit (a Dense layer of size 1) and a sigmoid activation. This unit will encode the probability
that the network is looking at one class or the other [25].
As shown in the Open Access Repository, using Keras, Tensorflow backend for the DNN and
OpenCV/Numpy for the image manipulation, a balanced dataset of 13,335 not-OK- and 13335
OK-cylinder examples is used, giving a total of 26,670. These were collected over a period of 14 months
from almost 4000 cylinder scans. The training part is mirrored vertically and horizontally resulting
in 85,344 training samples in total. All not-OK- cylinder examples are labeled 0 and all Ok examples
are labeled 1. As a standard procedure, the data is split into training dataset (80%), testing dataset
(10%) and validation dataset (10%). The training dataset is used to train the DNN throughout an
number of epochs as shown in Figure 9. It can be observed that both accuracy and loss do not increase
or decrease significantly after epoch number 10.
(a) DNN Model Training Accuracy (b) DNN Model Training Loss
Figure 9. DNN Training and Testing Results.
The testing dataset is subsequently used to test DNN performance. The confusion matrix is a
standard procedure to summarize the results of such a training by typically combining contingency
classes (TRUE, FALSE) and (OK, not-OK), hence building four categories: (1) True Negative (TN),
which is an error and has been predicted as an error; (2) False Positive (FP), which is an error but has
not been predicted as an error, and is by far the most damaging category; False Negative (FN) which
is not an error but has been predicted as an error; and (4), True Positive (TP) which is not an error
and has not been predicted as an error. Specifically, given the balanced dataset chosen, the accuracy
(ACC) delivered by the DNN soft sensor, defined by the expression ACC = ( TP + TN )/( TP + TN +
FP + FN ), is 98.4%. The TN rate is 97.85%, the TP rate is 99.01%, the FN rate is 2.15% and the FP
rate is 0.99%. These levels of ACC can be considered acceptable for such a complicated industrial
classification problem. The results are summarized in Figure 10.
Sensors 2019, 19, 3987 14 of 23
In Table 2 the DNN architecture shown in Figure 5 is described layer by layer by outlining
the rationale behind the choice of a layer rather than another. Going even further, to compare the
performance of the proposed soft DNN sensor, it has been compared with three similar architectures.
The result of this comparison is shown in Open Access Repository and summarized in Figure 11
in which it is clearly shown that the proposed DNN soft sensor has superior performance to other
alternative architectures.
Figure 11. Deep Learning Architecture Comparison. Time to Train vs. Accuracy.
Two parameters, accuracy and computational time, have been measured consistently with the
same training and test set, and then compared. First, it has been tested with an identical architecture
by adding a dropout, then it has been tested with a deeper architecture and finally with a more shallow
Sensors 2019, 19, 3987 15 of 23
DNN with fewer layers. The accuracy should be as high as possible in order to generate the lowest
possible error in data characterization, and the computation time should be as low as possible in order
to ensure that the soft DNN sensor can be effectively integrated into an Industry 4.0 environment,
thus ensuring maximum effectiveness and efficiency respectively. A smooth DNN sensor must be
not only accurate but also fast to ensure, among other things, a minimum Lead Time impact on the
value creation process and low CO2 emissions derived from the energy consumption associated with
the computation.
Table 2. DNN Architecture Detailed Description.
Layer Size Layer Name Layer Description and Rationale behind the Choice
As with the previous max pooling layer this layer is used to reduce
(23, 23, 64) max pooling2d 2
the complexity of the convoluted result.
As with the previous max pooling layer this layer is used to reduce
(10, 10, 64) max pooling2d 3
the complexity of the convoluted result.
As with the previous max pooling layer this layer is used to reduce
(4, 4, 32) max pooling2d 4
the complexity of the convoluted result.
(c) 3rd Layer Max Pooling-1 (d) 4th Layer Conv 2D-2
(a) Example Image of Error in not-OK cylinder (b) Activation Heatmap of Error in not-OK cylinder
Another option is to calculate an input image that gets the highest response from a layer. This
is done by displaying the visual pattern that each filter is meant to respond to. This can be done
with gradient ascent in input space: applying gradient descent to the value of the input image of
a convolutional network so as to maximize the response of a specific filter, starting from a blank
input image. The resulting input image will be one that the chosen filter is maximally responsive to.
An example is shown in Figure 14.
Finally, an alternative approach would be to show the outputs of all DNN layers as color-coded
images. Visualizing intermediate activations consists of displaying the feature maps that are output
by various convolution and pooling layers in a network, given a certain input (the output of a layer
is often called its activation, the output of the activation function). This gives a view into how an
input is decomposed into the different filters learned by the network. We want to visualize feature
Sensors 2019, 19, 3987 18 of 23
maps with three dimensions: width, height, and depth (channels). Each channel encodes relatively
independent features, so the proper way to visualize these feature maps is by independently plotting
the contents of every channel as a 2D image. For explanatory purposes, on the Open Access Repository,
four different examples, TP-TN-FP-FN, of such feature maps are depicted. These shall help the reader
better understand what the DNN sees and how it responds in different circumstances. One of these
examples, TN, is visualized in Figure 12.
1. Using the DNN fully automate OQC classification to predict the amount of errors a cylinder has.
The DNN only provides a successful result 98.4% of the time. To be sure that the wrongly classified
images are not big mistakes, human experts will review all possible errors. DNN has already
had a positive influence on the workflow, as we know how many errors are very likely an error:
DNN helps significantly in the planning of the next workflow step because it is known with a
high probability if the cylinder needs to go to the correction department or if it is very likely that
the product is an OK-cylinder.
2. Showing the error probability to the operator that is currently deciding if it is an error or if it
is not.
This gives a hint to the operator, who can give feedback if there are relevant mistakes that were
not predicted as mistakes. This can also help the operator to reduce the likelihood of missing an
error. Once this soft sensor was integrated in production, OQC productivity, measured in hours
per unit - time an operator spends in the OQC -, dramatically increased by 210% as decision about
defects is made in an automatic way.
3. Only showing possible errors that have been predicted by the DNN.
In the last step, the DNN could completely filter out errors that are not relevant. This can also
be used in multiple steps because it is possible to increase the threshold error probability for
the possible error to be shown. At some point a threshold will have to be chosen taking into
consideration the cost of checking a possible error and the cost of missing a error. This would
completely eliminate the step of checking the errors and the confirmed errors would only be
checked by the correction department.
5. Conclusions and Future Steps of Deep Learning in a Printing Industry 4.0 Context.
Although there has been an immediate performance increase in OQC error detection accuracy
and cost effectiveness, larger scope for improvement is down to the managerial dimension of such
a sensor. This is because it can be expanded to not only detect defects but also to classify them in
categories. Although this requires additional effort, it will enable the cause-effect analysis regarding
manufacturing conditions and defect frequencies.
Some of these efforts can be specifically targeted to achieve an improvement in the accuracy of
the model. For example learning from the false predictions: to further improve the correct prediction
Sensors 2019, 19, 3987 19 of 23
rate it is important to take a look at the examples that have not been predicted correctly. This could
potentially improve the understanding why the wrong prediction was made by the DNN:
• Not-OK examples that have been predicted as OK. Looking at the actual errors in the test data that
have not been predicted as errors, as in Figure 15a, a few issues could be the cause of the wrong
predictions. Some of the examples actually do not look like they are really not-OK. The cause of
this could either be, that the input data was not labeled correctly or that the error really is not
highly visible in the image.
• OK examples that have been predicted as not-OK. After looking at the visualization of the DNN,
it gets clear that the main focus for finding mistakes is looking for extreme edges. These can
be seen in a lot of the wrongly classified examples. Especially the first two examples seen in
Figure 15b have some extreme edges that are a result of a slight misalignment of the images in the
pre-processing. Therefore the image registration in the pre-processing part between the original
and the recording of the cylinder surface needs to be improved.
This technology could also be implemented at the customer side to increase defect detection
accuracy on the printed product itself. This strategic step is currently being discussed internally. Such
Sensors 2019, 19, 3987 20 of 23
analyses will provide sensitivity about operations and operational conditions, which will also impact
in value stream-related efficiency and effectiveness.
These aspects will probably be the next steps in further research actions to be developed within
an Industry 4.0 context. For instance, deep learning applied to manufacturing Industry 4.0 technology
will have an impact at various levels of aggregation in the printing manufacturing value chains:
1. Deep Learning at a shopfloor level shall impact quality, reliability and cost.
At the shopfloor level, this paper has shown an example of how deep learning increases the
effectiveness and efficiency of process control aimed at achieving better quality (e.g., with OQC)
and lower costs, allowing self-correction of processes by means of shorter and more accurate
quality feedback loops. This intelligence integrated in the value streams will allow many humans
and machines to co-exist in a way in which artificial intelligence will complement in many aspects.
In the future, significant challenges will still be encountered in the generation and collection of
data from the shopfloor.
The main challenge towards a fully automated solution is currently getting the Python DNN
integrated into the C++ cLynx program. After this is successfully completed, a testing phase with
the cLynx users is planned. If the results are satisfactory, the complete automatic process will
be started. If the results are not satisfying, further steps have to be taken so as to improve the
DNN further.
2. Deep Learning at a supply chain level shall impact lead time and on-time delivery.
At a higher level of supply chain, producing only what the customer needs, when it needs
it, in the required quality, the integration of deep learning technology will allow not only the
systematic improvement of complex value chains, but a better use and exploitation of resources,
thus reducing the environmental impact of industrial processes 4.0.
3. Deep Learning at a strategic level shall impact sustainable growth.
At a more strategic level, customers and suppliers will be able to reach new levels of transparency
and traceability on the quality and efficiency of the processes, which will generate new business
opportunities for both, generating new products and services and cooperation opportunities in a
cyber–physical environment. In a world of limited resources, increasing business volume can only
be achieved by increasing the depth of integrated intelligence capable of successfully handling
the emerging complexity in value streams.
To summarize, despite the "black box problem" and the challenge to have enough information
and labeled data available for learning, Deep Learning will probably conquer in the field of machine
vision, one country after another, and will act in the background without the user being aware of it.
The role that Deep Learning will play in the creation of cyber–physical systems will be adopted from a
strategic point of view, in which business leaders will tend to think of deep architectures as possible
solutions to problems.
Author Contributions: Conceptualization, J.V.-D., D.S., J.O.-M. and R.G.; methodology, J.V.-D. and D.S.; software,
J.V.-D. and D.S.; validation, J.V.-D., D.S., R.G. and J.O.-M.; formal analysis, J.V.-D. and D.S.; investigation, J.V.-D.;
resources, D.S.; data curation, D.S.; writing–original draft preparation, J.V.-D. and D.S.; writing–review and editing,
J.V.-D., D.S., M.B. and W.W.; visualization, J.V.-D. and D.S.; supervision, J.V.-D. and D.S.; project administration,
J.V.-D. and D.S.; funding acquisition, J.V.-D., J.O.-M. and R.G.
Funding: Authors would like to to recognise the support obtained from the EU RFCS program through project
number 793505’Ă’Ÿ4.0 Lean system 440 integrating workers and processes (WISEST)’Ă’Ź.
Acknowledgments: This research was partially supported by Matthews International GmbH, Gutenbergstrasse
1-3, 48691 Vreden. We specially thank Stephan Lammers and Tomas Sterkenburgh for comments on an earlier
version of the manuscript. We also thank Oliver Lenzen and Daniela Ludin from Hochschule Heilbronn,
and Martin Molina from Universidad Politécnica de Madrid for their institutional support.
Sensors 2019, 19, 3987 21 of 23
Abbreviations
The following abbreviations are used in this manuscript:
References
1. Ustundag, A.; Cevikcan, E. Industry 4.0: Managing The Digital Transformation; Springer Series in Advanced
Manufacturing, Springer: Cham, Switzerland, 2018.
2. Davis, J.; Edgar, T.; Porter, J.; Bernaden, J.; Sarli, M. Smart manufacturing, manufacturing intelligence and
demand-dynamic performance. Comput. Chem. Eng. 2012, 47, 145–156. [CrossRef]
3. Li, L. China’s manufacturing locus in 2025: With a comparison of ’Made-in-China 2025’ and ’Industry 4.0’.
Technol. Forecasting Social Change 2018, 135, 66–74. [CrossRef]
4. Shiroishi, Y.; Uchiyama, K.; Suzuki, N. Society 5.0: For Human Security and Well-Being. Computer 2018,
51, 91–95, [CrossRef]
5. Womack, J.; Roos, D. The Machine That Changed the World; Harper Perennial: New York, NY, USA, 1990.
6. Takeda, H. Intelligent Automation Textbook; Nikkan Kogyo Shimbun: Tokyo, Japan, 2009.
7. Nakabo, Y. Considering the competition and cooperation areas surrounding Industry 4.0. What will IoT
automate. J-Stage Top. Meas. Contr. 2015, 54, 912–917, [CrossRef]
8. Kuwahara, S. About factory automation and IoT, AI utilization by intelligent robot. J-Stage Top. Syst.
Contr. Inf. 2017, 61, 101–106, [CrossRef]
9. Villalba-Diez, J.; Ordieres-Mere, J. Improving manufacturing operational performance by standardizing
process management. IEEE Trans. Eng. Manage. 2015, 62, 351–360. [CrossRef]
10. Villalba-Diez, J.; Ordieres-Mere, J.; Chudzick, H.; Lopez-Rojo, P. NEMAWASHI: Attaining Value Stream
alignment within Complex Organizational Networks. Procedia CIRP 2015, 7, 134–139, [CrossRef]
11. Jimenez, P.; Villalba-Diez, J.; Ordieres.Mere, J. HOSHIN KANRI Visualization with Neo4j. Empowering
Leaders to Operationalize Lean Structural Networks. PROCEDIA CIRP 2016, 55, 284–289. [CrossRef]
12. Villalba-Diez, J. The HOSHIN KANRI FOREST. Lean Strategic Organizational Design, 1st ed.; CRC Press. Taylor
and Francis Group LLC: Boca Raton, FL, USA, 2017.
13. Villalba-Diez, J. The Lean Brain Theory. Complex Networked Lean Strategic Organizational Design; CRC Press,
Taylor and Francis Group LLC: Boca Raton, FL, USA, 2017.
14. Womack, J.; Jones, D. Introduction. In Lean Thinking, 2 ed.; Simon & Schuster: New York, NY, USA, 2003; p. 4.
15. Arai, T.; Osumi, H.; Ohashi, K.; Makino, H. Production Automation Committee Report: 50 years of
automation technology. J-Stage Top. Precis. Eng. J. 2018, 84, 817–820, [CrossRef]
16. Manikandan, V.S.; Sidhureddy, B.; Thiruppathi, A.R.; Chen, A. Sensitive Electrochemical Detection of Caffeic
Acid in Wine Based on Fluorine-Doped Graphene Oxide. Sensors 2019, 19, [CrossRef]
17. Garcia Plaza, E.; Nunez Lopez, P.J.; Beamud Gonzalez, E.M. Multi-Sensor Data Fusion for Real-Time Surface
Quality Control in Automated Machining Systems. Sensors 2018, 18, [CrossRef] [PubMed]
18. Han, L.; Cheng, X.; Li, Z.; Zhong, K.; Shi, Y.; Jiang, H. A Robot-Driven 3D Shape Measurement System for
Automatic Quality Inspection of Thermal Objects on a Forging Production Line. Sensors 2018, 18, [CrossRef]
[PubMed]
19. Weimer, D.; Scholz-Reiter, B.; Shpitalni, M. Design of deep convolutional neural network architectures for
automated feature extraction in industrial inspection. CIRP Annals 2016, 65, 417–420, [CrossRef]
20. Xie, X. A Review of Recent Advances in Surface Defect Detection Using Texture Analysis Techniques.
Electron. Lett. Comput. Vision Image Ana. 2008, 7, 1–22. [CrossRef]
21. Scholz-Reiter, B.; Weimer, D.; Thamer, H. Automated Surface Inspection of Cold-Formed MicroPart.
CIRP Ann. Manuf. Technol. 2012, 61, 531–534. [CrossRef]
Sensors 2019, 19, 3987 22 of 23
22. Rani, S.; Baral, A.; Bijender, K.; Saini, M. Quality control during laser cut rotogravure cylinder manufacturing
processes. Int. J. Sci. Eng. Comput. Technol. 2015, 5, 70–73.
23. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [CrossRef] [PubMed]
24. Jia, Y.; Shelhamer, E.; Donahue, J.; Karayev, S.; Long, J.; Girshick, R.; Guadarrama, S.; Darrell, T. Caffe:
Convolutional Architecture for Fast Feature Embedding. ACM Multimedia 2014, 675–678.
25. Chollet, F. Deep Learning with Python; Manning Publications Co.: Shelter Island, NY, USA, 2018.
26. Lin, T.; RoyChowdhury, A.; Maji, S. Bilinear CNN Models for Fine-Grained Visual Recognition. In
Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13
December 2015; pp. 1449–1457, [CrossRef]
27. Miskuf, M.; Zolotova, I. Comparison between multi-class classifiers and deep learning with focus on industry
4.0. In Proceedings of the 2016 Cybernetics & Informatics (K&I), Levoca, Slovakia, 2–5 February 2016.
28. Zheng, X.; Wang, M.; Ordieres-Mere, J. Comparison of Data Preprocessing Approaches for Applying Deep
Learning to Human Activity Recognition in the Context of Industry 4.0. Sensors 2018, 2146, 1–13, [CrossRef]
29. Aviles-Cruz, C.; Ferreyra-Ramirez, A.; Zuniga-Lopez, A.; Villegas-Cortez, J. Coarse-Fine Convolutional
Deep-Learning Strategy for Human Activity Recognition. Sensors 2019, 19, [CrossRef]
30. Zhe, L.; Wang, K.S. Intelligent predictive maintenance for fault diagnosis and prognosis in machine centers:
Industry 4.0 scenario. Adv. Manuf. 2017, 5, 377–387.
31. Deutsch, J.; He, D. Using Deep Learning-Based Approach to Predict Remaining Useful Life of Rotating
Components. IEEE Trans. Syst. Man Cybern. Syst. 2018, 48, 11–20. [CrossRef]
32. Shanmugamani, R., Ed. Deep Learning for Computer Vision; Packt Publishing-ebooks Account: Birmingham,
UK, 2018.
33. Wang, T.; Chen, Y.; Qiao, M.; Snoussi, H. A fast and robust convolutional neural network-based defect
detection model in product quality control. Int. J. Adv. Manuf. Technol. 2018, 94, 3465–3471. [CrossRef]
34. He, M.; He, D. Deep Learning Based Approach for Bearing Fault Diagnosis. IEEE Trans. Ind. App. 2017,
53, 3057–3065. [CrossRef]
35. Imai, M. KAIZEN: The Key to Japan’s Competitive Success; McGraw-Hill Higher Education: New York, NY,
USA, 1986.
36. Schmidt, D. Available online: https://patentscope.wipo.int/search/de/detail.jsf;jsessionid=
F4DFD8F2D86BB91896D53B4AB97E84A1.wapp1nC?docId=WO2018166551&recNum=871&office=
&queryString=&prevFilter=&sortOption=Ver%C3%B6ffentlichungsdatum+ab&maxRec=70951352 (accessed
on 15 September 2019).
37. Hinton, G.; Osindero, S.; Teh, Y. A fast learning algorithm for deep belief nets. Neural Comput. 2006,
18, 1527–1554. [CrossRef]
38. Zhang, K.; Zuo, W.; Gu, S.; Zhang, L. Learning Deep CNN Denoiser Prior for Image Restoration. In
Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu,
HI, USA, 21–26 July 2017; pp. 2808–2817, [CrossRef]
39. Zhang, K.; Zuo, W.; Chen, Y.; Meng, D.; Zhang, L. Beyond a Gaussian Denoiser: Residual Learning of Deep
CNN for Image Denoising. IEEE Trans. Image Process. 2017, 26, 3142–3155, [CrossRef] [PubMed]
40. van Rossum, G. Python Tutorial; Technical Report CS-R9526; Computer Science/Department of Algorithmics
and Architecture: Amsterdam, The Netherlands, 1995.
41. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. arXiv 2015,
arXiv:1512.03385.
42. Krizhevsky, A.; Sutskever, L.; Hinton, G. Imagenet classification with deep convolutional neural networks.
In Proceedings of the 25th International Conference on Neural Information Processing Systems, Lake Tahoe,
Nevada, 3–6 December 2012; pp. 1106–1114.
43. Alom, M.Z.; Taha, T.M.; Yakopcic, C.; Westberg, S.; Sidike, P.; Nasrin, M.S.; Hasan, M.; Van Essen, B.C.;
Awwal, A.A.S.; Asari, V.K. A State-of-the-Art Survey on Deep Learning Theory and Architectures. Electronics
2019, 8, [CrossRef]
44. Lecun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition.
Proc. IEEE 1998, 86, 2278–2324, [CrossRef]
45. Lin, M.; Chen, Q.; Yan, S. Network In Network. arXiv 2013, arXiv:1312.4400.
Sensors 2019, 19, 3987 23 of 23
46. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; D. Anguelov.; D. Erhan.; V. Vanhoucke.; A. Rabinovich.
Going deeper with convolutions. In Proceedings of the 2015 IEEE Conference on Computer Vision and
Pattern Recognition (CVPR), Bosto, MA, USA, 7–12 June 2015; pp. 1–9, [CrossRef]
47. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the Inception Architecture for Computer
Vision. arXiv 2015, arXiv:1512.00567.
48. Szegedy, C.; Ioffe, S.; Vanhoucke, C.; Alemi, A. Inception-v4, Inception-ResNet and the Impact of Residual
Connections on Learning. arXiv 2016, arXiv:1602.07261.
49. Huang, G.; Liu, Z.; Maaten, L.v.d.; Weinberger, K.Q. Densely Connected Convolutional Networks.
In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu,
HI, USA, 21–26 July 2017; pp. 2261–2269.
c 2019 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access
article distributed under the terms and conditions of the Creative Commons Attribution
(CC BY) license (http://creativecommons.org/licenses/by/4.0/).