A Survey On Deep Learning Methods and Tools in Image Processing
A Survey On Deep Learning Methods and Tools in Image Processing
Abstract: Deep Learning brings provocative light on various fields and research areas, such as image processing, Artificial Intelligence, and natural
language processing. it is an essential method applied to different applications in future. In recent years, the image processing and has been attracting
and increasing attention as it is one of the emerging areas of research. There are various applications for image processing in various domains such as
agriculture, space agencies, medical field, forensics and many others. As there is a progress in deep learning, many modern methods in deep learning
are recommended to improve image processing and image analysis performance. This document provides a broad overview on deep learning methods
used in image processing. Initially, an introduction on deep learning and image processing is given. Then, architectures of deep learning are introduced.
Then, the applications of deep learning on image processing like the image detection, image segmentation, and classification image classification are
explained. Thereafter, the benefits and weaknesses of deep learning tools common tools are mentioned along with the deep learning tools used in image
processing applications.
Index Terms: Artificial Intelligence, Convolution Neural Networks, Deep learning, Image Processing, Image segmentation, Machine Learning, Tools.
—————————— ——————————
image processing. Later the pros and cons of various tools y= f (x; ɵ). This mapping function y acquires the value of ɵ
are deliberated along with the tools that are applied to image settings to compute the best equivalent functions. Usually, a
processing. Finally the objective of deep learning in image network of direct action in extent consists of an input layer,
processing is expressed along with forth coming tools different hidden layers and an output layer. In addition to that
the information flow in deep feed forward network occurs in
2. ARCHITECTURES OF DEEP LEARNING one direction and does not move backward . This is depicted
in the sample figure . In this example the deep feed forward
2.1 Artificial neural networks (ANN) : network has few input layers, hidden layers and one output
ANN is important because just like human brain network it layer. Hence the deep feed forward network is one the ancient
grasps the information which is processed to establish various architecture of deep learning
models pertaining to neural network. The concepts of artificial
neural network evolved during 1980’s. These models of 2.3 Auto encoders Stacked
artificial neural network are calculation models having huge A simple deep feed forward network which is simple and
finite nodes and links among them. every node in an having input layers , hidden layers and an output layer is
distinguished ANN present an image of particular output referred as an auto encoder[6],[7],[8],[9] . A decoder and an
function . The link between the nodes communicate encoder are two parts of an auto encoder according to its
information between them . This information is represented as functionalities .An encoder is notated as f(x) that produces a
a weight . Hence ANN is an algorithm whose output rely upon decreased characteristic description from x,h where x is a
weights and activation function (output function) . In primary input h is a hidden layer. the notation of decoder is
accordance with, ANN algorithms are mainly classified into g(f(x)) . the decoder is familiar with reorganizing the basic
feedback neural networks and feed forward neural networks. input from the encoder output. the decoder minimizes the loss
Few of them are described below : Feed Forward neural function during reorganizing process:
networks: It is graph which is directed and having no cycles
and no feedback in the network. It is applied for the L(x,g(f(x))
conversion the input space information to output space. As it
is combined with various simple nonlinear functions; it has got The high-dimensional data is converted to low-dimension by
a good capability of processing information. its network encoding and decoding technique. Hence auto encoder is
topology is simple and easy to apply. much beneficial in image classification and various identical
Feedback neural networks: It is a complete graph having a tasks. The auto encoder is derived from sparse auto encoder,
feedback and no directions .it has state transitions while denoising auto encoder and contractive auto encoder. sparse
processing information. In order to handle the information auto encoder is dissimilar to auto encoders. the sparse
processing it uses dynamic system theory. the network constraint Ω(h) is added to the hidden layer. Hence its
cohesion is firmly connected to the function of associative restoration error is computed as
memory. It’s a decade that the research on artificial neural
network is getting deepened and a lot of advancement has L(x,g(f(x)))+ Ω(h)
been made in various domains in biology, smart robots,
finance and medical. In addition to these, artificial neural De noising auto encoder : the objective of denoising auto
networks have been strongly used to solve countless encoder is to minimize the loss function , they are dissimilar to
problems by showing many intellectual characteristics. sparse auto encoder
Commonly the attributes of artificial neural networks are L(x,g(f(x)))
primarily mirrored on associative memory, self-learning
capability and quick fixing the solution using optimization. Contractive auto encoder: Similar to sparse auto encoders,
they decrease the specific regularizer by accumulating Ω(h) to
hidden layer .The explicit regularizer is computed as as :
L(x,g(f(x)))+ Ω(h)
1058
IJSTR©2020
www.ijstr.org
INTERNATIONAL JOURNAL OF SCIENTIFIC & TECHNOLOGY RESEARCH VOLUME 9, ISSUE 02, FEBRUARY 2020 ISSN 2277-8616
1059
IJSTR©2020
www.ijstr.org
INTERNATIONAL JOURNAL OF SCIENTIFIC & TECHNOLOGY RESEARCH VOLUME 9, ISSUE 02, FEBRUARY 2020 ISSN 2277-8616
network. Here the final tier of CNN follows Deconvolutional occur simultaneously accompanying every class that has
neural network . Given a picture, image segmentation is very images in number 6000.the images in this dataset has no
essential to figure out varied things and relationship. The noise , small in size and are definitely labeled . Hence this is
utilization of image segmentation differs from simple an ideal dataset to finish the task significantly with a fewer pre
applications in image recognition to spatial images as they are processing steps. the following are some of the pictures
exemplary generous to know the texture . they are applied to collected from the above dataset:
medical images to identify cancer cells and even in surgery
preparations . This is an essential constituent in searching the
images in Fashion and retail enterprises
Here the computer is taught with the behavior of bird, dog, cat
etc, in order to find out the objects. The more cats the
computer gets trained, it recognizes the more cat objects. This
kind of training is said to be supervised learning. Here the
images are labeled to accomplish such task. Google has
given an open source framework for deep learning called as
Tensor Flow. The developer would have a better control at
micro level over each node (neuron) in Tensor Flow. To gain
the best performance the weights have to be set. For image Figure 5 : Image after applying above three steps
classification Tensor Flow has built in APIs. Convolution
neural networks is one of the well know techniques to improve Step 1: Pre-treatment: In first step of preprocessing, the data
image classification accuracy. In this technique the image is does not have any noise and it is organized so variance is
broken up into several parts where the machine tries to added to the data. The noise can be added using python
identify each part rather than sending the image array as an library named imgaug. Then a random combination of
input . this technique of sending the parts to machine allows to cropping the image, flipping and applying saturation and hue
accomplish the task parallel and identify the objects adjustments is done.
irrespective of the object location. Step 2: Splitting our data set: Computing gradient for the given
large dataset model is a time taking process . Hence the
3.4 . Sample data images are processed in tiny batches at every optimizer
In this example a dataset CIFAR-10[20] having images iteration. Generally the size of the batch is considered as 32
calibrated 60 000 in number each of size 32 X 32 pixels is or 64. In this example 64 is the batch size as it has big
taken. this dataset is categorized into ten classes that do not number of image. Further the dataset is broken up into
training set each having 50000 image and test set having 10
1060
IJSTR©2020
www.ijstr.org
INTERNATIONAL JOURNAL OF SCIENTIFIC & TECHNOLOGY RESEARCH VOLUME 9, ISSUE 02, FEBRUARY 2020 ISSN 2277-8616
1062
IJSTR©2020
www.ijstr.org