Techno-Societal 2020
Techno-Societal 2020
Pawar ·
R. Balasubramaniam ·
Babruvahan P. Ronge ·
Santosh B. Salunkhe · Anup S. Vibhute ·
Bhuwaneshwari Melinamath Editors
Techno-Societal
2020
Proceedings of the 3rd International
Conference on Advanced Technologies
for Societal Applications—Volume 1
Techno-Societal 2020
Prashant M. Pawar · R. Balasubramaniam ·
Babruvahan P. Ronge · Santosh B. Salunkhe ·
Anup S. Vibhute · Bhuwaneshwari Melinamath
Editors
Techno-Societal 2020
Proceedings of the 3rd International
Conference on Advanced Technologies for
Societal Applications—Volume 1
Editors
Prashant M. Pawar R. Balasubramaniam
SVERI’s College of Engineering Precision Engineering Center
Pandharpur, Maharashtra, India Bhabha Atomic Research Centre
Mumbai, Maharashtra, India
Babruvahan P. Ronge
SVERI’s College of Engineering Santosh B. Salunkhe
Pandharpur, Maharashtra, India SVERI’s College of Engineering
Pandharpur, Maharashtra, India
Anup S. Vibhute
SVERI’s College of Engineering Bhuwaneshwari Melinamath
Pandharpur, Maharashtra, India SVERI’s College of Engineering
Pandharpur, Maharashtra, India
© The Editor(s) (if applicable) and The Author(s), under exclusive license to Springer Nature
Switzerland AG 2021
This work is subject to copyright. All rights are solely and exclusively licensed by the Publisher, whether
the whole or part of the material is concerned, specifically the rights of translation, reprinting, reuse
of illustrations, recitation, broadcasting, reproduction on microfilms or in any other physical way, and
transmission or information storage and retrieval, electronic adaptation, computer software, or by similar
or dissimilar methodology now known or hereafter developed.
The use of general descriptive names, registered names, trademarks, service marks, etc. in this publication
does not imply, even in the absence of a specific statement, that such names are exempt from the relevant
protective laws and regulations and therefore free for general use.
The publisher, the authors and the editors are safe to assume that the advice and information in this book
are believed to be true and accurate at the date of publication. Neither the publisher nor the authors or
the editors give a warranty, expressed or implied, with respect to the material contained herein or for any
errors or omissions that may have been made. The publisher remains neutral with regard to jurisdictional
claims in published maps and institutional affiliations.
This Springer imprint is published by the registered company Springer Nature Switzerland AG
The registered company address is: Gewerbestrasse 11, 6330 Cham, Switzerland
Contents
v
vi Contents
1 Introduction
V. Patil (B)
Computer Engineering Department, K. J. Somaiya Institute of Engineering and Information
Technology, Mumbai, India
Y. Jadhav
Computer Science and Engineering Department, Amity School of Engineering and Technology,
Amity University, Mumbai, India
2 Literature Survey
Wei Wu, Chuanchang Liu, Zhiyuan Su have presented the result of the analysis
and the methodology of applying FaceNet. Recent studies show that deep learning
approaches to this challenge can produce impressive results. To overcome these chal-
lenges, this program uses a profound learning approach. This comes with a module
to detect the faces and identification of the faces. The MTCNN facial detection
module is very simple, accurate and robust enough for variations in lighting and
contexts. This module is based on FaceNet, which teaches you to map the facial
images directly at the position of the Euclidean space, where the two-pointed space
inside the Euclidean space corresponds directly to the similarity of two facial images.
When this Euclidean space is generated, facial images are transferred as facial vectors
to the FaceNet embedding system. Feature vectors are then listed in a SVM to help us
easily recognize the face images. The experimental findings indicate that the device
is highly accurate, low computational complexity, and has a strong implementation
benefit for real time identification [1].
Many businesses have widely adopted video monitoring devices to increase
protection in their workforce. There are few visible cameras because of their suscep-
tibility to eye strain A kind of control method is used in this paper to solve the
problem. The network is made up of wireless and video monitoring devices. When
security threats are detected, an alert may be activated by the wireless control device
and the video stream from the camera to the computer is enabled. Therefore, the
system avoids dangers resulting from machine breakdown or user error and thereby
increases industrial efficiency [2, 3].
Facial recognition is an essential process to identify and separate the facial region
from the majority of the photos of every face detection device. There are numerous
techniques proposed, ranging from the simple detection of edges to the advancement
of techniques like pattern recognition. The review contrasts two facial recognition
approaches the characteristics of them and the local binary patterns focused on impact
rate recognition and speed detection. The algorithms have been tested on Microsoft
Visual C++ 2010 Express with OpenCV software. Test findings show the reliability
and trust of the functionality of the local binary template for implementing a facial
recognition device in real time [4].
Paper 4 introduces the presentation on Jetson TX2 with FaceNet and MTCNN
algorithms of an advanced and multi-camera face recognition surveillance device.
Presents the analysis of the modern, multi-camera facial recognition tracking system
FaceNet and MTCNN algorithms on Jetson TX2. The suggested portable system
tracks the target or suspect along with the ID/location of the camera and documents
its presence through various camera installations in the database. Triplet loss is one of
the most significant things we find in this article. The gap between the anchor and the
result, which implies a separate identity, is reduced using Triplet loss procedure, if the
result is accurate and indicates the same identity; the gap between the anchor and the
sample is then maximized. Triplet loss is one of the easiest ways to practice effective
128-dimensional embedding for any image. The representation of the anchor here is
Office Monitoring and Surveillance System 7
the frame from which the triplet loss was measured. Three pictures are required to
quantify the triplet loss: an anchor, positive and negative [4, 5].
Study of paper 5 describes a way of monitoring individuals and obtaining canon-
ical frontal pictures that fit the sensor network model in 2D world coordinates. In front
photos, monitoring and identity management have a specific desirability because they
are essentially invariant with daily appearance changes. This method has been applied
and tested on the FaceNet wired camera prototype network. Their key contribution
is to demonstrate how to sense the trajectories of moving objects to acquire canon-
ical vision of high quality while maintaining node capacity. The authors present an
overview of the method and display the task algorithm on data obtained from the
FaceNet camera network [6–11].
The purpose of project implemented in paper 6 is to create a complete hardware
and software framework for the surveillance of entry into a secure facility with
modern facial recognition technology (FRT) and lightweight, cheap components.
For visualization in the operation, FRT senses and recognizes individuals as they
enter and exit the area protected. The project uses FaceNet FRT to boost facial
recognition accuracy by means of a deep-learning approach. The FaceNet FRT is
implemented on a Raspberry Pi using the Neural Compute device Intel Movidius
with input from two portable cameras [7].
Substantial advancement in the field of identity testing and facial detection has
been made lately, taking advantage of advances in the convolution neural networks
(CNN). The findings on human level have been exceeded by FaceNet’s performance
particularly in terms of accuracy in “Labeled Faces in the wild (LFW)” datasets
and “Youtube Faces in the wild (YTF).” The triplet loss in the FaceNet has been
proven successful for facial verification. However, if a large dataset is used to train
the model, the amount of triples available is explosive. In this article, we propose a
simple class-by-class triple loss based on intra-class distance metric learning which
can substantially reduce the number of potential triplets to learn, but simplifying the
classic three-point loss function hasn’t deteriorated the approach. Experimental tests
of the most commonly used LFW and YTF benchmarks show that the model with
a class-wise simple triplet loss can achieve the latest results. And the simulation of
the dispersal of the learned features based on the MNIST data set has shown the
effectiveness of the proposed method of better class separation and discriminating
against the other state-of-the-art loss functionality [12–14].
3 Proposed System
Our system provides automated surveillance and monitoring for offices. This system
will work on employee’s faces for distinguishing a person from a visitor and an
employee of the organization using Face Recognition. The project subject matter
is a solution to orthodox methods of logging in employees in physical handwritten
logbooks, keeping security guards on duty 24 × 7, relying solely on sensors to fire
breakouts, etc. So, to make up for these fairly unreliable methods, we have come
8 V. Patil and Y. Jadhav
up with a plan that this project will increase the quality and efficiency of security in
office premises. The main objective of our project will be to recognize an employee
to let them in the premises and greet visitors respectfully keeping security as a
higher priority. The model is a deep convolutional neural network, which is learned
by means of a three-fold loss function, which induces vectors for one identity to
become more identical (smaller distance). A significant breakthrough in this research
was the emphasis on teaching a computer to establish embedding directly (instead
of removing them from an intermediate layer of a computer).
The system will be fed live image frames through a CCTV [15], which will be
used to detect faces coming up in the frame. For the detection of faces, we are using
OpenCV. Multiple faces can be detected at once. Once the faces are detected, the
faces will be cropped to size and by analyzing the eye orientation, the faces will be
aligned to a proper axis. Once Face Alignment is achieved, these images will be fed
to Google FaceNet. FaceNet will then extract the features from the image received.
Then, these face embedding systems are used as the basis for training classifier
systems on standard face recognition benchmark datasets, achieving state-of-the-
art performance. These preprocessed images will be used by the Support Vector
Classifier to train the model. This training will be done for each new employee that
joins the organization. Once the training phase is completed, the preprocessed data
received realtime, will be fed to the Support Vector Classifier Inference model which
will recognize the face as an employee or not. Changes (recent scars, beard/non-beard,
different lighting conditions) to faces are anticipated to expand its classification
performance as illustrated in Fig. 2.
4 Experimental Results
We require 40–50 facial data of each employee. Above images are used for training
the system. Accuracy achieved ranges from 95 to 100% as shown in Fig. 3.
We have seen various kinds of techniques in this paper that allow us to detect the
face and recognize it in various situations such as dark and light and even in various
clothing where the face is partially visible. All the techniques such as viola jones
used for object detection in OpenCV and haar cascade classifier which is also used
by the help of OpenCV. However, we observed that FaceNet provide significantly
better results than the previously mentioned techniques. Additional features in this
project can be further added after successful implementation, for example, visitors
who have been granted access in the office can be monitored for behavioral analysis;
sentiment analysis to check whether the visitor becomes hostile and creates nuisance,
etc. Furthermore, additional features in which employees can be monitored to keep a
Office Monitoring and Surveillance System 9
check on the performance in work can be implemented. In restricted areas within the
company, only employees who are allowed access will be granted entry otherwise
no.
References
1. Wu W, Liu C, Su Z (2017) Novel real-time face recognition from video streams. In: Conference
on computer systems, electronics and control (ICCSEC), Dalian, pp 1149–1152. https://doi.
org/10.1109/ICCSEC.2017.8446960
2. Xiugang G, Yang L, Ruijin Z, Liang Y (2011) Study on monitoring systems with video capture
cards and wireless sensors. In: IEEE conference on electronics, communications and control
(ICECC), 03 November 2011
3. Saputra DIS, Amin KM (2016) Face detection and tracking using live video acquisition in
camera closed circuit television and webcam. In: 1st international conference on information
technology, information systems and electrical engineering (ICITISEE), Yogyakarta, pp 154–
157. https://doi.org/10.1109/ICITISEE.2016.7803065
4. Kadir K, Kamaruddin MK, Nasir H, Safie SI, Bakti ZAK (2014) A comparative study between
LBP and Haar-like features for face detection using OpenCV. In: 4th international conference
on engineering technology and technopreneuship (ICE2T), Kuala Lumpur, pp 335–339. https://
doi.org/10.1109/ICE2T.2014.7006273
5. Jose E, Greeshma M, Haridas MTP, Supriya MH (2019) Face recognition based surveillance
system using FaceNet and MTCNN on Jetson TX2. In: 5th international conference on advanced
computing & communication systems (ICACCS), Coimbatore, India, pp 608–613. https://doi.
org/10.1109/ICACCS.2019.8728466
6. Heath K, Guibas L (2007) Facenet: tracking people and acquiring canonical face images in a
wireless camera sensor network. In: First ACM/IEEE international conference on distributed
smart cameras, Vienna, pp 117–124. https://doi.org/10.1109/ICDSC.2007.4357514
7. Boka A, Morris B (2019) Person recognition for access logging. In: IEEE 9th annual computing
and communication workshop and conference (CCWC), Las Vegas, NV, USA, pp 0933–0936.
https://doi.org/10.1109/CCWC.2019.8666483
8. Chien Y-T, Huang Y-S, Jeng S-W, Tasi Y-H, Zhao H-X (2003) A real-time security surveillance
system for personal authentication. In: IEEE 37th annua international Carnahan conference on
security technology. Proceedings, Taipei, Taiwan, pp 190–195. https://doi.org/10.1109/CCST.
2003.1297558
9. Qezavati H, Majidi B, Manzuri MT (2019) Partially covered face detection in presence of
headscarf for surveillance applications. In: 4th international conference on pattern recognition
and image analysis (IPRIA), Tehran, Iran, pp 195–199. https://doi.org/10.1109/PRIA.2019.
8786004
10. Cao Y, Pranata S, Nishimura H (2011) Local binary pattern features for pedestrian detection
at night/dark environment. In: IEEE international conference on image processing, Brussels,
pp 2053–2056. https://doi.org/10.1109/ICIP.2011.6115883
11. William I, Ignatius Moses Setiadi DR, Rachmawanto EH, Santoso HA, Sari CA (2019) Face
recognition using FaceNet (Survey, Performance Test, and Comparison). In: Fourth interna-
tional conference on informatics and computing (ICIC), Semarang, Indonesia, pp 1–6. https://
doi.org/10.1109/ICIC47613.2019.8985786
12. Ming Z, Chazalon J, Luqman MM, Visani M, Burie J (2017) Simple triplet loss based
on intra/inter-class metric learning for face verification. In: IEEE international conference
on computer vision workshops (ICCVW), Venice, pp 1656–1664. https://doi.org/10.1109/
ICCVW.2017.194
13. Geng X, Zhou Z-H, Smith-Miles K (2008) Individual stable space: an approach to face
recognition under uncontrolled conditions. IEEE Trans Neural Networks 19(8):1354–1368
Office Monitoring and Surveillance System 11
14. Ibrahim R, Zin ZM (2011) Study of automated face recognition system for office door access
control application. In: IEEE 3rd international conference on communication software and
networks, Xi’an, pp 132–136. https://doi.org/10.1109/ICCSN.2011.6014865
15. Celine J, Sheeja Agustin A (2019) Face recognition in CCTV systems. In: International confer-
ence on smart systems and inventive technology (ICSSIT), Tirunelveli, India, pp 111–116.
https://doi.org/10.1109/ICSSIT46314.2019.8987961
Categorizing Documents by Support
Vector Machine Trained Using
Self-Organizing Maps Clustering
Approach
Abstract This paper mainly emphasis on the use of machine learning algorithms
such as self-organizing maps (SOM) and support vector machines (SVM) for clas-
sifying text documents. We have to classify documents effectively and accurately to
different classes based on their content. We tested classification of self-organizing
map on Reuters R-8 data set and compared the results to three other popular machine
learning algorithms: k-means clustering, k nearest neighbor searching, and Naive
Bayes classifier. Self-organizing map yielded the highest accuracies as an unsu-
pervised method. Furthermore, the accuracy of self-organizing maps was improved
when used together with support vector machines.
1 Introduction
It is very important to store information in a way that is easy for retrieval. This
becomes even more important in the case of electronic information. Every day we
search for documents on our computer is it at home and at work. With each passing
day the amount of information increases. The problem which we need to overcome is
manually storing all related information together and searching for that information.
This may take a lot of effort and time. The solution to this problem is to classify
and label documents. Once the documents are classified and ordered, we can go
V. Patil (B)
Computer Engineering Department, K. J. Somaiya Institute of Engineering and Information
Technology, Mumbai, India
Y. Jadhav
Computer Science and Engineering Department, Amity School of Engineering and Technology,
Amity University, Mumbai, India
A. Sirsat
Computer Engineering Department, St. John College of Engineering and Management, Mumbai,
India
through the documents and search for information according to our needs. Consider
a scenario where documents are already ordered and. In such a scenario, we can
browse through documents effectively. This is where machine learning comes into
focus.
Machine learning algorithms can classify documents automatically making it
easy to store them. These algorithms can be supervised or unsupervised. Supervised
machine learning algorithms involve human interference. Since it may not be the case
that class labels are already known for each document, this may give rise to human
error. Also, many a times the document has no title or has an ambiguous title, which
makes it difficult to understand the category of the document. Thus we have focused
on unsupervised machine learning algorithms for the task of document classification.
We have used Kohonen’s maps or self-organizing maps to classify documents. Self-
organizing maps (SOMs) [1] is an effective data Visualization method which can be
used to reduce the dimensions of data by using self-organizing neural networks [2].
It is a very popular method for clustering [3] and classification of text documents [4,
5]. We compared self-organizing map with one unsupervised [6] and two supervised
machine learning algorithms.
Self-organizing maps gave better performance as compared to k nearest neighbor
and k-means clustering with accuracies as high as 90–92%. However, the draw-
back of SOM is Curse of Dimensionality (COD) because of which its efficiency is
affected. Because of this, SOM is not able to differentiate between classes that give
the same Euclidean distance even though they have different features. Hence, we
have implemented a technique which uses SOM and SVM. SOM takes the unlabeled
text documents and forms clusters for these documents which form the training data
set for the Support Vect4or Machine (SVM) [7].
This method has a two-fold advantage; it reduces the effect of COD while also
eliminating human interference needed to label the training documents. Support
Vector Machine have been one of the most widely used machine learning algorithm
in most of the text classification application given as input for many a times for large
datasets manually classifying and labeling training documents in the preprocessing
stage leads to an increase in time and cost complexity. Thus, the unsupervised clus-
tering methods are used which eliminate the need of human involvement for grouping
together set of documents. The output of these methods is then Training phase of
SVM. This hybrid model gives excellent accuracy while also eliminating the effect
of COD. Moreover, we can see that the accuracy of text classification increases after
using this hybrid approach.
2 Literature Survey
Support vector machine is one of the most widely used supervised machine learning
algorithms. It is one of the most commonly used methods and has found use in a
variety of applications due to its excellent accuracy [8] (Fig. 1).
Categorizing Documents by Support Vector Machine … 15
However, for some of the classification domains, it might not always be possible to
label the training data set of Support vector machine. For certain domains, labeling the
training data might be difficult when human expertise is unavailable or when it may
be time-consuming and costly to involve humans. Hence, in cases where it may not
be possible to label training data, unsupervised clustering methods are introduced
as they can perform their task without training the dataset. For domains wherein
training data is unavailable, we can use unsupervised machine learning methods,
such as SOM and k-Means to group documents based on its content.
A SOM is a single layer neural network. The name artificial neural network is
because of the fact that they were originally inspired by the way how biological
neurons were thought to work. SOM is made up of multiple units lying along a grid
of n-dimensions. In the input layer for every unit has the number of weights and
input patterns are same. Training a Self Organizing Map when provided with an
input pattern requires calculating the distance between that pattern and every other
unit in the neural network [9]. A unit that has the least distance to Best Matching
Unit (BMU) is then selected and that pattern is then mapped onto that unit. There
will be a one to one correspondence between the mapping of patterns that lie nearby
in the input space and the units lying nearby in the output space. This will be visible
in the visual map generated by SOM [9, 10]. Thus, SOM is topology preserving as
neighborhoods of a node are kept intact through the mapping process (Fig. 2).
3 Proposed System
The data set we have selected is a variant of the Reuters- 21,578 collection. This
dataset from the year 1987 includes 21,578 English Reuters news documents. More
than one class labels to some the documents Therefore; we have discarded documents
16 V. Patil et al.
with multiple class labels and selected only those documents which have only one
class label.
We then selected the 10 largest classes to obtain the collection of 8008 documents,
collection of 8008 documents, consisting of 2254 test documents and 5754 training do
over the past two decades, many research works focusing on document classification
by machine learning algorithms have been carried out. A large number of these papers
focus on improving the efficiency of SVM. Many researchers have proposed a hybrid
method which combines unsupervised learning algorithms such as SOM with SVM
[7]. Wu’s research shows how we can improve the accuracy of SVM [11] analyzing
zebra fish gene expression [12] by analyzing data in the training data set by using
SOM. SOM filters the training data for Support vector machines thereby reducing
the train dataset by keeping only those neurons in every category of SOM that are
their BMUs while it eliminates the remaining we then performed conventional text
preprocessing on the dataset as shown in Fig. 3.
1. The dataset we have used is the Reuters R8 dataset having 5485 training and
2254 test samples respectively.
2. For preprocessing, firstly words we retransformed to their stems and then, stop
words were removed.
3. Next, we calculated the frequencies of resultant words and computed document
vectors for all documents.
4. Finally, tf-idf weights were calculated for all resulting word stems.
5. For using a self-organizing map in the document classification task we have to
label the map nodes with class labels of the training data set.
6. The labelled nodes then represent the document classes. The map can then
classify new documents by mapping them.
SOM is primarily used for labeling and organizing the data samples which are not
labeled. As mentioned earlier in this paper, the SOM [2] is affected by COD because
it measures similarity using Euclidean distance [7]. Reducing the dimensions of
original data in their principle components might be of some help. However, this
reduction is unsuitable for some datasets, because every feature consists of some
information. Using this may result in loss of information if any one of the dimensions
Categorizing Documents by Support Vector Machine … 17
is removed [7]. The obtained labels are coupled with the related original data samples
for creating a machine labeled training set for the SVM which was machine organized.
The SOM begins the step of clustering the documents by deciding which important
features are needed for efficient clustering of the data into groups [7]. A process of
statistical cluster analysis is usually used to better the feature extraction which is
iterative in nature. The dimensions of data are reduced to two dimensions which
simplify viewing and interpretation is obtained using the feature maps [7].
Self-organizing maps generate a map which represents the dataset [7]. When
different neurons are fired for each data point a greater number of neurons will give
larger number of classes. Thus, to prevent this, the neurons are first grouped in small
clusters using Euclidean distance. Many neurons together make a cluster [7]. Every
cluster is given a class number. Using the Euclidean distance all the data points are
grouped together around these clusters. A number of a particular cluster is assigned
to data points with the closest proximity to that cluster [7].
The combination of SOM and SVM which results in a method which uses unsuper-
vised learning and has better generalization ability [7]. This is assumed from the prop-
erties of the SVM. SVM is known to be excellent classification methods with respect
to performance [7]. The generalization attribute of Support Vector Machine because
we implement structural risk minimization which selects an optimal separating hyper
plane which guarantee classification with very high accuracy (Fig. 4).
First we compared different algorithms used for clustering based on their accuracy.
The results are as shown in the Table 1 and the graph in Fig. 6. We can see from the
Table 1 that the accuracy of SOM is 92.45% which is very good when compared to
methods such as k means and k nearest neighbor whose accuracies are 90.23% and
89.9% respectively. The accuracy of Naive Bayes is 95.24%. As seen in Fig. 6 the
accuracy of the hybrid model of SOM and SVM is 96.44% which is more than the
accuracy obtained from Naive Bayes (Fig. 5).
5 Conclusion
We tested Self Organizing Maps for classification of text documents and compared
it with some of the commonly used text classification methods. Naive Bayes had
the best accuracy overall, but in unsupervised algorithms, SOM turned out to have
the best accuracy for text classification. Sometimes, the knowledge and expertise to
label and organize documents into separate classes might be unavailable and thus
unsupervised methods viz. SOM proves to be useful. Self-organizing maps gave over
90% classification accuracy in the Reuters R-8 Data Set making it is effective for the
document classification. Also, the accuracy of self-organizing maps was comparable
Categorizing Documents by Support Vector Machine … 19
Table 1 Comparison of
Algorithm Accuracy (%)
clustering methods
k means 90.23
kNN 89.9
Naive Bayes 95.24
SOM 92.45
to some of the supervised machine learning methods that we tested. The training
phase of SOM generates an intuitive visual map, which is useful in a variety of
applications.
The learning procedure of this method though costly, is advantageous. Theo-
retically, the learning phase of SOM is more time consuming as compared to the
learning of k nearest neighbor. However, practically it is the exact opposite. The map
has significantly less nodes compared to the training dataset. This result to 10 time’s
quicker classification when compared to that obtained from k nearest neighbor using
20 V. Patil et al.
the same data set. We don’t have to build a new map whenever we get a new input
document in the dataset, we can map it and perform the learning phase afterwards
when there is no new input data available. The results presented above show that
when SOM is used along with SVM the accuracy obtained is near to that which we
get when the data used for training is classified by supervised machine algorithms.
Categorizing Documents by Support Vector Machine … 21
References
1 Introduction
Figure 1 shows the proposed E-shaped mulilayer MSA with capacitive feed. The
design Model consists of a copper ground plane, radiating Patch, air gap, and FR-4
as a dielectric substrate. The Capacitive probe feed technique has been used for the
impedance matching with the radiating patch.
Bandwidth Improvement of Multilayer Microstrip … 25
The air gap acts as a dielectric substrate having a dielectric constant is equal to
one. Here the proposed model has become Multilayer consisting of a copper ground
plane, air gap with a height of 6 mm, a dielectric substrate with a height of 1.6 mm
easily available in the market, and an E-shaped radiating patch [11]. The permittivity
can be selected near to 1 to obtain a wide bandwidth. If the dielectric substrate has
larger thickness then it gives rise to an increase in probe reactance for the excitation
of surface waves and coaxial feed, which reduces the efficiency of the antenna so that
dielectric height should be selected properly. This design model is used to investigate
the antenna parameters like Operating Frequency, Bandwidth, Return Loss, VSWR,
Directivity, Total Gain, Radiation pattern, and E-field.
Figure 2 represents the Bandwidth and Return loss of E-shaped Suspended MSA
which gives the bandwidth of 238 MHz and Return Loss −12.98. The return loss of
Antenna is in acceptable range.
26 A. K. Rathod et al.
Figure 3 represents the VSWR of E-shaped Suspended MSA. It shows the VSWR
1.5778 must be less than 2 as per the standards. The given antenna can be radiate
and fabrication process can be done.
Figure 5 represents the Gain of E-shaped Suspended MSA is 6.0196 dB, which is
better than the conventional Antena.
See Fig. 6.
Table 1 consists of result analysis table for Hexagonal MSA & Suspended Hexagonal
MSA with performance Parameters like Operating Frequency in GHz, Bandwidth in
28 A. K. Rathod et al.
MHz, Return Loss in dB, VSWR, Directivity in dB and Total Gain in dB as shown
in Table 1.
The Bandwidth can be defined as a reappearance extends over which VSWR is
always under two (which relates to an arrival loss of 9.5 dB or 11% reflected force).
The VSWR or impedance BW of the MSA is characterized as the recurrence extend
over which it is coordinated with that of the feed line inside determined cutoff points.
The Bandwidth of proposed E-shaped multilayer MSA is 238 MHz, which is
within the specified limits as shown in Fig. 2. Return Loss has been observed in an
acceptable range. VSWR can be defined in terms of the input reflection coefficient
(G) as:
1 + ||
V SW R = (1)
1 − ||
5 Conclusion
Reference
1. Kumar G, Ray KP (2003) Broadband micro strip antennas. Artech House Boston, London
2. Al-Halawani MG, Al-Najjar M, Abdelazeez MK (2016) Microstrip antenna design for UWB
applications. IEEE. 978-1-5090-2886-3/16/$31.00
3. Faouri YS, Awad NM, Abdelazeez MK (2019) Hexagonal patch antenna with triple band
rejections. In: Proceedings on 2019 IEEE Jordan International Joint Conference on Electrical
Engineering Information Technology JEEIT 2019, pp 446–448
4. Deshmukh AA, Ray KP (2013) Analysis of broadband Psi ()-shaped micro strip antennas.
IEEE Antennas Propag Mag 55(2)
5. Tiwari K, Pujara D (2015) Multilayer slotted microstrip antenna for Wi-Fi application. IEEE.
978-1-4799-7815-1/15/$31.00
6. Kachare AB, Mathpati MS, Rathod AK (2015) Effect of size and location of the DGS on char-
acteristics of rectangular micro strip patch antenna. Int J Comput Appl (0975–8887), Confer-
ence on Emerging Applications of Electronics System, Signal Processing and Computing
Technologies (NCESC 2015)
7. Liyakath RA, Takshi A, Mumcu G (2013) Multilayer stretchable conductors on polymer
substrates for conformal and reconfigurable antennas. IEEE Antennas Wireless Propag Lett 12
8. Saha R, Maity S, Trigunayat N (2015) Enhancement of gain, bandwidth and directivity of a
patch antenna by increasing dielectric layers of the substrate through micromachining technique
for RFID application. In: 2015 international conference on advances in computer engineering
and applications (ICACEA), IMS Engineering College, Ghaziabad, India, IEEE. 978-1-4673-
6911-4/15/$31.00
9. Nageswara Rao L, Santhosh Kumar V (2020) A novel multiband micro strip patch antenna for
satellite applications. Int J Adv Sci Technol (IJAST) 29(3s):177–183. ISSN: 2005-4238
10. Ali KM, Nataraj C et al (2020) Bandwidth analysis of microstrip antenna with fractal design
for 2.4 GHZ wireless applications. Int J Adv Sci Technol (IJAST) 29(1):1344–1358. ISSN:
2005–4238
11. Hong S, Kim Y, Won Jung C (2017) Transparent microstrip patch antennas with multilayer
and metal-mesh films. IEEE Antennas Wireless Propag Lett 16(c):772–775
Use of Median Timbre Features
for Speaker Identification of Whispering
Sound
Abstract Identifying speaker from the whispered voice is difficult task contrasted
to neutral as voiced phonations are absent in the whisper. The accomplishment of
the speaker identification system for the most part relies on the selection of proper
audio features reasonable for the type of database and type of application. This paper
examines the various audio features available and emphasizes on the use of selected
timbrel features which are sorted by Hybrid Selection Algorithm. The limited number
of timbrel features namely MFCC, Roll-off, Brightness, Roughness, and irregularity
which are found outperforming when tested on CHAIN database. Likewise, the
possibility of using the MEDIAN based features is investigated by analysis. The
use of Median timbrel features reported an enhancement in speaker identification
accuracy by 2.4% compared to timbrel features only in whisper train-whisper test
scenario.
1 Introduction
Speaker identification requires the main processes as (i) Pre-processing, (ii) Feature
extraction (iii) Training and (iv) Testing. Pre-processing may involve noise reduc-
tion, pre-emphasis, silence removal and equalization so forth. Audio features are
the speaker-specific information represented in a compact form. They will make it
possible to compare the speakers for identification purpose with minimum processing
speed and data. In the training phase, all extracted audio features are utilized for
modelling the speakers which will be used as the templates in the testing phase.
V. M. Sardar (B)
Jayawantrao Sawant College of Engineering, Pune, MS, India
M. L. Jadhav
MET’s Institute of Engineering, Nasik, MS, India
S. H. Deshmukh
Maharashtra Institute Of Technology, Aurangabad, MS, India
While testing phase, features of query speaker is compared for the closest speakers
among the database. There are two types of variants in the speaker identification
system, in particular, closed-set and open set. In the open set scenario, the unknown
speaker not available in the database may also be used, but that task will be speaker
verification [1]. The speaker verification is a binary decision to whether a claimed
speaker is the same or not by comparing with the single audio template. While in
an identification task, every speaker is compared with every other speaker in the
database [2].
Identifying the person from the whispered voice is a challenging task due to
its hidden and multi-dimensional attributes. Unmistakable separation ability of the
rich phonation is misplaced from a whisper [3]. The ADs applicable for the neutral
database is not found effective for the whispered database. Hence, the adapted
approach as shown in Fig. 1 where a selection of the proper ADs is carried out
in the beginning only. This process of selecting the appropriate ADs reasonable for
the type of database and application will enhance the identification rate.
A majority of low-level audio features are discussed in the literature. It includes
primary features such as the zero-crossing rate, speech bandwidth, the spectral
centroid, and energy [4, 5]. A second high level and complex feature used for speech
and speaker recognition are like Mel-frequency cepstral coefficients (MFCC), roll-
off and brightness etc. Such features use the parametric analysis of the spectral
envelope [6]. It is expected that the features used for the classification need to be
more uncorrelated. Typical classification results using better non-correlated MFCC
features found correct up to 95% for music and speech classification which is a rela-
tively less complex task [7]. Besides, these results have reported for clean speech
scenario. For other databases including noise, music background, and telephone
speech etc. the performance degrades considerably [8, 9]. The block diagram of
speaker identification system is as shown in Fig. 1.
This paper is organized as follows. We portrayed the classification of audio
features, timbrel audio descriptors, MIR toolbox, and the impact of timbrel feature
2 Audio Descriptors
We can classify the audio features comprehensively in two classes as follows. The
global descriptor is the class of feature where computation is done on the complete
signal as a whole. E.g. attack-time of a sound is known from the complete duration
of an audio signal. Instantaneous descriptors is another class which computes on a
short period (say 40 ms) of audio signal called a frame. The spectral centroid of a
signal can vary with the time, hence it will be termed as an instantaneous descriptor.
As an instantaneous descriptor produces multiple values for the number of frames,
use of statistical operations (like the mean or median, the standard deviation, and or
inter-quartile range etc.) is essential to derive a single value representation. In the
CUIDADO project [10], a listing of 166 audio features is provided.
Depending upon the type of process used for extraction of the feature, we can
further differentiate:
• Features that are directly extracted from the time domain waveform of audio signal
like the zero-crossing rate.
• A transform like FFT, wavelet etc. is applied on a signal to extract the features.
E.g. MFCC.
• Feature extracted based on a signal model like the source or filter model.
• Features which converges to the human auditory response (response on bark or
ERB scale).
Timbre is the perceptual and multi-dimensional feature of sound. As the exact defi-
nition of timbre is very difficult, it can be analyzed by following the attributes
[11].
1. Harmonic analysis: The number, relative strengths, structure of harmonics.
2. Partial analysis: Phase, inharmonic partials, the content of partials.
3. Time-related parameter like rise-time.
4. Steady-state and attack slices.
34 V. M. Sardar et al.
Musical Information Retrieval (MIR) toolbox is for the most part intended to enable
the study of the relation between musical attributes and music-tempted sensation.
MIR toolbox uses a modular outline. It is well known that the common algorithms are
used in audio processing like segmentation, filtering, framing etc. with an addition of
one or more distinguished algorithms at some stage of processing. These algorithms
are available in a modular form and the individual blocks can be integrated to capture
some features [12].
The philosophy to integrate the appropriate modules is proposed in Fig. 2. For
example, to measure irregularity and brightness, we need the implement the algo-
rithms like reading audio samples, segmentation, filtering, and framing as the
common processes between them. In the final stage, due to inherent differences, irreg-
ularity needs peaking algorithm and brightness needs spectrum analysis (Fig. 2).
Even, the integration of different stages depends upon parameter variations. E.g.
mirregularity (…, ‘Jensen’), where the adjoining partials are taken into considera-
tion and mirregularity (…, ‘Krimphoff’) which considers the mean of the preceding,
same and next amplitude [13]. The flow diagram of algorithm modules and their
integration to extract the selected timbrel features are shown in Fig. 2.
Roll-off frequency: Roll-off is assessed from the foremost energy (85% or 95% as
a standard) contained below the predefined frequency.
Roughness: It estimates the average disagreement between all peaks of the signal.
It is also an indicator of the presence of harmonics generally higher than the 6th
harmonic.
Brightness: It is the measure of the percentage of energy spread above some cut-off
frequency.
N −1
−2ikn
Xk = xne N k = 0, . . . , N − 1 (1)
n=0
3 System Description
This undertaking makes use of the CHAIN database developed at the School of
Computer Science and Informatics University College Dublin [15]. It consists of
a total of 36 speakers with 33 samples each, with a good mix of male and female
voice samples (20 males and 16 females). The speech samples of 2–3 s duration are
recorded at a sampling frequency of 44.1 kHz. The phrases/sentences are selected
from CSLU and TIMIT database which ensures the phonetic balance in the corpus.
Hybrid selection is the iterative process which starts with the targeted timbre class
of ADs and further investigates the AD which maximizes the identification result
[16]. This algorithm also used for liver tissue pathologicssal image classification in
[17]. After every iteration, the feature which maximizes the classifier accuracy is
appended in combination with the previous feature/s. The process continues until no
further increase in accuracy is observed. The algorithm proved that MFCC, Roll-off,
Roughness, Brightness, and Irregularity are found to be best suitable.
Feature
Hybrid
Database
Extraction Classifier
Selection
Speaker
ID
K-NN is a simple and non-parametric algorithm which separates the data points
into several classes and a new sample point is used to anticipate the class. The non-
parametric technique of classification does not make any assumptions on the distri-
bution of data. This methodology is helpful as the greater part of the real world data
which does not obey the generally assumed pattern (e.g. linear regression models).
While managing with this classifier, the following parameters are used: the number
of nearest neighbors (k), a distance function (d), decision rule and n labelled samples
of audio files Xn . A class label to the unknown sample is identified based on the
minimum distance from the training classes. In another word, K-NN calculates a
posteriori class probabilities P (wi|x) as
ki
P( wi |x) = .P(wi ) (2)
k
where ki is the number of vectors which belongs to class within the subset of k vectors
[18]. For our system, all the variants of KNN classifier are verified to maximize the
identification accuracy. The variations tested for the number of nearest neighbor
are 1-NN, 2-NN, and 3-NN. Two distance functions Euclidean and City-block are
investigated. The rules namely nearest and consensus are also tested. After a variety
of experiments, it is concluded that a combination of 3-NN neighbors, City-block
distance and Nearest rule give the maximum identification accuracy in [19].
usually defined to be the MEAN of the two middle values. So MEDIAN may be
considered as the fully sheared mid-range.
MEDIAN can be calculated by the following formula which will be generalized
for odd and even number of values in the set:
a# 2x + a# 2x /0.5
MEDIAN = (3)
2
Comparing the feature values of query (test) speaker with the MEDIAN values
of the particular assigned class (at the time of training), distance from MEDIAN
is recommended. As the intra-speaker features are distributed in some range of
its variability, the median as a reference of class value will certainly improve the
identification accuracy. The following analysis will confirm our conclusion.
In the above illustration, five samples (1_1, 1_2, 1_3, 1_4, and 1_5) of the same
speaker (speaker_1) are listed for the intra-speaker City-block distance. Then in
another approach, the distances of the speakers from the MEDIAN of a class are
Use of Median Timbre Features for Speaker Identification … 39
Table 3 Comparative identification accuracy by using MFCC, timbre only and timbre (Median)
Speech mode % Accuracy
Training–Testing MFCC Timbre Timbre (Median)
Neutral–Neutral 91.5 95.0 95.6
Whisper–Whisper 78.8 86.5 88.1
Neutral–Whisper 66.0 73.0 76.4
CHAIN database of 35 speakers, MFCC only/timbre feature/timbre (median), and K-NN
Classifier
calculated (shown in bracket). It is seen that the distances of the inter-speaker samples
are less when calculated concerning MEDIAN as compared the absolute values. It
means that the possibility of identification of all samples of speaker_1 increases as the
intra-speaker distances are reduced when calculated from the reference of MEDIAN
values.
5 Results
This section presents the results with the following parameters. Total 35 speakers
with whispered samples from a CHAIN database are used for the experiments. The
samples of each speaker are selected with a choice of 70% samples for training
and 30% for testing. KNN classifier settings are as—Rule: nearest, Neighbor: 3,
and distance-Metric: City-Block distance. The selected Timbre features by Hybrid
Selection Algorithm are Roll-off, Brightness, Roughness and, irregularity.
Mel Frequency Cepstral Coefficient (MFCC) is widely used features for speaker
identification task. Hence, the performance using MFCC features is compared with
timber features and the Median values of it (Table 3).
The above results in Fig. 4 are compared with a baseline speaker identification
system for whispered speech using the CHAIN database [20]. Further, an enhance-
ment of 2.75% in the identification accuracy system in whisper train-whisper test
scenario is reported in [21] are reproduced in Table 4.
The proposed timbre features using Median value shows that the identification
accuracy is increased by 2.4% in whisper train-whisper test scenario which is the
focus of this undertaking.
6 Conclusion
A large number and variety of audio descriptors are available which are selected
according to the application. As the attributes of whisper are drastically change
compared to the neutral voice, the perceptually motivated, multidimensional timbre
40 V. M. Sardar et al.
100
80
60
40
20
0
Neutral-Neutral Whisper-Whisper Neutral-Whisper
% Accuracy MFCC % Accuracy Timbre % Accuracy Timbre(Median)
References
1. Hourri S, Kharroubi J (2020) A deep learning approach for speaker recognition. Int J Speech
Technol Springer 23:123–131
Use of Median Timbre Features for Speaker Identification … 41
2. Jahangir R et al (2020) Text-independent speaker identification through feature fusion and deep
neural network. IEEE Access 8:32187–32202
3. Singh A, Joshi AM (2020) Speaker identification through natural and whisper speech signal.
In: Janyani V, Singh G, Tiwari M, d’Alessandro A (eds) Optical and wireless technologies.
Lecture Notes in Electrical Engineering, vol 546. Springer, Singapore
4. Fan X, Godin KW, Hansen JHL (2011) Acoustic analysis of whispered speech for phoneme
and speaker dependency. In: Proceedings of the annual conference of the international speech
communication association, INTERSPEECH, pp 181–184
5. Bhattacharjee M, Prasanna SRM, Guha P (2018) Time-frequency audio features for speech-
music classification. Project: Broadcast Video Analytics
6. Davis SB, Mermelstein P (1980) Comparison of parametric representations for monosyllabic
word recognition in continuously spoken sentences. IEEE Trans Acoust Speech Signal Process
ASSP s28:357–366
7. Hermansky H, Malaya N (1998) Spectral basis functions from discriminant analysis. In:
International conference on spoken language processing
8. Toonen Dekkers RTJ, Aarts RM (1995) On a very low-cost speech-music discriminator.
Technical Report 124/95, Nat. Lab. Technical Note
9. Dobrowohl FA, Milne AJ, Dean RT (2019) Timbre preferences in the context of mixing music.
Appl Sci (2076–3417) 9(8):1695–1695
10. Peeters G (2004) A large set of audio features for sound description (similarity and
classification) in the CUIDADO project
11. Park TH (2004) Towards automatic musical instrument timbre recognition, PhD thesis, the
department of music, Princeton University
12. Albert-Ludwigs-Universität Freiburg (2007) A Matlab Toolbox for Music Information. In:
Proceedings of the 31st annual conference of the Gesellschaft für Klassifikation e.V., March
7–9, pp 261–268
13. Lartillot O (2011) MIR toolbox 1.3.3 (Matlab Central Version)—User’s Manual. Finnish centre
of excellence in interdisciplinary music research, University of Jyväskylä, Finland
14. Sardar VM, Shirbahadurkar SD (2019) Timbre features for speaker identification of whispering
speech: selection of optimal audio descriptors. Int J Comput Appl (Taylor Francis, U.K.). ISSN
1206–212X
15. Cummins F, Grimaldi M, Leonard T, Simko J (2006) The chains speech corpus: characterizing
individual speakers. School of Computer Science and Informatics University College, Dublin
16. Deshmukh S, Bhirud SG (2012) A hybrid selection method of audio descriptors for singer
identification in North Indian Classical Music. In: Fifth international conference on emerging
trends in engineering and technology, pp 224–227
17. Liu H, Jiang H, Zheng R (2016) Computational and mathematical methods in MedicinePB—
Hindawi Publishing Corporation
18. Shah JK, Smolenski BY, Yantorno RE, Iyer AN (2015) Sequential k-nearest neighbor pattern
recognition for usable speech classification. Signal Processing Conference, IEEE Xplorer
19. Sardar VM, Shirbahadurkar SD (2018) Speaker identification of whispering sound: effective-
ness of timbral audio descriptors. In: International conference on power, communications and
sustainable energy system, Chennai
20. Wang J-C, Chin Y-H, Hsieh W-C, Lin C-H, Chen Y-R, Siahaan E (2015) Speaker identification
with whispered speech for the access control system. IEEE Trans Autom Sci Eng 12:1191–1199
21. Sardar V, Shirbahadurkar S (2018) Speaker identification of whispering speech: an investigation
on selected timbrel features and KNN distance measures. Int J Speech Technol 21. https://doi.
org/10.1007/s10772-018-9527-4
Intelligent System for Engine
Temperature Monitoring and Airbag
Deployment in Cars Using
Abstract The CAN bus has emerged as vital means of communication within the
automotive sector. Intrinsically varied applications are being implemented using the
Controller Area Network Protocol. In this paper, two major automotive applications
namely engine temperature monitoring and control and airbag deployment mecha-
nism are implemented with the assistance of ARM7 based microcontroller nodes.
Thus the essential CAN protocol has been implemented for real time embedded
automotive applications.
Keywords ARM7 · Basic CAN Protocol · CAN Bus · Nodes · Real-Time data
1 Introduction
node consists of the ARM7 based microcontroller (with inbuilt CAN Controller) and
may Transceiver each. The important time data from sensors of the sensing node
is processed by the microcontroller LPC2129. The CAN Controller takes care of
the CAN broadcast message data. This data is then sent onto the CAN Bus. This
data is then received by the controlling node from the CAN bus. The controlling
node then further takes appropriate action. For controlling the temperature, a lover
is often used; while the airbag deployment mechanism is often achieved by turning
on a relay connected to the controlling node. This ensures the engine temperature
being monitored and controlled further and therefore the driver’s safety; that has
drawn special attention thanks to increasing number of accidents day by day caused
by high risky speeds. Thus the essential CAN protocol has been implemented for
real time embedded automotive applications.
2 System Design
A CAN bus automotive electronic system consists of every controller through the
CAN bus interconnected together to exchange information (such as speed, engine
speed, engine ambient temperature etc.). This mutual connection helps the controllers
to use the info at an equivalent time. On the opposite hand it helps to extend the
throughput of the system, to develop updated features within the system and reduce
the system cost. Electronic automotive instrument cluster node can fully reflect the
car factor, product design and technical standards. Intelligent sensor nodes accurately
receive various sorts of signals and at an equivalent time can eliminate interference
signals [2].
1. CAN Bus Electrical Characteristics: CAN transmission medium formed by
the two, One is called high-level transmission line CANH and the other one is
called low-level transmission line CANL.
2. Hierarchical structure of CAN protocol based BUS: According to CAN BUS
network structure, Classified into the following five levels, as shown in Fig. 2.
3. CAN Message Transmission and Frame Type: CAN message use frame as
transmission unit. In the CAN 2.0B specification is provided in standard format
data frame, Given in two different 2.0B frame format, the difference is the length
of different identifiers: Frame with n-bit identifier called the standard frame, the
frame contains 29 identifiers is called extended frame. Message transmission
has the following four different types of frames: data frame, remote frame, error
frame, overload frame [3].
4. Can Bus Node Design: Figure 3 shows the CAN Bus node bus which consists
of two nodes namely sensing node and the controlling node. The sensing node
consists of following sensors: temperature sensor and shock sensor. The real
time data from the sensors is sent to the ARM7 based LPC2129 microcon-
troller and then to the CAN Bus via CAN Transceiver MCP2551. This data is
then received by the controlling node and appropriate action is taken. Thus the
ambient temperature of engine can be monitored. Also to control it further, a
cooling fan can be started. The other application is of Airbag deployment mech-
anism. As soon as the data is received from the shock sensor, the controlling
node takes appropriate action. The action can be making a relay on as way to
deploy the airbag further.
The sensor used for sensing temperature is LM35 while the sensor used for shock
sensing is 801S. The shock sensor is micro vibration detection sensor and is useful
for the demo purpose. The microcontroller used for the node is LPC2129 (ARM7
based and Philips made microcontroller). It has several advantages including inbuilt
CAN Controller, supports high speed CAN operation, low power consumption etc.
The CAN Transceiver plays a major role here. It is used for communication between
the node and the CAN bus and then the other nodes. The data is sent onto the bus
via the high speed CAN Transceiver. The CAN Transceiver MCP2551 is compatible
with CAN 2.0B. It can provide differential transmit capability to the bus and differ-
ential receive capability to the CAN controller. The chip was originally designed for
automotive high-speed communications is fully compatible with ISO/DIS standard,
and the moment of interference in the anti-car environment, protecting the ability
of the bus. It is compatible with ISO-11898 standard at speeds up to 1 Mb/s, can
resist the strong interference the moment of interference such as the automotive
environment. General internal bus of circuit protection and limiting circuit, and has
a low current standby mode and slope control to reduce radio frequency interfer-
ence [4]. The sensing node may control the temperature by turning a cooling fan
on. Also the airbag deployment is depicted by turning on a relay. Thus two appli-
cations namely engine temperature monitoring and airbag deployment mechanism,
have been implemented.
Intelligent System for Engine Temperature … 47
The system software is written in embedded C and is compiled using ARM Keil Micro
vision v4 software. Flash Magic is used to upload the program to the Microcontroller
chip. The software flow consists of CAN transmission and Reception of messages.
The flow is depicted below with the aid of flowcharts as shown in Figs. 4, 5.
See Fig. 4.
The CAN Controller is initialized first. The data from one node is sent onto the bus
and then received by the other node. The flowcharts of transmission and reception
of data are shown in figs. 4 and 5 [5]. The received messages are then monitored on
the terminal of a personal computer (Fig. 6).
4 Results
As shown in Fig. 6 the real time data can monitored through UART on a personal
computer. The messages with corresponding ids can be seen on the terminal. A
specific message can be dedicated to certain sensor data. The real time data thus
can be monitored and corresponding controlled actions can be taken further. The
engine temperature can be controlled by turning on a cooling fan; while the airbag
deployment mechanism can be achieved by turning on a relay.
Intelligent System for Engine Temperature … 49
References
1. Piao C-H, Chen L, Cao J (2010) A design for controller area network bus real-time monitoring
system. In: 2011 International conference on computer science and network technology, pp
1621–1624
2. Bo L, Tao J (2012) The design of monitoring system based on CAN bus. In: International
conference on measurement, information and control (MIC), pp 137–140
3. Ran L, Junfeng W, Haiying W, Gechen L (2010) Design method of CAN BUS network
communication structure for electric vehicle. In: IFOST Proceedings. 978-1-4244-1849-7/08/
4. Li R, Liu C, Luo F (2008) A design for automotive CAN bus monitoring system. In: IEEE
vehicle power and propulsion conference (VPPC), Sept 3–5, Harbin, China
5. Quanqi W, Jian W, Yanyan W Design of vehicle bus data acquisition and fault diagnosis system.
978-1-61284-459-6/11/
Analysis and Prediction
of Temporomandibular Joint Disorder
Using Machine Learning Classification
Algorithms
1 Introduction
Today, the issue of TMJ disorder has been growing systematically across all ages
of the human population. Early diagnosis of this musculoskeletal condition will
significantly aid in maintaining the mental and physical health of the subject [1–
3]. With the increase in the use of machine learning-driven models in the forecasts
of various human diseases, their early detection now seems to provide promising
results based on different health and physiological constraints. We were inspired
by this reason to raise awareness in the diagnosis and study of TMJ condition to
R. B. Kakkeri (B)
AISSMS Institute of Information Technology, Sinhgad Academy of Engineering, Kondwa, Pune,
India
D. S. Bormane
AISSMS College of Engineering, SPPU, Kennedy Road, Pune, India
enhance improved technique of treatment. The round end of the lower jaw, known
as condyles, slips around the socket of the temporal bone joint when we open our
mouth [3]. TMJ disease diagnosis becomes a problem when there are many other
facial and muscle conditions whose few signs are very similar to those with signs
of TMJ disease, which is why it is considered a complex disorder. That makes this
job a challenging one. Early detection and treatment [4] are most imperative steps
to be taken to decrease the symptoms of TMJ disorder problem and to improve the
quality of life of suffering people. However, there is no procedure of medical test
for detection of TMJ disorder. In dentistry, One of the most common disorder in the
oral cavity that mostly affects women is temporomandibular joint disorder. Popular
Temporomandibular Joint Disorder (TMD) symptoms include,
• Problems with opening your mouth wide as you try.
• Jaws in the open-mouth or closed-mouth role that get “stuck” or “stall”.
• When you open or shut your mouth or chew, cracking, ringing, or grating sounds
in the jaw joint.
• Inflammation or swelling on the side of your face [1–3].
• Dental pain, headaches, neck aches, dizziness, earaches, hearing issues, upper
shoulder pain, and ear ringing (tinnitus) can also be present in patients [4].
The method of diagnosing this is conducted manually with the use of clinical tools
that with improper use becomes unsafe for the patient. The existing clinical methods
are also affluent for or common citizens. Arising technologies aided with applications
through the use of machine learning algorithms. Signal processing is practically
applied in medical diagnosis Digital acquisition techniques with processing of signals
becomes an important option especially in medical diagnosis [3, 5]. Moreover, the
clinical diagnostic procedure for TMJ disorder is costly. In the long run, lack of
awareness or neglecting of early signs and symptoms of TMJ disorder may lead to
even costlier clinical treatment. The general objective of the study is
To detect Temporomandibular Joint Disorder (TMJD) using surface electromyo-
graphy of masticatory muscles.
To introduce current prospects of diagnosis and treatment of Temporomandibular
Joint Disorder.
To mention advantages of using surface electromyography of masticatory
muscles for detection of Temporomandibular Joint Disorder using machine leaning
classification algorithms.
The current paper aims to predict TMJ disorder. This paper is arranged in the
following sections: Sect. 1 gives introduction, Sect. 2 discuss materials and method-
ology. In Sect. 3, experimental results are described. Finally, the conclusions and
future scope of improvement are presented in Sect. 4.
Analysis and Prediction of Temporomandibular … 53
The first part in the research process conducted is the data assembly [3, 8]. Literature
survey is carried out by reading of articles and related journals from the internet
and clinical study was conducted to get details about TMJ disorder, its signs and
symptoms with radiographic evaluation through Orthopantomogram, Conversation
with Neurologist and dental surgeons were directed to alleviate the early findings
in selecting the subjects. A consent form is taken from all the subjects to willfully
participate in the study. Following Fig. 1 shows block diagram of proposed work.
The present study was conducted with total 84 subjects. Forty one subjects were
included in the control group and the rest were in experimental group. While
selecting the subjects, those with the history of dysfunction syndrome and orthodontic
management, lip incompetence, soreness in any muscle of mastication and previous
rebuilding in tooth were excluded.
A complete dental history is made for each subject and intraoral photographs
(OPG) are taken to identify the TMJ parameters [6]. Frontal belly of masseter and
frontward belly of temporalis of both sides (which are for convenience mentioned
here as masseter and temporalis, respectively) were studied.
For all the subjects, bipolar surface electrodes were used. The EMG recording
of these muscles was done in both groups at most calm position of jaw i.e., at
54 R. B. Kakkeri and D. S. Bormane
rest position and during maximum clenching at maximum intercuspal position [9–
11]. In both control group and experimental group, during the activity of maximum
clenching and during rest, the recording was done. Before recording, volunteers were
seated comfortably on a normal chair. Position of the head was kept erect and no
undertaking was allowed as jaw muscles respond to change in crown position. The
subject was settled and not able to see the computer screen. Facial areas were prepared
using a Medi-swab. EMG signals were composed using bipolar throwaway surface
electrodes of 6 mm diameter (Blue Sensor A-50-VS, Medico test, lstykke, Denmark)
placed on the volunteer’s temporalis and masseter muscle areas on both sides of
the face, thus giving four data collection channels (temporalis-right, masseter-right,
masseter-left, temporalis-left). Electrode cream (Grass EC2, USA) was functional
to give a flat surface on the electrode face and then the protective support detached.
Two surface electrodes were applied to the muscle, next to each other, the centers
16 mm apart laterally along each muscle, which was located by palpation. Reference
electrode was involved to an ear lobe.
The electrodes were associated via a Polygraph amplifier (Model 7E, Grass Instru-
ment Co., Massachusetts, USA) to an ADC interface comprising 12-bit, 333 kHz
and sampled using Spike2 data collection software (Cambridge Electronic Design
Ltd., Cambridge, UK), sampling at a rate of 1 kHz. The Polygraph 7P3 preamplifier
modules used filters (26 dB) of 3 Hz and 20 kHz, thence to 7DA driver amplifier
modules with output frequency response to 40 kHz (26 dB). The 7DA module had a
notch 50 Hz filter to suppress mains electrical interference. Recording was continued
in each case for up to 1 min.
Pre-processing of data is a method in which the raw data is changed into a substantial
and practical format. Real-world evidence is typically partial and unreliable since
it includes a lot of mistakes and negligible values [5, 9]. A pre-processed upright
details still vintages to a decent performance. Various types of data pre-processing
[8] are used to adjust inaccurate and conflicting data, such as managing missing
information, minimizing data, etc.
3 Methodology
The next move was to develop the algorithm for the diagnosis of TMJ disorder
with obvious results on how to diagnose TMJ disorder by feature selection [12]. A
series of Computer Vision outlines is referred to and the probability of using them
to create the algorithm was checked. In the development of the algorithm [6, 7],
python libraries were attempted. This system libraries and documents were analyzed
Analysis and Prediction of Temporomandibular … 55
on which of these may provide the simplest but most efficient method for obtaining
the outcomes.
The whole dataset has been split into two parts, i.e. one part is the training dataset
other part is the dataset testing with a ratio of 70:30 respectively. Training data has
once again been broken into two parts for cross-validation purposes. One component
is the preparation dataset, and another component is the 70:30 ratio evaluation dataset.
Fig. 2 shows training & testing.
SVM is an algorithm for supervised machine learning which can be used for problems
of classification or regression [13, 14]. Fig. 3 shows SVM To transform the results,
it utilizes a technique called the kernel trick, and then finds an optimal boundary
between the possible outputs based on these transformations. Simply put, it performs
some highly complex transformations of information, then works out how to isolate
the information depending on the labels or outputs you have identified. We used a
linear model of the kernel [12].
56 R. B. Kakkeri and D. S. Bormane
3.4 Adaboost
The result is measured in terms of specificity, sensitivity, and accuracy by using the
confusion matrix and classification report. The result depends on how accurate the
model is trained. The most relevant and important features selected were Masseter
and Temporal muscles for left and right during rest and Maximal voluntary clenching
(MVC).
Gauging performance is key to check how fine a classification model work to attain
an objective. Performance valuation metrics are used to evaluate the efficacy and
58 R. B. Kakkeri and D. S. Bormane
On all but one (k − 1) of the subclasses, train ML model, and then test the model
on the sub-set that was not used for training [14]. This procedure is repeated k times,
each time with a different sub-set intended for evaluation (and omitted from training).
4.3 Results
5 Conclusion
References
1. Uddin F, Baten RBA, Rita SN, Sadat SA, Chowdhury NM (2017) Management of temporo-
mandibular joint dysfunction syndrome: an overview. J Bangladesh Coll Physicians Surg
35(3):133–141
2. Klasser GD, Okeson JP (2006) Electromyography in the diagnosis and treatment of temporo-
mandibular disorders. Oral Health 137:763–771
3. Mapelli BC, Zanandréa Machado LD, Giglio CS, De Felício CM (2016) Reorganization of
muscle activity in patients with chronic temporomandibular disorders. Arch Oral Biol 72:164–
171
4. Bianchi J et al (2020) Osteoarthritis of the temporomandibular joint can be diagnosed earlier
using biomarkers and machine learning. Sci Rep 10(1):1–14
5. Spiewak C (2018) A comprehensive study on EMG feature extraction and classifiers. Open
Access J Biomed Eng Biosci 1(1):1–10
6. Khezri M, Jahed M (2007) Real-time intelligent pattern recognition algorithm for surface EMG
signals. Biomed Eng Online 6:1–12
7. Sweeney EM et al (2014) A comparison of supervised machine learning algorithms and feature
vectors for MS lesion segmentation using multimodal structural MRI. PLoS ONE 9:4
8. Chowdhury RH, Reaz MBI, Bin Mohd Ali MA, Bakar AAA, Chellappan K, Chang TG
(2013) Surface electromyography signal processing and classification techniques. Sensors
(Switzerland) 13(9):12431–12466
9. Latif R, Sanei S, Shave C, Carter E (2008) Classification of temporomandibular disorder
from electromyography signals via directed transfer function. In: 30th annual international
conference of the ieee engineering in medicine and biology society, vol 2, no 3, pp 2904–2907
10. Pinho JC, Caldas FM, Mora MJ, Santana-PenÍn U (2008) Electromyographic activity in patients
with temporomandibular disorders. J Oral Rehabil 27(11):985–990
11. Suvinen TI, Kemppainen P (2007) Review of clinical EMG studies related to muscle and
occlusal factors in healthy and TMD subjects
12. Khammas BM, Monemi A, Bassi JS, Ismail I, Nor SM, Marsono MN (2015) Feature selection
and machine learning classification for malware detection. J Teknol 77(1):243–250
13. Khan MMR, Arif RB, Siddique AB, Oishe MR (2019) Study and observation of the variation
of accuracies of KNN, SVM, LMNN, ENN algorithms on eleven different datasets from UCI
machine learning repository. In: 4th international conference on electrical engineering infor-
mation and communication technology iCEEiCT 2018, pp 124–129
Analysis and Prediction of Temporomandibular … 61
14. Kucuk H, Tepe C, Eminoglu I (2013) Classification of EMG signals by k-nearest neighbor
algorithm and support vector machine methods, pp 1–4
15. Al-Faiz MZ, Ali AA, Miry AH (2010) A k-nearest neighbor based algorithm for human arm
movements recognition using EMG signals, EPC-IQ01 2010. In: 1st conference energy, power
control, vol 6, no 2, pp 159–167
Machine Learning Approach
in Cooperative Spectrum Sensing
for Cognitive Radio Network: Survey
1 Introduction
Fig. 1 Classification of
spectrum sensing
techniques [1] in which the harmful interference is prevented with licensed users and
the available spectrum is identified to improve the spectrum’s utilization.
In cognitive radio network one of the important techniques is Spectrum sensing
and is also the primary task for establishing a cognitive radio system [1].
Spectrum Sensing is of three types. Interference-based sensing, non-cooperative
sensing and Cooperative sensing.
In cognitive radio networks, machine learning methods of Cooperative spec-
trum sensing is the major leading technique. The capability of decision making can
be enhanced by combining machine learning with cognitive radio networks from
the previous experiences [2]. Compared to traditional methods unstable cognitive
spectrum resources can be used effectively.
Figure 1 shows the classification of Spectrum sensing. Non-cooperative sensing,
interference-based sensing and Cooperative sensing Spectrum sensing are the
classification of spectrum sensing.
The performance of detection of spectrum holes is enhanced by cooperative
sensing in which secondary user senses the spectrum with the collaboration with
primary user. Also known as receiver detection technique.
In non-cooperative spectrum sensing, individual CR scans for primary signal and
based on its detection CR decides the presence or absence of primary user.
Cognitive radio measures the interference environment and their transmissions
are adjusted such that the interference to PU is not above the regularity limit in
interference-based sensing [3].
Cyclo-stationary feature detection, energy detection and match filter detection are
the classification of non-Cooperative sensing.
One of the intelligent cooperative sensing methods is machine learning which has
the special feature of self-adaptation to environment, high sensing.
Table 1 shows the comparative advantages and disadvantages of spectrum sensing
methods [4]. The author has created this table to familiarize spectrum sensing and for
better understanding the state-of-art techniques to the researchers. The performance
of any spectrum sensing algorithm is determined by the probability of detection,
probability of false alarm, the energy detection for cognitive radio network follows
the following formula [5, 13].
Machine Learning Approach in Cooperative Spectrum … 65
1. Let time of sensing be τs, at the sensing node, when the PU is absent, the signal
that is received is given by
where g(u) is gain of the channel, r(u) is primary signal which is received.
3. Assuming some of the parameters, detection of energy is given by
τ s
1
ED = |s(u)|2du (3)
P0
0
2 Literature Survey
Zan Li et al. [6] proposed CSS model of machine learning in which user grouping
concept is shown. CSS framework based on optimization model which had four
modules. Group scheduling module, SVM classification module, user grouping
module and SVM training module are the four modules of grouping concept. Without
the reduction in accuracy of sensing and the functions like grouping of optimized
users, redundant and abnormal are achieved by using three algorithms of grouping.
In the algorithm of first grouping there are groups of two which are normal and
abnormal. Then in second grouping algorithm redundant and non-redundant users
are distinguished and finally optimized model is established. As per the requirement
of cooperative sensing optimization of specific number of groups are done where
the users are divided, selecting each time. Finally, the user group are non-redundant
and secured that involve in the recognition of pattern. Optimization problem can be
solved by Binary particle swarm optimization. By using this grouping algorithms,
the CR networks can be avoided the harm of abnormal and the redundant users.
Assigning silent state for long term or temporarily for many users reduces the unnec-
essary energy consumption. The efficiency of sensing can be improved by increasing
the grouping. This is done by increasing the speed of classification in model of SVM
in machine learning. Centralized CSS model network consisting of fusion center is
used in the simulation with 10 or 20 CRs. Here the simulation result shows that there
is 17.98% rate of improvement for the speed of average classification. Along with the
entire network security they also showed the operational efficiency of SVM model
increased using group algorithm 1. Few non-highly redundant users and highly redun-
dant users are used in grouping algorithm 2 with network of higher signal quality.
Machine Learning Approach in Cooperative Spectrum … 67
SVM achievement 100% accuracy. In algorithm 3, the users of CR are divided into
cooperation groups of two or more which are optimized. Cooperative sensing can be
performed independently by each group. The classification speed increases with the
number of groups being more.
Yingqi Lu et al. [7] proposed machine learning based classification which uses the
probability vector of dimension being low. This method has the time for classifica-
tion shorter and duration for training smaller. So, energy vector with N-dimensions
is not used here. Keeping the same spectrum performance and by converting a high
dimensional feature vector, a feature vector of constant two-dimensional is achieved.
With M secondary users and primary user selecting single SVM and K-means algo-
rithms of machine learning are used. Small scale CRN uses primary user as single
and M = 2SUs and large-scale CRN using single primary user (PU) and M = 9SUs
are considered.
In Small scale CRN, the techniques of machine learning are trained and then
new feature vectors are tested, which gives the result of (PFA) probability of false
alarm and (PD) the probability of detection. They showed that The accuracy of
detection compared to algorithms of AND rule and OR rule which are hard fusion,
classification based on machine learning is better. In terms of energy vector, there is
better performance by SVM based classification using probability vector. With the
highest transmitting power of PU, performance is better.
In large scale CRN, through plot of Receiver operating characteristics curves
accuracy is shown better in both SVM-ploy and SVM-linear when the probability
vector is combined. Here energy vector is replaced by probability vector and proved
the accuracy to be better.
Through simulation they proved that
1. With the equal training samples, the duration of training for K-means clustering
is longest.
2. SVM-ploy takes more training time than SVM-linear as more time is spent on
feature vector mapping by functions of polynomial kernel.
3. SVM-linear is more efficient than SVM-ploy.
4. With the probability vector, the performance is best of SVM linear compared
to various decision-making scheme such as K-means, SVM-ploy. These deci-
sions making schemes have duration of training and classification delay low,
probability of detection high.
Xiaolin Ma et al. [8] proposed Extreme Learning machine (ELM) for numerous
primary users to implement the algorithm of sensing cooperative spectrum with the
network of cognitive radio. ELM which is an algorithm of machine learning that
has three step free tuning process is used to make very accurate the channel sensing
and identification which involves. In this paper initially the identification of multiple
primary users is established by spectrum sensing model. A classification scheme
of channel pattern can be obtained by ELM with the help of energy detection and
channel model. Channel model as well as sensing model, energy model, fusion center
is also implemented.
68 V. S. Kulkarni et al.
In channel model, the primary user which uses the channel detects the signal
consists of an authorized signal and noise which is called Rayleigh channel. Using
multiple primary users, the signals can be detected by secondary user in the network
if the number of primary users is more than one. In energy model, usage of channel
by primary users can be sensed by energy. The fusion center aims at determination
of usage of channel. Based on the transmission of energy vectors from secondary
user channel usage is determined.
Simulation is carried out with multiple primary users and the results are compared
with SVM, it was shown that training time of ELM is shorter. The detection
probability being higher the probability of overall detection will also be better.
SVM and ELM algorithms were proposed to train the samples with 1000 and 2000
as the sample size. It is proved that ELM needs only 92 ms to reach probability of
detection compared to SVM which requires 8 times more time than ELM. Accuracy
of SVM is also less than ELM.
Olusegun Peter Awe et al. [9] worked in cognitive radio network in order to
solve the problem of sensing the spectrum on the condition of multiple primary
users and proposed algorithm based on based multi-class support vector machine
(SVM) which is ECOC (error correcting output codes). Spectrum holes is detected
by Joint spatial–temporal detection which is implemented using this algorithm. To
study the attributes of each state, through one-versus-all (OVA) and one-versus-
one (OVO) Multi-class support vector algorithms are used [10–12]. Error correcting
output codes (ECOC) scheme is used to solve the optimization problem. The multi
class problem can be dealt with ECOC which is a framework that creates multiple
binary classification task by breaking it [9] and [10]. Initially the condition is consid-
ered as detection of multiple class signal in which more than one sub-class are
formed from each class. Then multi-class SVM algorithms checks for performance
of the energy-based features and ECOC (error correcting output codes). Detection
of performance is mainly based on accuracy of classification and curves of receiver
operating characteristics. They proved that unused spatial spectrum can be detected
by using ECOC.
Karaputugala Madushan Thilina et al. [1] proposed technique of machine learning
for cognitive radio in which both supervised and un-supervised methods are used.
weighted K-nearest-neighbor and support vector machine learning is used in super-
vised and gaussian mixture model, K-means clustering is used in unsupervised where
cooperative spectrum sensing is implemented. Initially by using the energy levels of
vector the availability is checked for channel which is considered as feature vector.
The decision is taken by feeding this to the classifier. A classifier categorizes into
two categories the availability of channel with available class and the unavailable
class for each feature vector. The training phase has to be covered by classifier before
the classification being online. Partitioning of training feature vectors into K clus-
ters is done by implementing the algorithm of K-means in which test energy vector
class and primary user state represented by each cluster is determined by classifier.
Mapping of each cluster determines the available and unavailable class of channel.
Training feature vectors are obtained from Gaussian mixture distributions in GMM
which corresponds to a cluster. The SVM and the KNN is also proposed due to the
Machine Learning Approach in Cooperative Spectrum … 69
capability of higher prediction. In the SVM, the margin between separating hyper-
planes and feature vectors are maximized to obtain support vectors. Performance is
determined by evaluating all the classification techniques such as the ROC curve, the
classification delay and training time.
The result shows that
1. Capability of classification type K-nearest neighbor for 1000 energy vectors is
shown by the time taken for uploading the training energy vector to the classifier
as 50 μsas compared to GMM having high training duration of 12,796 s for
1000 samples and SVM having highest training duration of 1. 65,817 s for 1000
samples
2. KNN classifier has high classification delay even though training time is lowest
compared to Fisher linear discriminant which shows lowest classification delay.
Also, no change is observed in classification delay in GMM, K-means clustering
and Fisher linear discriminant classifiers with the training energy vector with
different batch.
3. K-means clustering is better approach compared to another classifier because of
capability PU detection being higher and classification delay and lower training.
4. Compared to other algorithms performance of detection is found highest in
SVM classifier by using Kernel functions like polynomial kernel and linear
kernel which is used to map feature space with higher dimensional space.
They also concluded that the classifiers can be trained by obtaining the energy
vectors one to one by improving the CSS approaches. Comparative Table Method-
ologies.
Table 2 shows the comparative methodologies used by various authors and their
findings which led to a better result.
3 Conclusion
References
1. Thilina KM, Choi KW, Saquib N, Hossain E (2013) Machine learning techniques for
cooperative spectrum sensing in cognitive radio networks. IEEE J Sel Areas Commun
31(11):2209-2221
2. Kim KJ, Kwak KS, Choi BD (2013) Performance analysis of opportunistic spectrum access
protocol for multi-channel cognitive radio networks. J Commun Netw 15(1):77–86
3. Arjoune Y, Kaabouch N (2019) Review-a comprehensive survey on spectrum sensing in cogni-
tive radio networks. In: Recent advances, new challenges, and future research directions,
sensors
4. Bkassiny M, Li Y, Jayaweera SK (2013) A survey on machine-learning techniques in cognitive
radios. IEEE Commun Surv Tutor 15(3):1136–1159 (Third Quarter)
5. Bae S, So J, Kim H (2017) Article on optimal cooperative sensing with energy detection in
cognitive radio
6. Li1 Z, Wu W, Liu X, Qi P (2018) Improved cooperative spectrum sensing model based on
machine learning for cognitive radio networks. J Inst Eng Technol 2485–2492. ISSN 1751–8628
Machine Learning Approach in Cooperative Spectrum … 71
7. Lu Y, Zhu P, Wang D, Fattouche M (2016) Machine learning techniques with probability vector
for cooperative spectrum sensing in cognitive radio networks. In: IEEE wireless conference
and networking conference track 1: PHY and fundamentals, pp 1–6 (WCNC)
8. Ma X, Ning S, Liu X, Kuang H, Hong Y (2018) Cooperative spectrum sensing using extreme
learning machine for cognitive radio networks with multiple primary users. In: 2018 IEEE 3rd
advanced information technology, electronics and automation control conference, pp 536–540
9. Awe OP, Lambotharan S (2015) Cooperative spectrum sensing in cognitive radio networks
using multi-class support vector machine algorithms. In: IEEE 2015, pp 1-7
10. Arthy A, Periyasamy P (2015) Review on spectrum sensing techniques in cognitive radio
network. In: Proceedings of UGC ponsored national conference on advanced networking and
applications, pp 80–83
11. Hsu C, Lin C (2002) A comparison of methods for multiclass support vector machines. Neural
Netw IEEE Trans 13(2):415–425
12. Allwein E, Schapire R, Singer Y (2001) Reducing multiclass to binary: a unifying approach
for margin classifiers. J Mach Learn 1:113–141
13. Yucek T, Arslan H (2009) IEEE communications surveys & tutorials
Study on Detection of Leukemia
in Human Blood Pattern Based
on Microscopic Photographs Using
Image Processing Techniques
Abstract For the time being, blood disorders are defined by examination of micro-
scopic images of blood cells. It may lead to identity the blood disorders class of
blood-associated diseases. This paper discusses an initial look at the creation of
Microscopic Blood Pattern Images to detect leukemia forms. Pixel reading can be
very critical since it is possible to detect and classify pixel diseases at an earlier stage.
From there, more steps, such as disease control, surveillance and prevention, can be
completed. Photos are used because they can be fairly priced and do not require
costly check-out and laboratory equipment. White blood cell deficiency, leukemia,
will be identified by the gadget. The computer can use microscopic image capabilities
to look at texture, geometry, colour and statistical evaluation changes. Changes in
these characteristics may be used as feedback to a classifier. A literature review has
been undertaken and it is suggested to categories types of leukemia by reinforcement
learning to know. A brief debate about the problems involved has also been prepared
by researchers.
1 Introduction
Over the past decade, scientific imaging has proved to be one of the most vital
visualisation and analysis techniques in biology and medicine. A remarkable creation
of new, powerful devices for the identification, storage, transmission, reading and
display of scientific images has occurred at this period. This has contributed to a
substantial increase in the utility of virtual photo processing strategies [1] for clinical
problem solving. In advisement of System used for medical field, the toughest part of
medical imaging lies. A close interdisciplinary partnership between physicians and
engineers involves the design, implementation and evaluation of complex clinical
frameworks. Reading via pix is most importantly aimed at accumulating information,
detecting diseases, analyzing diseases, manipulating and remedying, monitoring and
evaluating [2]. Meanwhile, blood disorders are classified by visual examination of
microscopic blood cell snap shots. It could contribute to a group of safe blood-
associated diseases from the detection of blood disorders. Most cancers are one of
the most feared via The Human Disease. Leukemia is a form of blood cancer, and it
can result in death if it is discovered past due. Leukemia takes place when the bone
marrow contains a number of strange white blood cells. If there are lots of exceptional
white blood cells, the blood system’s equilibrium can be interrupted. When the blood
sample is obtained and analyzed by haematologists, the life of ordinary blood may
be identified. Hematologists must visually examine the microscopic pix, and the
procedure is time consuming and tiring [3–5].
The method needs human professional and sensitive to errors due to emotion
Disruption and Course’s human physical capacity have its own limit. In addition, it is
difficult to obtain a steady impact from visible inspection [3]. For additional research,
visual inspection may also have the simplest qualitative effects [3]. Research shows
that all data is used for maximum of recent techniques Approximately blood for e.G.
Red blood cell variety, degree of haemoglobin, stage of hematocrit, mean corpuscle
extent and much greater as the parameter for classifying diseases like thalassaemia,
most cancers, and many others. The early and fast detection of the type of leukemia
greatly facilitates the presentation of the correct treatment for the precise form of
leukemia [6]. The diagnostic strategies currently used depend on the reading of
Immunephenotyping, Fluorescence In Situ Hybridization (FISH), DanCytochem-
istry Cytogenetic Assessment [6, 7]. State-of-the-art and high-priced laboratories are
required on the way to run the diagnostic techniques and a high ratio of misidentifica-
tion was mentioned; as suggested in [8]: “The first-rate laboratories rely on as many
as a dozen distinct, hard-work-intensive technologies, all of which require relatively
educated professionals.” Having said that in relation to the subtype, maybe 50% of
sufferers are misdiagnosed [6]. With this unit, higher images can be processed, study
time reduced, the impact of subjective elements removed and the accuracy of the
identity system at the same time increased [9]. The leukaemia category and inspec-
tion will be focused on the texture, form, length, coloration and white blood cell
statistical analysis. This research is hoped to help boom productivity globally and
can benefit and be a huge contribution in the field of medical and trend popularity at
the same time. The primary aim is to decorate algorithms that can extract information
from human blood Where human blood is the primary source for detecting diseases
at an earlier stage and can save it quickly [10]. This system should be powerful for
the diversity that exists among individuals, protocols for pattern selection, time and
many others [10]. In this article, in classifying leukaemia types, we will advise on
the use of encouragement to learn (RL). As in [11], a variety of medical picture
problems can be solved by Reinforcement Mastering. Among the objects concerned,
Study on Detection of Leukemia in Human Blood … 75
clinical photographs have a very similar degree of grey and texture. Additionally,
segmentation blunders can appear and boom. Any other issue could be the lack of
a sufficient number of samples of education if a Supervised Mastering Method is
hired. Background. Blood is the main source of records that provide an indicator of
fitness changes and specific disease improvements. Adjustments to the amount or
presence of materials that have been produced can direct a person’s health situation.
1.1 Leukemia
Most blood cells are formed from cells called stem cells within the bone marrow. A
soft material contained in the center of any bone can be the bone marrow. Stem cells
are going to expand and become some pretty blood cells. Each blood group has a
role of their own. Blood components are made up of:
(a) Red blood cells (erythrocytes)—It humps oxygen to the tissues and carbon
dioxide back to the lungs.
(b) White blood cells (leukocytes)—To Contaminate-Safeguard the organism.
Several forms of white blood cells are available.
(c) Platelets—helps control bleeding via blood coagulation.
(d) Plasma—The dissolved ion-containing fluid in the blood needed for cell
activity, consisting of sodium, potassium, chloride, hydrogen, magnesium and
iron. The cells will die when the blood cells are old or damaged and born cells
will replace them [12].
Figure 1 demonstrates how damaged cell got replaced with new cells and grown
into multiple blood components. They grow into either somatic myeloid cells or
somatic lymphoid cells. Myeloid stem cells have now matured and become myeloid
blast cells. This explosion would produce a red blood corpuscle, a platelet, and a
number of other white blood corpuscles. Lymphoid stem cells may also develop and
can form a lymphoid blast, and many kinds of white blood cells may ultimately form
this blast. The lymphoid blast differs from the white blood cells in the myeloid blast.
Since the disease is dangerous and may cause death, the research would specialize in
leukaemia. Bone marrow produces irregular white blood corpuscles for someone who
has leukaemia. Abnormal white blood cells will not die until they have to, contrary
to normal cells. Therefore, there are several irregular white blood corpuscles that
conflict with normal white blood cells in order to perform their functions. It constructs
a blood system imbalance within the human body. Leukemia is also clustered to
promote the rapid progression and severity of this disease. Leukemia is either acute
or chronic.
(a) Leukemia-Leukemic cells may perform tasks like normal white blood cells at
an earlier level. They will eventually develop into serious leukaemia.
(b) Leukemia—Leukaemia cells are unable to perform functions such as normal
white blood cells. Leukemia cell count can develop rapidly and become serious
76 S. M. Karve et al.
within a short period of time. Typically, leukaemia is also classified into four
types [8].
(c) Acute leukemia (ALL)—In children aged 2–10 years, acute leukaemia (ALL)
typically occurs. This is the most common form of leukaemia. In adults, they
still appear.
(d) Figure 2 shows acute chronic myelocytic leukaemia (AML) is common in
children under 1 year of age with this form of leukaemia. It’s very unusual for
adolescents. Yet it’s mostly in Adults who are 40 years of age.
Some study has been wiped out to automate the process of detection of blood
corpuscles and the patient can be correctly diagnosed next. A lot of them are [13]
to establish a method for the identification and classification of plasmodium by
microscopic images of blood cells. They use the morphology approach, so the key
criteria for improving this technique are the best methods for segmenting images
of blood corpuscles [14]. Identify patients with thalassemia using genetic program-
ming and neural networks. Instead of microscopic pictures, matured blood corpuscle,
platelet and reticulocyte data is used for eg cell percentage, hemoglobin level, hema-
tocrit level, mean volume corpuscle, distribution of hemoglobin distance, etc. to
spot patients with thalassemia, thalassemia characteristics, and usual. The multi-
layer perceptron (MLP) with 2 hidden layers is the result obtained. Training data
outcomes are 99.27% maximum accuracy, 98.16% mean accuracy, and 0.64% vari-
ance. Although the results of data testing are 88.72% overall accuracy, 84.44% mean
accuracy and 2.41% variation. While [4] a method called Leuko is being developed
and textural information is used to increase differences between leukocytes. They
used textural parameters, which are energy, inertia, homogeneity and correlation,
based on grey level occurrence matrices (GLCM). In designing the Leuko method,
Due to tedious process of features selection data reductioncan be achieved in the
view of learning the classification algorithms quickly and accurately. Classifier can
help standardize from available data, it is easier to interpret results as well as reduce
the time [15]. A device was then developed to identify leukemia cells by using
photographs of the bone marrow. Using the Support Vector Machine (SVM) clas-
sifier, the framework was designed to exploit features related to texture, geometry
and statistical analysis in blood cell images. In order to get the best recognition, they
stress on producing and selecting characteristics. They only pick 30 best features
and this generates a 13.07% training data error and 18.71% testing data errors. In
medical imaging, there are many applications that use reinforcement learning [11]
and the In their work, Kubota et al. [16] have used reinforcement learning (RL). They
use RL in order to Beat the medical photos with some issues. Among the objects
concerned, medical images have a very similar degree of grey and texture. Errors
in segmentation can occur and increase. If a supervised learning technique is used,
another concern may also be the lack of a sufficient number of coaching samples. A
minimum training dataset is required using the RL method [11]. Using Q-learning to
segment images from computed tomography (CT). They use images of cranial CT.
They noticed that they were ready to segment a picture into certain distinct regions
at the same time. The images are split into multiple sub images. An intervention
for sub images to differ and change the Qmatrix is selected by the Reinforcement
Learning (RL) agent. There is an evaluator who compares the outcomes and gives a
gift to the RL agent. The precision of segmentation that they achieve is above 95%
[16]. Apply the RL method to segmental prostate ultrasound images. They still use
Q-learning and the findings indicate a high potential for reinforcement learning to
be implemented in medical segmentation. By using an RL agent, their technique is
78 S. M. Karve et al.
to control the local threshold and hence the post-processing parameter [17]. Extract
the area of the kidney as a preprocessor for the diagnosis of renal disease. They use
X-ray CT abdominal images. Inside the rough kidney area, Q-learning is used and
the edge of the kidney contour is detected. However, for an actual contour, there are
a few error margins and it is corrected by the snake process. The chance of success
is a sort of that’s a poor 53%, another application for RL is used by [18].
2 Research Methodology
Research methodology and steps which will be utilized in this research includes
(Fig. 3):
1. Image Acquisition: With efficient magnification, blood images from slides will
be collected from the nearby hospital.
2. Preprocessing: The photos will be acquired during image processing and unnec-
essary staining during you’ll be distracted by noise. As our ROI would be white
blood cells, the history is going to be removed. Image enhancement will be
performed during this preprocess because the contrast enhancement technique
is capable of enhancing the quality of the medical image [19].
3. Segmentation: White blood corpuscle (WBC) segmentation and determination
of the nucleus ROI for WBC only. This is because the cytoplasm is sparse in
leukaemia cell images [7]. So the emphasis will only be on the nucleus of the
WBC. Determination of the WBC forms from the nucleus should be completed.
Only lymphocytes and myelocytes should be considered, and whether or not
they are blast cells or not should be determined. Others should be omitted, such
as neutrophils, basophils and eosinophils.
3 Classifications
aim to expand the general amount of incentive values should also be selected. The
agent would use a strategy that we have called as a policy of action to decide on
an environmental action. Otherwise the trend would be reduced by the mechanism
generating this behavior [22]. Q-learning can be a soft strategy that suggests that its
Q-values, regardless of discovery, approximate the optimum Q-values that are Q(s,a).
There will be Q-values contained in the Q-matrix. Q(s,a) is that r, obtained by taking
action, a from the economy, s, the estimated amount of potential payoffs. The one
with the best Q value is going to be the optimum action. Q(s,a) will be revised to
support the experience as follows:
Q(s,a) = (1 − α)Q(s, a) + α r + γmaxQ(s , a )
where an is the rate of learning and 0 < γ < 1 is the discount factor. As a conclusion,
as in Fig. 5, the research methodology can be viewed.
4 Discussion
Some blood cell concerns need to be addressed in order to solve them. One of the
issues is the matter of the blood corpuscle itself. Capell [6] say that their method
fails to classify some of the blood cells in the process of classification. Thanks to
environmental strain, a number of cells are also deformed into an arbitrary shape [3].
The Notes on their algorithms that do not distinguish cells that overlap. Overlapping
cells may also be joined by disease-induced cells [23].
He argued that identification between two neighboring cells in their development
line is the most difficult problem since the cells are very close and thus the boundary
point between two neighbors is not well defined. But overlapping problems are also
solved using the watershed approaches used in their studies [6]. Another challenge
is information collection. Blood sample images should be adequate to ensure that
generalization properties are always expressed and that unseen data can be identified
correctly [6]. Lacking samples means that the info [5] can only be represented by
a few key components [24]. Believes that separate data sets are the perfect route
should be used for each point. However, cross validation or bootstrap sampling can
also be used as it is difficult to collect a sizeable amount of samples. It is hoped that
the approach of RL in the classification phase would reduce the issue of inadequate
data. Once we are to build the method, all the problems posed by scientists have to
be taken into account. By applying effective strategies, we should always strive to
resolve them.
5 Conclusion
This study includes using microscopic blood sample images to classify the forms of
leukaemia. By using features in microscopic images, the device will be constructed by
analysing changes in texture, geometry, colours and statistical analysis as a classifier
input. The system should be effective, accurate, less time interval, smaller error, and
high precision, cheaper cost and robust for individual varieties, sample collection
protocols, time and so on. Knowledge derived from microscopic blood sample images
will support individuals by rapidly predicting, resolving and treating blood disorders
for a particular patient.
References
1. Valencio CR, Tronco MN, ACB Domingos, Domingos CRB (2004) Knowledge extraction
using visualization of hemoglobin parameters to identify thalassemia. In: Proceedings of the
17th IEE symposium on computer based medical systems, pp 1–6
2. Adollah R, Mashor MY, Nasir NFM, Rosline H, Mahsin H, Adilah H (2008) Blood cell image
segmentation: a review. In: Proceedings of biomedical, vol 21, pp 141–144
82 S. M. Karve et al.
3. Ritter N, Cooper J Segmentation and border identification of cells in images of peripheral blood
smear slides. In: 30th Australasian computer science conference, conference in research and
practice in information technology, vol 62, pp 161–169
4. Sabino DMU, Costa LDF, Rizzatti EG, Zago MA (2004) A texture approach to leukocyte
recognition. Real Time Imaging 10:205–206
5. Colunga MC, Siordia OS, Maybank SJ (2009) Leukocyte recognition using EM algorithm.
MICAI, LNAI 5845. Springer Verlag, Berlin Heidelberg, pp 545–555
6. Capell K (2012) Meeting leukemia’s diagnostic challenge. Business Week, The Mc
Graw Hill Companies, London. https://www.businessweek.com/magazine/content/05_36/b39
49007_mz001.htm
7. Halim NHA, Mashor MY, Hassan R (2011) Automatic blasts counting for acute leukemia based
on blood samples. Int J Res Rev Comp Sci 2(4):971–976
8. Bergen T, Steckhan D, Wittenberg T, Zerfab T (2008) Segmentation of leukocytes and erythro-
cytes in blood smear images. In: 30th annual international IEEE EMBS conference, Vancouver,
Canada, pp 3075–3078
9. Shitong W, Min W (2006) A new detection algorithm (NDA) based on fuzzy cellular neural
networks for white blood cell detection. IEEE Trans Inf Technol Biomed 10(1):5–10
10. Shin H, Markey MK (2006) A machine learning perspective on the development of clinical
decision support system utilizing mass spectra of blood samples. J Biomed Inform 39:227–248
11. Chitsaz M, Woo CS (2011) Software agent with reinforcement learning approach for medical
image segmentation. J Comput Sci Technol 26(2):247–255
12. Bergen T, Steckhan D, Wittenberg T, Zerfab T (2008) Segmentation of leukocytes and erythro-
cytes in Blood Smear Images. In: 30th annual international IEEE EMBS conference, Vancouver,
Canada, 20–24 August, pp 3075–3078
13. May P, Ehrlich HC, Steinke T (2006) ZIB structure prediction pipeline: composing a complex
biological workflow through web services. In: Nagel WE, Walter WV, Lehner W (eds) Euro-Par
2006. LNCS, vol 4128. Springer, Heidelberg, pp 1148–1158
14. Foster I, Kesselman C (1999) The grid: blueprint for a new computing infrastructure. Morgan
Kaufmann, San Francisco
15. Piurri V, Scotti F (2004) Morphological classification of blood leukocytes by microscope
images. In: IEEE international conference on computational intelligence for measurement
systems and applications, Boston, MA, USA, pp 103–108
16. Kubota Y, Mitsukura Y, Fukumi M, Akamatsu N, Yasutomo M (2005) Automatic extraction
system of a kidney region based on the Q-learning. Springer-Verlag, Berlin Heidelberg, pp
1261–1267
17. Markiewicz T, Osowski S, Marianska B, Moszczynski L (2005) Automatic recognition of
the blood cells of myelogenous leukemia using SVM. In: Proceedings of international joint
conference on neural networks, Montreal, Canada, pp 2496–2501
18. Sadeghian F, Seman Z, Ramli AR, Kahar BHA, Saripan M (2009) A framework for white blood
cell segmentation in microscopic blood images using digital image processing. Biol Proced
Online 11(1):196–206
19. Sahba F, Tizhoosh HR, Salama MMA (2006) A reinforcement learning framework for medical
image segmentation. In: International Joint Conference on Neural Networks, Vancouver,
Canada, 16–21 July, pp 511–517
20. Qiang W, Zhongli Z (2011) Reinforcement learning, algorithms and its application. In: Inter-
national conference on mechatronic science, electric engineering and computer, Jilin, China,
pp 1143–1146
21. Cseke I (1992) A fast segmentation scheme for white blood cell images. IEEE, pp 530–533
22. Birndorf NI, Pentecost JO, Coakley JR, Spackman KA (1996) An expert system to diagnose
anemia and report results directly on hematology forms. Comput Biomed Res 29:16–26
23. Using visualization of hemoglobin parameters to identify thalassemia (1981) Proceedings of
the 17th IEE symposium on computer based medical systems, pp 1–6. M.S.: 2004 identification
of common molecular subsequences. J Mol Biol 147:195–197
Study on Detection of Leukemia in Human Blood … 83
24. Srinivisan KS, Lakshmi D, Ranganathan H, Gunasekaran N (2006) Non invasive estimation
of hemoglobin in blood using color analysis. In: 1st international conference on industrial and
information system, ICIIS 2006, Sri Lanka, 8–11 August, pp 547–549
Brain Tumor Detection Using Deep
Neural Network
Abstract Brain tumor identification is an essential task for assessing the tumors
and its classification based on the size of tumor. There are various types of imaging
strategies such as X-rays, MRI, CT-scan used to recognize brain tumors. Computed
Tomography (CT) scan images are used for in this work for Brain tumor Image Iden-
tification. CT-scan images are used, because as it gives size, shape and blood vessels
detailing and is non-invasive technique. CT-scan is commonly utilized because of the
superior quality of image. Deep learning (DL) is the most recent technology which
gives higher efficiency results in recognition, classification. In this paper, the model
is developed by using Convolution neural network to detect the tumor of brain image
from a dataset from Kaggle. The dataset contains near about 1000 images. Tumor is
identified by image processing algorithm using CNN, time complexity is 90 m sec,
and the accuracy of the present system is 97.87%.
1 Introduction
Brain Tumor can be characterized as an artificial and abnormal increase in brain cells.
The human skull is inflexible, and its size is restricted in this manner any undesired
development in the tumor may influence on human brain activity. Besides, it gets
increased into other body organs and impact human body functionality [1, 2]. The
World Health Organization (WHO), says that the brain malignant growth influences
under 2% of the human population, which causes serious complications and intrica-
cies [3]. A brain tumor can be gathered from multiple points of view, primary and
secondary tumor. The previous findings show 70% of all brain tumor are primary,
while secondary (optional) tumor are the residuals 30%. This categorization of the
primary tumor first originates in the brain at the early stage. On the contrary, tumors
that first emerge in some other pieces of the body and afterwards moved to the cere-
brum are called optional tumors, and most of them are dangerous [4]. A brain tumor
is an aggregation of irregular cells in the brain. Any development inside the skull,
such a restricted space can cause issues. Brain tumors can be harmful (malignant) or
non-harmful (benign). At the point when the harmful tumor is developed, it increases
the pressure inside the skull. This may cause brain damage, and life expectancy is
very low [2]. Various imaging strategies can be utilized to distinguish and arrange a
brain tumor. In imaging modalities, MRI is widely recognized non-invasive methods
[5–7]. CT scan shows a tumor shape, size and location, blood vessels that feed the
tumor. X-ray images show different types of body parts in different shades of black
and white; the X-ray images are widely used to checking fractures (broken bones)
[8].
At an early stage, tumor detection is done by using image processing provides a
fundamental understanding of the presence of tumor and its type which is advanta-
geous for the treatment strategy and diagnosis to save patients life [9]. Moreover, the
detection of a stage of tumor is a difficult task for radiologists and physicians. This
needs specialists to detect tumor by comparing tumor tissues and by areas, to make
Images clear for human vision, image processing filters are used. Thus, it is a require-
ment for Computer-Aided Diagnosis (CAD) framework to early diagnose brain tumor
in reasonably less time to reduce radiologist burden [2]. Machine Learning (ML) is the
science that allows software applications to predict more accurate outcome depend
on training without being specifically programmed [10]. ML algorithms play a vital
role in the medical imaging field as a part of artificial intelligence [11]. Supervised
and unsupervised learning are two types of ML. In supervised learning techniques,
the outcome is based on input variables it analyses the training data and produces
output from the labelled data, in unsupervised learning, the data is not labelled, and
the algorithm is applied on input data without any prior training of data. Artificial
Neural Network (ANN) is based on a collection of connected artificial neurons. It
is used to simulate input database internal weighting system [12], Support Vector
Machine (SVM) is used for characterization problem it uses the kernel to transform
input data based on these transformations it gives optimum boundary between the
possible outputs, and K-Nearest Neighbors (KNN) is the non-parametric approach
used for the classification. Classification is based on nearest neighbour rule [13]. In
comparison, the outcome of the unsupervised technique is dependent only on the
input variables [14] and Self-Organization Map (SOM) [15]. For segmentation of
tumor the image is to be converted into grayscale used this information for detection
of tumor. From the handcrafted features, meaningful information can be extracted
by the expert in the medical field and which requires a considerable time and may
cause an error while handling large database [16].
Feature learning and accuracy is achieved by increasing training samples instead
of machine learning and neural systems which are the main advantages of Machine
learning which promote extremely powerful and exact model [17, 18]. The approach
Brain Tumor Detection Using Deep Neural Network 87
to train and developed the classification of brain tumor using CNN architecture was
developed connected convolution layers and achieved an accuracy of 91.43% [19].
In this paper, the identification of brain tumor is obtained using CNN classifier
the paper is focused on pre-processing, segmentation and feature extraction. For
simulation Python programming language is used. It is used for general purpose,
and it is a high-level programming language, and it’s more efficient than any other
languages. Accuracy of the proposed method is of 97.87%.
2 Methodology
Figure 1 shows the block diagram for Brain tumor identification using Convolution
Neural Network (CNN). Convolution Neural Network (CNN) is used in Medical
image processing field [2]. A four-layer CNN is introduced for tumor detection, as
shown in Fig. 1. Following is the block diagram with an explanation.
2.1 Pre-processing
There is a lot of non-essential noise in the image, so we need to improve the quality of
the images in pre-processing ways. The main function of the processing is to improve
the single-volume ratio, improve image appearance, remove noise and undesirable
backgrounds, smoothing the interior parts and keeping the edges.
The convolution neural network contains inputs and output layers, as well as many
hidden layers. CNN’s hidden layers usually include a series of convolution layers
that convince duplication or other dot product. Four types of layers available on CNN
as represented in Fig. 2, layers are
(1) Input layer, (2) Convolution Layer, (3) Pooling Layer and (4) Fully Connected
layer.
Input Layer: The input of the network is a 2 Dimensional digital image. The
network has an input layer which takes the image as the input.
Convolution layer: These are the most important blocks used in convolution
neural networks. Convolution is a filter applied to an input that results in an activation
function. Neural Networks consists of a series of hidden layers that transform the
input image. Every layer is a collection of a set of neurons, where each layer is fully
connected to all neurons in the next layer.
Pooling layer: Pooling is the process of extracting the features from images output
of a convolution layer. Maximum pooling operation calculates the maximum value
in every patch of the feature map.
Fully connected layer: It forms the last few layers in a neural network. The input
to the fully connected layer is the output from the final Pooling layer. This process
determines the most accurate weights. The output Eq. 1 below,
Brain Tumor Detection Using Deep Neural Network 89
The primary CNN shown in Fig. 2. CNN advantage is it performs well, and it
gives better accuracy. It is covering local and global features. It also learns different
features from images. In algorithm-based image classification, we need to select the
features (local, global) and classifiers. Other learning algorithms or models can also
be used for image identifications and classification.
Convolution Neural Network is used in object recognition and biological image
segmentation which enhances the performance of deep learning methods. In tradi-
tional methods, features are to be extracted first and then these features are fed
into the training model for the desired result of classification. CNN’s automatically
understands complex characteristic features directly from the data given.
Thus, in this paper, the main focus is on CNN architecture design instead of image
processing for extracting features. CNN takes patch image from the entire image as
input and uses convolution filters which are flexible and trainable; also, it makes use
of local subsampling to extract features.
Figure 3 shows the four-layer CNN model gives us a remarkable result for the
detection of the tumor. We observe the performance in two ways based on the sepa-
ration of the dataset. First, we tried to partition data in 70:30 formats that means 70%
data goes to the training and remaining data goes for testing by using this splitting
ratio, the accuracy is 97.5% [20]. And secondly, 80% of the images are used for
training, and some of the images are given for testing where we get 97.87% accuracy
and 98.47% training accuracy. In line with these lines, the proposed model provides
a positive effect on a scale of 80:20. Table 1 shows the above details in the form of
a table.
Fig. 3 A convolution
network architecture
Table 2 Performance
Methodology Accuracy (%)
Comparison
Seeta et al. [20] 97.5
Hussain [17] 96.87
Proposed CNN model 98.87
3 Flowchart
4 Results
In this work, CNN is used, which provides image data classification. The reason why
CNN is best is CNN doesn’t require feature extraction because CNN layers extract
features from an image by itself. CT scan image data set, which is made available
from Kaggle open-source dataset portal.
The system captures the brain images and verifies the brain tumor is present or
not in that particular brain image. Then it will find a brain tumor is present or not.
Figure 5 shows the GUI system of the complete process, in that process first we
Train the model using CNN algorithm after completing the training process select
the image of brain tumor and then press the prediction button. Finally, the system
Display the result in that result display a message whether the tumor is detected or
not.
Figure 6 shows the system GUI. There are 3 button search buttons have a different
functionality and Image Process windows for image preprocessing technique. In that
GUI first train the model after completing training procedure, select brain tumor
image and after that click the prediction button then we get results on above image
Brain tumor is not detected.
Figure 7 shows the result of the proposed system. It shows the detected tumor,
and Fig. 8 shows the non-tumor image detection.
Train Here CNN Module (Fig. 9) will be training with an accuracy of 98.87%.
Y- Y- Y-
N- N- N-
5 Conclusion
The main aim of this research project is to design an effective system for identifying
brain tumors with high accuracy, performance and low complexity. This paper shows
tumor identification. The Kaggle database contains tumor and non-tumor CT scans
Brain Tumor Detection Using Deep Neural Network 93
6 Future Scope
(a) By using Web Application patient can upload their brain image and get the
result, whether the tumor is present or not.
(b) By developing Mobile Application it’s easy to use them for the evaluation for
all other Kind of diseases.
(c) In Feature, this technique can be developed to classify the tumors based on
Feature Extraction.
References
12. Louis DN, Perry A, Reifenberger G, von Deimling A, Figarella-Branger D, Cavenee WK,
Ohgaki H, Wiestler OD, Kleihues P, Ellison DW (2016) The 2016 world health organiza-
tion classification of tumors of the central nervous system: a summary. Acta Neuropathol
131(6):803–820
13. Bishop C (2006) Pattern recognition and machine learning. Springer-Verlag, Berlin, Germany
14. Machhale K, Nandpuru HB, Kapur V, Kosta L (2015) MRI brain cancer classification using
hybrid classifier (SVM-KNN). In: Proceedings of international conference on industrial
instrumentation and control (ICIC), pp 60–65
15. Shasidhar M, Raja VS, Kumar BV (2011) MRI brain image segmentation using modified fuzzy
C-means clustering algorithm. In: Proceedings of international conference on communication
systems and network technologies (CSNT), pp 473–478
16. Goswamiand S, Bhaiya LKP (2013) Brain tumour detection using unsupervised learning
based neural network. In Proceedings of international conference on communication
systems and network technologies (CSNT), pp 573–577
17. Hossain T, Shishir FS, Ashraf M, Al Nasim MDA, Shah FM (2019) Brain tumor detection using
convolutional neural network. In: 2019 1st international conference on advances in science,
engineering and robotics technology (ICASERT)
18. Siar M, Teshnehlab M (2019) Brain tumor detection using deep neural network and machine
learning algorithm. In: 2019 9th international conference on computer and knowledge engi-
neering (ICCKE), Mashhad, Iran, 2019, pp 363–368. https://doi.org/10.1109/iccke48569.2019.
8964846
19. Paul J (2017) Deep learning for brain tumor classification. In: 13 March 2017 SPIE 10137,
Medical imaging 2017: biomedical applications in molecular, structural, and functional
imaging, 1013710. doi: https://doi.org/10.1117/12.2254195
20. Seetha J, Selva kumar Raja S (2018) Brain tumor classification using convolutional neural
networks. Biomed Pharmacol J 11:1457–1461. https://doi.org/10.13005/bpj/1511
Design and Simulation of Different
Structures of Micro Strip Patch Antenna
for Wireless Applications
Abstract The need for multiband, bigger addition and low profile radio wires to help
numerous remote applications prompted the plan of Microstrip reception apparatuses.
Microstrip radio wires because of their little profile configuration take less zone. This
paper presents a straightforward rectangular Microstrip Patch Antenna, E-Shaped,
U-Shaped, +-Shaped radio wires work at 2.2 to 3.8 GHz. The Proposed reception
apparatus will be in lightweight, keen and conservative unit contrast and comprises
of metallic fix and ground between which is a dielectric medium called the substrate.
This various structures of MSA are utilized for military, remote and common appli-
cations. The CADFEKO programming is utilized to register the increase, power,
radiation example and S11 of receiving wire.
1 Introduction
Taking into account the advancement of the ongoing remote correspondence frame-
works and its application, more extensive data transfer capacity, multiband and low
profile radio wires are in incredible interest for both business and military applica-
tions. The quick increment of remote interchanges prompts an enormous interest in
the planning of a multiband radio wire. Expectedly, every reception apparatus works
at single or double recurrence groups, where distinctive receiving wire is utilized for
various applications. The plan of proposed receiving wires is utilized for the fast,
The term fractal was begotten by the French mathematician B.B. Mandelbrot during
the 1970s after his spearheading research on a few normally happening sporadic
and divided calculations not contained inside the domains of customary Euclidian
math [5]. The term has its underlying foundations in the Latin word fractus, which
is identified with the action word finger (which means: to break). These calculations
were commonly disposed of as amorphous; however, Mandelbrot found that specific
uncommon highlights can be related to them. He found a typical component in huge
numbers of these apparently sporadic calculations and figured speculations depen-
dent on his discoveries [6]. Two instances of normally happening fractal calculations
are snow-chips and limit of geographic mainlands. The fractal receiving wires are
not the same as customary radio wires since it is fit for working at various frequencies
all at once. The vital favourable circumstances of fractal radio wires are diminished
reception apparatus size, uphold multiband and wideband activity with improved
receiving wire execution. These can be accomplished utilizing fractal calculation
like Hilbert, Sierpinski, Koch and Minkowski are the different kinds of fractal calcu-
lations [7]. All these fractal calculations are utilized to plan a little size multiband
and wideband reception apparatuses.
Design and Simulation of Different Structures … 97
3 Methodology
The patch antenna has been designed for the following dimensions to achieve the
requirements of wireless communication applications.
Resonant frequency (fr ) = 2.4–4.8 GHz, Dielecrtric constant (ε) = 4.4.
The radio wire is planned at 2.4 GHz recurrence and built as a fix on the substrate.
Model number one in Fig. 1 shows a rectangular fix receiving wire configuration to
work near 3.8 GHz will be demonstrated. The model is first developed as a fix on
an infinitely huge substrate since it rushes to make and to recreate. Figure 2 shows
an E-formed fix radio wire plan fractal math is in gendered in an iterative style,
prompting self-comparative structure.
4 Simulation Results
5 Conclusion
It would have been a major bit of leeway to know the CADFEKO 14.0 reproduction
programming already as a lot of estimations might have been applied. Miniature strip
radio wires and their hypothesis get substantially more perplexing as you need to
make more proficiency. The various sorts of reception apparatuses are explored and
effectively reproduced in this paper. The reproduced reflection coefficient, impedance
and radiation design demonstrated well execution. Miniature strip reception appa-
ratuses have become a quickly developing territory of examination. Their poten-
tial applications are boundless, on account of their lightweight, conservative size,
and simplicity of assembling. E-formed micro strip reception apparatus is entirely
100 Anil. J. Kokare et al.
References
1. Ranjan A, Singh M, Sharma MK, Singh N (2015) Analysis and simulation of fractal antenna
for mobile Wi-max. IEEE Trans Future Gener Commun Netw 7(2):57–64
2. Jannani A (2011) Design of E-shape fractal simple multiband patch antenna for S-band LTE
and various mobile standards. IEEE Trans Antennas Propag 8:126119–126126. https://doi.org/
10.1109/access.2020.3006831
102 Anil. J. Kokare et al.
3. Permana CG, Munir A (2011) Printed multiband antenna for mobile and wireless communica-
tions. IEEE Trans Antennas Propag 236–240
4. Sreelaxmi K & Mithun Megha Sarkar TP (2014) Multiband miniaturized fractal antenna for
mobile communications. Int J Res Eng Technol (IMPACT: IJRET) ISSN(E): 2321–8843;
ISSN(P): 2347–4599 2(4)
5. Khidre et al (2012) Presented U slot micro-strip antenna for higher mode applications. IEEE
Access Digit Object Identifier 8:112840–112845. https://doi.org/10.1109/access.2020.3002789
6. Dhillon SS, Marwaha A, Nagpal A (2013) Multiband E-shaped fractal microstrip patch
antenna with DGS for wireless applications. In: Proceedings of international conference on
computational intelligence and communication networks, Mathura, pp 22–26
7. Lotfi P, Azarmanesh M, Soltani S, Bayatmaku N (2016) Design of simple multiband patch
antenna for mobile communication applications using new E-shape fractal. IEEE Antennas
Wirel Propag Lett 10
State Context and Hierarchical Trust
Management in WSN for Intrusion
Detection
R. B. Kagade (B)
Department of Computer Science and Engineering, SVERI’s College of Engineering, Pandharpur,
Maharashtra, India
PAH Solapur University, Solapur, Maharashtra, India
R. B. Kagade · J. Santhosh
Department of Computer Science and Engineering, Veltech Rangarajan Dr Sagunthala R & D
Institute of Science and Technology (Deemed to Be University), Avadi, Chennai, Tamilnadu, India
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 103
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_11
104 R. B. Kagade and J. Santhosh
1 Introduction
Wireless sensor network (WSN) is collection of sensor nodes that are equipped
with environmental sensors for heat, moisture, humidity, pressure, air/water quality,
weight, sound etc. WSN does not have fixed topology. WSN has wide range of
applications in field of agriculture, health, home, industrial, military and natural
calamities for monitoring and data collection purpose. The advantage of WSN is
that it is very easy to install in critical regions where normal network cannot be set.
Sensor nodes contain five components trans-receiver, processor, battery, hardware
and memory. The wide range of applications can be elaborated as follows:
Wild Habitat Monitoring
Sensors can be used to monitor the conditions of wild animals or plants in wild
habitats. Sensors can be used to collect information in the water and above water.
Disaster Monitoring
Information collection in natural and non-natural disaster areas is very important task.
A normal network cannot be set in such areas. WSN becomes prominent solutions
in such situations.
Warfield Monitoring
Sensors can be deployed in a war field to monitor the presence of forces and vehicles,
and track their movements, enabling close surveillance of opposing forces.
Sensitive Place Protection
Sensor nodes can be deployed around sensitive objects, for example, atomic
plants, strategic bridges, oil and gas pipelines, communication centers, and military
headquarters, for protection purpose.
Vehicle Monitoring
Sensors can be mounted on unmanned robotic vehicles, tanks, fighter planes,
submarines, missiles, or torpedoes to guide them around obstacles to their targets
and lead them to coordinate with one another to accomplish more effective attacks
or defenses.
Remote Sensing
Sensors can be deployed for remote sensing of nuclear, biological, and chemical
weapons, detection of potential terrorist attacks, and reconnaissance [1]. Due to its
vital range of applications it has several ranges of attackers at different levels. In
WSN sensor nodes have limited battery power, limited power to communicate and
State Context and Hierarchical Trust Management … 105
NIDS will monitor the traffic, malicious changes that are happening in the network
which causes changes that leads to system degradation.
HIDS may also be able to make out malicious business trade that starts with malware
and is attempting to put out on top for other systems.
Signature-based detection systems guide all the small data traversing the network
and make a comparison against a knowledge-base of signatures or given properties
of experienced violent behavior, like antivirus software.
Anomaly-based detection systems guide network business trade and make a compar-
ison against a started baseline, to come to a decision about what is taken into account
normal for the network with respect to bandwidth, protocols, harbors and other
devices [3, 4].
Trust is the degree of belief that a node can have on another node in the network
based on trust metric or trust rating o the node. Different types of attacks due to
intrusion are illustrated as follows [5].
GTMS: group based trust management scheme GTMS works in two types of
topologies intergroup topology and intra-group topology. In intergroup topology
106 R. B. Kagade and J. Santhosh
2 Related Work
clustering of WSN and deploying cluster based intrusion detection system. The mech-
anism works in static and heterogeneous network, hierarchical and clustering struc-
ture. Hierarchical cluster based and leach based routing protocol is used for efficient
communication. The mechanism is proved to be as robustness and fault tolerant
design. The research provides reliable service but is an expensive mechanism. The
intrusion detection architecture supports real-time detection property almost 80.6%.
The mechanism results 80.6% and 55.7% accuracy for content based detection and
context based detection capabilities respectively.
Daojing He proposed a distributed trust evaluation model for medical sensor
networks [5]. The traditional cryptographic methods are not sufficient for trust
evaluation in medical sensor networks. The research work uses transmission rate
and leaving time into trust evaluation to detect malicious nodes. They proposed an
application-independent and distributed trust evaluation model for MSNs. The trust
management is carried out through the use of simple cryptographic techniques [5].
A novel distributed trust evaluation model for MSNs, where each node manages
trust records of other nodes about performing some activities. Centralized malicious
node detection and secure unicast routing is presented. It proved to be as improved
packet delivery with effective malicious node identification mechanism. IDS mech-
anism depends on transmission time, relieving time and packet dropping ratio. 88%
accuracy is shown by researchers with limited number of malicious nodes in WSN.
Fenye Bao proposed hierarchical trust management for WSNs and applied it
to routing and intrusion detection to detect selfish or malicious nodes [12]. The
work focused on multidimensional trust attributes and the trust value was calculated
through social trust and QoS trust, including intimacy, honesty, energy, and unselfish-
ness; meanwhile, subjective trust and objective trust were taken into consideration
to validate the proposed protocol [1]. However, the node with the maximum number
of interactions with neighbors was considered as the most trustworthy in the process
of the calculation of the intimacy trust inspired by social networks. The difference is
the consideration of the reasonable range of the maximum number of interactions,
as interaction that exceeds the range indicates malicious behavior. The mechanism
is proved to have IDS detection capabilities with 90% accuracy with false positive
probability is zero and method is scalable. Trust based geographic routing is used in
WSN.
Xiaoyong Li put forward a lightweight and dependable trust system for clustered
WSNs, which aims at decreasing resource consumption and enhancing the relia-
bility of CHs’ trust evaluation [13]. A self-adaptive weighting mechanism is used
to calculate trust value of CH which is better than subjective weight method. A
series of theoretical proofs were given in the research to verify the effectiveness of
the mechanism. In the process of trust evaluation, only successful and unsuccessful
interactions were taken into consideration, with no other trust evaluation factors taken
into account. The mechanism takes interactive trust, honesty trust and content trust
into account, addressing problems of consuming energy maliciously and tampering
multidimensional observing data with lower resource overhead, which is described in
108 R. B. Kagade and J. Santhosh
the performance evaluation [1]. The research is focused on minimizing memory over-
head and transmission overhead. It works better to protect from garnished attack and
bad mouthing attack. Results show that intrusion detection capabilities are 87.5%.
Shaikh proposed GTMS; a group-based trust management scheme for clustered
WSNs. GTMS evaluates the trust of a group of nodes in contrast to traditional
trust schemes that always focus on the trust values of individual nodes [14]. In this
approach WSN requires small memory to store trust value at each node. The mech-
anism achieves significant reduction of the cost associated with the trust evaluation
of distant nodes. But it depends on a broadcast-based strategy to collect feedback
from the CMs of a cluster, which requires a significant amount of resources and
power. This mechanism worked for wired and wireless mechanism. It focused on
reducing cost of trust evaluation. Trust value is calculated depending on time based
past interaction. Timing window is used to measure the number of successful and
unsuccessful interactions. Trust evaluation cost is minimized by 14.6–15.7%.
Ismail Butun has presented intrusion detection system for mobile Ad Hoc
networks [15]. Agent based distributed and collaborative IDSs are emphasized in
research. Two types of classifiers are used for detection of intrusion; Decision Tree
and support vector machine. Dynamic Source Routing, Ad hoc On-demand Distance
Vector and Destination Sequenced Distance Vector protocol is used for routing data
in WSN.
Limitations of Classical Approaches
Signature-based detection approaches are relatively easy to implement, require no
learning curve. This eliminates the risk of over-training or voluntary deformation
of the profile that can be observed in behavior-based approaches. However, these
approaches require an active maintenance and very frequent updates of the signa-
ture database to integrate any new attack discovered. Indeed, the update cannot be
performed automatically as in the case of behavior-based detection. This fact implies
a higher rate of false negatives.
The problem arises especially with very recent attacks for which signatures have
not been included in the database yet. Also, the absence of a standard pattern descrip-
tion language limits the usefulness of signatures described in a given language since
interoperability between different detectors is probably not possible. If signature
values are too simplified it can lead to detection of false intrude which corresponds
to legitimate actions and therefore to trigger false positives.
Anomaly-based detection approaches have several interesting features. First, as
the hypotheses are made only on the normal behavior of the system and not on
possible attacks, detection is exhaustive. Indeed, “the system allows a prior to detect
all that differs” from established normal behavior. Thus, it becomes possible to
envisage detection of unknown attacks and no specific knowledge about the attacker
is required. All necessary information is collected within the system. On the other
hand, once the learning phase terminates, the IDS does not require particular update.
The definition of normal behavior evolve only slightly if any.
Nevertheless, a high rate of false positives is the main weakness of these
approaches because it is sometimes difficult to define the “normal behavior”. Sudden
State Context and Hierarchical Trust Management … 109
changes in the environment can have an impact on behavior. This sudden change in
behavior will be considered as an anomaly and an alert will be generated. Also,
since the first phase is dedicated solely to the development of the definition of
“normal behavior”, this one is particularly vulnerable to attack. Indeed, the pres-
ence of signals related to an attack in the learning trace will result in skewing the
definition of behavior. Thereafter, any similar attack will be treated as a normal
behavior. The information used during this first phase in the optimal condition must
be totally free from damage. In practice, it is frequently impossible to have such
perfect environment.
3 Proposed Methodologies
Cluster Creations
In wireless sensor network large numbers of application specific sensor nodes are
connected with each other on ad hoc purpose. If each node starts communication,
computation and routing then energy consumption, collision and congestion can
occur. Hence it may become reason for performance degradation. Node clustering
can solve these issues. In clustered network wireless sensor network is divided into
small units. Cluster head is elected. Sensor nodes in each cluster will communicate
with respective cluster head CH and CH aggregates data and will transfer data to
a central base station. The cluster head is considered to have maximum energy. To
maintain the maximum energy at cluster head re-election after certain interval of
time is best possible solution.
LEACH [16] is the first clustering scheme. It is used for periodical data gathering
in WSNs. It assumes the sensor nodes will remain static and communicate with
each other by single-hop only, and they can transmit data to the sink node directly.
110 R. B. Kagade and J. Santhosh
Its operation is divided into rounds and each round is composed of two phases
[17]. In cluster formation phase, LEACH elects some cluster heads according to the
probability shown below.
k
N −k(r mod N /k)
, Ci (t) = 1
Pi(t) =
0, Ci (t) = 0
where k is the desired number of cluster heads, Ci (t) is the indicator function deter-
mine whether or not node i has been a cluster head in most recent (r mod N/k) rounds.
Pi (t) is the probability for node i to become a cluster head. The rest of sensor nodes
join the proper cluster according to the signal strength from the cluster heads. In
the data transmission phase, the cluster heads aggregate the data from their cluster
members. Since cluster head is chosen by probability in each round, the load is
balanced to certain extent.
Trust Calculations
Interactive trust refers to the trust value computed by the number of interactions
between nodes, and an interaction means a node sending/receiving a packet or a
request to/from another node. The trust value is mapped to the integer number in
the range of [0, 10], where 0 demonstrates the most distrustful, while 10 implies the
most trusted, and 5 is the medium trust [1] shown in Fig. 1.
Sensor nodes trust is evaluated by the CH in a cluster, i.e., CH-to-SN trust,
which considers multidimensional trust, including interactive trust, honesty trust
and content trust, during the procedure of trust calculation.
Interactive trust SITij (t) is calculated by the number of interactions between
node j and its CH i in t. In the proposed method, interaction refers to all commu-
nication behavior including sending and receiving of request and data packets. The
greater the number of interactions of two nodes, the higher is the trust value [18].
However, in WSNs, if the number of interactions exceeds a threshold, the trust value
will decrease because there may exist malicious interactions such as attacks that send
a large amount of packets or requests to exhaust the energy of the node. Therefore,
unlike trust evaluation in social networks, the interactive trust evaluation method in
where x denotes the largest integer that is equal to or less than x, μ is the mean value
of the number of interactions between CH and SNs in the same state, λμ is taken as
the threshold of the interaction range, in which λ is a parameter used to define the
upper limit of normal interactions, and θ is a significant factor, which values 1, 10
and 100 when wij is a single digit, tens digit or hundreds digit, correspondingly, and
so on.
Example:
The WSN is assumed to contain eight nodes and one cluster head. Cluster head is
selected depending on computation power, energy, honesty and distance between
CH and base station (Fig. 2).
(c) For node H. The weight is 16; it is greater than threshold λμ which is 14. As
the number of is greater than 14 it is considered as distrustful. θ is 10 as the
wij is two digit number. The trust value is calculated as follows (Table 1)
When there are no interactions between active members, i.e., s = f = 0, the trust
value is 5. If there are unsuccessful interactions, the honesty trust value will decrease
sharply because of the punishment executed by [13]. For non active members, they
inherit the trust value of their lasno hibernating state.
Content Trust of SNs
Content trust is the trust evaluation based on observing data, which is data-oriented
trust calculated by CH. Content trust is introduced because the WSN is a data-
centric network and the observing data are the factor of most concern for applications.
Tampering attacks often occur in WSNs to interfere with the network and applications
and can be identified by content trust.
1
dm 2
Di j = (Xik − X jk)2̇
k=1
S O T i j = αS I T i j + β S H T i j + (1 − α − β)SC T i j
Parameters α, β ∈ [0, 1] are weights for each sub trust value. The higher the
weight, the more important that sub trusts is to overall trust and vice versa.
114 R. B. Kagade and J. Santhosh
Detection rate
It is defined as ratio of total number of attacks detected and total number of
attacks appeared [19].
P K T detect
W S N acc =
P K T detect + F P K T detect
FP
F Pdetect =
FP + T N
5 Conclusion
This paper addresses the security issues of wireless sensor network. A more promi-
nent intrusion detection system based on context level and trust level is introduced.
This mechanism achieves more than 90% accuracy in detection of routing attack and
sinkhole attack. The architecture suggested in this paper is used to develop two level
of trust model. Where first trust level is sensor node to sensor node communication
and second trust level is between sensor node and cluster head. The overhead of
cluster head and base station is minimized in this way. Architecture will be tested
on simulator ns2. Accuracy of 90% and 95% is expected in intrusion detection and
context text detection respectively (Table 2).
References
1. Zhang Z, Zhu H, Luo S, Xin Y, Liu X (2017) Intrusion detection based on state context and
hierarchical trust in wireless sensor networks. IEEE Trans Content Min
2. Benisha RB, Raja Ratna Dr S (2018) Prevention of cyber attacks in control systems: a review.
In: Proceedings of the 2nd international conference on trends in electronics and informatics
(ICOEI 2018)
3. Ntalampiras S (2014) Detection of integrity attacks in cyber-physical critical infrastructures
using ensemble modelling. In: IEEE transactions on industrial informatics, vol 11
4. Sadi Y, Ergen SC, Park P (2014) Minimum energy data transmission for wireless networked
control systems. IEEE Trans Wirel Commun 13(4):2163–2175
5. He D, Chen C (2012) A distributed trust evaluation model and its application scenarios for
medical sensor networks. In: IEEE transactions on information technology in biomedicine, vol
16, no 6
6. Boukerche A, Li X, El-Khatib K (2007) Trust-based security for wireless ad hoc and sensor
networks. Comput Commun 30:2413–2427
7. Maddar H, Kammoun W, Youssef H (2017) Trust intrusion detection system based on location
for wireless sensor network. Springer International Publishing AG
8. Liu K, Abu-Ghazaleh N, Kang K-D (2007) Location verification and trust management for
resilient geographic routing. Parallel Distrib Comput 67(2):215–228
9. He T, Stankovic JA, Lu C, Abdelzaher T (2003) SPEED: a stateless protocol for real-time
communication in sensor networks. In: Proceedings of the 23rd international conference
ondistributed computing systems (ICDCS’03), Washington, DC, USA, pp 46–55
10. Zhou M-Z, Zhang Y, Wang J, Zhao S-Y (2009) A reputation model based on behavior trust
in wireless sensor networks. In: Eighth IEEE international conference on scalable computing
and communications
11. Jadidoleslamy H (2011) A hierarchical intrusion detection architecturer for wireless sensor
networks. Int J Netw Secur Its Appl (IJNSA) 3(5)
12. Bao F, Chen I, Chang M, Cho J (2012) Hierarchical trust management for wireless sensor
networks and its applications to trust-based routing and intrusion detection. IEEE Trans Netw
Serv Manage 9(2):169–183
13. Li X, Zhou F, Du J (2013) LDTS: a lightweight and dependable trust system for clustered
wireless sensor networks. IEEE Trans Inf Forensics Secur 8(6):924–935
14. Shaikh RA, Jameel H, d’Auriol BJ, Lee H, Lee S, Song Y (2009) Group-based trust management
scheme for clustered wireless sensor networks. IEEE Trans Parallel Distrib Syst 20(11):1698–
1712
15. Butun I, Morgera SD, Sankar R (2014) A survey of intrusion detection systems in wireless
sensor networks. IEEE Commun Surv Tutor 16(1):234–241
16. Heinzelman WR, Chandrakasan A, Balakrishnan H (2002) An application-specific protocol
architecture for wireless microsensor netw orks. IEEE Trans Wirel Commun 1(4):660–670
17. Ye M, Li C, Chen G, Wu J (2005) An energy efficient clustering scheme in wireless
sensor networks. In: PCCC 2005, 24th IEEE international performance, computing, and
communications conference
18. Hao F, Min G, Lin M, Luo C, Yang L (2014) MobiFuzzyTrust: an efficient fuzzy trust inference
mechanism in mobile social networks. IEEE Trans Parallel Distrib Syst 25(11):2944–2955
19. Kalnoor G, Agarkhed J (2018) Intrusion threats and security solutions in wireless sensor
networks. Int Robot Autom J
Portable Camera Based Assistive Text
and Product Label Reading from Hand
Held Object by Using Android App
for Blind Person
Abstract We propose a camera-based mechanical man app. This app helps the blind
persons to browse the text on explicit objects. In this system the camera captures the
actual text on the object. Multiple techniques square measure applied to its text.
Such as Optical Character Recognition that supply the operation of scanning and
recognition of text and a few have integrated voice output. From a grayscale image,
thresholding are often accustomed produce binary pictures i.e. image with solely
black or white colors, Filtering are often accustomed cut back the noise of image, Next
image segmentation technique is employed to perform the method of partitioning a
digital image into multiple segments. The goal of segmentation is to modify and/or
amendment the illustration of a picture into one thing that’s a lot of significant and
easier to analyses. Image scaling is the method of resizing a digital image. Next
technique employed in this project is template matching. Temples matching is a way
in the digital image process for locating tiny components of a picture that match
a template image. Also template extraction are often employed in producing as a
vicinity of internal control, some way to navigate a mobile golem or as some way to
notice edges in images then finally voice output are going to be generated then blind
man will simply listen to the text on it explicit object.
1 Introduction
The 314 million visually impaired folks worldwide, forty five million square measure
blind handicap that was free by “World Health Organization” in ten facts concerning
sightlessness. Even during a developed country just like the US, the 2008 National
Health Interview Survey (NHIS) reportable that AN calculable 25.2 million adult
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 117
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_12
118 S. Thigale and R. B. Kagade
Americans (over 8%) square measure blind [1, 2]. The human baby generation ages
square measure quickly increasing. all over written text square measure obtainable
as an example receipts, bank statements, building menus, room handouts, report,
product packages, directions on medication bottles, room handouts etc. While screen
readers, optical aids and video magnifiers will facilitate blind users and people
with low vision to access documents. The few devices offer sensible access to
common hand-held objects like product packages and objects written with text as an
example prescription medication bottles. The visually handicapped person to scan
written labels and merchandise packages can enhance freelance living and foster
economic and social self-reliance. Today, there square measure already has some
systems for transportable use however they can’t handle product labeling. The trans-
portable Universal Product Code readers designed to assist blind people identify
totally different products in an in depth product info will change the user for visu-
ally handicapped person to access data regarding these products through speech and
Braille. during this system an enormous limitation is that it’s terribly exhausting for
blind users to seek out the position of the Universal Product Code and to properly
purpose the Universal Product Code reader at the Universal Product Code however
Some reading-assistive systems like pen scanners could be used in these and similar
things. This technique integrates OCR computer code used for scanning and recog-
nition of text and a few have integrated voice output. The OCR is optical character
recognition. These systems square measure typically designed and perform best with
document pictures with appropriate backgrounds, commonplace fonts, a tiny low
vary of font sizes, and well-organized characters. This technique is used instead of
business product boxes with multiple decorative patterns. The OCR computer code
cannot directly handle scene pictures with complicated backgrounds and therefore
the document to be scan should be nearly flat, placed on a transparent, dark surface
and contain principally text [3, 4].
Even though variety of reading assistants square measure designed specifically for
the blind folks, to our knowledge, there’s no existing scanning assistant will read text
from the varieties of difficult patterns and Even though variety of reading assistants
square measure designed specifically for the blind folks, to our knowledge, there’s
no existing scanning assistant will read text from the varieties of difficult patterns
and backgrounds that found on many everyday business product. Like text data will
obtainable in varied scales, fonts, colors, and orientations. Images. Initial is Milk
box (Top) and Men lavatory signage (Bottom).
Above Figs. 1 and 2 shows (a) camera captured images. (b) Localized text regions
that show in fig mark blue. (c) Text region scropped from image. (d) Text codes
recognized by OCR (optical character recognition). Text at the top-right corner of
bottom image is shown during a increased callout. Mobile accurately reads black print
text on a white background however there some issues to recognizing colored text
or text on a colored background. It cannot scan text with complicated backgrounds
as a result of they can’t simply detected the text from background. Text printed
on cylinders with crooked or incomplete pictures as an example as soup cans or
medication bottles. These systems require a blind user to manually localize areas
of interest and text regions on the objects. As shown in Fig. 1, such text data will
Portable Camera Based Assistive Text … 119
Fig. 1 Examples of printed text from hand-held objects with multiple colors backgrounds
appear in multiple scales, fonts, colors. to help blind persons to scan text from hand-
held objects, we have planned of a camera-based helpful text reading frame work
to track the item of interest at intervals the camera read and extract print text data
from the item. the prevailing rule will effectively handle complicated background and
multiple patterns and extract text data from each hand-held objects and assemblage as
shown in Fig. 2. The hand-held object extract camera image and we tend to develop
a motion-based technique to obtain region of interest (ROI) of the item. During
this ROI perform the text recognition technique. The localization of text regions
in scene pictures they’re divide in 2 classes: rule-based and learning–based. Rule
primarily based rule to used component level image process to extract the text data
120 S. Thigale and R. B. Kagade
Fig. 2 Two examples of text localization and recognition from camera captured Images. First is
Milk box (Top) and Men bathroom signage (Bottom)
from predefined text layout like character size, ratio, edge density, character structure,
color uniformity of text and learning primarily based rule square measure used model
text structure and extract representative text options to make text Classifiers [5, 6].
2 Existing System
In these days society, there square measure already a couple of systems that square
measure moveable use for blind persons. For instance, moveable code readers. The
blind those who want to access data regarding these product, the moveable code reader
helps the blind individuals to spot totally different product in an intensive product
information. However properly position the code reader at the code is incredibly
arduous task for blind users.
Con jointly there some system that uses camera-based helpful text reading frame-
work. Foremost object of interest inside the camera read is track so written text data
from the item is extracted by this framework. During this framework, the item of
interest is positioned at the middle of camera’s read. This object ought to be posi-
tioned specified, it ought to be seem within the camera read. These wide angles
of this camera accommodate the users with solely particular/approximate aim. The
system extract hand-held object from the camera image. To get the region of interest
(ROI) of the item this framework uses the motion primarily based technique. Then
text recognition is performed just for that region of interest (ROI) [7, 8].
Portable Camera Based Assistive Text … 121
Sometimes multiple scales, fonts and colors square measure won’t to write the
text characters. Conjointly immense quantity noise is contained within the captured
pictures that contain text instead of industrial product boxes with multiple ornamental
patterns the pictures with straightforward background, commonplace fonts, tiny vary
of font sizes and well organized character square measure used for these systems,
systems provide higher performance however scene pictures with advanced back-
grounds aren’t directly handled some optical character recognition(OCR) software
system shown in Fig. 3.
3 Proposed System
In existing system camera capture the text on image and so process is finished in
portable computer and voice output is created however it’s not possible for visually
handicapped person to hold the portable computer every time. To overcome the
matter of existing system the new system is introduced that’s transportable camera
primarily based golem app to helps the visually handicapped person. We projected
golem app this app facilitate the visually handicapped person to browse the text on
explicit object. In this system camera capture the actual text on image and multiple
techniques square measure applied on it text and finally voice outputs are generated
and so visually handicapped person will simply listen the text on it explicit object.
122 S. Thigale and R. B. Kagade
Gray scale
In Fig. 4 photography and computing, a grayscale or gray scale digital image is a
picture during which the worth of every component could be a single sample, that is,
it carries solely intensity data. Pictures of this type, conjointly referred to as black-
and-white, area unit composed solely of reminder grey, varied from black at the
weakest intensity to white at the strongest.
How does one convert a color image to grayscale? If every color component is
represented by a triple (R, G, B) of intensities for red, green, and blue, however
does one map that to one range giving a grayscale value? There area unit following 3
algorithms. The lightness methodology averages the foremost distinguished and least
Original image
Lightness
Average
Luminosity
Fig. 5 Flower
124 S. Thigale and R. B. Kagade
5 Conclusion
In this paper we’ve represented a system that scans written text on hand-held objects
for aiding visually handicapped people. So as to resolve the common aiming draw-
back for blind users. We tend to plan mobile apps. The mobile camera captured
specific images and applied the text recognition on the image to extract and localize
the written text. We planned an OCR software package. This OCR software package
performs varied techniques for manufacturing the audio output for the visually hand-
icapped person. Then visually handicapped people simply scan the text from the
object.
References
1. Rajkumar N, Anand MG, Barathiraja, Portable camera-based product label reading for blind
people. Department of electrical and electronics engineering, The Kavery Engineering College,
Mecheri, Tamil Nadu, India
2. Yi C, Tian Y, Arditi A Portable camera-based assistive text and product label reading from
hand-held objects for blind persons
3. Text detection in natural scene images by Stroke Gabor words, Yi C Dept of Computer Science,
The graduate center, City Univ of New York, New York, U.S.A., e-mail: [email protected].
Tian Y Dept of Electrical Engineering, The City College and Graduate Center, City Univ of
New York, New York, U.S.A., e-mail: [email protected]
4. Chen X, Yang J, Zhang J, Waibel A Automatic detection and recognition of signs from natural
scenes
5. Kumar S, Gupta R, Khanna N, Chaudhury S, Joshi SD Text extraction and document image
segmentation using matched wavelets and MRF
6. Yang X, Tian Y, Yi C, Arditi A Context-based indoor object detection as an aid to blind
persons accessing unfamiliar environments. Dept of Electrical Engineering, The graduate Center
Arlene R Gordon Research Institute, the city college of New York. The City University of New
York Lighthouse International NY 10031, USA NY 10016, USA New York, NY 10222, USA
{xyang02,ytian}@ccny.cuny.edu [email protected] [email protected]
7. Experiments with a New Boosting Algorithm., Yoav Freund and Robert E. Schapire,
AT&T Laboratories 600 Mountain Avenue, Murray Hill, NJ 07974-0636, yoav,
[email protected]
8. Assistive Text Reading from Complex Background for Blind Persons, Yi C Dept. of
Computer Science, The Graduate Center, City Univ of New York New York, U.S.A. e-mail:
[email protected], Tian Y Dept of Electrical Engineering. The City College and Graduate
Center, City Univ of New York, New York, U.S.A
Automatic System for Identifying
Cholesterol
Mohua Biswas, Pragtee Tathe, Geeta Unhale, and Papiya Biswas Datta
Abstract At the outer edge a solid white ring circling iris exist which is known
as sodium ring or deposit normally located around iris of an eye which symbolizes
the existence of high level cholesterol in the human body. There is a half-circle of
gray, white and yellow deposits in the outer edge of cornea known as Arcus senilis
or arcus senilis cornea occurred because of fat and cholesterol deposits in people
under 45 years of age. The high cholesterol is also an indication of the presence of
hyperlipidemia which indicates the increased amount of fats in the blood. There is a
risk of developing heart disease which leads to stroke and death. There is a chance to
occur a common type of heart disease known as coronary artery disease (CAD) when
the arteries that supply blood to heart muscle become hardened and narrowed. This
is also due to the buildup of cholesterol. Iridology approach is the different type that
also helps to identify diseases with the help of pattern of an iris. Whenever there is a
excess deposition of the cholesterol occur in the body it will create a whitish sodium
ring around the iris, By analyzing the sodium ring we can identify the existence of
cholesterol in the human body.
1 Introduction
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 125
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_13
126 M. Biswas et al.
approach based on laser helps to calculate the cholesterol in human blood through
skin. In order to classify the cholesterol in the skin the infrared (IR) absorption spec-
troscopic is proposed [1]. Generally, there is a presence of cholesterol of approx-
imately 11% in skin because of weight. It gets increased if there is a symptom of
coronary artery disease. In that case for analyzing coronary artery disease palm test
is not so effective as a screening tool. The blood test or lipid profile test is the modern
method to analyze the level of cholesterol. Before lipid profile test there is a need of
12 h fast. This test is tested by observing variations in the pattern of iris which helps
to identify the increase level of cholesterol known as Arcus senilis. Arcus senilis is
a half-circle of gray, white and yellow deposits in the outer edge of cornea. Arcus
senilis is caused due to the lipid deposit in the inner layer of the cornea peripheral.
Because of increase amount of cholesterol there is chance of occurring arcus juve-
nilis problem which indicates discoloration of an eye [2]. Thus for analyzing and
detecting the level of cholesterol in the body of humans’ iris pattern is the alternate
and effective approach known as iridology which was found by Bernard Jensen’s.
Many iridologists admitted that the occurrence of heart disease is due to cornea
having yellowish-white deposit. The high level of cholesterol in the blood is known
as Hypercholesterolemia. It is necessary to check the level of cholesterol regularly
which helps to get rid of the formation of gallstones with bile pigments that normally
occurs because of high level of cholesterol. Protein in human body helps cholesterol
to flow in human body and it is known as lipoprotein. Cholesterol is distinguished in
two ways depend upon the amount of protein present: (i) Low Density Lipoproteins
(LDL) and (ii) High Density Lipoproteins (HDL).
2 Human Iris
The iris of the eye is a thin having 12 mm in diameter and rounded in shape consists of
connective tissue and muscle that encircles the pupil. The Fig. 1 shows the schematic
diagram of a human eye. There is a lens behind the iris of the eye. At the back of the
eye on the inside there are linings covered by cells for light-sensing known as retina
[3]. Retina converts light into electrical impulses. These impulses are transferred to
the brain by optic nerve. The coloring portion of the human eye manages the light to
Fig. 1 Schematic
representation of human eye
Automatic System for Identifying Cholesterol 127
enter. The iris of the eye shut the pupil if the light is bright and because of that less
amount of light passes. If the light is low then iris opens the pupil that allows light to
pass. The size of the pupil is managed by the sphincter and the dilator muscles that
vary from 10% to 80% [4].
3 Iridology
4 Methodology
The following the approaches that are supported by the automated system in this
paper are shown in Fig. 4:
Fig. 4 Methodology
Automatic System for Identifying Cholesterol 129
B. Pre-processing:
The second phase is image preprocessing. The Fig. 5 shows captured image for
second phase. As we normally gathered images from various sources there may
be existence of distortion. So the purpose of image pre-processing is to improve
the distorted images to get better quality image with the help of enhancement
techniques. In this phase histogram equalization method is used to enhance the
contrast of an image.
C. Image Localization or Segmentation:
The image localization or image segmentation is the most critical phase. It is
a method of detecting interior and exterior border regions of the iris. By using
subtraction approach circular shaped iris is detected and extracted from it.
D. Image Normalization:
As the phase of segmentation ends the image normalization phase starts where
it reconstruct the image of an iris. In this phase it unwrap the circular region
of the iris to rectangular block with the help of Daugman’s rubber sheet model
method. Evaluation can be done either from top to bottom or from bottom to
top.
E. Region of Interest:
In this phase of extraction the portion of the iris image is extracted which we
want to refine by simply cropping the normalized iris image. The region of
sodium ring is only extracted.
F. Feature Extraction:
After the region of interest is carried out the next phase is to find out the various
features such as Mean, Entropy, Standard Deviation, Smoothness, Kurtosis,
Variance and Homogeneity related to this region. The results are displayed in
two ways; one is the normal conditions and other is the exceptional cases of
cholesterol.
G. Classification: (Keep text style as normal)
Figure 6 show the classifier used for next stage of image processing based on the
result of feature extraction we classify the level as normal level and cholesterol
level which is beyond normal level [8].
5 Result
In Fig. 7, the above result shows the implementation of our methodology tested on
patient suffering from cholesterol. In this two images of an iris are captured; one is
colour image and other is black and white. In the case of localized iris image the
130 M. Biswas et al.
white portion i.e. sodium ring is present in brown fiber like area. The sodium ring is
segmented and by analyzing the ROI we get the conclusion that there is a presence
of cholesterol in human body beyond the safe level.
In Fig. 8, there is absence of sodium ring in brown portion which means no
cholesterol in the body.
Automatic System for Identifying Cholesterol 131
6 Conclusion
References
1. Li L-H, Dutkiewicz EP, Huang Y-C, Zhou H-B, Hsu C-C (2019) Analytical methods for
cholesterol quantification. Sci Direct 27(2)
2. Arora A, Paggowal N, Chawla D, Sharma E (2018) A non-invasive framework for detecting
hypercholesterolemia using computer vision. IJEDR 6(4). ISSN: 2321–9939
3. Songire SG, Joshi M (2016) Automated detection of cholesterol presence using iris recognition
algorithm. Int J Comput Appl
4. Shanker NR, Ezhil A, Archana S (2012) Non-invasive method of detection of cholesterol using
image processing. Int J Med Eng Inform
5. Silambarasan T, Sethupriyan P (2015) Identification and cholesterol estimation of skin using
hand pattern image. Int J Adv Res Electr Electron Instrum Eng 4(8)
6. Mahesh Kumar SV, Gunasundari R (2013) Comparative Analysis and Earlier Detection of
Cholesterol Based on the Measurement of Various Lipid Level Deposits in the Eye. CIIT 5(7)
7. Adi KG, Rao PV (2016) A non-invasive method of detection of cholesterol by level set segmen-
tation and ANN as classifier using MLP-BP and RBF. Int J Adv Electr Comput Eng (IJAECE)
3(3)
8. Adi KG, Rao PV, Adi VK (2015) Analysis and detection of cholesterol by wavelets based and
ann classification. In: 2nd International conference on nanomaterials and technologies (CNT
2014)
Design of an IoT Based System
for Monitoring and Controlling
the Sub-Station Equipment
1 Introduction
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 133
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_14
134 P. Bodke and A. A. Kalage
the Internet of Things (IoT) technology [1, 2]. IoT is the network of physical devices
embedded with electronics, software, sensors, actuators and network connectivity
which have the ability to identify, collect and exchange the data. Each thing is
uniquely identifiable through its embedded computing system and able to inter-
operate within the existing internet infrastructure [3]. An IoT based network strategy
for monitoring and controlling the sub-station equipment provide efficient time and
resource management [4, 5]. This project mainly focused on to indicate the quality
and present amount of oil in the transformer & oil circuit breaker (OCB) at a remote
location without being physically present so that proper corrective actions must be
taken by the operator. This proposed system represents an alternative way to provide
differential protection where the distance between two sides of the equipment does
not affect system performance.
This system’s objectives are to proposed IoT-based smart monitoring & control-
ling system using Raspberry-Pi module, which will enable the large data storage on
the cloud at a very low cost so that live tracking can be done.
Figure 1 shows the block diagram of the proposed system. The system aims to provide
an operator at a remote location with sufficient information to determine the quality
and present amount of oil in the transformer & oil circuit breaker (OCB).
Even though monitoring & controlling, the system represents an alternative way to
provide differential protection where the distance between two sides of the equipment
has no effect on system performance. An alarm message with what corrective action to
be taken will be provided to the operator at any dangerous condition. Any authorized
person can get mail alert & perform the same tasks by visiting the power system’s
official webpage. It uses the different sensors to sense the transformer oil quality and
quantity as well as the current sensor has been used, and the data is sent to the cloud
server through Wi-Fi network. The advantage of this system is that the data can be
saved and the real-time data can be obtained at any time. Once the data is received
it analyses and checks for the respective action and signal are sent to the actuator.
The major part of the system is as follows:
A. Sensor
The sensors are used here to sense the transformer/OCB oil quality and oil level
also the current sensor to sense the current of a current transformer.
• Transformer or OCB Oil Level Sensor: The HC-SR04 Ultrasonic distance
sensor is used to sense the oil level of the transformer or OCB.
• Transformer oil quality sensor: The TSSP40 infrared sensor is used. An IR
sensor consists of an IR LED and an IR Photodiode. When the IR transmitter
emits radiation, it reaches the object, and some of the radiation reflects back
to the IR receiver.
• Current sensor: The ACS712 is a fully integrated, hall effect-based linear
current sensor with 2.1kVRMS voltage isolation and an integrated low-
resistance current conductor.
B. Actuators
Actuator performs the tasks that are required by the operator at a remote loca-
tion. Servo motor is connected to Raspberry-pi module which performs the
operation as per the instructions received.
C. IoT based differential protection
The differential protection is one of the most commonly used protection
schemes for electrical equipment in power system substation. But due to the
cost of pilot wire, this protection scheme can’t be useful for every system.
One of the major disadvantages of pilot wire differential protection is that as
the length of pilot wire increases the effect of capacitance on system perfor-
mance increases. IoT based differential protection is a smart solution where
two modules send the secondary current data to the webserver. When the data
coming from two different terminals of the protected zone differs from each
other beyond permissible limit, the webserver sends a command to an internet
module which initiates the CB operation to protect the system. So thus, the
problem associated with a pilot wire can be eliminated. To implement practi-
cally IoT based differential protection, two current sensors readings are taken
and the difference of this current is continuously monitoring.
136 P. Bodke and A. A. Kalage
3 IoT Technology
Internet of things IOT consists of two words Internet and Things. The term things
in IoT refers to various IoT devices having unique identities and have capabilities to
perform remote sensing, actuating and live to monitor of certain sort of data [6, 7].
IoT devices are also enabled to have a live exchange of data with other connected
devices and application either directly or indirectly, or collected data from other
devices and process the data and send the data to various servers.
In this system, we are using a Raspberry-Pi 3 Model B+ controller with IoT tech-
nology using Wi-Fi 8266 module. From Fig. 2, the ultrasonic sensor and current
sensor, both data’s are stored in the cloud, and these data’s are controlled using
Raspberry-Pi. Finally, the sensors provide the data continuously to the consumer [8].
5 Proposed Methodology
the operator. After getting the correct data from the sensors, they can easily able to
operate substation parameters such as CB and relay [10]. For data exchange MQTT
protocol is used. MQTT is fast becoming one of the main protocols for IoT (Internet
of Things) deployments. The Linux operating system is used.
The proposed system uses different sensors to sense the transformer oil quality
and quantity as well as the current sensor has been used, and the data is sent to the
cloud server through Wi-Fi network. The advantage of this system is that the data
can be saved and the real-time data can be obtained at any time. Once the data is
received, it analyses and checks for the respective action and signal are sent to the
actuator.
6 Results
After the successful connection of communication devices to the server, oil tank
vacancy level and CTs current are sent to the webserver in an automatic manner. At
the violation of their normal values, an alarming message with a corrective step to be
taken to solve the specific problem was sent to the webserver. We can continuously
monitor the substation parameters on ThingSpeak UI.
Figures 3 and 4 show the command Email in order to control the servo motor (CB
state) and to change the tap position of the transformer from the webpage. Based on
the experiment results, it is terminated that the proposed new IoT based sub-station
controlling and monitoring system worked successfully.
7 Conclusions
In this paper, the IoT based smart system has been developed for remote monitoring
and controlling the entire sub-station equipment, which is very reliable, user-friendly
and low cost as compared to the conventional system. The proposed approach is
an absolutely automatic system that includes self-checking of oil level from the
transformer/oil circuit breaker, continuous sensing of two CTs secondary current,
sending data to the webserver, storing and displaying data in the web page, and
sending a comment to an internet module for performing the specific task such as
initiating the CB operation, tap changing the transformer, and so on.
Also, it created a new pathway to take necessary measures priory in case of any
emergencies that may occur in substations. In future, the system can be enhanced
with additional features for automation of the tasks at the substation.
References
1. Hossain MS, Rahman M, Sarker MT, Haque ME, Jahid A A smart IoT based system for
monitoring and controlling the sub-station equipment
2. Balamurugan S, Saravanakamalam D (2017) Energy monitoring and management using
internet of things. In: Proceedings of the international conference on power and embedded
drive control (ICPEDC), Chennai, pp 208–212
3. Morello R, Capua CD, Fulco G, Mukhopadhyay SC (2017) A smart power meter to monitor
energy flow in smart grids: the role of advanced sensing and IoT in the electric grid of the
future. IEEE Sens J 17(23):7828–7837
4. Hlaing W, Thepphaeng S, Nontaboot V, Tangsunantham N, Sangsuwan T, Pira C (2017)
Implementation of Wi-Fi-based single-phase smart meter for the internet of things (IoT). In:
Design of an IoT Based System for Monitoring … 139
Abstract Human body is the magical creation of god. It carries many intercon-
nected systems. Changes in one system replicate changes in another system. Due to
such interconnection of systems, we can analyze one system by observing changes
in another system. Iridology supports the same theory. Iridology tells the relation
between iris and other systems present in the body. Evaluation can be done in the
form of the iris that speaks about the physical condition of different body parts. In
the proposed method by observing different iris images without doing any compli-
cated and time-consuming test, we can perform diagnosis for the brain tumour. This
method can be used as a pre diagnosis tool.
1 Introduction
One of the best creations of god is the human body. The human body contains many
organs, and these organs are linked with each other. One of the important organs in
the human body is the eye which contains many parts like sclera which is whitish
outer background, iris which is middle central ring may be brown, black or blue and
the black centre pupil. The branch of science which focuses on the study of the iris is
called Iridology. Iridology, also known as iridodiagnosis, is a technique in which the
iris images are analyzed with different aspects. The technique of iridology is based
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 141
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_15
142 P. Tathe et al.
on the theory that different body parts are related to different regions in the iris. The
human iris is categorized into 90 different regions, and each region replicates the
status of different human body parts. Whenever there is a change in the body organs,
it is reflected in the iris. The form of the iris varies according to the condition of body
parts. The right eye iris replicated the status of right side body organs and left eye
iris replicates the status of left body organs [1, 2]. With several nerves connected,
the iris is linked with body parts as shown in the iris chart in Fig. 1. It indicates the
distribution of different organs over the iris circle.
The well solidly organized fibers of iris are the indication of a healthy body. The
unhealthy body has disturbed solidly organized fibers and makes them slackly as
shown in Fig. 2.
A brain tumour is a major disease that has affected over many million people across
the globe; the rate of people getting affected will exponentially increase in the coming
years. There are different methods for detecting the brain tumour like MRI scan, CT
scan etc.; these techniques are time-consuming; it also needs specially trained persons
to handle it. The major thing related to these techniques is that the patients at the initial
stage are not going to look forward towards these techniques. The proposed project
aims to construct a graphical user interface that allows any physician to predict brain
health with user-friendly and fast diagnosis techniques. These techniques integrate
various image processing steps together in order to complete the process of diagnosis
as shown in Fig. 3.
Implementation of Iridology for Pre Diagnosis of Brain Tumor 143
2 Methodology
3. Sectionalisation
5. Taking ROI
6. Feature Extraction
7. Classification
144 P. Tathe et al.
another part contains images of iris of brain tumour suffered person. The database
can also be formed by using images collected from an online database such as CASIA
database, MMU database, and UBIRIS database [1]. Normally, in-hospital ophthal-
mology department is dealing with the iris images. For the proposed work database
is created by taking the total 60 images out of those 30 images are of a normal
person, and 30 images are of persons with a brain tumour. The below figure shows
the captured image of an eye.
B. Pre-processing of eye images
The pre-processing is the next step in the methodology of the proposed system. In this
stage, whatever the captured image is there, it has to pass through various filtering
algorithms. After passing through these filtering algorithms, the image quality gets
improved because this stage scales back the presence of noise within the iris image.
This stage is very important to make end results more relevant than first. For the iris
images, adaptive filtering is used [1, 3–5].
C. Sectionalisation
Sectionalisation is the process to seek out interior and exterior precincts of the iris.
The circular Hough transforms often won’t to comprehend the radius and centre
coordinates of the pupil and iris sections. By subtracting pupil from the sclera, we will
be able to fix the iris. Once the iris region is segmented from an eye, the subsequent
step is to modify the iris region into fixed dimensions. After subtraction, we will get
the iris pattern into a circular shape ring [1, 6, 7].
D. Standardization of the iris image
Daugman’s rubber sheet model (as represented in Fig. 5) is employed for standard-
ization of Iris ring after standardization of iris ring; the ring is transformed into a
rectangular shape were to represent the points on the ring we can use x and y coordi-
nates. The relation between preceding and the following image is specified with the
relation between the coordinate systems [1, 3, 8].
With reference to the point mentioned in the introduction part that different sections
of the iris are related with specific body parts so for performing analysis of brain
tumour we need to crop the area from the normalized iris. As we have divided the
iris area into sectors, we only need to crop the sector, which represents the brain
area. In the circular iris, if we consider a specific point as a starting point then with
respect to that, we have to select coordinates for a brain tumour. Depending upon
which point we have selected as a starting point to convert into a rectangular shape,
the coordinates values are going to change. If we compare the iris with the image of
a clock, then the zone from 11 to 1 represents the brain area. So accordingly, we can
take the region of interest [2, 9] as shown in Fig. 6.
F. Feature extraction
Once we get the significant region subsequently, we can find out different
features for that region. The extracted features are as follows [1, 3].
1. Mean: The average level of intensity of the image.
G−1
Mean: μ = i p(i)
i=0
G−1
V ariance: σ 2 = (i − μ)2 p(i)
i=0
G−1
Skewness: μ3 = σ −3 (i − μ)3 p(i)
i=0
G−1
Entr opy: H = − p(i)ln[ p(i)]
i=0
G−1
K ur tosis: μ4 = σ −4 [(i − μ)4 p(i)]−3
i=0
G−1
Energy: E = [ p(i)]2
i=0
G. Classification
Once we have calculated different features for normal as well as abnormal iris ROI
images, the next step is to train the classifier. Nowadays there are different classi-
fiers which we can use, but out of all support vector machines is more preferable
because it has the kernel trick to transform our data and with the help of which an
optimal boundary is detected for preferable outcomes. So nonlinear SVM classified
the images into two classes; one indicates healthy human iris, and other indicates iris
affected by a brain tumour [1, 10].
3 Result
Figure 7 shows the various operations which are mentioned in the methodology on
normal eye images. If we consider the sequence of operations, then in the first step
eye image is taken with the help of the high-resolution camera. Then that image is
converted into grayscale or in black and white format. Then using the segmentation
method iris area is detected. Once the iris ring is detected, then it is converted into
a rectangular shape, and then the specified area for a brain tumour is separated
by the cropping method. Once we get the ROI then the different features listed in
methodology are calculated, Table 1 shows the values of different features.
In Fig. 8 once our features get calculated depending upon the values, classifier
gives the result. Figure 9 shows the result of the classifier for normal Iris images.
Figure 10 shows the various operations which are mentioned in the methodology
Implementation of Iridology for Pre Diagnosis of Brain Tumor 147
Fig. 7 Various operations on input eye image to find out ROI of brain tumour
on brain tumour affected people eye images. If we consider the sequence of opera-
tions, then in the first step eye image is taken with the help of a high-resolution camera.
Then that image is converted into grayscale or in black and white format. Then using
the segmentation method iris area is detected. Once the iris ring is detected, then it is
converted into a rectangular shape, and then the specified area for a brain tumour is
148 P. Tathe et al.
Fig. 8 Colour feature and texture feature distribution for normal iris images
separated by the cropping method. Once we get the ROI then the different features
listed in methodology are calculated, Table 2 shows the values of different features.
Figure 11 shows the plot of values of different features calculated for abnormal
iris ROI image.
Once our features get calculated depending upon the values of features, classifier
gives the result. Figure 12 shows the result of the classifier for brain tumors affected
person’s Iris image.
Once we train the classifier with 30 iris images and we check the system for the
remaining 30 images then out of that for 25 images, we will get accurate results. So,
we can say that for the above system we get about 83% of accuracy.
Implementation of Iridology for Pre Diagnosis of Brain Tumor 149
Fig. 10 Various operations on input eye image to find out ROI of brain tumour
4 Conclusion
About brain tumour as pre diagnosing stage in an easy and accurate way. As this
technique is non-invasive and less time consuming, it is user friendly to the patients.
If we train the classifier with different variety of iris images, we can also move
towards further accuracy. The proposed system is developed for the pre diagnosing
150 P. Tathe et al.
Fig. 11 Colour feature and texture feature distribution for brain tumour affected iris image
Fig. 12 Decision of
classifier
of brain tumour using Iris images by using MATLAB simulation software. This
method allows any physician to predict.
References
1. Chaskar UM, Sutaone MS (2012, Apr 3) On a methodology for detecting diabetic presence
from iris image. IEEE Xplore. ISSN 12654029
2. https://www.irisdiagnosis.org. Accesses May 2010
3. Tathe PB, Patil MM (2015) Analysis of health condition based on iris image. IJIERT
ICITDCEME-2015. ISSN: 2394-3696
4. Zheng J (2018) An iris recognition algorithm for identity authentication. IEEE Xplore. https://
doi.org/10.1109/ICITBS.2018.00162
5. Kaur N, Juneja M (2014) A review on iris recognition. IEEE Xplore, INSPEC Accession
Number: 14254396
6. Lozej J, Štepec D (2019) Influence of segmentation on deep iris recognition performance. IEEE
Xplore. https://doi.org/10.1109/IWBF.2019.8739225
Implementation of Iridology for Pre Diagnosis of Brain Tumor 151
Abstract In current period, Light fidelity network is becoming very famous and
catching the interest of some of customers with that one advanced technology-based
totally features. It brings a brand new decision inside the Wi-Fi communication
exchange and those are showing so much curiosity to recognize approximately it.
Light fidelity is a wireless communication-primarily based expertise which transmits
records over the community through a light source like led’s as opposed to radio
frequency indicators (RF) with very excessive information charge. The aim of paper
Wireless and optical networks are widely used nowadays because the network perfor-
mance is a very important issue to supply services to a good variety of users whereas
reassuring users’ quality of service necessities. It aims to analysis the wireless and
optical networks performance. The learning toward research the act as an Light
fidelity Network system in wireless communication consumes remained created and
achieve a knowledge speed of 10 k rate over 40 cm distances.
1 Introduction
Inside the period of overloaded (statistics conversation) global, light fidelity stands
an innovative technique of wireless communication exchange like stands usages
managed lighting to transfer information wireless. The communication of statistics
remains one among maximum essential each day accomplishments within imme-
diate developing international. The present wireless communication networks that
is connect North America Country toward net square measure very gradually while
multiple gadgets. Also using amount of tools which get entry to the remaining, the
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 153
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_16
154 N. Deval et al.
2 Working
To purpose of paper toward study the presentation optical and wireless network
established on optical transmission method analyze the parameter of performance
network in Light Fidelity as shown in Fig. 1 [3].
2.2 Methodology
Li-Fi primarily built communication method is totally different from light communi-
cation system as a result of Visible Light Communication is simply a point-to-point
communication system. If Li-Fi technology into pictures then by using sender to
receiver through digital transmission [4].
The working principle is simple are using optical communicate to sender and
receiver through messages shown in Fig. 2.
There running of li-fi as following. There are two simple components are used:
led light source and light sensor (picture-detector). The light source is picture and
detector on every other end. When led bulb is on it starts off evolved sparkling and
the light sensor senses the light from light source and gets signals now the shape of
binary signals i. e. 1 or 0 [5]. When a few statistics is transmitted over the community
from the net to the consumer tool, it transmits over the network and flashing of led
bulb is a sign of the message after which picture detector feel mild and get hold of
the message and ahead to its vacation spot end by way of displayed in Fig. 3.
• LED, Photo Detector (Light Sensor), Optical Wireless Communication (OWC),
End-user Devices (Laptops, computer), Radio Frequency Signal’s Amplifier, Line
of sight.
More the brighter LED bulb, More frequently data will transmit light signals over
the network and highly reliable.
Table 1 Parameter
Parameter Wireless fidelity Light-fidelity
Speed 54–250 Mbps 1–3.5 Gbps
Ranges 20–100 m 10 m
IEEE standard 802.11b 802.15.7
Network topologies Point to multipoint Point to point
2.5 Speediness
For Wireless Fidelity, we have rate margins for information transmission Although
Light Fidelity offers exceptionally high-speed of the network, and that we resolve
transfer large records in exactly limited minutes of your period. Quickness for Li-Fi is
10,000 creases quite wireless fidelity abundant more than 1 Gbps are often achieved.
2.6 Security
Radio waves will penetrate the complete manner via walls. These ends in varied
safety worries as they’ll be intercepted while not issue. At an equivalent time as facts
switch aimed at light fidelity is extremely enclosed then innocent. As we to tend
to use seen light-weight not at all sign separates via partitions. Such seen delicate
communication may be used firmly in airplanes without worrying airways indicators.
3 Related Work
The author described this report possesses statistics approximately the visible mild
communicate the use of li-fi technology. This modern-day generation li-fi describes
transmission of “records via illumination” taking the fiber out of fiber optical by way
of sending records through a controlled light bulb that varies in intensity quicker than
the human eye can observe. Discuss of this paper on a Visible Light Communication
(Li-Fi) performed work done by Wate [3].
The author described li-fi is a label for wireless-communication structures using
mild as a service rather of traditional radio frequencies, as in wireless. Li-fi has the
benefit of being able to be used in touchy regions along with in aircraft without
inflicting interference. Discuss of this paper on a The New Era of Transmission and
Communication Technology: Li-Fi (Light Fidelity) LED and TED Based Approach
performed work done by Prakash [7].
The author defined the short boom in devices utilization and application devel-
opments has brought on many researchers to appearance past wireless. The oppor-
tunity of the usage of visible mild for brief variety verbal exchange is explored with
158 N. Deval et al.
4 Li-Fi Network
The resultant great radio rate of recurrence flickers which remain abundant devel-
oped than the recharge amount of a laptop display are not evident to the residents of
the area.
Influence as fine as records dismiss remain condition to each light fixture using a
quantity of dissimilar methods.
5 Features of Li-Fi
5.1 Volume
5.2 Effectiveness
Everywhere stand an out sized variety of cellular wireless improper positions that
put away large amount of drive. Record of the drive is working aimed at conserving
depressed the end position rather than communication. So, effectiveness of such
Radio corrupt positions is extremely little.
5.3 Safety
6 Applications
Extra ability difficulties of cell systems could also remain fresh burdened to light
fidelity networks wherever offered. That’s especially powerful at the allow wherever
restricted access tend to sit down [9].
160 N. Deval et al.
Some personal or community light beside road spots may remain used to deliver
li-fi hotspots and then the equal transportations then method foundation could also
remain accustomed show and manipulate illumination and information.
Since of sturdy sign interest in seawater, roof use is impractical. Audio waves must
extraordinarily small bandwidth then interrupt sea life. Li-fi presents an answer aimed
at quick-variety infrastructures.
Workstations, functional phones, doses then different cell devices will interconnect
directly the employment of li-fi. Fast selection hyperlinks provide terribly excessive
records quotes and additionally offers security.
Wireless Communication Using Light Fidelity Network 161
8 Conclusion
Light fidelity has standard generation in the area of Wi-Fi information transmission.
It’s miles superior techniques of Wi-Fi Communications that use light as facts carrier.
This paper is part of advent of model stand simulation stage of optical spread method
in matlab Simulink model. To investigate maintain and enhance their community
overall performance light fidelity network.
References
Abstract The system “Smart Trolley with Automatic Master Follower and Billing
System”, has been developed for shopping malls. Once the client enters the mall, the
trolley follows the client with the help of Bluetooth and Ultrasonic sensors placed
in the cart. All the trolleys are designed with a RFID reader, digital display screen,
Bluetooth and ultrasonic sensors. All the products within the shopping complex
are connected with RFID tags. Once the customer puts any item into the cart, its
distinctive ID range is detected. At the same time, the chosen product name with the
cost, expiry date, discount amount if any available are displayed on the liquid digital
display screen, thereby the value gets added to the whole bill. If the client wants to
take away the item from the cart, the client must press the button which is available
on the trolley and then the client has to scan the product once again and therefore
the cost of that specific product gets subtracted from total quantity bill. Once the
purchase is completed, the customer has to select the upload button and the whole
quantity bill is send to the billing counter. With the assistance of payment app the
payment is completed. Once the bill amount is paid, the data is distributed to the exit
gate and the trolley gets disconnected to the customer and the client can exit the gate.
1 Introduction
In Indian cities one can see a very big queue at shopping malls on holidays, festi-
vals and weekends. This queue becomes a lot longer throughout large offers and
discounts. As the client enters the mall one must carry the trolley until the end of
R. Arpita · K. S. Shashidhara
Department of ECE, Nitte Meenakshi Institute of Technology, Bangalore, India
e-mail: [email protected]
V. Dakulagi (B)
Department of E&CE, Guru Nanak Dev Engineering College, Bidar, Karnataka, India
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 163
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_17
164 R. Arpita et al.
the shopping that becomes difficult for older individuals. When purchase of all the
things is completed the client must reach the billing counter for billing purpose. If
the barcode scanner techniques are used to scan the items, then this may be a time
intense method. So this leads to long queues at the billing counter. Therefore this
system provides solutions to all these problems. The main objective of this system
is to help the older individuals and to reduce the time spent by the customer in the
shopping mall.
2 Related Work
The automatic billing system has been implemented with RFID technology [1, 2].
The add-on features such as included product detection, product recommendation,
budget setting and automatic billing [1]. Further in [3], the android application has
been developed, where barcode scanner was used to scan the products. This paper
projected a way to decrease the time of the client. In [4], the aim of this paper was
to cut back the shopping time. The barcode scanner and the touchscreen display
were used and the payment was done using payment app like Paytm, UPI. In the
paper [5], the aim of the paper was to modify and accelerate the method by the
utilization of RFID tags on every product and Android application connected to the
microcontroller employing a Bluetooth module [6, 7].
3 Proposed System
The complete system contains of AT mega 328P small controller, RFID reader, RFID
tags, 16*2 LCD, ultrasonic sensing element, Bluetooth, WIFI module is as shown
in Fig. 1. This projected system works as client in the shopping mall, she/he first
takes a trolley car. Each trolley car has ultrasonic device and Bluetooth, as he/she
comes closer to the trolley car, the trolley car gets connected to explicit client and
also the trolley car follows the client. Each cart is connected with a RFID reader,
a microcontroller and LCD screen. Once the client starts dropping product into
the trolley car, tags are scan by the reader and also the reader sends the data to the
microcontroller. The microcontroller compares the data with the information already
keep in it. If the information matches then the price of that product are displayed on
the LCD screen for user. If the user would like to get rid of any product from the cart
then they’ll subtract that product from trolley car and value of that specific product
are subtracted from the overall quantity in a flash and the product information with
total quantity gets transmitted to the central Billing system through wireless network
module. With the assistance of the payment app, the overall bill quantity is often paid.
Once the payment is completed, the data is shipped to the exiting gate. Therefore the
gate opens and also the trolley car gets disconnected.
Smart Trolley with Automatic Master Follower and Billing System 165
Arduino, UNO, Bluetooth HC-05, RFID reader module, RFID tags, Ultrasonic
sensor HC-SR04, Power supply, WIFI module. Arduino IDE, Google Firebase, MIT
Inventor.
The functionality of the system has been explained with the help of flow chart shown
in Fig. 2. Initialize the system by scanning the RFID tags to the RFID reader. Once
the tag is read, it reads the memory of the tag from the data which is already assigned
to the particular tags. The cost, name, the expiry date and the discount of the products
gets displayed on the LCD. All the selected products are scanned in the similar order
and the respective cost of all the items are added to the total bill amount. If any items
are to be removed from the cart, the particular item should be scanned again and
hence the price of the product gets reduced from the total bill amount. At the last,
the final amount is displayed and the total bill is generated.
5 Experimental Results
The Drone unit and the Dropping Mechanism unit have been developed and the
system is successfully coordinated. Initially the drone without payload as shown in
Fig. 3 has been tested successfully.
166 R. Arpita et al.
The client enters the shopping complex and he/she picks the smart trolley. Each
smart trolley has ultrasonic sensors and Bluetooth placed in it so the smart trolley
follows that individual client. Every product within the shopping mall is hooked up
with the RFID tags and every RFID tags have distinctive number that is employed
to unambiguously establish the merchandise. Once the merchandise placed upon the
RFID reader the tag gets known by the reader once it comes inside the antenna vary
and so the product-ID beside the end date and value is shipped via server. Google
Firebase is employed to form and maintain the information at the server finish.
At the server finish, the information is compared with the RFID ID range that
is placed on the smart trolley and therefore the details of those things like Tag ID,
price, and expiry date and discount percentage are compared. If it matches then
the information gets displayed on the liquid crystal display screen. The information
communicated is saved for a flash within the small controlled memory. The client
is able to choose the things with the assistance of the information displayed on the
screen.
If the client would really like to get rid of any item then he/she must scan the
merchandise once more that removes the actual item. The actual item and its price
are aloof from the whole bill quantity. The ultimate quantity displays on the liquid
crystal display.
After finishing the shopping, the client must press the “upload” button. Once the
button is ironed, the whole bill quantity of all things gets generated. The client must
select the payment technique and so created invoice is transferred to payment web
site and so the payment is finished.
Once the payment is finished, the knowledge is transferred to the exit gate. Thus
the client has completed the shopping while not standing in long queues. When
the system is initialized the first message displayed is “SHOPPING CHART” as its
initialization message. When the RFID tags of the products are read by the reader,
the details of the product gets displayed and the amount of each product gets added
to the total bill as shown in Fig. 3.
The Firebase Database is used for assigning and storing the value of the products.
Google Firebase is used to create and maintain the database at the server end as
depicted from Fig. 4.
With the help of ultrasonic sensors, the calculation of the distance between the
trolley and the Customer is done and the calculated value gets displayed on the LCD
screen.
The Ultrasonic sensor situated on the trolley handles establishes the connection
between client and trolley with the help of Bluetooth module. Further with the help
of ultrasonic sensors, the distance between the trolley and the Customer is calculated
and value gets displayed on the LCD screen as shown in Fig. 5. The trolley follows
the customers when the Bluetooth gets connected to that particular person and the
trolley follows that client till the end of the shopping with the aid of ultrasonic sensor
as shown in Fig. 5. In Fig. 6, the calculated distance value is displayed on LCD.
The mobile application shown in Fig. 7 has been developed can be used by both
the shopkeeper and the Customer. Using this app, the seller can update the details of
any products whereas the customer can use this app to make the payments.
168 R. Arpita et al.
The trolley follows the customers when the Bluetooth gets connected to that
particular person and the trolley follows that client till the end of the shopping. As
shown in Fig. 5. The mobile application shown in Fig. 7 has been developed can
be used by both the shopkeeper and the Customer. Using this app, the seller can
update the details of any products whereas the customer can use this app to make the
payments.
Once the purchase of items is completed, then the bill can be paid using the mobile
app developed. The Bill summary is displayed in the app as shown in Fig. 8.
The items being purchased is logged along with details such as Product name,
Quantity and Total cost as shown in Fig. 8.
6 Conclusion
The objective is effectively attained in the prototype model developed. The developed
product is of low cost, amiable to use and does not require any specific practice.
The ability to take a decision can be done in the cart itself which can be used in
the shopping complexes for effortless and clever way of purchasing items to save
vitality, time and money of the customers. The project is evaluated with different
trial cases with distinct items assessed for all the practical trials. The system ought to
170 R. Arpita et al.
be generated that scale back the time of the customers in looking out the placement
of the product. The client simply sorts the name of the item; the trolley ought to
mechanically guide them to the location of the item. The system ought to be hooked
up with buzzer; if the load of trolley exceeds certain weight unit then it will facilitate
the older customers. Additional economical and long vary RFID readers are often
used. Thefts are often avoided by use of robotic ARM for choosing the items. Since,
it’s a prototype it has limited load carrying capacity around 3 kg.
References
1. Prasiddhi K (2017) Innovative shopping cart for smart cities. In: 2017 2nd IEEE interna-
tional conference on recent trends in electronics information and communication technology
(RTEICT), 19–20 May, India
2. Ravindranath KR, Sanjay AI, Prashant CM (2017) RFID based supermarket shopping system.
In: 2017 international conference on big data, IoT and data science (BID) Vishwakarma Institute
of Technology, Pune, 20–22 Dec
3. Wankhede SS, Nikose A, Radke DP, Khadse DB, Tiwari S (2018) Electronic shopping trolley
for shopping mall using android application. In: Proceedings of the international conference on
communication and electronics systems (ICCES 2018)
4. Viswanadha V, Pavan Kumar P, Chiranjeevi Reddy S (2018) Smart Shopping Cart. 2018 IEEE
5. Kumar A, Srivastava S, Gupta U (2019) Internet of things (IoT) based smart shopping centre
using RFID. 2019 IEEE
6. Raghunandan, Mallikarjun, Satheesh Rao (2017) Billing automation for supermarkets. In: Inter-
national conference on current trends in electronics information and communication technology
(ICCTCEEC-2017), India
7. Sutagundar A, Ettinamani M, Attar A (2018) Iot based smart shopping mall. 2018 IEEE
Analysis and Design of E Shaped Dual
Band Microstrip Textile Antenna
for Wireless Communication
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 171
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_18
172 H. Bhaldar et al.
1 Introduction
The microstrip antenna is widely used for wireless application due to its lightweight,
compact size. Due to use of textile material as in the development wearable antennas,
wireless body area network (WBAN) has been rapidly increasing in the wearable
devices which are used for wireless communication in the field of medical, sport wear,
civil and satellite applications. The wearable antennas easily fabricated into clothes.
In this paper E shaped microstrip textile antenna is proposed for dual frequencies for
wireless application at 3.7 and 7.4 GHz. The wearable antenna can be designed by
using various textile materials such as jeans, cotton, polyester, leather and PDMS.
The relative permittivity (Er ) of these textile materials varies from 1.25 to 1.7 except
PDMS material. It is easier now days to study the performance characteristics such
as bending and crumpling conditions of antenna, due to use of textile material in
designing of wearable antenna.
The two frequency bands of 3.7 and 7.4 GHz; these frequencies come under S
band of frequency range 2–4 GHz and the C band of frequency range 4–8 GHz in the
electromagnetic spectrum. The application of S band is used for airport surveillance
system and to control the air traffic. The S band also used for satellite communication
as downlink frequency of range 3.4–4.2 GHz from space to earth station. The C
band is mostly used over Ku band for satellite communication. This band is also
used for multiple applications such as weather forecasting radars and Wi-Fi devices
as IEEE802.11a. The medical accelerators are operated at the S band frequencies.
They are designed at high frequencies to reduce the size and weight. The medical
accelerators are used in radiation therapy field and it has been implemented in the
Intra Operated Radiation Therapy (IORT).
The multiband E shaped microstrip antenna has been designed with ground and
defective ground structure. The proposed antenna is operating at 3 frequency bands
of 3.7, 6.7, 7.9 and 8.7 GHz and antenna provided the 120, 495, 225 and 315 MHz
bandwidths at these resonating frequencies [1].
The proposed antenna has been designed with the rectangular shape which is
resonating at 2.45 GHz frequency and provides the high gain of 8.05dBi. The
designed microstrip textile antenna with jeans as dielectric material been used for
the Wi-Fi modem and achieved the return loss of −15.76 dB [2]. TERA founda-
tion has designed the antenna which is resonating at S and C bands. The designed
antenna used for medical accelerators in hydrotherapy which radiates the protons
or light ions. The designed medical accelerators have been used at 3 and 5.7 GHz
frequencies [3].The designed antenna has simulated at 2.45 GHz with various dielec-
tric substrate materials such as Cotton, Polyester, Cordura and Lycra. The values of
return losses at 2.45 GHz frequency for respective textile material have studied −
32, −35, −29 and −31 dB [4, 5].
The multiband microstrip antenna has designed with moon structure patch on top
and proposed antenna is resonating at the frequencies from 5.3 to 10.15 GHz and
provided the bandwidth of 62.78%, return loss of −25 dB at 5.44 GHz and −24 dB
at 8.05 GHz. The author used the moon strip line structure of patch [6]. The Idellyse
Analysis and Design of E Shaped Dual Band Microstrip Textile … 173
and Rama Reddy have proposed the circular polarized textile antenna and U slot
conical antenna which is designed at ISM band of frequency range 2.4–2.45 GHz,
the values of S11 parameter −35 and −20 dB are observed at resonating frequency
[7, 8].
The proposed multiband wearable antenna has been radiated at frequencies of
3.42, 9.73 and 11.17 GHz with directivities of 3.35, 4.23 and 5.19 dBi [9]. The
author has observed the miniature feeding network for aperture coupled antennas
designed at frequency of 2.4–2.483 GHz with gain of 5.6 dBi. The author have
studied various shapes of coupling aperture such as ring, H-shaped, rectangular, E
shaped, cross etc. at various frequencies in ISM band [10].
The proposed microstrip antenna have designed with circle shaped radiator with
tapered microstrip feed line and has rectangular ground plane. The proposed antenna
used at ultra-wide band (UWB) application for frequency range of 2.8–16 GHz [11].
From the literature survey, it is stated that the gain and directivity of proposed
antenna is enhanced in this paper. The proposed antenna can be used for high
directional antenna for S and C band application.
The microstrip textile antenna is most preferred antenna over the printed antenna due
to ease in fabrication and light in weight. The microstrip textile antenna is similar
in structure with simple microstrip patch antenna. The microstrip textile antenna
consist of top patch made with copper foil and bottom ground plane made with
same copper foil of thickness 35 micron. These two planes are separated by textile
material as dielectric substrate as shown in Fig. 1. This methodology of antenna
design is demanding for flexible textile (fabric) antennas, in which designed antenna
can easily attached to clothes and antenna can be wore on the body. This technology
introduced the one body to other body communication called Wireless Body Area
Network (WBAN).
In this proposed study, The rectangular E shaped microstrip textile antenna is
designed for 3.7 GHz frequency also the proposed antenna is resonating at 7.4 GHz
additional frequency band due to structure designed E shaped. The inset line feeding
method used proposed antenna and excited with voltage source connected to the port
of an antenna. The microstrip antenna is designed at the frequency band of 3.7 GHz
for S band and 7.4 GHz for C band wireless communication. The top patch of antenna
consists of E shaped rectangular structure of dimension 35 mm × 31 mm of width
and length. The E shaped slot is designed by dividing the patch width in equal 5
slots which is shown in Fig. 2. The bottom ground plane is designed with full ground
and made of copper foil. The cotton shirt is used as substrate material with dielectric
constant of 1.6 with substrate height of 1 mm, so that designed antenna is wearable,
light in weight and compact in size.
The width and length microstrip patch antenna is calculated for resonant frequency
of 3.7 GHz and dielectric constant of 1.6 by using standard equations of rectangular
microstrip patch antenna design [2]. The Table 1 shows that the list of antenna
parameter used design E shaped rectangular microstrip textile antenna.
a b
c d
Fig. 2 Proposed antenna a E shaped patch of simulated antenna b Ground plane of proposed
antenna. c E shaped patch of fabricated antenna d Ground plane of fabricated antenna
Analysis and Design of E Shaped Dual Band Microstrip Textile … 175
The proposed microstrip antenna resonating at two frequency bands 3.8 and
7.49 GHz. These frequency bands are used for S band and C band communica-
tion. The downlink frequency of satellite communication used in this frequency S
band 3.4–4.2 GHz and the C band mostly used for radar communication 4–8 GHz.
The medical accelerators also used in the S band in the radiation filed theory.
The various parameters of antenna has been analyzed at these two frequency
bands such as return loss, VSWR, directivity and bandwidth as shown in Table 2.
The designed antenna is excited with voltage source connected to microstrip feed
line of thickness around 2 mm of impedance 50 . The proposed antenna have been
designed and simulated in CST Microwave Studio the various antenna parameters
have been observed.
(I) Return Loss: The return loss should be kept as low as below −10 dB, to achieve
the best radiation. From Fig. 3a, it has been observed that designed antenna
have two frequency bands at 3.86 and 7.49 GHz with return losses of −33.877
and −19.652 dB. The S11 parameter values of proposed antenna has been
improved and compared with paper implemented. The tested results of fabri-
cated antenna provide the two bands at 2.45 and 3.5 GHz, the wide bandwidth
of 1.9 GHz is obtained at 3.5 GHz frequency band which is shown in Fig. 3b.
The fractional bandwidths are also calculated at these frequency bands using Eq. 1.
where FH = High cut off frequency, FL = Low cut off frequency and Fc = Centre
frequency.
The % fractional bandwidths have been calculated 2.33 and 1.6 at these
frequencies which is shown in Table 2 of simulated results.
Fig. 4 3D radiation pattern the antenna at 2 frequency bands: a 3.7 GHz and b 7.4 GHz
(II) 3D Radiation pattern: From the Fig. 4, it has been observed that 3D radi-
ation pattern of theantenna at resonating frequencies provides the directivity
of 8.74 dBi at 3.7 GHz frequency and 7.32 dBi at 7.4 GHz frequency. These
values of directivity have been improved 7.02–8.74 and 6.37–7.4 compared
with paper implemented and these values fulfil the need of wireless commu-
nication. The proposed antenna can be used for high directional antenna for S
and C band application.
(III) 2D Radiation pattern of proposed antenna: From Fig. 5, it has been
concluded that polar plot of designed antenna has the main lobe magnitude of
8.74 dBi and angular bandwidth of 68.8° at 3.7 GHz frequency and with the
main lobe magnitude of 6.02 dBi and angular bandwidth of 44.1° at 7.4 GHz
frequency.
It has also observed from Fig. 5, that the designed antenna also provided the back
radiation of −21.6 and −15.3 dB at these frequencies.
(IV) VSWR of antenna: From Fig. 6, it has been observed that the value of
VSWR for designed microstrip textile antenna lies in between 1 to 2, which
states the proper impedance matching between excited source and the feed-
line. The values of VSWR at resonating frequencies are 1.14 and 1.28 at 3.7
Fig. 5 2D Radiation pattern of the proposed antenna at 2 frequency bands a 3.7 GHz and b 7.4 GHz
178 H. Bhaldar et al.
and 7.4 GHz. These vaules of VSWR have been compared with the paper
implemented and it is has been observed values are better.
(V) Efficiency: The efficiency of anenna isdefiend as the ratio of radiated powerof
anantenna to the applied power of antenna. The radiation efficiency antenna
differs from the total efficiency of antenna due to loss takes place because of
the impedance mismatching.
The total efficiency is given by Eq. 2.
Et = Zl ∗ Er (2)
4 Conclusion
In this proposed study, the E shaped microstrip textile antenna has been designed for
3.7 GHz frequency. The proposed prototype is simulated and analyzed for the various
antenna parameters VSWR, Return Loss and Bandwidth. The gain of antenna has
been improved to good scale at resonating frequencies, if we compared with base
paper implemented. The designed antenna is used for S band and C band commu-
nication. From the simulated results, it has been observed that, antenna is radiating
very properly at 3.7 and 7.4 GHz frequency bands. The total efficiency is reduced
due to impedance loss in comparison of radiated efficiency.
Analysis and Design of E Shaped Dual Band Microstrip Textile … 179
Acknowledgements The author would like to thank the review committee and my guide Dr. Sanjay
Kumar Gowre for his valuable guidance and support. The author also thankful to SVERI’s College
of Engineering Pandharpur for providing support to use antenna laboratory and Research center,
Department of Electronics and Communication, BKIT Bhalki for support and motivation.
References
1. Nagpal A, Dillon SS, Marwaha A (2013) Multiband E-shaped fractal microstrip patch antenna
with dgs for wireless applications. Int Conf Comput Intell Commun Netw 978-0-7695-5069-
5/13 $26.00 © 2013 IEEE. https://doi.org/10.1109/CICN.2013.14
2. Bhaldar H, Goware SK, Mathpati MS, Jadhav AA (2020) Design of high gain wearable rect-
angular microstrip textile antenna for wireless application. Int J Innov Technol Exploring Eng
(IJITEE) 9(5). https://doi.org/10.35940/ijitee.C8478.039520 ISSN: 2278-3075 (March 2020)
3. Mendes C, Peixeiro C (2018) On-body transmission performance of a novel dual-mode
wearable microstrip antenna. IEEE Trans Antennas Propag 66(9) (September 2018)
4. Chen SJ, Ranasinghe DC (2018) A robust snap-on button solution for reconfigurable wearable
textile antennas. IEEE Trans Antennas Propag 66(9) (September 2018)
5. Potey PM, Tuckley K (2018) Design of wearable textile antenna with various substrate and
investigation on fabric selection. In: 3rd international conference on microwave and photonics
(ICMAP 2018), 9–11 Feb 2018
6. Saxena A, Singh VK (2018) A moon-strip line antenna for multi-band applications at 5.44 GHz
resonant frequency. In: 4th international conference on advances in electrical, electronics,
information, communication and bio-informatics (AEEICB-18)
7. Martinez I, Werner DH (2018) Circular-polarized textile based antenna for wearable body area
networks. IEEE Trans 978-1-5386-7102-3/18/$31.00© 2018 IEEE
8. Reddy RS, Kumar A (2018) Dual band circular polarized wearable antenna for military appli-
cations. In: 2018 international CET conference on control & communication, July 2018.
978-1-5386-4966-4/18/$31.00 ©2018 IEEE
9. Purohit S, Raval F (2014) Wearable textile patch antenna using jeans as substrate at 2.45 GHz.
Int J Eng Res Technol 3(5). ISSN 2278-0181 (May 2014)
10. Zhang J, Yan S, Vandenbosch GAE (2017) Miniature network for aperture-coupled wearable
antennas. IEEE Trans Antennas Propag 65(5) (May 2017)
11. Sun Y, Cheung SW, Yuk TI (2014) Design of a textile ultra-wideband antenna with stable perfor-
mance for body-centric wireless communications. Published in IET Microwaves, Antennas &
Propagation. Accepted on 3rd July 2014. https://doi.org/10.1049/iet-map.2013.0658
Enhancement Technique for Early Stage
Lung Cancer Detection Using Foldscope
Methodology
Abstract Due to ever increasing population and growth in pollution there has been
considerable increase in the cancer-causing agents due to this there is a need to
detect cancer at its early stages so that it becomes easier to cure it. The precautions
to be taken to cure or prevent cancer or other tumor is that cause death very less or
completely negligible this is why there is a need to create a system that can suggest
the early stages of cancer and the precautions that will be taken in order to prevent
this cancer from growing more. Cancer as it is cannot be cured hundred percent
when it is in the moderate stages and above as there are very few treatments for
undertaking or curing cancer completely. We have undertaken the job of detecting
lung cancer as it is the major cause for death in Cancer diseases. Lung cancer when
it is in the initial stages is very accurate and is very difficult to detect and it spreads
rapidly and the chances to stop this growth are very less or completely negligible so
to avoid this, we have decided to create a system that can detect lung cancer at its
early stages. Lung cancer comes in stages and to cure this we need to understand
each and every stage precisely the lung cancer has around Four stages. After looking
at the stages of cancer we can clearly see that the first stage is very important in the
recovery process and for cancer treatment. So, if we detect the cancer at its early
stages it becomes easier for us to cure it and the patient suffers very less due to
this cancer. In some cases, the microscopic examination of this cancer shows wrong
results as the cancer is very less and cannot be easily detected even at microscopic
level. If we record such occurrences it becomes easier for us to detect for the next
time and the chances of error reduces. Here comes the role of Technology and our
system. As we are using machine learning for the development of a system which
is Computer aided design for early lung cancer detection it becomes easier to train
the machine for intangible occurrences of results which cannot be neglected as life
of each and every patient is precious. Our system is completely based on the data
analysis of multiple cases of cancer where lung cancer differs from first stage to
V. D. Jadhav (B)
CSE Department, SVERIs COE, Pandharpur, India
e-mail: [email protected]
L. V. Patil
CSE Department, SKN COE, Sinhgad Road Vadgaon Bu, Pune, India
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 181
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_19
182 V. D. Jadhav and L. V. Patil
last stage. By considering the occurrences of results we can analyze and design the
pattern for this disease and can easily predict whether the underlying patient being
monitored is suffering from cancer. We are using machine learning based on Python
programming language as it becomes easier to monitor data as well as the detection
process of cancer. Python makes it easier to use machine learning and we can train
the system for the future. Python has multiple directories which we can use for our
project and which makes our project very simple and accurate as we are going to
take the data set which includes arrays of data.
1 Introduction
Machine learning is a branch of science that deals with training of the machine for
future occurrences. For some cases the cancer detection process gives wrong results
as the cancer occurrence is very rare and for this, we use machine learning which
helps us to detect the cancer for such occurrences in future proceedings. Even for the
microscopy detection of Cancer lungs we need some predetermined data set which
helps us in detection of that cancer so that we can clearly say that it is a normal cell
or paranormal cell. We have taken images of multiple cancer as well as normal cells
and taken the data into consideration of a data set. Due to this multiple use of data
we can clearly say that we have covered multiple cases as the cancer which we are
going to detect is intangible to the system. We have covered around 1000 different
cases where cancer is in early stages.
The data is it is completely unusable when it is not properly managed and precisely
developed. We have taken search algorithms that are very precise in itself and create
very accurate data that can be used easily and precisely to take tests under serious
conditions. This data from the data set is used for matching the occurrences of given
sample images for the system which will detect the occurrences of Cancer if present
or absent. Before creating the detected the images collected of the cells need to be
processed at, they need some sort of image segmentation and image processing as
it contains noise as well as lots of error which are due to the microscopic detection
using human help. Machine learning makes it easier to calculate the situation and
monitor the rare case executions and enter it into the data set as the rare cases should
be properly managed in the future occurrences. In some cases, the cancer in the cells
will not be detected due to its slightly lesser intense present in the cancer cells. These
rare occurrences will be recorded and sent to the data set so that it can store these
occurrences for further convenience and avoid for the inconvenience due to absence
of the data for calculation [1]. Our system will record every occurrence of the result
and send it to the data set so that the system for every next calculation becomes more
pieces and continuously gets more Precision for this testing process. This system will
work fine and will definitely be a boon to the medical world as it is more precise and
can detect cancer at its early stage and make it easier to cure the patient and treat the
Enhancement Technique for Early Stage Lung Cancer Detection … 183
patient easily. This is the reason we have selected this task and problem statement
for making the world a better place.
Foldscope is an origami based paper microscope which can assemble from
punched sheets of card-stock and lens [2]. The foldscope is firstly printed on a
card sheet and according to the marking on the card sheet the foldscope is assembled
the same way as in origami craft works. The lens is also assembled in the fold-
scope. It is used for the magnification required for getting a microscopic view of
the subject or sample being tested by the foldscope user. There are also magnetic
couplers used for connecting a sophisticated camera or other imaging device which
is used for capturing images of videos of the sample. The magnification of the fold-
scope ranges from 140X to 2000X and 2 micron resolution. We made samples of
specimens of different test subjects for the purpose of microscopic testing of the
subjects under foldscope. The created specimen samples were loaded on slide and
then were mounted on foldscope for getting a microscopic view of the samples.
The magnification is 140x+ and resolution is 2 microns [3]. We can actually see and
examine microscopic cellular activities of the test specimens. By this we can propose
the actual underlying abnormality of the cells of the specimens.
Lung cancer remains the leading cause of Cancer related problems in the world.
Early diagnosis can improve the effectiveness of treatment and increase the patient’s
chance of survival. Positron emission tomography (PET), computed tomography
(CT) [4], low dose computed tomography (LDCT), and contrast-enhanced computed
tomography (CE-CT) are the most common noninvasive imaging modularities for
detecting and diagnosing lung nodules [5]. Pet CT scans can be done at various
centers available in the country but due to machinery requirement and cost of the
Machines, very few centers are available [6, 7].
Due to the cost of the machinery and the required technology used for the detection
the cost of this testing is very high and cannot be done by a normal middle class family.
With high cost level how the current systems available make this system completely
invisible to every common man living in the country. Is there a need to make this
system somewhat cheaper so that every citizen in the country being rich or poor can
use this system for their help. Also the reach of the centers is limited to cities and
metropolitan areas. Due to this the people from rural areas and the people living in
the villages and sub urban areas are not benefited by these centers. The average cost
of the pet CT scan at any Centre across India ranges from 15,000 to 20,000 rs for
tests and this does not assure the complete detection of lung cancer [8]. This is just a
preliminary examination done in order to check whether the person’s lung is affected
with cancer or not. Further is the test negative in searching for lung cancer or in case
of unavoidable circumstances which lead to inconsistent results the patient is sent
for other scans which cost lakhs and lakhs of rupees. Even if the patient manages to
collect this money the result would never be different if cases of cancer are very rare.
184 V. D. Jadhav and L. V. Patil
For such cases the money as well as the time that a patient should be treated is wasted
which time could prove to be crucial in the recovery of the patient. To avoid such
circumstances, we have decided to build projects in such a way that every person
living in this country would be benefited by this system. Power system is not much
cost efficient as it requires just one cellular Level Examination of the patient and the
rest of the work will be done by the machine. The system which we are going to
develop will tell you the result of the comparison of the cellular Level examination
of the patient with the predefined data set we have developed in order to detect if
the patient is suffering from lung cancer or not. The cost of the system will not be
more and it will be beneficiary for every patient had the patient will require no direct
visiting to the center of examination [9].
Because the patient can visit the nearby facilities in order to take the samples of
the cells microscopic level and directly send this to the centers so that we can check
the samples and tell whether the patient samples of specimens are suffering from
cancer or affected with any other cases.
This could even be a remote management system for cancer detection and can
prove to be a simple and efficient way for the people living in rural areas or remote
areas where small medical facilities are also not available for cancer detection. The
fight against cancer will become easier and more widely available due to this system.
The reach of cancer awareness and recovery will increase day by day, if this system
comes into existence this is the need for the development of the system.
1.2 Objectives
(a) To implement methodology that works with cellular level detection and
determination of cancer.
(b) To predict the possibility of lung cancer using the data-set facilitated by
machine learning [10].
(c) To compare between cancerous and non cancerous nodules [11].
(d) To create awareness among the masses for early screening of the disease.
(e) To reduce the need for higher-level machinery implementation in order to
develop low cost systems with high output.
2 Methodology
With reference to Fig. 1, we are going to provide the solution to resolve the above
stated problem by designing and developing a CAD system for early lung cancer
detection by inculcating python integrated with machine learning as a key technique
for evaluation of the collected data [12]. Our system has different modules for clear
segmentation of the detection of cells specimen which are stated below:
Enhancement Technique for Early Stage Lung Cancer Detection … 185
Fig. 1 Architecture of
proposed system
A. Search Module
This module is specifically designed and propagated to locate the testing specimen
which is going to be submitted to the system for further evaluation procedure. This
module is a key benefit which facilitates the user to easily locate the test image and
submit it to the system.
B. Image Processing Module
The image of the cell specimen that is passed by the user through the Search Module
is later entered into this module for further processing. The image that is passed by
the search module is not ready for evaluation and needs precise processing. In this
module the image is assessed for noise as well as other impairments and are removed
from the image set. The image is later contrast adjusted for sharpening and better
clearance. Later on this processed image is passed to the next module.
C. Thresholding Module
The image that is received from the previous module is parsed by Otsu’s thresholding
algorithm. Image segmentation takes place in this algorithm and now the image is
ready to be examined and compared with the data-set.
D. Classification Module
Here, the processing ready image is examined and tallied with the data-set. The
image is completely compared with all the existing data-set images and the result
is sent back to the data-set in order to increase the accuracy for further evaluation.
In case, if the result is not found satisfactory then this situation is recorded so that
186 V. D. Jadhav and L. V. Patil
it does not repeat for the next time. In this way the module never makes the same
mistake twice.
Details of the Proposed Implementation
• Search Module
This module is specifically designed and propagated to locate the testing specimen
which is going to be submitted to the system for further evaluation procedure.
• Image Processing Module
In this module the image is assessed for noise as well as other impairments and is
removed from the image set. The image is later contrast adjusted for sharpening and
better clearance.
• Thresholding Module
The image that is received from the previous module is parsed by Otsu’s thresholding
algorithm. Image segmentation takes place in this algorithm and now the image is
ready to be examined and compared with the data-set.
• Classification Module
The image is completely compared with all the existing data-set images and the result
is sent back to the data-set in order to increase the accuracy for further evaluation.
Working
The user needs to browse through the files and select the target image that is to be
entered in the system for the evaluation process. The image that is taken initially
has certain agents of error that need to be processed in order to get a more feasible
image for the testing process. The image also needs to be in a required format as the
prescribed data-set is built of.
Later, when the image is brought in the system then it needs to be further processed
in order to remove the noise and the errors. The noise may be of different sorts like
salt noise, smoke, etc. This noise is removed by using certain techniques that enhance
the image by removing such noise and the errors. The next parameter that is to be
enhanced is the sharpening and the contrast that is to be maintained and enhanced
accordingly in order to make the image more reliable and perfect for evaluation.
After the image processing the, image is next taken into the algorithmic evaluation
so as to increase its compatibility with the system. Otsu’s Thresholding technique is
used for making the image more and more suitable to work with the system. This
algorithm does significant scaling and other required jobs that make the image more
precise and more efficient in terms of details.
This processed image is then taken to evaluate and determine the cancer causing
agents presence in the sampling image. If the given image has some features that are
similar to that of any of the details in the data-set then the system records that data
Enhancement Technique for Early Stage Lung Cancer Detection … 187
and enters it into the data-set for next successive evaluation. All the results that are
evaluated by the system are successively sent to the data-set so as to improve the
quality of the data-set and increase the precision of the system. The more results we
evaluate the more the system works effectively.
The entire development process has been subdivided into two: the front end devel-
opment and the backend development. The front end comprises the visually visible
parts such as the Browse Button, Image Path, Predict Button and the Result. The
back end contains the data-set and its interaction with the trained data.
Front End Development
The front end is coded using PAGE Python GUI Generator. PAGE is a cross-platform
drag-and-drop GUI generator, bearing a resemblance to Visual Basic. It allows one
to easily create GUI windows containing a selection of Tk and ttk widgets. Required
are Tcl/Tk 8.6 and Python 2.7+. I am actually using Tcl/Tk 8.6 and Python 2.7. PAGE
springs from Virtual Tcl, a Tcl/Tk program, forked to generate Python modules that
realize the desired GUI. Tcl is required for running PAGE but is not required for
executing the generated Python code.
PAGE is not an end-all, be-all tool, but rather one that attempts to ease the burden
on the Python programmer. It is aimed at the user who will put up with a less than
completely general GUI capability in order to get an easily generated GUI. A helper
and learning tool, it does not build an entire application but rather is aimed at building
a single GUI class and the boilerplate code in Python necessary for getting the GUI
on the screen.
Back-end Development
The back-end development is totally done with Python. Some of the image pre-
processing techniques like Noise Removal, Gray Scaling and Otsu’s Thresholding
are used. One of the fundamental challenges in image processing and computer
vision is image denoising. What denoising does is to estimate the original image by
suppressing noise from the image. Image noise may be caused by different sources
(from sensors or from environment) which are often not possible to avoid in practical
situations. Therefore, image denoising plays an important role in a wide range of
applications such as image restoration, visual tracking, image registration, and image
segmentation. While many algorithms have been proposed for the purpose of image
denoising, the problem of image noise suppression remains an open challenge, espe-
cially in situations where the images are acquired under poor conditions where the
noise level is very high.
In global thresholding, we used an arbitrary value for threshold value. So, how
can we know if a value we selected is good or not? Answer is trial and error method.
But consider a bimodal image (In simple words, bimodal image is an image whose
histogram has two peaks). For that image, we can approximately take a value in
the middle of those peaks as a threshold value. That is what Otsu binarization does.
So in simple words, it automatically calculates a threshold value from an image
histogram for a bimodal image (For images which are not bimodal, binarization
won’t be accurate.). For this, our cv2.threshold () function is used, but passes an
188 V. D. Jadhav and L. V. Patil
extra flag, cv2.THRESH OTSU. For threshold value, simply pass zero. Then the
algorithm finds the optimal threshold value and returns you as the second output,
retVal. If Otsu thresholding is not used, retVal is the same as the threshold value you
used.
The above Fig. 2 is the unprocessed image that is selected by the user. The image
contains sources of error and noise and needs to be processed.
The above Fig. 3 is the processed image after noise removal. Salt-noise and other
types of noise are removed from the image. The image is also sharpened so as to
make it compatible with the data-set.
The above Fig. 4 is gray-scaled to hide the uncertain parameters and con-
concentrate on the required part of the image. This image comes under 2-D image
spectroscopy of array data.
The Fig. 5 shown above is the processed image by the algorithm we have used
known as Otsu’s Thresholding Technique. The algorithm enhances the image details
and makes the image more accurate and precise [13]. This algorithm does image
enhancement.
4 Conclusion
learning. Helps to create awareness among the masses for early screening of the
disease. Reduces need for higher-level machinery implementation in order to develop
low cost systems with high output.
190 V. D. Jadhav and L. V. Patil
References
1. Gonzalez Rafael C, Woods Richar E, Eddins Steven L Digital image processing using MATLAB
2. Cybulski JS, Clements J, Prakash M (2014) Foldscope: origami-based paper microscope. PLoS
ONE 9(6):e98781. https://doi.org/10.1371/journal.pone.0098781
3. Ferlay J, Soerjomataram I, Ervik M, Dikshit R, Eser S, Mathers C, Rebelo M, Parkin DM,
Forman D, Forman D (2015) [online]. Available: www.wcrf.org
4. Grove O, Berglund AE, Schabath MB, HJWL Aerts, Dekker A, Wang H, Gillies RJ (2015)
Data from: quantitative computed tomographic descriptors associate tumor shape complexity
and intratumor heterogeneity with prognosis in lung adenocarcinoma. The Cancer Imaging
Archive [online]. Available: https://doi.org/10.7937/K9/TCIA.2015.A6V7JIWX
5. Grove O, Berglund AE, Schabath MB, Aerts HJWL, Dekker A, Wang H, Gillies RJ, Muoz-
Barrutia A (2015) Quantitative computed tomographic descriptors associate tumor shape
complexity and intratumor heterogeneity with prognosis in lung adenocarcinoma. PLOS ONE.
Public Library of Science (PLoS)
6. Lee JH et al (2017) Anatomically and functionally distinct lung mesenchymal populations
marked by Lgr5 and Lgr6. Cell 170(11491163):e1112
7. Storvall H, Ramskold D, Sandberg R (2013) Efficient and comprehensive representation of
uniqueness for next-generation sequencing by minimum unique length analyses. PLoS ONE 8
8. Clark K, Vendt B, Smith K, Freymann J, Kirby J, Koppel P, Moore S, Phillips S, Maffitt D,
Pringle M, Tarbox L, Prior F (2013) The cancer imaging archive (TCIA): maintaining and
operating a public information repository. J Digital Imaging 26(6):1045–1057
9. Armulik A, Genove G, Betsholtz C (2011) Pericytes: developmental, physiological, and
pathological perspectives, problems, and promises. Dev Cell 21:193215
10. Simon H (1994) Neural networks. Mac Millen College Pub co, New York
11. Marques S et al (2016) Oligodendrocyte heterogeneity in the mouse juvenile and adult central
nervous system. Science 352:13261329
12. Zepp JA et al (2017) Distinct mesenchymal lineages and niches promote epthelial self-renewal
and myofibrogenesis in the lung. Cell 170(11341148):e1110
13. Vanlandewijck M et al (2011) A molecular atlas of cell types and zonation in the brain
vasculature. Nature 554, 475480
Foldscope to Detect the Growth
of Microorganisms on Various Materials
and Vessels
Vanita D. Jadhav, Richa Tamhane, Kiran Kedar, Shruti Kawade,
and Aboli Gaikwad
1 Introduction
The easiness of Foldscope attracts the attention of many of us. The educationalists
to researchers and the common people to scientists. The thing that pulls them is this
modern origami based microscope with its wide applications. Foldscope, however,
has a big hand in solving the microbial problems. During the twenty-first century,
Manu and his team put forth a modern hand microscope—Foldscope and it helps
lots of young minds to solve the mysteries and puzzles of the microbial world [1].
This modern world microscope is sure to take up other dimensions of the microbial
world with its applications [2]. In this paper, we shall discuss the use of foldscope
to detect the growth of microorganisms on different construction materials and on
vessels [3].
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 191
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_20
192 V. D. Jadhav et al.
This study consists of the samples from Cement, Wood, Iron plates, Gravel Asbestos
sheets, brick, glass and tiles. The considered sample was diluted one by one using
sterile distilled water. The diluted sample is now treated with agar medium by
pour plate technique and is incubated for 24–48 h. Petri plates were considered
for enumeration and finally, foldscope was used to detect the bacteria on the sample.
1432 microbes were found enumerated among 174 samples. From the considered
sample material, 6 bacterial genus and 3 fungal genus were found isolated. Foldscope
images were later compared with the microscopic images and it was found that it
was the best alternative, cheap and durable. Detection of microbes on construction
materials had always been a challenging task in the science out there. But foldscope
has now proved to be the best possible solution for such problems.
Foldscope to Detect the Growth of Microorganisms on Various … 193
Microbes can be found almost everywhere in nature including air, water and land.
They are ubiquitous in nature. In the other cause, microbes are also important in our
day to day life and culture in ways to treat sewage, produce fuels, enzymes and even to
ferment the food. They play a role as an essential biological compound. On the other
hand, microbes are those pathogens responsible for spreading diseases and polluting
the hygiene [2]. Many disease causing pathogenic microorganisms are present in
the environment and they are easily exposed to the food products by handling and
growing on the surface of vessels used for food preparation [10].
Foldscope can be used to detect different types of microorganisms and bacteria
but can’t help to detect the viruses. The percentage of bacteria is highest with respect
to other reasons to cause illness and is up to 66%. Table 1 shows some of the bacterias
affecting food.
Moulds like Aspergillus produce aflatoxin, ochratoxin, citrinin and patulin are
vulnerable to humans. However, these are even difficult to get rid of. To prevent the
growth of such microbes, cleansing and sanitizing of utensils, vessels and equipment
must be done from time to time to avoid consequences [4].
Detection of microbes
As the studies have clearly proved that foldscope has a major role in the application to
detect microbes in cooking vessels. Beyond utensils, growth of microbes is found in
kitchen sponges and dish-clothes too. If it is not cleaned timely, it may lead to cross-
contamination and thus contaminate the food too. This study consists of different
types of cooking vessels such as copper, brass, stainless steel, tin, plastic.
3 Procedure
Starch of 1000 ml was prepared and sterilized. The sterilized starch solution was
poured onto the vessels 100 ml each and kept for incubation at room temperature 27
°C. This starch has been applied over all the cooking vessels.
194 V. D. Jadhav et al.
It was made of sterile distilled water with the composition of nutrient agar 14, 2.5 g in
500 ml water. A plate was kept over it and it was incubated for about 24 h. Hence, the
presence and absence of colonies above it was monitored. This showed the growth
of microbes on the plate.
The slides were prepared by simple staining of selected colonies. The cell wall
composition made it clear about the bacteria whether it was Gram positive or gram
negative. Thus, gram staining was done.
Staining was performed at various time intervals to differentiate between gram posi-
tive and gram negative. The slide was placed in the paper based microscope interfaced
with a mobile phone. After staining the cells were observed under foldscope, the
recorded images were further analyzed using a microscope. Morphological obser-
vations were done upon simple staining methods and microbes were seen using
foldscope. Foldscope images clearly have shown the presence of bacteria and fungi
from the collected samples (Refer Figs. 1 and 2).
4 Conclusions
We came to know that foldscope unlike other microscopes is durable, affordable and
easy to carry anywhere. It is a best choice to detect microbes and other micro particles
for educational or research purposes. It gives best possible results and proves to be
an ideal tool to detect and monitor them. Experiments clearly showed that foldscope
gives upto 140X magnifications. Low cost and high outputs, made this tool a best
choice among the society. It has various applications and its specific advantages in
each new application.
Foldscope to Detect the Growth of Microorganisms on Various … 195
References
1. Jadhav VD (2018) Foldscope: a low cost magnification device and its applications in various
fields
2. Rathod D. Foldscope: a educational cum research tool used in identification of microorganisms
3. Cybulski JS. Foldscope: origami based microscope
4. Calder R. Preliminary studies in the use of foldscope paper microscope for diagnostics analysis
of crystals in urine
5. Muthukumaran E. An experimental investigation on the quality of air and growth of microor-
ganisms in various construction materials at various conditions and in various cooking
vessels
6. Ephraim RKD. Diagnosis of Schistosoma haematobium infection with a mobile phone-mounted
foldscope and a reversed-lens CellScope in Ghana
7. Sharmila S. Anatomical Characterization of medicinal plants using foldscope: a paper based
origami microscope
8. https://en.wikipedia.org/wiki/Foldscope
9. Raghavamma STV. Applications of paper microscope for the early detection of parasitic
pathogens in blood smear
10. Banerjee S. Foldscope, the frugal innovation and its application in food microscopy—a review
Implementation of A* Algorithm
for Real-Time Evacuation During Fire
Situation
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 197
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_21
198 S. K. Rudrawar et al.
1 Introduction
Living in skyscrapers has caused significant issues and difficult and unpleasant chal-
lenges in occupant evacuation during fires or any other emergencies. Predefined
evacuation maps are static in nature, because they showcase some main evacuation
routes which could be prone to blockage as a fire hazard evolves [1]. Evacuation
management regards the exit choice to be critical; some prominent evacuation strate-
gies generally consider egress by shortest distance, or balanced use of exits/stairs for
proper crowd dispersion, or prioritization of Fire floor or Fire layer evacuation [2].
To alleviate casualties, it is important to have emergency evacuation plans in place
before fire outbreaks occur. However, there considerably too many factors which
involve uncertainties and complexities in evacuations. Previous studies unfold that
amongst others, prominently, building structures, fire scenarios and evacuees share
a common physical and social interaction during evacuation activities for fire inci-
dents. Some factors why these components indulge in the evacuation are listed in
Table 1.
Smoke inhalation kills many, compared to the fire. The prominent dangers are
smoke and high temperatures, too that accompany fires, potentially of causing large
scale casualties. Therefore, critical signaling to the evacuees needs ensuring that
evacuation routes are safe and free from smoke to the maximum possible extent [3].
Hence the main aim of this paper is to inculcate such a mechanism which ultimately
is a real-time fire-safe pathfinder that helps improving the efficiency of evacuation.
Many projects in the past limit the real-time delivery of fire incident information
to fire services providers and first responders only.
But the efficiency of evacuation can be improved if evacuees are enabled to take
the correct decision on time through digital signage/mobile devices. It is aimed to
be a guidance system for fire evacuation that will employ smart phones to provide
appropriate as well as timely information. This approach will be more reliable than
just some static data on walls shown by guide boards or say any announcements.
A revised system is needed that can reorganize the existing safety infrastructure
of large buildings to become effective in saving people, reliably, by fast response.
The proposed system is not meant to replace existing fire-safety systems, but
rather it can incorporate evacuation intelligence into them. And has better future
scope due to its efficiency.
3 Methodology
Fire sensing nodes have Node MCU to command and are prototyped as a module
consisting of 3 main sensors: a fire-detecting flame sensor, a smoke and gas MQ-2
sensor and aDHT-11 temperature sensor.
Congestion monitoring nodes also have Node MCU to command are additionally
consisting of 4th main sensors: IR (infrared) sensor for congestion detection, along
with others: a fire-detecting flame sensor, a smoke and gas MQ-2 sensor, and a
DHT-11 temperature sensor. Hence respective sensors are used at each node for data
acquisition by the Node MCU [11].
An algorithm was to be developed that would majorly take 2 primary inputs: Building
Map and the Fire situation. The system needed to be triggered by fire situation from
Implementation of A* Algorithm for Real-Time Evacuation … 201
the network meshed; and should reliably identify location of fire based sensor data.
Then using strategies in path planning, we aimed to develop evacuation maps starting
at various points as needed from inside the building. Hence sensor inputs have been
routed to MATLAB via Thingspeak server and A* is used for sensor data optimization
in evacuation path planning [12].
A* algorithm calculates the cost function f(n), as in Eq. 1 where h(n) is the distance
between the start position and node n and g(n) is the distance between n and target.
To have the input of Sensor Node onto the on-line server Thingspeak Platform, a
bread-board implementation of the sensors is connected to a Wi-Fi. This is depicted
by Fig. 1.
Once Wi-Fi connection is sensed by the nodes, the monitored data is transmitted to
the Thingspeak Server. Hence, these sensor nodes continuously send the data online
to Thingspeak. This is shown in Fig. 2.
This data is further imported in MATLAB and incorporated in to the A* algorithm.
For instance, data captured by Node1, Node2 and Node3 is imported in MATLAB
Script by the following lines of code from the respective thingspeak channel.
Presence of fire (flame) brought nearer to the Flame Sensor, sends a digital 1 or
0 to Node MCU. It in turn analyses this data by appropriate thresholding, for the
Fig. 3 Android
application—front page
204 S. K. Rudrawar et al.
changed reading, packages this data and sends it to Thingspeak channel. Node MCU
by its ESP8266 Wi-Fi transceiver sends this data wirelessly out to the Wi-Fi network
aligned field-wise to the Thingspeak Channels with its information. Hence fire is
imitated by candle flame as sensed by fire sensor. For the prolonged presence of fire,
it is simulated to aggravate, and this spreading of fire is assumed as the dispersion
Wi-Fi specifications over which the nodes were tested is described below:
• Security: WPA2 PSK
• Frequency: 2.4 Ghz band
• Identification: Pixel Graphics
• SSID: uitech333.
Test Case 1:
Path calculated in case of absence of fire danger: In this instance, there was no fire
present and hence no fire detected by the sensor nodes. So the evacuee was provided
with the shortest path to escape from the room which he queried over the app from.
MATLAB calculation of optimized route is shown in Fig. 5a and b.
Test Case 2:
Decision of Correct Choice of Available Exits: In this instance, fire was present
at one door of a room. Even if that path was shortest, but it was overridden by a
comparatively safer path. So the evacuee was provided with the shortest path to
escape from the room which he queried over the app from. MATLAB calculation of
optimized route is shown in Fig. 6a and b.
In this instance, fire was now detected at another door of the same room. The
second door had no fire. Hence the shortest and safest path was suggested by opti-
mization. So the evacuee was provided with the shortest path to escape from the
Implementation of A* Algorithm for Real-Time Evacuation … 205
Fig. 5 a Path calculated in case of absence of fire danger. b Real time Output on Mobile App
Fig. 6 a Decision of correct choice of available exits. b Real time output on mobile app
room which he queried over the app from. MATLAB calculation of optimized route
is shown in Fig. 7a and b
Fig. 7 a Decision of correct choice of available exits. b Real time output on mobile app
5 Conclusion
Proposed work gives a framework for calculating better evacuation routes for occu-
pants of an indoor scenario threatened by a fire caused, and a case of single occu-
pant considered. The A* algorithm via MATLAB uses in-formation collected from
the sensor network. Information from the sensors indicates door area blockage. A
Implementation of A* Algorithm for Real-Time Evacuation … 207
network provides indication on the direction to exit to the occupant for a quick evac-
uation. We tested a version of routing using the information about transit able spaces
and fire location and another that adds a distance as safety margin between the occu-
pant path and fire to keep the occupant far from the fire. The prototype results were
satisfactory, as aimed through the design framework.
Each sensor node can be enabled to be decisive. The end user interface can be
developed further as active Exit Signs/digital Sign Boards spread across facility that
consider themselves as starting point, and detect dangerous areas around them in real
time. Such a UI can direct evacuees to the safest evacuation path from that room to the
nearest exist. Eventually data of both, the spread of the hazard and of the safe-egress
accomplished, can be shared to the Hub computer for calibrating the management
strategy decisions taken up each time by the nodes, thereby avoiding congestion.
References
11. Ronchi E, Uriz FN, Criel X, Reilly P (2016) Modeling large-scale evacuation of music festivals.
Case Stud Fire Saf 5:11–19 (Elsevier)
12. Radianti J, Granmo O-C, Sarshar P, Goodwin M, Dugdale J, Gonzalez JJ (2015) A spatiotem-
poral probabilistic model of hazard and crowd dynamics for evacuation planning in disasters.
Appl Intell 42:3–23
Involuntary Traffic Control System
1 Introduction
Development of an intelligent remote control system for street light and traffic signal
control system needed because of present traffic light controllers based on old micro-
controllers such as AT89C51 which has very less internal memory and no in-built
ADC. These systems have limitation because it will use the predefined program that
does not have the flexibility of modification on real-time application. The present
traffic system has a fixed time interval for a green and red signal, which does not
provide the flexibility to the system and street lighting system public sector are design
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 209
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_22
210 S. V. Darshane et al.
according to the old standards. The intelligent remote control system for street light
and traffic signal control system consists of high-performance, low cost, low power.
The system will deal with two basic problems: (i) Detection of traffic volume by
using genetic algorithm (ii) automatic control of street light using the sensor. The
traffic signal management is a very important facet in electronic equipment town
traffic system as we have a tendency to all glorious, traffic systems area unit time-
varying, random system. Thus lots of standard strategies for traffic signal manage-
ment primarily based precise models fail to deal efficiently with the advanced and
ranging traffic things. During all amongst in every of the most options of contem-
porary cities is that the permanent growth of population in a comparatively little
space. The consequence of this truth is that the increase within the range of cars and
additionally the requirement of movement and transport of individuals and product
in urban town networks. Traffic congestion in main road networks is one amongst the
most problems to be addressed by today’s traffic management schemes. Automation
combined with the increasing penetration of on-line communication, navigation,
and advanced driver help systems can ultimately lead to intelligent vehicle main
road systems (IVHS) that distribute intelligence between margin infrastructure and
vehicles which above all on the long term are one amongst the foremost promising
solutions to the traffic congestion drawback. During this paper, we have a tendency
to gift a survey on traffic management and management frameworks for IVHS. First,
we have a tendency to provide a short summary of the most presently used traffic
management strategies for freeways. Next, we have a tendency to discuss IVHS
based traffic management measures. Then, varied traffic management architectures
for IVHS like PATH, Dolphin, Auto21 CDS, etc. are mentioned, and a compar-
ison of the varied frameworks is conferred. Finally, we have a tendency to sketch;
however, existing traffic management methodologies may fit in Associate in Nursing
IVHS-based traffic management set-up.
Fast transportation system and rapid transit system are nerves of economic devel-
opment for the nation. All developed countries have a well-developed transportation
system with efficient traffic control on the road in, rail, and air transportation of good,
industrial products, human resources, and machinery are the key factors which influ-
ence the industrial development. Mismanagement and traffic congestion result in
long waiting time loss of fuel and money. It is therefore utmost necessary to have a
fast, economical and efficienttraffic control system for nation development.
2 Literature Review
In current system traffic get collected at particular places. Now days there are lack
of traffic analysis which results into heavy Traffic. Many time emergency services
are trapped. We refer various papers.
The author Mu et al. [1] describes A camera-based rule for period of time durable
stoplight detection and recognition was planned. This rule is supposed chiefly for
autonomous vehicles. Experiments show that our rule performs well in accurately
Involuntary Traffic Control System 211
investigating targets and in determinative the gap and time to those targets. However,
the current methodology planned here can have some drawbacks. First, the maneuver
performs well within the daytime but not additionally within the dead of night. The
warning rate can increase within the dead of night as results of lots of light-weight
interference. Whereas the maneuver can discover every circulartraffic light and other
people with arrows, exclusively the classical suspended, vertical traffic lights were
detected. Detection and recognition of lots of types of traffic lights will meet an
important house for future work.
Shu-Chung [2] proposed driver assistant system design supported image process
techniques. A camera is mounted on the vehicle front window to sight the road
lane markings and confirm the vehicle’s position with regard to the lane lines. A
changed approach is projected to accelerate the HT method during a computationally
economical manner, thereby creating it appropriate for time period lane detection.
The no heritable image sequences are analyzed and processed by the projected system
that mechanically detects the lane lines. The experimental results show that the system
works with success for lane line detection and lane departure prediction.
Alcantarilla [3] presents degree automatic road traffic management and looking
system for daytime sequences using a B&W camera. Necessary road traffic data
like mean speed, dimension and vehicles numeration are obtained practice laptop
computer vision methods. Firstly, moving objects are extracted from the scene by
suggests that of a frame-differencing algorithm and texture data supported grey scale
intensity. However, shadows of moving objects belong to boot to the foreground.
Shadows are far away from the foreground objects practice silk hat transformations
and morphological operators. Finally, objects are tracked in AN extremely Kalman
filtering technique, and parameters like position, dimensions, distance and speed of
moving objects are measured. Then, per these parameters moving objects are classi-
fied as vehicles (trucks or cars) or nuisance artifacts. For results mental representation,
a 3D model is projected onto vehicles among the image plane. Some experimental
results practice real outside sequences of images ar shown. These results demon-
strate the accuracy of the planned system to a lower place daytime interurban traffic
conditions.
Hussian [4] proposed system involves use of Wireless sensing element network
technology to sense presence of Traffic close to any circle or junction and so able to
route the Traffic supported Traffic handiness or we will say density in want direction.
this method doesn’t need any system in vehicles therefore are often enforced in
any Traffic system quite simply with less time and fewer pricey additionally. this
method uses Wireless sensing element networks Technology to sense vehicles and
a microcontroller based mostly routing formula programmed for wonderful Traffic
management [12].
Sinhmar [5] proposed system records vehicle count in its memory at user prede-
fined recording interval on real time basis. This recorded vehicle count information
is utilized in future to research traffic condition at various traffic lights connected to
the system. For acceptable analysis, the recorded information is downloaded to pc
through communication between microcontroller and also the computer. Adminis-
trator sitting on laptop will command system (microcontroller) to transfer recorded
212 S. V. Darshane et al.
3 System Architecture
4 Proposed Algorithm
5 Conclusion
Thus in this we have made an analysis on Automatic Traffic Control Signal in which
our motive is to reduce the Traffic on road in peak hours and make a way for emer-
gency situations.The method presented in this paper is simple and there is no need
to use sensors that have been commonly used to detect Traffic in the past. However,
one of the most important disadvantages of this method is extreme sensitivity to
light. For example, when installed in the road, changes in sun light potentially cause
interference with the camera. This problem can be overcome by using specific filters
during Image Processing or changes in Matlab code. With some improvements, this
method can be used to detect road accidents and identify violations of the spiral
movements of cars.
References
1. Mu G, Xinyu Z, Deyi L, Tianlei Z, Lifeng A (2015) Traffic light detection and recognition for
autonomous vehicles 22(1):50–56
2. Shu-Chung Y-C, Chang C-H (2015) A lane detection approach based on intelligent vision.
Comput Electr Eng 42:23–29
3. Alcantarilla PF, Sotelo MA, Bergasa LM (2008) Automatic daytime road traffic control and
monitoring system
4. Hussian R (2013) WSN applications: automated ıntelligent traffic control system using sensors.
J IJSC E 3(3)
5. Sinhmar P (2012) Intelligent traffic light and density control using Ir sensors and microcon-
troller 2(2)
6. Teodorescu H et al (1999) Fuzzy and neuro-fuzzy systems in medicine. CRC Press, Boca
Raton, FL, USA
7. Danescu R, Nedevschi S, Meinecke M, Graf T (2007) Stereo vision based vehicle tracking
in urban traffic environments. In: IEEE conference on ıntelligent transportation system, pp
400–404
8. Ferrier NJ, Rowe SM, Blake A (1994) Real-time traffic monitoring. In: Proceeding of the
second IEEE workshop on applications of computer vision, pp 81–88
9. Vu A, Barth M (2019) Catadiopetric omnidirectional vision system integration for vehicle-
based sensing. In: Proceeding of IEEE ıntelligent transportation system conference
10. Cao M, Vu A, Barth M. A novel omni-directional vision sensing technique for traffic
surveillance. In: Proceeding of IEEE
11. Jin B, Zou D, Gan Y (2010) Research and design of traffic detection based on GPRS. In: IEEE
conference on advanced computer control (ICACC)
12. https://www.researchgate.net/publication/337086901_Energy_Efficient_Routing_Prot
ocol_for_IOT_Based_Application
Early Detection of Diabetic Retinopathy
Using Machine Learning
Abstract Early detection of Diabetic Retinopathy shields patients from losing their
vision because Diabetic Retinopathy may be a typical eye disorder in diabetic
patients. The elemental explanation for a visual deficiency within the populace. Thus,
this paper proposes an automated method for image-based classification of diabetic
retinopathy. The technique is separated into three phases: image processing, feature
extraction, and image classification. The target is to naturally group the evaluation
of non-proliferative diabetic retinopathy at any retinal image. For that, an underlying
image preparing stage separates blood vessels, microaneurysms, and hard exudates,
so on extricate highlights utilized by a calculation to make sense of the retinopathy
grade.
1 Introduction
Diabetic Retinopathy (DR) may be a standout among the foremost successive reasons
for visual debilitation in created nations. The primary source of the latest instances of
visual deficiency within the working-age populace. By and huge, almost 75 individ-
uals go dazzled every day as an outcome of DR. A viable treatment for DR requires
early finding and consistent checking of diabetic patients, however, this can be a
testing undertaking because the malady indicates few manifestations until it’s past
the purpose where it’s possible to allow treatment.
As shown in above Fig. 1, Diabetic retinopathy is an eye issue that can cause
Visual deficiency—little veins within the back of the attention called retinal veins
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 215
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_23
216 V. V. Bandgar et al.
[1]. Indications of Diabetic Retinopathy are gliding spot in vision, obscured vision
and blocked vision. At the purpose when the sugar level in the blood builds, blood
vessels within the back of the attention finally end up frail, and seeable of this vessel
releases the blood and lipoproteins liquid, at the moment liquid winds up skimming
spot in vision with the goal that Diabetic patient cannot see anything totally through
the vision, within the event that we do not do the treatment of this ailment on the
time then it’d be conceivable of complete vision misfortune or visual deficiency [2].
On the off chance that we distinguished early the indication of Diabetic Retinopathy,
it’s conceivable to stay extra loss of vision.
to both patients further because of the resources of the healthcare system. The accu-
racy, sensitivity and specificity of the algorithm for detecting diabetic retinopathy
(DR) can help ophthalmologists, and physicians raise the red flag and thus provide
early treatment to patients and convey in exceedingly more preventive care which
might bring down the burden on a healthcare Machine learning can thus help the
old adage-prevention better than cure, by predicting who is more vulnerable to be in
danger of DR or not [5].
Four stages of Diabetic Retinopathy are as follows
The first stage is known as Mild Non-Proliferative Diabetic Retinopathy (Mild
NPDR). In this stage, there will expand like swelling in the veins in the retina and
little inflatable like swelling in the veins known as Microaneurysms. The second stage
is known as Moderate Nonproliferative Diabetic Retinopathy (Moderate NPDR). In
this stage, a portion of the veins in the retina will end up blocked. The third stage is
known as Severe Non-Proliferative Diabetic. This method detects and classifies the
diabetic retinopathy. Preliminary results show that k-nearest neighbours obtained the
best result with 68.7% for the dataset with different resolutions.
Advantages
Perform automated classification of Diabetic Retinopathy and component analysis
in less time.
Disadvantages
The paper does not include testing the methods with larger data sets and classifying
the subtypes of the retinopathy.
The author has mainly focused on the detection of Glaucoma and Diabetic
Retinopathy. Glaucoma can be detected by the cup to disc ratio (CDR). Diabetic
retinopathy can be detected by Exudates, Hemorrhages, Microneurysms and Cotton
Wool Spots. RGB images are converted into YCbCr. Y plane is used for detection of
blood vessels, optic disc and exudates. After candy edge detection, the image will be
converted into binary to perform Skeletonization operation. DCT is used for feature
extraction. The author has proposed DCT (Discrete Cosine Transform) for feature
extraction [6].
Baraman, “Automatic Exudate Detection from Non-dilated Extracted feature goes
to SVM classifier. After that Extracted feature goes to SVM classifier, after that
Diabetic Retinopathy-Retinal Images using Fuzzy C-means Clustering” [7].
3 System Architecture
Fig.2 Purposed systems for detection and classification of different stages of diabetic retinopathy
Early Detection of Diabetic Retinopathy Using Machine Learning 219
The Messidor database [9] includes 1200 eye fundus shading numerical photos of the
back shaft obtained by 3 ophthalmologic workplaces using a shading video 3CCD
camera on a Topcon TRC NW6 non-mydriatic retinograph with a 45° field of view.
The pictures were caught implementing 8 bits for each shading plane at 1440 × 960,
2240 × 1488 or 2304 × 1536 pixels.
800 images were acquired with pupil dilation (one drop of Tropicamide at
0.5%) and 400 without dilation. The 1200 pictures are bundled in 3 sets, one for
every ophthalmologic division, utilizing the TIFF position. What’s more, an Exceed
expectations document with therapeutic judgments for each picture is given.
In this work, we utilize the pictures of only one ophthalmologic division containing
152 pictures without retinopathy (grade 0), 30 with mellow NPDR (grade 1), 69 with
moderate NPDR (grade 2), and 149 with serious NPDR (grade 3).
The above Fig. 3 shows the flow of the project. It is divided into two parts Training
Phase and Testing Phase. It shows the steps performed during the implementation.
The image is preprocessed first, then from that feature extraction is done, and on the
basis of that, the image is classified. The same steps are performed for the test image.
The above Fig. 4 shows the steps and methods of the project. The activities performed
during implementation are Preprocessing, feature extraction and classification [10].
See Fig. 5.
4.2 Microaneurysms
See Fig. 6.
220 V. V. Bandgar et al.
See Fig. 7.
5 Conclusion
An efficient method for the detection of microaneurysms, hard exudate sand blood
veins has been presented. The classifier gives an average accuracy of 88%. We
conclude that image processing plays an important role in the diagnosis of DR. Future
works are to detect soft exudate to improve the accuracy of retinopathy detector.
The sensitivity and specificity of binary classification are 0.8930 and 0.9089,
respectively, which could be a satisfactory result. Furthermore, we developed an
automatic inspection app that may be utilized in both personal examination and
remote treatment. With more image data collected, we expect the accuracy is often
even more enhanced, further improving our system.
Early Detection of Diabetic Retinopathy Using Machine Learning 221
Fig. 7 Extraction of exudates (conversion from RGB to CMY and binarization operation)
References
1. https://algoanalytics.com/diabetic-retinopathy-machine-learning/
2. Panse ND, Ghorpade T, Jethani V (2007) Glaucoma and diabetic retinopathy diagnosis using
image mining. Int J Computer Appl 5 (May 2015). MPI Forum: Message Passing Interface.
https://www.mpi-forum.org
3. Anisur Rahman Khan (2013) 3.2 million people in Bangladesh suffer from diabetes,
Arrkhan.blogspot.com [Online]. Available: https://arrkhan.blogspot.com/2013/10/32-million-
people-in-bangladesh-suffer.html. Accessed: 01 Apr 2016
4. Decenci ere E, Zhang X, Cazuguel G, Lay B, Cochener B, Trone C, Gain P, Ordonez R, Massin
P, Erginay A, Charton B, Klein J-C (2014, Aug) Feedback on a publicly distributed database:
the Messidor database. Image Anal Stereol 33(3):231–234
5. Ong G, Ripley L, Newsom R, Cooper M, Casswell A (2004) Screening for sight-threatening
diabetic retinopathy: comparison of fundus photography with automated colour contrast
threshold test. Am J Ophthalmol 137(3):445–452
6. Sophark A, Uyyanonvara B, Baraman S (2007) In automatic exudate detection from non-
dilated diabetic retinopathy—retinal images using Fuzzy C-means clustering; Barney B (2007)
Introduction to parallel computing. Lawrence Livermore National Laboratory
Early Detection of Diabetic Retinopathy Using Machine Learning 223
7. Gurudath N, Celenk M, Riley HB. Machine learning identification of diabetic retinopathy from
fundus images. School of Electrical Engineering and Computer Science Stocker Center, Ohio
University Athens, OH 45701USA OpenMP, The OpenMP ARB. https://www.OpenMP.org
8. Zhang F (2010) Research on parallel computing performance visualization based on MPI.
International conference. IEEE explorer
9. Gandhi M, Dhanasekaran D (2013) Diagnosis of diabetic retinopathy using morphological
process and SVM classifier. Int Conf Commun Signal Process
10. Walter T, Klein JC, Massin P, Erginay A (2002, Oct) A contribution of image processing to
the diagnosis of diabetic retinopathy–detection of exudates in colour fundus images of the
human retina. IEEE Trans Med Imaging 21(10):1236–1243. In: Klepacki D, Watson TJ (eds)
Mixed-mode programming. Research Center presentations, IBM
The Effective Use of Deep Learning
Network with Software Framework
for Medical Healthcare
Padmanjali A Hagargi
1 Introduction
Humans learn from experience. The more experience you have, the more knowledge
you can learn. In the field of deep learning in the field of artificial intelligence (AI),
this principle is also the same, that is, machines powered by artificial intelligence
P. A. Hagargi (B)
Shri Vithal Education & Research Institute’s College of Engineering, Pandharpur, Maharashtra,
India
e-mail: [email protected]
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 225
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_24
226 P. A. Hagargi
software and hardware to learn knowledge from experience. These experiences for
the machine to learn from it are determined by the data collected by the machine,
and the amount and quality of the data determines the amount of knowledge that
the machine can learn [1]. Deep learning is a branch of machine learning. Many
traditional machine learning algorithms have limited learning ability. The increase
in data volume does not continuously increase the amount of knowledge learned.
The deep learning system can improve performance by accessing more data, that is,
the machine pronoun of “more experience.” Once the machine has gained enough
experience through deep learning, it can be used for specific tasks such as driving a
car, identifying weeds between field crops, confirming a disease, detecting machine
malfunctions, and more. The deep learning network learns by discovering the intricate
structure of empirical data. By building a computational model that contains multiple
processing layers, a deep learning network can create multiple levels of abstraction
layers to represent data. For example, a convolutional neural network deep learning
model can be trained using a large number (e.g., millions) of images, such as images
of these cats. This type of neural network typically learns from the pixels contained
in the acquired image. It can group the physical features of the cat in the image
(such as paws, ears, and eyes) and classify the pixels that represent these physical
features into groups. There is a fundamental difference between deep learning and
traditional machine learning. In this example, the domain expert takes a considerable
amount of time to engineer the traditional machine learning system to detect the
physical characteristics of a cat. For deep learning, only a very large number of cat
images need to be provided to the system, and the system can learn to form the
physical characteristics of the cat [2]. For many tasks (e.g., computer vision, speech
recognition, machine translation, androbotics), deep learning systems outperform
traditional machine learning systems. This is not to say that building a deep learning
system is much easier than building a traditional machine learning system. Although
feature recognition is performed autonomously in deep learning, we still need to
adjust thousands of hyper parameters (buttons) to ensure the validity of the deep
learning model [3].
2 Literature
Deep learning is effective in medical imaging because of the high quality of data
and the ability to classify images by convolutional neural networks. For example,
deep learning is comparable to dermatologists in skin cancer classification, and even
better. Several vendors have obtained FDA approval for the use of deep learning
algorithms for diagnostic purposes, including imaging analysis for oncology and
retinal diseases. By predicting medical events from electronic medical record data,
The Effective Use of Deep Learning Network with Software … 227
in-depth learning has helped to make significant progress in improving the quality
of healthcare [4, 5].
years, many experts have summarized, reviewed and discussed the research status and
problems of deep learning in medical image analysis. Recently, the review published
in Medical Image Analysis provides a comprehensive summary of deep learning in
medical image classification, detection and segmentation, registration and retrieval
[9].
3 Methodology
Neural networks are an important machine learning technique in the field of machine
learning and cognition. A mathematical model or computational model that mimics
the structure and function of a biological neural network to estimate or approximate
a function. The neural network is calculated by a large number of artificial neuronal
connections. In most cases, artificial neural networks can change the internal struc-
ture based on external information. It is an adaptive system in general, it has a learning
function [10]. A typical neural network has the following three parts: Architecture:
Structures specify variables in the network and their topological relationships. For
example, the variables in the neural network can be the weights of the neuron connec-
tions and the activities of the neurons. Activity Rule: Most neural network models
have a short time scale dynamics rule that defines how neurons change their stimulus
values based on the activity of other neurons. The general incentive function depends
on the weight in the network Learning Rule: Learning rules specify how weights in
the network adjust over time. This is generally seen as a dynamic rule for a long time
scale. In general, the learning rules depend on the excitation values of the neurons.
Usually a simple neural network consists of three parts: the input layer, the hidden
layer, and the output layer. Generally, the nodes of the input layer and the output layer
are fixed, and the hidden layer can be freely specified in Fig. 1 is a schematic diagram
of a simple neural network structure. The circles in the figure represent neurons, and
the arrows represent the flow of data. Corresponding to a different weight, the weight
is obtained by the network by learning itself [11].
Each neuron contains three parts: input, output, and computational model. The
neuron can be regarded as a calculation and storage unit. The calculation is the
calculation function of the input of the neuron. Storage is a neuron that temporarily
stores the result of the calculation and passes it to the next layer. Figure 2 shows a
typical neuron model with 3 inputs, 1 output, and 1 computational function [12].
Where x is the input, w is the weight, and a directed arrow indicating the connection
can be understood as follows: at the beginning, the transmitted signal is still a, and
there is a weighting parameter w in the middle, and the weighted signal becomes x.
w, so at the end of the connection, the size of the signal becomes x w. The weighted
signal summation at the neuron yields the output Z through the function f, as follows:
n
w i · xi
i=1
In fact, in order to achieve better results for neurons, each input is usually weighted
with an offset b, which is calculated as follows: The function f is also called the acti-
vation function. The common activation functions are sigmoid function, Hyperbolic
Tangent (tanh) function and ReLU function.
Deep Neural Network (DNN) DNN is a fully connected network with a very large
number of layers. The number of layers determines the ability of the neural network
to characterize data using fewer neurons per layer to fit more complex functions. Note
that there is no fixed definition of “depth” here. In the field of speech recognition,
the 4th layer is considered to be relatively deep; and the 20-layer model in the field
of image recognition is not uncommon.
230 P. A. Hagargi
MRI image A magnetic resonance image (MRI), which is a measure of the magnitude
of a magnetic resonance signal generated by a hydrogen atomic nucleus in human
tissues and organs and under an external strong magnetic field, and is received by a
computer for an in-.vitro nuclear magnetic resonance signal detector. The information
data is subjected to 3Dimage reconstruction. It provides very clear images of human
soft tissue anatomy and lesions as shown in Fig. 3 [14].
CT images computed tomography (CT) scans a section of a certain thickness of a
human body with a precisely collimated X-ray beam, and receives X-rays transmitted
through the section by a detector rotating with the ray beam, and finally the computer
reconstructs a 3D image of the corresponding human body section based on the X-
ray signal data received by the detector. It has a sub-millimeter spatial resolution
and provides clear human anatomical structure and lesion imaging [9]. It has been
widely used in a variety of clinical disease examinations and auxiliary diagnosis as
shown in Fig. 4 [16].
Medical X-ray images are electron density metric images of different tissues and
organs of the human body. X-ray based imaging including 2D computer radiography,
digital X-ray photography, digital subtraction angiography and mammography, and
The Effective Use of Deep Learning Network with Software … 231
Fig. 3 MR image
Fig. 4 Computed
tomography
3D spiral computed tomography, etc., have been widely used in orthopedics, lungs
Clinical disease detection and auxiliary diagnosis such as breast and cardiovascular,
but 2DX ray images cannot provide three-dimensional information of human tissues
and organs as shown in Fig. 5 [16].
Ultrasound imaging The ultrasound beam is used to scan the human body, and
the image of the internal organs is obtained by receiving and processing the reflected
signals. In recent years, ultrasound imaging technology has been continuously devel-
oped, and new ultrasound imaging technologies such as 3D color ultrasound, ultra-
sound holography, intracavitary ultrasound imaging, color Doppler imaging and
ultrasound biological microscopy have emerged as shown in Fig. 6 [16].
Positron emission tomography (PET) uses positron emission information
emitted by the tracer labeled with radioactive elements such as F18 to image [16].
Therefore, PET image is a measure of the activity of the corresponding tracer and can
provide tumor biology. Information (such as glucose metabolism, hypoxia, prolifer-
ation, etc.) information, its standard intake value can be used to clinically identify
the tumor good/malignant. PET provides more intuitive and accurate visualization
of biological and radiobiological properties than CT and MRI as shown in Fig. 7.
such a scientific computing tool chain. Regardless of the number of people in the
community or the level of activity, Keras is currently growing faster than Torch.
MXNet is an open source, lightweight, portable and flexible deep learning library
developed by DMLC (Distributed Machine Learning Community), which allows
users to mix symbolic programming patterns and instructional programming patterns
to maximize Efficiency and flexibility are now the deep learning framework recom-
mended by AWS. Many of MXNet’s authors are Chinese, and their biggest contribu-
tion is Baidu. At the same time, many authors come from deep learning projects such
as cxxnet, minerva and purine2. It is the first to support multiple GPUs and decen-
tralized in each framework, while its decentralized performance is also very high.
Its upper computational graph optimization algorithm allows symbolic computation
to be performed very quickly, and saves memory. Turning on the mirror mode saves
memory, and even training other frameworks on some small memory GPUs because
of insufficient video memory [15]. MXNet’s various levels of system architecture
(below the hardware and operating system bottom layer, layer by layer is more and
more abstract interface) as shown in Fig. 8.
AlexNet is a deep network that was first applied to ImageNet, and its accu-
racy is greatly improved compared to traditional methods. It starts with 5 convo-
lutional layers, followed by 3 fully connected layers, as shown in the Fig. 9: Alex
Krizhevs proposed AlexNet to use the ReLU activation function instead of the anh
or Sigmoid activation function used in the early days of traditional neural networks.
ReLU mathematics is expressed as:
f (x) = max(0, x)
The advantage of ReLU over Sigmoid is that it is faster to train because the
derivative of Sigmoid is very small in the stable zone, so the weight is basically
no longer updated [16]. This is the problem of gradient disappearance. So AlexNet
uses ReLU behind both the convolutional layer and the fully connected layer [10].
Another feature of AlexNet is that it reduces the overfitting of the model by adding
a dropout layer.
Dropout layer behind each fully connected layer [11]. The Dropout layer randomly
closes the neuron activation values in the current layer with a certain probability, as
shown in the Fig. 10.
DenseNet The structure has been developed by linking each layer to all other
layers in the style of forward feed, adopting the fact that there are shorter connections
between the intersecting networks near the input and those close to the output, which
can be deeper, more accurate and efficient to train. For each layer, the property
maps of all previous layers as input, their feature maps are also used as input to
all subsequent layers. DenseNets achieves significant improvements to the latest
technology in most areas, requiring less memory and calculation to achieve higher
performance [6]. In the study, the pretrained models were tested by making them
suitable for the classification task. Each model was operated with a maximum of
1812 iterations and a single GPU [11, 12]. The operating performances of the models
are as in Table I. Validation saccuracy as a model performance criterion has been
shown. According to the results, the highest validation accuracy was realized with
the DenseNet201 network. When the operating times are considered, DenseNet201
has the longest working time due to the number of parameters [16].
236 P. A. Hagargi
Python programming language that has the most popularity in recent times is a
Python. Python, in 1991 by Guido van Rossum, it is high level, an open source
general aim programming language. This dynamic programming language is hold
The Effective Use of Deep Learning Network with Software … 237
5 Conclusion
Using the best combination of programming languages and images attempts are made
to get best results for diagnosis. The field of medical image analysis has begun to
pay attention to the development of deep learning techniques on these key issues.
However, the transition from an artificial-based system to a system that learns features
from data are gradual reviewed the application of deep learning in medical image
analysis. Although they have done a lot of work, it seems to be flawed only from a
computer perspective. In recent years, new advances in deep learning techniques have
provided new insights into medical image analysis that allow only the morphological
and/or texture patterns in images to be found from the data. Although deep learning
techniques seem to have achieved the most advanced performance in many medical
application studies, few have shown that it can go beyond traditional methods. At
the same time, the current research on deep learning in the field of medical image
analysis is mainly based on technology, and the evaluation indicators used are also
evaluation indicators in the computer field. For any medical application, we would
prefer to see that the relevant technology is evaluated according to medical rules, for
example, through multi-center, randomized, controlled research methods to prove
that the technology has a more significant advantage than the previous technology.
Furthermore, an important challenge in medical image training is that the number of
training samples for most depth models is related to the number of learning param-
eters. Therefore, how to reduce over-fitting has always been a problem. When deep
learning training results are sent to new central applications, the models need to
migrate learning to maintain performance, which undoubtedly leads to low repro-
ducibility of clinical applications. Stability and reproducibility are the basic premise
that the technology can be widely used in clinical practice. Therefore, researchers
in the field of deep learning should pay attention to how the repeatability of the
algorithm in the prospective sample is guaranteed. In addition, the quality of the data
used for training is another cause of catastrophic results.
The Effective Use of Deep Learning Network with Software … 239
References
1. Arevalo J, González FA, Ramos-Pollán R, Oliveira JL, Lopez MA G.( 2016) Representation
learning for mammography mass lesion classification with convolutional neural networks.
Comput Methods Programs Biomed 127:248−257
2. Gao XT, Lin S, Wong TY (2015) Automatic feature learning to grade nuclear cataracts based
on deep learning. IEEE Trans Biomed Eng 62(11):2693–2701
3. Hagargi PA, Kumbhar S (2020) Combined effects of IoTs and medical sensors for effective
application of smart health care system. J Eng Comput Archit 10(4):11–19. ISSN NO: 1934-
7197
4. Hagargi PA, Danake PV, Gaikwad SN, Mahajan DS, Thite KG, Supekar PS (2020) Internate of
things based patient health monitoring system. J Aegaeum 8(4):221–228. ISSN NO: 0776-3808
5. Hagaragi PA, Shubhangi DC (2018) Brain tumor MR image fusion using most dominant
features extraction from wavelet and curvelet transforms. J Emerg Trends Innov Res (JETIR)
5(5):3–38. ISSN-2349-5162
6. Hagaragi PA, Shubhangi DC (20181) An efficient approach for brain tumor detection and
classification in multi-model MR brain images using RPCA and QT decomposition fusion and
ANN classification technique. Int Res J Eng Technol (IRJET) 5(5):26–40, e-ISSN: 2395-0056,
p-ISSN: 2395-0072
7. Hagaragi PA, Shubhangi DC (2018) Brain tumour detection and ART classification technique
in MR brain images using RPCA QT decomposition. Int Res J Eng Technol (IRJET) 5(4):1717–
1725, e-ISSN: 2395-0056, p-ISSN: 2395-0072
8. Hagaragi PA, Shubhangi DC (2017) A decomposition technique for image fusion to detect
brain tumor using T1, T2 weighted MR images. Int J Curr Eng Sci Res (IJCESR) 4(12):62–73
(Part-VII) https://doi.org/10.21276/ijcesr, o-ISSN: 2394-0697, p-ISSN: 2393-8374
240 P. A. Hagargi
9. Hagaragi PA, Shubhangi DC (2017) A survey on fusion based methods using MR image
analysis for disease diagnosis. Int J Eng Dev Re (IJEDR) 5(4):328–333. ISSN-2321-9939
10. Hagaragi PA, Shubhangi DC (2016) Brain tumor MR image fusion using wavelet and
curvelet transforms. In: Proceedings of 2nd IEEE international conference on engineering
and technology ICETECH-2016. IEEE, 978-1-4673-9916-6/16
11. Hagaragi PA, Shubhangi DC (2016) Multi model MRI brain image fusion technologies and
graph cut segment. Int J Eng Future Technol 1(1):55–65. ISSN 2455-6432
12. Hagaragi PA, Shubhangi DC (2014) Comparisons of different methods of image edge detection
with morphologic method. J Innov Comput Sci Eng 4(1)
13. Hagaragi PA, Shubhangi DC(2012) Image edge detection using morphological operation. Int
J Res Comput Appl Manag. ISSN-2231-1009
14. Fakoor R, Ladhak F, Nazi A, Huber M (2013) Using deep learning to enhance cancer diagnosis
and classification. In: International conference on machine learning. Omnipress
15. Lee LK, Liew S-C (2015) A survey of medical image processing tools. https://doi.org/10.13140/
RG.2.1.3364.4241. In: 4th international conference on software engineering and computer
systems, Kuantan, Malaysia
16. He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: IEEE
conference on computer vision and pattern recognition. IEEE Comput Soc 770–778
Detection of Brain Tumor Using Image
Processing and Neural Networks
1 Introduction
In the past few decades medical studies have adapted a lot of changes in their approach
toward treating a patient which has improved the standards of living and provided
comfort. Biomedical tools have been developed which are used by the doctors to
analyze and study about a particular disease or disorder. Even after so much thrive
there are some life taking afflictions like cancer, leukemia or brain tumor which if
not detected on time can be proved as fatal.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 241
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_25
242 V. Dhillo et al.
As per a survey taken by the National Brain Tumor Foundation (NBTF) the
death rate due to brain tumor has been increasing. The increase in deaths is caused
due to many causes like treatment not given on time, tumor not detected due to
poor image quality or contrast and many more. The idea of this project is to avoid
such situations and improve the quality of treatment. Machine learning has high
dimensional features which can definitely be utilized in such conditions. This will
not only help in increasing accuracy but also help the doctors to cure and recover the
patient within the required time. The target is to build a self-learning system which
scans the images and detects the tumor. Neural networks are nothing but a multilayer
architecture which is utilized for quantitative analysis. For example, the neurons in
our body they are a network of cells which perform functions as directed. The tasks
performed by neural networks is similar to the task of neurons.
2 Literature Survey
Image processing is being used in the medical field for a very long time and for many
tests. In order to detect the brain tumor doctors, suggest for tests like the positron
emission tomography (PET), compute tomography (CT) and excessively used is the
magnetic resonance image (MRI). MRI is preferred by the majority because it does
not emit harmful rays and has no side effects on the body on the other hand it gives
a precise view of the disorder. When it comes to image processing there are various
types of methods which are used to analyze or detect an image.
At the time of detecting any image many parameters are taken into consideration
like the color, contrast, pixels, intensity, boundaries, dimension etc. According to
these parameters any image can be categorized and the optimal required function
can be performed. In order to detect the brain tumor many processes were adapted
and with the development of technology many other processes and tests are being
developed [1]. Mentioned are some of the processes which have been used like
thresholding, region based, contour and shape, statistical based and machine learning
but these methods came with their own demerits. In the process of thresholding the
pixels separated and then grouped together on the basis of their intensity values [1].
For some cases, this method proved to be useful but, in some cases, where the tumor
is at its very first stage and the intensity or the color contrast of the lump formed
is very fade and the pixels could be grouped wrong showing no signs of tumor [2].
The second method in the list which was tested was the contour and shape method
in which boundaries and segments are detected in this the boundaries and segments
of the lump are detected but as the lumps are of irregular sizes and sometimes, they
are formed near the inner lining of the brain due to which boundaries could merge
and accuracy wasn’t gained [3]. For the third method which was the region-based
method groups of similar regions and made and then further detected. Last method in
the list was statistical based which was based on probable values, labels and optimal
Detection of Brain Tumor Using Image Processing … 243
distribution [4]. Machine learning when is further sub divided into categories in
which one of them is called as deep learning which will be used in this project. Used
methods, the approach, its merits and demerits will be discussed further.
3 Methodology
a universal set which includes everything or it can be a sub set which is dedicated
for a particular task [7]. To detect any image, it’s very important to train the network
and then assign proper functions for the required task. In order to train a network for
a set of images the pre-processing on the images necessary [8].
(2) Image Pre-processing
As mentioned earlier images can be classified into many sections and on the basis of
many parameters. In order to train a network all, the images which are being used or
which are being stored should be uniform [2]. Images can be divided on the basis of
RGB and Gray-scale. MRI scans are often in the gray scale format so all the images
stored in the folder should preferably be in the gray scale format. Dimensions of the
image should be uniform as well for which one should resize the image and change
its dimensions to ‘227 × 227’ which is the required dimension when a network is
being formed. Figure 2 shows the image pre-processing for RGB image hence it
shows ‘m × n × 3’ in which ‘m and n’ are the dimensions of the image [9].
3. Transfer Network
While we use deep learning, we often come around pre-trained or trained networks
which already have a defined data storage. For updating the network or add new data
a process is used which is called transfer network. In order to perform this process,
changes are made in the input layer and the layer just before the outcome. Another
important factor which should be taken into consideration is the memory manage-
ment [6]. The method of transfer network also helps in maintaining the storage. As
mentioned before neural networks are a self-learning system and it stores only the
required information in the dedicated folders and you don’t land up with double
images in the folder. When you compare Figs. 2 and 3 you can notice the changes in
networks. Figure 3 is nothing but the changes in the last layers of the same system
which is showed in Fig. 2. After you have completed the process of transferring the
network you have to again train the network with the new stored images.
(4) Network Training
The last task to perform is to train the network with the given data set and the
given commands. In order to detect the brain tumors, the network had to be trained
Detection of Brain Tumor Using Image Processing … 245
on numerous MRI scans and perform segmentation in order to detect the tumor
accurately. Tumors are nothing but lump like structures which are formed in the
brain. At times multiple tumors are developed within the brain. Human brain can
be divided into three main parts which is the cerebrum, the brain stem and the
cerebellum. After which it is further sub-divided into hemispheres which are the
frontal lobe, parietal lobe, temporal lobe and the occipital lobe. These lobes working
conjunction to maintain the functions of the body and each lobe carries out multiple
functions. Tumor can be formed in any of these lobes. The network can be trained
for detecting the tumor and also detect in which part of the lobe it is formed. The
input given has to be a data set of multiple scans and the region of the lobes has to
be defined.
4 Results
When compared to other methods listed above like manual inspection, digital photog-
raphy and automated analyzer the use of neural networks in combination with data
sciences has proved to provide more accuracy with a greater speed. The reason why
this method works exceptionally is because of its trained network. CNN (convo-
lutional neural network) is a mathematical process mainly used for image analysis
and has an adaptive behavior. The network gets trained over a thousand pictures and
improves its accuracy after each run. One can not only compare its speed or accu-
racy but also the how it maintains the data set and memory. The intension of this
project was to obtain results similar to the results given by a specialist. There are soft-
wares available to perform deep learning but the most feasible would be MATLAB
as there are inbuilt functions available like the image processing tool box and the
deep learning tool box. One can practice or test functions on MATLAB’s pre-trained
networks such as ‘alexnet’. The circled part in Figs. 4 and 5 is the tumor which was
246 V. Dhillo et al.
detected after segmentation. The results conclude that with the help of this system
single as well as multiple tumors can be detected with greater accuracy.
5 Conclusion
Image processing has found its way in the healthcare stream and will continue to grow.
There is a broad perspective of using image processing for many other tests as well
like detecting the hemoglobin, WBC and RBC in the blood. This method can also be
used for retinopathy which can detect diabetes and detect cataract as well. Adaptive
methods like this have proved to be useful and can be further developed as well.
With advancing technology and systems in industries the style of living is changing.
These new systems can be a little complicated to implement when compare to the
earlier methods but are more effective and life changing. Biomedical imaging hand
in hand with data science will not only make the process simpler but systematic as
Detection of Brain Tumor Using Image Processing … 247
well. Reduction of errors, less human efforts and advancement in the life expectancy
due to early detection are the aims to be achieved to further which we can add a more
digitalized system.
References
1. Abdalla HEM, Esmail MY (2018) Brain tumor detection by using artificial neural network.
In: International conference on computer, control, electrical, and electronics engineering
(ICCCEEE), Khartoum, pp 1–6. https://doi.org/10.1109/ICCCEEE.2018.8515763
2. Fathurachman M, Kalsum U, Safitri N, Utomo CP (2014) Heart disease diagnosis using extreme
learning based neural networks. In: International conference of advanced informatics: concept,
theory and application (ICAICTA), Bandung, pp 23−27. https://doi.org/10.1109/ICAICTA.
2014.7005909
3. Abdullah AA, Chize BS, Nishio Y (2012) Implementation of an improved cellular neural network
algorithm for brain tumor detection. In: International conference on biomedical engineering
(ICoBE), Penang, pp 611–615. https://doi.org/10.1109/ICoBE.2012.6178990
4. Jemimma TA, Vetharaj YJ (2018) Watershed algorithm based DAPP features for brain tumor
segmentation and classification. In: International conference on smart systems and inventive
technology (ICSSIT), Tirunelveli, India, pp 155–158. https://doi.org/10.1109/ICSSIT.2018.874
8436
5. Yazid MHA, Satria MH, Talib S, Azman N (2018) Artificial neural network parameter tuning
framework for heart disease classification. In: 5th International conference on electrical engi-
neering, computer science and informatics (EECSI), Malang, Indonesia, pp 674–679. https://
doi.org/10.1109/EECSI.2018.8752821
6. Ai H, Fang M, Lvov YM, Mills DK, Jones SA (2002) Applications of the electrostatic layer-
by-layer self-assembly technique in biomedical engineering. In: Proceedings of the second joint
24th annual conference and the annual fall meeting of the biomedical engineering society]
[engineering in medicine and biology, Houston, TX, USA, vol 1, pp 502–503. https://doi.org/
10.1109/IEMBS.2002.1136916
7. Joshi S, Borse M (2016) Detection and prediction of diabetes mellitus using back-propagation
neural network. In: International conference on micro-electronics and telecommunication
engineering (ICMETE), Ghaziabad, pp 110–113. https://doi.org/10.1109/ICMETE.2016.11
8. Lekha S, Suchetha M (2018) Real-time non-invasive detection and classification of diabetes
using modified convolution neural network. IEEE J Biomed Health Inform 22(5):1630–1636.
https://doi.org/10.1109/JBHI.2017.2757510
9. Banowati C, Novianty A, Setianingsih C (2019) Cholesterol level detection based on iris recogni-
tion using convolutional neural network method. In: IEEE Conference on sustainable utilization
and development in engineering and technologies (CSUDET), Penang, Malaysia, pp 116–121.
https://doi.org/10.1109/CSUDET47057.2019.9214690
Automatic Guided System for Image
Resolution Enhancement
Abstract High-resolution images decide the performance of the system. Image reso-
lution enhancement is the process to extend no of pixels. During this, we increase
the quality of the image to form it suitable for all applications. Within the proposed
method, we’ve used a combination of DWT, SWT and Interpolation, i.e. combina-
tion of 3 traditional methods. DWT has edge loss problem, so alongside it, SWT
is employed to beat edge loss problem. Initially, DWT of a low-resolution image is
taken. The LL band of this is often interpolated by a factor of two and it’s combined
with SWT of the input low-resolution image. Again inverse DWT is taken of the
resultant image.
1 Introduction
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 249
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_26
250 N. P. Kulkarni et al.
Characteristics of the image and to restrain some useless attributes of the image.
The image, which has been improved, can satisfy some unique analysis better than
the primary one. It’s the method of manipulating a picture in order that resultant
image is more suitable than the first one for a specific application.
2 Motivation
In recent years there’s an increase in the demand for better quality images within
the various applications like medical, astronomy, visual perception. Image resolu-
tion improvement is to boot wide helpful for satellite image applications that embody
building construction, bridge recognition, in GPS technique. For image improvement
method, there a time-invariant transform. To restore the interpretation invariance is to
average some slightly different DWT, called decimated DWT, to define the stationary
wavelet transform (SWT). The Stationary wavelet transform (SWT) could also be a
wavelet transform algorithm designed to beat the shortage of translation-invariance
of the discrete wavelet transform (DWT). Translation invariance is achieved by
removing the down samplers and up samplers within the DWT and upsampling
the filter coefficients. The SWT is an inherently redundant scheme because the
output of every level of SWT contains an equivalent number of samples because
the input, so for the decomposition of N levels there’s a redundancy of N within the
wavelet coefficients. Hasan Demirel and Gholamreza Anbarjafari propose a picture
resolution enhancement technique [1] supported interpolation of the high-frequency
sub-band images obtained by discrete wavelet transform (DWT) and therefore, the
input image. The sides are enhanced by introducing an intermediate stage by using
stationary wavelet transform (SWT). During this correspondence, one level DWT
[2] (with Daubechies 9/7 as wavelet function) is employed to decompose an input
image into different sub-band images.
Another transform is the stationary wavelet transform (SWT) [3]. The Discrete
Wavelet Transform isn’t a time-invariant transform. To restore the interpretation
invariance is to average some slightly different DWT, called decimated DWT, to
define the stationary wavelet transform (SWT). In DWT [4] basic computational,
the step may be a convolution followed by decimation. The decimation retains even
indexed elements. But the decimation might be administered by choosing odd indexed
elements rather than even indexed elements. This selection issues each step of the
decomposition method, and thus at each level, we tend to selected oddly. It will thus
by suppressing the down-sampling step of the decimated algorithmic program and
instead up-sampling the filters by inserting zeroes between the filter coefficients.
Automatic Guided System for Image Resolution Enhancement 251
3 Method Adopted
The planned technique uses a combination of DWT, SWT and Interpolation to rein-
force low-resolution image to high-resolution image. The brief steps within the
method are as given below. The proposed method is as shown in Fig. 1.
Step 1: Take standard benchmark high-resolution image (say 512 × 512).
Step 2: Downsample the high-resolution image 2 times by using DWT. During
this project, the discrete wavelet transform algorithm is meant using Haar [5] wavelet
function for DWT decomposition and IDWT. An obtained low-resolution image is
of size 128 × 128.
High-Resolution
Low-
resolut
ion
SWT
DWT
LL LH HL HH
LL LH HL H
IDWT
High-
resolution
The proposed method has been tested with a standard set of test grey images and
colour images. The image enhancement is carried out in the following steps.
1. As shown in Fig. 2 the high-resolution standard benchmark input images (size
512 × 512) are taken.
2. As shown in Fig. 3, these high-resolution input images are downsampled to
128 × 128 and low-resolution images are obtained.
3. As shown in Fig. 4, this low-resolution image is enhanced to 512 × 512 by the
proposed method. First, take level 2d DWT of low-resolution image.
4. Then take 1 level 2d SWT of the low-resolution image. It is as shown in Fig. 5.
Automatic Guided System for Image Resolution Enhancement 253
Fig. 2 High-resolution
image
Fig. 5 2D SWT of a
down-sampled
low-resolution image
Automatic Guided System for Image Resolution Enhancement 255
Fig. 6 Enhanced
high-resolution resolution
Proposed work gives the better result as compared to traditional methods. The
performance is calculated by calculating Peak Signal to Noise ratio (PSNR) as given
in Eq. (1) Signal to Noise Ratio (PSNR) is used to analyse the quality of image,
sound and video files in dB (decibels). The original image and resultant image are
having different PSNR values.
Table 1 shows that PSNR value has been improved by using the proposed method
as compared to benchmark methods.
5 Conclusion
tested for colour images and give higher results. This is a good method for image
resolution enhancement.
References
1. Demirel H, Anbarjafari G (2010) Satellite image resolution enhancement using complex wavelet
transform. IEEE Geosci Remote Sens Lett 7(1):123–126
2. Zhao S, Han H, Peng S (2003) Wavelet domain HMT-based super image resolution. In:
Proceedings IEEE international conference image processing, vol 2, pp 933–936
3. Demirel H, Anbarjafari G (2011) Image resolution enhancement by using discrete and stationary
wavelet decomposition. IEEE Trans Image Process 20(5):1458–1460
4. Demirel H, Anbarjafari G, Izadpanahi S (2009) Improved motion-based localized super-
resolution technique using discrete wavelet transform for low-resolution video enhancement. In:
Proceedings 17th European signal processing conference, Glasgow Scotland, pp 1097−1101
5. Temizel A, Vlachos T (2005) Wavelet domain image resolution enhancement using cycle-
spinning. Electron Lett 41(3):119–121
6. Phanindra P, Chinna Babu J, Usha Shree V (2013) VLSI implementation of medical image
fusion using Haar transform. Int J Sci Eng Res 4(9). ISSN 2229-5518
7. Anbarjafari G, Demirel H (2010) Image super resolution based on interpolation of wavelet
domain high frequency sub-bands and the spatial domain input image. ETRI J 32(3):390–394
Residual Network for Face Progression
and Regression
Abstract In computer vision application, the style transfer is a most active area,
where deep generative networks have been used to achieve desired results. The
development of adversarial networks training produces a high-quality image result in
terms of face age progression and regression that is face aging and de-aging. Inspired
by Ian Goodfellow, in this paper, we have designed the combinational network using
the residual block, convolution and transpose convolutional in CycleGAN for face
age progression and regression. Face aging is an image to image translation concept
which is used in many applications such as cross-age verification and recognition,
entertainment, in smart devices like biometric system for verification purpose etc.
The proposed architecture preserves the original identity as it is and converts young
people to old and vice versa. The network consists of residual blocks to extract
deep features. The UTKFace unpaired image dataset is used to do experiments. The
qualitative analysis of proposed methods in terms of performance metrics which
gives better results. The performance metrics calculated such as Mean Squared Error
(MSE), Root Mean Squared Error (RMSE) and Structured Similarity Index (SSIM)
to the quality of image.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 257
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_27
258 D. V. Atkale et al.
1 Introduction
The sketch of ‘Afghan Girl’ earned global recognition when it was highlighted on the
cover page of National Geographic Magazine in 1985, considering that the person in
imaginary which remains unidentified for next few years until she was found in 2002
[1]. In such thousands of similar examples of finding the lost person, the key to find
the missing person is an old photo of that person. Recently, deep learning models
are widely used. The reason behind this is that deep learning models give faster and
more precise results as well as computational complexity is less compared to existing
machine learning techniques. To work on face age which is helpful in the industry
as well as academics. Consider an example of movies, for the post-production event
many of the actors were reworked either for the beautification of face or texture is
prepared as per the expectations of the producer. More precisely, face aging and
de-aging is nothing but effects generated by applying different cosmetics on the face
that is makeup or some lighting effects.
Face aging is the process which changes according to the hormonal (gens) and
environment throughout our life. The main goal of face aging is to transform age of
the photo based on the ground truth image or conditions given by users. Face Aging
is used for many applications such as entertainment [2], finding missing persons
[3], cross-age face recognition [4], etc. Recently many researchers have worked
on face aging problems and designed the number of their models. Still, the face
aging is a challenging problem in terms of accuracy, performance measures, cross-
age verification and recognition, identity missing, etc. Due to the limitations of the
feature representation, classification and illumination variation, we are going to use
the deep generative network. In this paper, we designed a deep generative adversarial
network for face aging. The proposed network consists of an encoder and decoder
concept using residual blocks. Next section is introduced to get a brief idea about
the existing approaches for face age progression and regression.
2 Literature Survey
Traditional face aging has two approaches prototype model and physical model:
prototype model determines the average of a face age group and to construct the aged
face the difference between the average faces taken to generate an age a pattern. This
model generates the unrealistic face images the information would be lost [3–5]. The
physical model deals with facial structure and shape. Aging is done by changing the
muscles, wrinkles, size of chin expanded, hair color etc. The restrictions of this model
require a large number of paired datasets during training, and the computational rate
is too high [6, 7].
Later, for face age progression, deep neural networks were used for feature extrac-
tion. As compared to previous methods, automatically the features were extracted
which gives better aging effects using deep neural network algorithms such as Deep
Residual Network for Face Progression and Regression 259
Restricted Boltzmann Machine [8, 9], Recurrent Neural Network [2], etc. Still the
originality, age information missing and at the time of training the same person many
face images with different ages.
Afterwards, Generative Adversarial Networks (GANs) [10], works successively in
different tasks such as image synthesis [11–14], image super resolution [15], image-
image translation [16–18], image inpainting [19]. The GAN models give realistic
images with aging effects. However, GANs still have instability during training. In
conditional generative adversarial networks (cGANs) [11] the generator learns based
on conditions such as information related to the labels [20] used for the multi-domain
image to image translation and language descriptions [21] in case of text to image
synthesis, etc. But the discriminator was unable to distinguish between a real face
and generated face.
Later to overcomes on this issue the identity-preserved conditional generative
adversarial network (IPCGANs) [22] comes in existence, they were added the percep-
tual loss [23] to keep original information, this model is useful in case of the paired
image dataset. The conditional adversarial auto-encoder (CAAE) [24] first time
successfully generates face age progression and regression using unpaired image
dataset.
It has the least learning ability of generators as well discriminator due to that only
rough wrinkles were added on the face. Recently, the style transfer [25] concept is
a famous research area where the image is translated from one domain to another
domain. The cycle consistent adversarial network (Cycle GAN) [26] which consists
of pairs of generator and discriminator network to generate targeted age face image.
This CycleGAN has achieved better performance in image generation tasks. We
inspired from this network, and we have designed our model, the details of the
proposed method are given below.
The drawback of only residual blocks, it creates the artifacts, so we have designed
a combinational network.
The combinational network consists of three blocks residual, convolution and
Transpose convolution to improve the quality of the generated image; the quality is
checked by using performance metrics.
3 Proposed Approach
The overall block diagram of the proposed approach shown in Fig. 1. It consists of
several blocks such as image dataset to call the database images for further processing,
the task of preprocessing unit is to remove the unwanted noise from the image, or it
enhances some features of image which is required for further image analysis and
processing. The pre-processing based on the requirements such as rescaling, reshape
etc., network has mainly two sub-parts generator and discriminator. The generator
network generates the images based on the loss functions definition, whereas the
discriminator checks whether the generated image is real or fake. Training means
there is learning relation data and attributes out the whole training dataset some
images were considered. The training block is used to train the system neurons
based on the user’s conditions, at last, a testing block which helps to check trained
module performance by applying the trained module weights to any external image
or any image from the same dataset.
Fig. 2 The generator network architecture which consists of three blocks such as Conv2D, Residual
and Transpose Conv2D
Linear Unit (Relu). The convolutional layer extracts the features; residual block is
used for skip connection to reduce the complexity of the network. The activation
function used in the generator network is Rectified Linear Unit (Relu), which helps
to improve our results. The residual block has stride 2 for down-sampling, the stride
is nothing but the number of pixels by which the window moves after each operation
at the time of training the data.
262 D. V. Atkale et al.
This section consists of summaries of the network object functions. Basically, the
proposed network is based on the generative adversarial networks (GANs), which
consists of a generator (G) and discriminator (D). This basic network is trained using
the min–max game equation as,
Ex∼pdata log logD(x) + Ez∼p(z) log log(1 − D(G(Z))) (1)
where z represents the noise vector, the first part of the equation gives information
related to the original dataset images are identical or not whereas the second part
tells about the generator images are real or fake, this all decisions were given by
discriminator network.
The loss functions used in the proposed approach as Adversarial loss function,
LGAN G, Dy, X , Y = 0.5 ∗ Ex∼Pdata(y) DY (y) + Ex∼Pdata(x) [1 − DY (G(x))] (2)
The X is related to young people, Y for old people, G generator and DY differ-
entiates between the original image and generated image based on age domain that
is young or old. Cyclic Loss function, the cyclic Loss is used to convert one domain
Residual Network for Face Progression and Regression 263
image to another domain that is in face aging. The young age group is converted to
the older age category and vice versa.
The Instance Normalization and layer normalization operations are similar to each
other. The layer normalization is used to normalize the input features, whereas the
Instance Normalization is used for normalization through each channel in training
example. Mostly, the normalization operation uses the mean and standard deviation
formulas. The Instance Normalization is expressed as,
1 1
W H W H
xtijk − μti
ytijk = , μti = xtilm , σti2 = (xtilm − muti )2 (3)
σti + ε
2 HW HW
l=1 m=1 l=1 m=1
The overall experiments have been conducted on UTKFace dataset. All the exper-
iments performed on Google Colab platform with Tesla P100-PCIE-16 GB GPU
machine. The results of the proposed approach are as shown in Fig. 3. It has divided
into two age groups as (20–30) as young people and (50–60) as old people age group
and conversion of them shown in results. The results are shown in two categories first
where we are transforming (20–30) age group to (50–60) and again reconstructed it
back to its original age group that is (20–30) this our young people face aging and
de-aging. Secondly, the (50–60) age category is converted into (20–30) and trans-
forming it back to its original age group that is (50–60) which an old to young face
image generation. In both the cases face aging and de-aging operation takes place.
4.1 Dataset
The UTKFace dataset is used, which has around 20 k face images. The age ranging
from 0 to 116 years of different races and genders. The dataset image resolution is
256 × 256 pixel. In the proposed approach, we are only concerned about the two age
categories as the young age group is (20–30) and older (50–60) remaining images
are filtered out.
Mean squared Error (MSE) is calculated by performing the arithmetic operation that
is the difference between the real image pixel and predicted image pixel. Lower the
value of MSE denotes higher the quality of image. The value of MSE is noted in
264 D. V. Atkale et al.
Fig. 3 The results of our combinational network, which consists of two age group (20–30) and
(50–60). The first row of the table indicates the face aging of young to old that is (20–30) age
category face image translated to (50–60) and again reconstructed it back to its original age group.
Similarly, the second row, old to young where (50–60) age group converted to (20–30) age category
and again reconstructed it back to its original form that is (50–60) age group
Table 2 MSE is calculated by using the observed value and predicted for an N number
of data points as,
1
N
2
MSE = Yi − Ỹi (4)
N i=1
Residual Network for Face Progression and Regression 265
Table 2 The quantitative analysis of the proposed method by comparing the original image and
reconstructed image
Root Mean squared Error (RMSE) which is similar to the MSE only the root has
been taken to get the accurate results. The values of RMSE are given in Table 2. The
formula is as given below,
(1/2)
1
N
RMSE = MSE = Yi −Yi )2 (5)
N i=1
Structured Similarity Index (SSIM) is the similarity distribution between the real
image and reconstructed image. The SSIM range is from 0 to 1. SSIM is one of
the best ways to check the quality of an image. The SSIM is represented in Table
2. The formula for calculating the SSIM between the two images that is original is
considered as A and reconstructed image as B given below,
(2μA μB + C1 )(2σAB + C2 )
SSIM (A, B) = 2 (6)
μ A + μ2 B + C1 σ 2 A + σ 2 B + C2
266 D. V. Atkale et al.
5 Conclusions
Acknowledgements The authors wish to thank the department of Electronics and Telecommu-
nication Engineering, SVERIs College of Engineering Pandharpur for the support during this
research.
References
13. Makhzani A, Shlens J, Jaitly N, Goodfellow I, Frey B (2015) Adversarial autoencoders. arXiv
preprint arXiv:1511.05644
14. Mirza M, Osindero S (2014) Conditional generative adversarial nets, arXiv preprint arXiv:
1411.1784
15. Ledig C et al (2017) Photo-realistic single image super-resolution using a generative adversarial
network. Proc IEEE Conf Comput Vision Pattern Recogn
16. Huang X et al (2018) Multimodal unsupervised image-to-image translation. Proc Eur Conf
Comput Vision (ECCV)
17. Kim T et al (2017) Learning to discover cross-domain relations with generative adversarial
networks. Proc 34th Int Conf Mach Learn 70. JMLR. Org
18. Liu M-Y, Breuel T, Kautz J (2017) Unsupervised image-to-image translation networks. Adv
Neural Inf Process Syst
19. Yeh R et al (2017) Semantic image in painting with perceptual and contextual losses. arXiv
preprint arXiv:1607.07539
20. Choi Y et al (2018) Stargan: unified generative adversarial networks for multi-domain image-
to-image translation. Proc IEEE Conf Comput Vision Pattern Recogn
21. Zhang H et al (2017) Stackgan: text to photo-realistic image synthesis with stacked generative
adversarial networks. Proc IEEE Int Conf Comput Vision
22. Wang Z et al (2018) Face aging with identity-preserved conditional generative adversarial
networks. Proc IEEE Conf Comput Vision Pattern Recogn
23. Dosovitskiy A, Brox T (2016) Generating images with perceptual similarity metrics based on
deep networks. Adv Neural Inf Process Syst
24. Zhang Z, Song Y, Qi H (2017) Age progression/regression by conditional adversarial
autoencoder. Proc IEEE Conf Comput Vision Pattern Recogn
25. Zhang S et al (2019) Stylistic scene enhancement GAN: mixed stylistic enhancement generation
for 3D indoor scenes. Vis Comput 35(6–8):1157–1169
26. Zhu J-Y et al (2017) Unpaired image-to-image translation using cycle-consistent adversarial
networks. Proc IEEE Int Conf Comput Vision
27. Thengane VG et al (2018) Cycle face aging generative adversarial networks. In: 2018 IEEE
13th international conference on industrial and information systems (ICIIS). IEEE.
Design and Simulation of 2-Element,
Circular Shaped MIMO Antenna
for C-Band Frequencies
Abstract This paper summarizes the novel approach of designing MIMO antenna
working at dual-band frequencies with wide bandwidth. The 2-Element, circular-
shaped MIMO antenna proposed here is operating at C-band frequencies which
can be used for satellite communication, cellular communication such as 4G and
5G, Wi-MAX, and WBAN wearable applications. Two circular shaped microstrip
patches fed with the tapered feed line, are placed diagonally opposite of each other.
The designed MIMO antenna resonates at 3.1 and 6.9 GHz frequencies. The first
operating band is from 2.83 to 5.58 GHz with a maximum bandwidth of 2.75 GHz
(88.6% Bandwidth) and the second operating band is from 6.19 to 7.95 GHz with
a maximum bandwidth of 1.76 GHz (29% Bandwidth). Reflection Coefficients at
those frequencies are less than −10 dB and VSWR value is between 1 and 2 which is
necessary and sufficient condition for any antenna to radiate efficiently. The radiation
efficiency of this antenna is 58% and observed directivity is 3.4 dBi.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 269
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_28
270 A. Jadhav et al.
1 Introduction
2 Literature Review
Details of work carried out by other researchers along with methodology and
outcomes are compared with the work proposed here; given in Table 1.
In [1], the design of Wideband GNSS Antennas has been reported for the Global
Positioning System (GPS) antennas. RT6010 substrate with εr = 10.2 and loss tan δ
= 0.0023 is used as substrate material. Bandwidth enhancement is done up to 38%
(1.16–1.61 GHz).
In [2], Integrated Duplexing and Multi-Band Filtering are implemented for
triple-band frequency (2.46–2.56, 3.45–3.56, and 5.01–5.26 GHz). The reflection
coefficients for all those three bands of frequencies are between −20 to −30 dB.
Design of Vertical Metallic Strip Pairs Feeding Structure is described in [3], in
which Rogers RT6010 and Rogers RO4003 are used as substrate materials and it is
designed to radiate at Five frequencies (2.4, 3.5, 4.1, 4.8, 5.2 GHz).
The novel triple-band Rectifier antenna reported in [4], in which FR-4 substrate
with εr = 4.4, δ = 0.02, and a thickness of 1.6 mm is used. It is designed for
Energy Harvesting Applications in GSM-900, GSM-1800, and Wi-Fi-2450 having
dimension 54 mm × 42 mm.
To improve the bandwidth of the antenna, Defected Ground Structure is used in
[5], due to which bandwidth enhancement is achieved up to 27.2, 21.2, and 17.5%.
In [6], Triple-Band Monopolar Antenna is designed using Rogers Duroid 5880LZ
εr = 2.0 for Triple Band: 2.51, 3.56, and 4.62 GHz. A circular patch with a diameter
Table 1 Comparison of this proposed work with existing results
References Technique Substrate Operating Overall S-parameters Bandwidth/% Reference VSWR Applications
and εr frequency band(s) dimensions Bandwidth impedance mentioned
(mm)
[1] Wideband 1.27 mm 1.16–1.61 GHz 112 × 112 Return loss 1.16–1.61 GHz 25–50 NA Global
GNSS RT6010 × 30 mm better than (38%) Navigation
antenna substrate 10 dB Satellite
with εr = Systems
10.2 (GNSS)
antennas
[2] Integrated εr = 2.55 2.46–2.56, 45 × 50 × −20 to − 5.7 and 10.4% NA NA NA
duplexing 3.45–3.56, and 0.8 mm 30 dB
filtering 5.01–5.26 GHz
[3] Metallic Rogers Five frequencies 30 mm × Up to −20 dB 150, 360, 150, 50, NA NA WLAN,
strip pairs RT6010 and (2.4, 3.5, 4.1, 4.8, 30 mm × 55 MHz WiMAX, and
feeding Rogers 5.2 GHz) 9.813 mm 5G
structure RO4003
[4] Triple-band FR-4 with εr 0.9, 1.8, and 54 mm × −20 to − NA 50 NA Energy
rectifier = 4.4, δ = 2.45 GHz 42 mm 30 dB Harvesting
0.02 and Applications
thickness of GSM-900,
1.6 mm GSM-1800,
and
Design and Simulation of 2-Element, Circular Shaped MIMO Antenna …
Wi-Fi-2450
[5] Defected FR4 with 7.317–9.58, 144 mm × Below − 27.2, 21.2, 17.5% NA NA X, Ku, and
ground thickness 25.45–31.4, and 144 mm 10 dB Ka bands
structure 0.3 mm 35.37–42.167 GHz
(continued)
271
Table 1 (continued)
272
References Technique Substrate Operating Overall S-parameters Bandwidth/% Reference VSWR Applications
and εr frequency band(s) dimensions Bandwidth impedance mentioned
(mm)
[6] Triple-band Rogers Triple Band: 2.51, Circular Below − 22.80% 50 1–2 Not
monopolar Duroid 3.56, and 4.62 GHz patch 10 dB mentioned
antenna 5880LZ εr diameter of
= 2.0 150 mm
[7] Parasitic 1.0 mm 960,1860, and 118 mm × Return loss 25.6% for NA 3:1 Wi-Fi, RFID,
shorted strip thick FR-4 2590 MHz 42 mm less than − 850–1100 MHz, and LTE
substrate εr 6 dB 13.3% for
= 4.4 1750–2000 MHz
and 22.7% for
2190–2750 MHz
This Work 2 elements, FR-4 3.1 and 6.9 GHz 55 mm × −15 dB 88.6% 56 1.03 4G and 5G,
circular substrate 80 mm Bandwidth (2.83 Wi-MAX
shaped εr = 4.4 to 5.58 GHz with and WBAN
MIMO a maximum wearable
antenna bandwidth of applications
2.75 GHz), and
29% Bandwidth
(for 6.19 to
7.95 GHz with a
maximum
bandwidth of
1.76 GHz)
A. Jadhav et al.
Design and Simulation of 2-Element, Circular Shaped MIMO Antenna … 273
of 150 mm is used to get the reflection coefficient Below −10 dB and bandwidth of
22.8%.
In [7], Parasitic Shorted Strip is designed for 960, 1860, and 2590 MHz using
1.0 mm thick FR-4 substrate εr = 4.4 with dimensions of 118 mm × 42 mm. This
antenna has a return loss of less than −6 dB and bandwidth 25.6% for 850–1100 MHz,
13.3% for 1750–2000 MHz, and 22.7% for 2190–2750 MHz which can be used for
Wi-Fi, RFID, and LTE.
3 Methodology
Fig. 1 Design of 2-element, circular shaped MIMO antenna. a Design of 2-elements, b parameters
274 A. Jadhav et al.
The proposed antenna is designed with EM solver software and the results obtained
for various parameters are plotted. Many antenna parameters can be identified to
improve. But here the focus is mainly on scattering parameter, bandwidth, VSWR,
and reference impedance. Scattering parameters are the performance metric of any
antenna. For the MIMO antenna using 2 Patches, scattering parameters will be S11,
S12, S21, and S22.
S11 is an indication of the reflection coefficient of patch-1. At 3.1 GHz frequency,
the value of the observed S11 parameter is −11.04 dB and At 6.9 GHz frequency,
the value of the observed S11 parameter is −13.36 dB as shown in Fig. 3 and Table
2. S22 is an indication of the reflection coefficient of patch-2. At 3.1 GHz frequency,
the value of the observed S22 parameter is −11.03 dB and At 6.9 GHz frequency,
the value of the observed S22 parameter is −13.62 dB as shown in Fig. 4 and Table
2.
Bandwidth is the range of frequencies at which the communication can be carried
out. As far as antenna design is concerned, is the necessary and sufficient condition
for the antenna to radiate effectively is its reflection Coefficient should be less than
−10 dB and its VSWR value should be between 1 and 2. The antenna bandwidth is
calculated below −10 dB line. The difference between higher frequency (FH) and the
lower frequency (FL) is referred to as bandwidth. In this, the bandwidth is expressed
Design and Simulation of 2-Element, Circular Shaped MIMO Antenna … 275
to 95% with Fc = 3.1 GHz. For 2nd band of patch 2, bandwidth is measured from 6.24
to 7.99 GHz. So, the frequency difference obtained here is 1.75 GHz and percentage
band with approaches to 25.4% with Fc = 6.9 GHz action in Fig. 6 and Table 3.
Voltage standing wave ratio (VSWR) is a term that is used to represent the reflec-
tion of the signal into the system. It also indicates the amount of impedance mismatch.
If the impedance of the transmitter circuit does not match with the impedance of the
antenna then most of the transmitted waves are reflected. And those reflected waves
will set up standing waves. VSWR is also called a standing wave ratio. Ideally, the
value of voltage standing wave ratio is 1 and practically it may vary from 1 to 2.
For the proposed antenna patch 1, the VSWR value for 3.1 GHz is 1.78 and for
6.1 GHz it is 1.54. For Patch 2, he VSWR value for 3.1 GHz is 1.77 and for 6.1 GHz
it is 1.52 as shown in Figs. 7 and 8; Table 4.
5 Conclusion
References
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 281
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_29
282 A. S. Vibhute et al.
1 Introduction
There are two forms of visual-based defect detection; the manual inspection and
also the automated one [1]. In many industries, the quality of products is tested by
manual review with the assistance of gauges, and if it doesn’t fit on gauge properly,
then the job is taken into account to be faulty. Also in some industries, the standard
testing is completed by a human eye wherein human observes the form and size of
the output products are starting of the product line. But in practice, just some pieces
are taken and verified for dimension and orientation of the shapes. This manner of
manual testing could be a tedious process and at risk of human errors which reduce
the standard of the products.
This approach has been utilized in the industry before the existence of automated
visual inspection. The activities during this manual inspection are to search the
defect, recognize the fault and make the sole decision. Thus, the training for quality
inspection is crucial to boost the skill of examination and to attenuate errors during
the manufacturing process. They have to look out a way to identify the task with
defects than to form clear decision either to accept or to reject or to remodel the
defected part. Moreover, human inspectors are slow and have become ineffective
after completing the task that required an extended time. They’re either full of fatigue
or sickness or human weaknesses. Hence, they have frequent rest to want care of a
high-performance level. So we want to travel for an automatic visual inspection [2].
Today there is no field of technical endeavour neglected without the impact of
Digital Image Processing. Digital Image Processing may well be a way for an auto-
matic visual inspection [3]. We proposed here the attribute inspection technique that
uses a camera which captures a digital image of every job. The captured image is
preprocessed, filtered. Then the size or attributes are extracted, measured using edge
detection [2, 4] and segmentation techniques [5, 6]. Finally, the output is compared
with a reference or actual dimension in drawing using feature matching. The image
processing task is to look out the faulty piece and to form a decision whether to accept
or reject manufactured product using classification [7]. This increases the speed and
accuracy and avoids human errors which are common in quality testing and also
increases productivity. Quality testing using DIP performs acceptable range.
Some researchers illustrated previous research works which are studied to beat
the restrictions of subjective Evaluation in the visual quality inspection by a human
inspector. They developed an automatic procedure replacement by using computer
vision and image processing technologies to automate the method. These attempts
are to figure out the defects in manufacturing by using digital image processing [8].
Here, we proposed a system which helps to keep up the count of excellent products
furthermore as faulty products produced within the entire day [9] because it identifies
the defect and makes a choice to accept or reject the manufactured product on the
premise of attribute inspection [10].
Attribute Inspection of Product Using Image Processing 283
2 Methodology
See Fig. 1.
Images of the desired product are acquired through digital cameras. Photos are usually
obtained by one or more cameras placed at the position under inspection. The func-
tions of the cameras are typically fixed. In most cases, industrial automation systems
are designed to inspect only known objects at fixed positions [3]. The picture needs
to be adequately illuminated and arranged to facilitate good image acquisition. The
data flow starts from the image acquisition module by connecting the digital camera
with laptop pc (Fig. 2). The image acquisition module captures and transfers Image
into the computer for processing [8], in our paper input image shown in Fig. 3.
Once images are acquired, they’re filtered to get rid of background or unwanted
reflections from the illumination system. Image restoration can also be applied to
boost image quality by correcting geometric distortions introduced by the acquisition
system i.e. camera. The acquired input images are preprocessed by using multiple
operations like grayscale conversion, threshold effect and noisy objects elimination
which are present in the pictures, as shown in Fig. 4.
Attribute Inspection of Product Using Image Processing 285
Sequence of input images is acquired from a digital camera, and then it is converted
into a grayscale image.
After Grayscale conversion threshold effect is applied with the particular value. As
a result, we get Binary Image from the Gray Scale Image as shown in figure.
The gradient magnitude is given by |G| = Gx 2 + Gy 2 .
Typically, [15] an approximate magnitude is computed using:
2.4 Segmentation
Image segmentation may be a necessary technique used for image analysis. It’s the
tactic of partitioning a digital image into multiple segments (sets of pixels, also
called image objects). This step tries to partition the image into regions of interest
that correspond to part or whole objects inside the scene [12]. The varied segmen-
tation techniques used are EM algorithm, OSTU algorithm and Genetic Algorithm
[6]. Threshold selection is employed in OTSU algorithm. Compared with all other
segmentation methods, the Otsu method is one of the only successful ways for image
Thresholding because of its simple calculation. Thresholding creates binary images
from grey-level ones by turning all pixels below some threshold to zero, and each
one pixel this threshold to a minimum of one [16]. If g(x, y) might be a threshold
Attribute Inspection of Product Using Image Processing 287
(x, y) = 1; (x, y) ≥ T
(x, y) = 0; other wise
We have used Euclidean distance for feature measurement like inner diameter, outer
diameter and have matched. The Euclidean distance gives straight line distance
between two points. It’s appropriate once we have continuous numerical variables
and need to reflect absolute distances. This distance takes into consideration every
variable and doesn’t remove redundancies. Moreover, this distance doesn’t scale-
invariant, so generally must scale previously to use the gap. The Euclidean distance
[10] between two weight vectors provides a measure of similarity between the corre-
sponding images Imgref and Imgtest, as shown in Fig. 10. The measurement output
is shown in Fig. 9. The formula for which is,
2
Euclidean Distance (Imgref, Imgtest) = (Imgref, Imgtest)
(2|A ∩ B|) (2|A ∩ B|)
Matching = Mismatching = 1 −
|A| + |B| |A| + |B|
In case of visual inspection, the system has to decide if the result of manufacturing
meets the quality standards, by matching [18] the computed features with a known
model. If the product satisfies the matching criterion, it is considered to be accepted
or else rejected.
A graphical program (GUI) may be a visual interface to a plan. An honest GUI can
make schedules more comfortable to use by providing them with a uniform appear-
ance and with intuitive controls like push buttons, list boxes, sliders, menus, then
forth. A graphical-based language allows the user to figure directly with graphics.
288 A. S. Vibhute et al.
The developed GUI window gives a matching result for the acceptance and rejection
process [19]. From the all MATLAB functions that are presented and implemented,
all of them combine in the GUI screen to be the last final model for our proposed
system [20] as shown in Fig. 11.
Initially, we have captured 100 images of the required job (Ring Gear). The photos
are taken through all angles in such a way that all dimensions are covered. While
acquiring images from different angles, care is being carried by maintaining the same
distance from every angle (top view, side view and 45° angle). We formed three
databases for three different perspectives. Then we have taken new input images
one by one (around 100 test images) and compared with the reference database for
getting a final result of acceptance or rejection of input test job image.
After preprocessing dataset with the help of image processing techniques such
as Thresholding, denoising using the median filter, we need to do edge detection.
Initially, used canny method and results are as shown in Fig. 5. So, to get the more
precise output, we have gone for Sobel edge detection whose products are better as
shown in Fig. 6, than the Canny edge detection.
4 Conclusion
In this study, the image processing technique is used for automatic inspection and
internal quality control. Although lot of research carried by different researcher
doing research in images processing, there’s scope to use image processing tech-
niques for internal control of commercial product. The image processing techniques
290 A. S. Vibhute et al.
makes inspection automatic and fast. Also improves quality and production rate of
an industry. Algorithm is proposed for real time quality monitoring of manufactured
product. This proposed system can replace manual inspection of commercial product.
Result will indicate product is appropriate or not. Using this automatic inspection
system cost of inspection is reduced as we require only 1 time installation cost. Also
accuracy of inspection will increase.
292 A. S. Vibhute et al.
References
1. Rahman NNSA, Saad NM, Abdullah AR, Ahmat N (2019) A review of vision based defect
detection using image processing techniques for beverage manufacturing industry. J Teknol
81(3):33–47
2. Canny J (1986) A computational approach to edge detection. IEEE Trans Pattern Anal Mach
Intell 6:679–698
3. Malamas EN, Petrakis EG, Zervakis M, Petit L, Legat J-D (2003) A survey on industrial vision
systems, applications and tools. Image Vis Comput 21(2):171–188
4. Threshold HNN (2016) A survey on relevant edge detection technique for medical image
processing. Eur J Biomed 3(7):317–327
5. Fouzia MT, Nirmala K (2010) A literature survey on various methods used for metal defects
detection using image segmentation. Evaluation 8
6. Ramadevi Y, Sridevi T, Poornima B, Kalyani B (2010) Segmentation and object recognition
using edge detection techniques. Int J Comput Sci Inf Technol (IJCSIT) 2(6):153–161
7. Lu D, Weng Q (2007) A survey of image classification methods and techniques for improving
classification performance. Int J Remote Sens 28(5):823–870
8. Peansupap V, Laofor C (2010) Digital image processing for evaluating defect level in visual
quality inspection. In: W078-special track 18th CIB world building congress, Salford, May
2010, p 110
9. Watpade AB, Amrutkar MS, Bagrecha NY, Vaidya A (2014) Qcuip: quality control using image
processing. Int J Eng Res Appl 4:15–18
10. Chavan HL, Shinde SA (2016) Defective product detection using image processing. Int J Sci
Res (IJSR) 5(6)
11. Shinde MD (2016) To detect and identify the defects of industrial pipe. Int J Sci Res (IJSR)
5(7):1561–1564
12. Swarnalakshmi R (2014) A survey on edge detection techniques using different types of digital
images. Int J Comput Sci Mob Comput 3(7):694–699
13. Ramesh TP, Bisht Y (2014) Detection and classification of metal defects using digital image
processing. Int J Adv Res Sci Eng 3
14. Senthilkumaran N, Rajesh R (2009) Image segmentation—a survey of soft computing
approaches. In: 2009 international conference on advances in recent technologies in commu-
nication and computing. IEEE, pp 844–846
15. Image filtering median filtering. https://www.cs.auckland.ac.nz/courses/compsci373s1c/Patric
esLectures/Image%20Filtering_2up.pdf
16. Cai H, Yang Z, Cao X, Xia W, Xu X (2014) A new iterative triclass thresholding technique in
image segmentation. IEEE Trans Image Process 23(3):1038–1046
17. Vala HJ, Baxi A (2013) A review on Otsu image segmentation algorithm. Int J Adv Res Comput
Eng Technol (IJARCET) 2(2):387–389
18. Karami E, Prasad S, Shehata M (2017) Image matching using SIFT, SURF, BRIEF and ORB:
performance comparison for distorted images. arXiv preprint arXiv:1710.02726
19. Kiran R, Amarendra H, Lingappa S (2018) Vision system in quality control automation. In:
MATEC web of conferences, vol 144. EDP Sciences, p 03008
20. Humod AH, Leman Z, Samin IR, Ashrafi N (2015) Improving productivity and quality in
manufacturing by applying computer vision systems. Image Process 10(5):1–9
A Proposed Method for Audio
Steganography Using Digital Information
Security
Abstract In the current era of digital technology, the information security is the
challenging task. For the secrete communication information hiding is an essential
element. The current information steganography system uses objects like audio,
image and video. The audio steganography is the technique that convey hidden
message by modifying an audio signal in an unnoticeable manner. It is a technique for
the hiding secret message in the host audio signal. The original audio message before
steganography and after encoding message is having uniform characteristics. The
embedding secrete audio message in the original audio file is a more challenging and
difficult task. This paper presents the comprehensive survey of audio steganography
techniques for information security. The experiment was tested using proposed LSB
technique for audio steganography. This paper extended towards quality measure of
steganography message. The quality of audio steganography measures using energy
score, Mean square error, Peak signal to noise ratio. From this experiment the quality
of audio steganography is observed as 92.759% for MSE and 94.971% for PSNR
technique. Audio information hiding is the one of the robust and dynamic ways of
protecting the privacy and secretes communication.
1 Introduction
In this era of rising technologies, digital communication has become an integral and
significant part of everyone’s life. In the rapid development of digital communica-
tion the information security becomes an important concern. The methods and algo-
rithms available for digital data security uses a cryptographic primitive for secure
data transmission and secrete communication. With the advent of the technology,
people started to use private communication for sharing and transmission using the
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 295
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_30
296 P. Kurzekar and S. Darshane
2 Audio Steganography
Audio steganography is the ability and science of thrashing digital information such
as text messages, documents, and binary files into audio files. The primary message
is known as the carrier signal or message and the secondary message is known as to
the payload signal or message. Characteristically normal audio file and carrier file
are same and not recognized in tapping communication technology [8]. The general
steganography technique is shown in Fig. 1.
Steganography is achieved by means of three varieties of techniques: injection,
substitution, and generation. The insertion technique is embedded in the data to
cover in the insignificant part of the carrier file, which is normally unseen by oper-
ating systems and application software. The substitution technique substitutes the
insignificant bits in the original carrier message with the bits of the data to secrete.
Insignificant bits are those bits that can be modified without destroying the eminence
or destroying the reliability of the carrier message. This technique takes advantage of
A Proposed Method for Audio Steganography Using Digital … 297
the limited abilities of the human auditory system; which cannot identify two sounds
that are slightly not alike. The generation technique examines the data to cover and
produces out of them a new set of data. It is a dynamic method of creating a carrier
message based on the information enclosed in the data to cover [9].
3 Review of Literature
capacity. A filter is added to restrict the changes in the stego file. Obtained stego files
are used to generate unique keys. The filtered file and the generated key sent to the
receiver. The key is used to extract the correct message at the receiver side [12].
Sridevi proposed a useful method of audio steganography by customizing the
LSB algorithm and a strong encryption key with enhanced security is suggested.
Enhanced Audio Steganography (EAS) is a combination of audio steganography
and cryptography. EAS works in two steps: it uses the most effective encryption
algorithm in the first level and in the second level it uses a modified LSB (Least
Significant Bit) algorithm to enclose the message into audio [13].
Submission technique is also a good choice for audio steganography. Message bits
are placed into multiple and higher LSB layer values using genetic algorithms ensuing
in enhancement of robustness. The robustness of the system should be increased
against intruders which try to exhibit the secret message and also some involuntary
attacks like noise etc. [14].
Mane suggested a method known as Least Significant Bit (LSB) method. In this
method, consecutive least significant bits are replaced with private message bits from
each sample of cover audio. LSB method is very simple but less powerful. This paper
differentiated the spectrum of original audio before embedding and audio signal after
embedding [15].
A new 4th bit rate LSB audio steganography method is a proposed new approach in
the current era. This method minimizes the embedding distortion of the host audio. In
this method, message bits are fixed into the 4th LSB layer. It leads in high robustness
against noise addition. As compared to standard LSB method, the perceptual quality
of audio is more in proposed method [16].
The substitution method is having some limitations for audio steganography. The
main problem of this technique is that it is less significant against attack. There are
two types of attack: One, it tries to extract the private message and other tries to
destroy it. As in the standard LSB method, the secret message is stored into the least
significant bit, so this method is more susceptible to attack. Therefore for security
purposes, the message is stored into a bit other than LSB. If a message is stored into
deeper bits, the system will become more powerful. But the main disadvantage is
that when the message is stored into MSB, the host audio signal gets altered. So,
by using an intelligent algorithm this problem is solved where the message bits are
embedded into MSB and other bits are altered to decrease an error. The message is
stored into multiple MSB to make a system more robust and high capacity [17].
Divya proposed a method where multiple LSB bits are used for hiding a text in
an audio signal using steganography and cryptography is used for security purposes.
For LSB audio steganography, the maximum number of bits is altered from 16 bit
audio samples. They use two novel approaches for substitution technique of audio
steganography improving capacity of cover audio for storing additional data. In
this technique, the message bits are stored into 35–70% compared to standard LSB
technique which uses 4 LSBs for data storing [18].
The researcher proposed a Genetic algorithm is also a good choice for audio
steganography. They studied the various audio steganography techniques using
A Proposed Method for Audio Steganography Using Digital … 299
genetic algorithms and LSB approaches. They tried some approaches which help
in audio steganography. In these techniques, hidden messages are written in such a
way that only the sender and corresponding receiver are able to see the message [19].
Researchers currently turn towards hiding the high quality secret message in audio
files. The abundance of audio messages makes them eligible to convey secret informa-
tion. Many researchers started to explore how the audio signals and audio properties
can be exploited in the era of information security [20]. Several approaches were
considered, the most robust and significant ones are Least Significant Bit [21], Echo
hiding [22], Hiding in Silence Interval [23], Phase Coding [24], Amplitude Coding
[25], Spread Spectrum [26], and Discrete Wave Transform [27].
Fundamentally, the LSB technique depends on implanting every bit from the data
to hide into the rightmost bits of every audio sample of the carrier audio message.
The LSB technique proved as advancement towards a HAS unable to understand
the slight variation of audio sampling frequency towards the high frequency region
of the audible spectrum. The LSB technique allows high embedding rate without
impairing the quality of the audio file. This technique is robust and dynamic for
audio steganography.
The technique uses the very fact that the majority of the knowledge in a very
sample in any audio file is contained within the MSBs instead of LSBs. In the LSB
coding approach the slandered data broadcasting rate is 1 kbps per 1 kHz. In some
implementations of LSB coding, the two smallest significant bits of a sample are
substituted with two message bits. This implementation increases the amount of data
that will be determined but also enlarges the amount of resultant noise in the audio
file as well. The representation of LSB coding technique is shown in Fig. 2.
In this echo hiding technique, the secret data are embedded into the audio signals as a
short acoustic echo. In fact, an echo is a reproduction of sound, however, received by
the listener some time after the original sound. The echo is perceptible; its amplitude
must be reduced and undetectable. In order to hide data, bits whose values are 0, it
is characterized by an echo overdue 1 ms; bits whose values are 0 are represented by
an echo delayed 2 ms.
300 P. Kurzekar and S. Darshane
In this technique, the original signal is divided into chunks before the encoding
process. Once the encoding process is finished, the blocks are concatenated back
together to create the final signal [28]. The complete block diagram of echo hiding
technique is shown in Fig. 3.
Amplitude coding technique conceals secret data in the magnitude speech spectrum
while not distorting the carrier audio signal. It is based on finding a safe spectral area
in the signal whose magnitude speech spectrum is below a certain value. In addition,
the carrier locations are preferred based on how much they can badly affect the audio
signal [25].
A Proposed Method for Audio Steganography Using Digital … 301
Spread Spectrum technique scatters the secret data over the frequency spectrum of
the audio file using a precise code independent of the original signal. Fundamentally,
secret data are multiplied by a code known to the corresponding level only, and then
implanted in the carrier audio message. In this technique the data is generated by m-
sequences code known as sender and receiver for the secret communication [29]. To
control stego speech distortion, [30, 31] have proposed an embedding method where
splitted data is secreted under a frequency cover. The spread spectrum is combined to
phase changing to increase the strength of the transmitted data against additive noise
and allows easy detection of the fixed data. In this method, a reliable hiding capacity
of 3 bps was attained. The graphical representation of spread spectrum information
encoded in the original message is described in Fig. 4.
Discrete Wave Transform technique private messages are encoded in the smallest
significant bits of the wavelet coefficients of the audio signals. Often, private data
is chosen to be secreted in the wavelet coefficients and not in silent sections of the
audio signal so as to promote the imperceptibility of the audio file [27].
5 Proposed Approach
From the enriched literature observed the limitation of available techniques and
their procedure. The human ear is highly sensitive and can regularly notice even the
slightest bit of noise introduced into a sound file. The main limitation associated with
parity coding is not much closure making introduced noise inaudible. The limitation
of phase coding towards data transmission rate because of this technique message is
encoded in the first segment of signal only. Phase coding technique recommended
only when the small amount of data considered for a steganography approach. Least
significant bit (LSB) coding is the robust way to encode messages in a digital audio
302 P. Kurzekar and S. Darshane
file. Substituting the least significant bit to each frequency point with a binary message
allows for a large amount of data to be encoded [21]. In the available data hiding
methods in enriched literature the proposed method for embedding secret messages
within audio file, LSB is the simple and robust method for inserting messages in audio
signal towards noise free environment. It embeds secret message-bits in a subset of
the LSB levels of the audio message.
The following steps are:
a. Receive Audio file
b. Convert the file into bit Pattern
c. Each character of secrete message convert into the bit pattern
d. Replaces the LSB bit from audio with LSB bit from character in the message.
The proposed system is to provide a good, robust and dynamic method for infor-
mation hiding in audio from hackers and sent to the destination for secret commu-
nication. The advantages of the proposed system does not change the size of the
original audio file even after encoding and also suitable for all audio formats.
6 Experimental Analysis
For these experiments a real example was tested to illustrate how the proposed
steganography technique works. It involved using the various steps of the algorithm
to envelop a secret text message into a 16-bit WAV carrier audio file. For this exper-
iment we selected the different 10 secret messages. The secret data to hide is a text
message shown in the experiment that says “we will kill you”. The private message
is preprocessed and converted into a binary form. Ten audio samples are randomly
selected for this experiment. The chunks obtained in implementation step one are
stored into the three LSBs of the audio samples selected in step two. The final output
is the carrier WAV audio file now carrying the secret data. The graphical represen-
tation of original audio message “you are very nice” is shown in Fig. 5. Figure 6
represents the audio file with encoded secrete message. The spectrogram of audio
file with encoded message is shown in Fig. 7. Figure 8 represents the spectrogram
of original audio message after secrete message extracted.
Fig. 5 Original audio signal message for the “you are very nice”
A Proposed Method for Audio Steganography Using Digital … 303
Fig. 6 The audio file with encoded message “we will kill you”
The mean square error (MSE) of associate reckoner calculates the common of the
squares of the errors, that is, the variation between the estimator and what is estimated.
MSE is a risk function, equivalent to the predictable value of the squared error loss
or quadratic loss. The distinction happens attributable to uncertainty or as a result
of the reckoner does not account for data that might manufacture a lot of correct
estimation of speech synthesis [32].
Peak Signal to Noise Ratio, often abbreviated PSNR, is the ratio between the most
probable influence of a signal and the power of corrupting noise that affects the
quality of its illustration. PSNR is usually articulated in terms of the logarithmic
decibel scale. PSNR is most ordinarily accustomed live the standard of reconstruction
of signal and image. The signal in this case is the original data, and the noise is
the error introduced by synthesis [33]. The PSNR and MSE method was used for
quality measure of steganography audio based on proposed LSB technique. Table 1
represents the MSE and PSNR values for LSB technique.
A Proposed Method for Audio Steganography Using Digital … 305
Table 1 MSE and PSNR results for audio steganography using proposed LSB technique
Sr. No. Original audio signal Stegno audio signal MSE PSNR
1 S001 N001 5.23 1.26
2 S002 N002 8.9 7.56
3 S003 N003 7.61 1.29
4 S004 N004 5.32 3.24
5 S005 N005 7.61 7.32
6 S006 N006 9.1 9.78
7 S007 N007 8.06 11.23
8 S008 N008 5.32 1.29
9 S009 N009 8.06 3.24
10 S0010 N0010 7.2 4.08
Average 7.241 5.029
Quality (100 − average) 92.759 94.971
7 Conclusion
References
6. Lee YK, Chen LH (2000) High capacity image steganographic model. IEEE Proc Vis Image
Signal Process 288–294
7. Alvaro M, Guillermo S, Gadiel S (2005) Is image steganography natural? IEEE Trans Image
Process 14
8. Cvejic N, Seppben T (2002) Increasing the capacity of LSB-based audio steganography. Finland
9. Johnson NF, Jajodia S (1998) Exploring steganography: seeing the unseen. Comput J 31(2):26–
34
10. Asad M, Gilani J, Khalid A (2012) Three layered model for audio steganography. In:
International conference on emerging technologies (ICET)
11. Rana L, Banerjee S (2013) Dual layer randomization in audio steganography using random
byte position encoding. Int J Eng Innov Technol 2(8)
12. Gandhi K, Garg G (2012) Modified LSB audio steganography approach. Int J Emerg Technol
Adv Eng 3(6):158–161
13. Sridevi R, Damodaram A, Narasimham S (2009) Efficient method of audio steganography by
modified LSB algorithm and strong encryption key with enhanced security. J Theor Appl Inf
Technol 771–778
14. Bankar Priyanka R., Katariya Vrushabh R, Patil Komal K (2012) Audio steganography using
LSB. Int J Electron Commun Soft Comput Sci Eng 90–92
15. Mane A, Galshetwar G, Jeyakumar A (2012) Data hiding technique: audio steganography using
LSB technique. Int J Eng Res Appl 2(4):1123–1125
16. Gadicha AB (2011) Audio wave steganography. Int J Soft Comput Eng (IJSCE) 1:174–177
17. Zamani M et al (2009) A secure audio steganography approach. In: International conference
for internet technology and secured transactions
18. Divya SS, Ram Mohan Reddy M (2012) Hiding text in audio using multiple LSB steganography
and provide security using cryptography. Int J Sci Technol Res 1:68–70
19. Nehru G, Dhar P (2012) A detailed look of audio steganography techniques using LSB and
genetic algorithm approach. Int J Comput Sci (IJCSI) 9:402–406
20. Djebbar F, Ayad B, Abed-Meraim K, Hamam H (2011) A view on latest audio steganography.
In: 7th IEEE international conference on innovations in information technology
21. Gopalan K (2003) Audio steganography using bit modification. Proc Int Conf Multimed 1:629–
632
22. Gruhl D, Bender W (1996) Echo hiding. In: Proceeding of information hiding workshop, pp
295–315
23. Shirali-Shahreza S, Shirali-Shahreza M (2008) Steganography in silence intervals of speech.
In: Proceedings of the fourth IEEE international conference on intelligent information hiding
and multimedia signal, pp 605–607
24. Qi Y-C, Ye L, Liu C (2009) Wavelet domain audio steganalysis for multiplicative embedding
model. In: Proceedings of the 2009 international conference on wavelet analysis and pattern
recognition
25. Djebbar F, Ayad B, Abed-Meraim K, Habib H (2012) Unified phase and magnitude speech
spectra data hiding algorithm. J Secur Commun Netw
26. Khan K (1984) Cryptology and the origins of spread spectrum. IEEE Spectr 21:70–80
27. Cvejic N, Seppanen T (2002) A wavelet domain LSB insertion algorithm for high capacity
audio steganography. In: Proceedings 10th IEEE digital signal processing workshop and 2nd
signal processing education workshop, p 5355
28. Gopalan K, Wenndt S (2004) Audio steganography for covert data transmission by impercep-
tible tone insertion. In: WOC 2004, Banff, 8–10 July 2004
29. Czerwinski S, Fromm R, Hodes T. Digital music distribution and audio watermarking. https://
reference.kfupm.edu.sa/content/d/i/digital_music_distribution_and_audio_wat_1045219.pdf
30. Queirolo F. Steganography in images. Final communications report. https://eric.purpletree.org/
file/Steganography%20In%20Images.pdf
31. Cox IJ, Kalker T, Pakura G, Scheel M (2005) Information transmission and steganography, vol
3710. Springer, pp 15–29
A Proposed Method for Audio Steganography Using Digital … 307
32. Lehmann EL, Casella G (1998) Theory of point estimation, 2nd edn. Springer, New York.
ISBN 0-387-98502-6. MR 1639875
33. Huynh-Thu Q, Ghanbari M (2008) Scope of validity of PSNR in image/video quality
assessment. Electron Lett 44(13):800. https://doi.org/10.1049/el:20080522
Spine Diseases Detection Using SVM
Abstract This paper addresses the use of HOG (Histogram of Oriented Gradi-
ents) and SVM (Support Vector Machine) for boundary regression for biomedical
image segmentation and comparing the output with image processing segmenta-
tion techniques. Here we will use MATLAB for computation of HOG and SVM.
MATLAB image processing toolbox will be extensively used for reading, processing,
visualizing and saving the images.
1 Introduction
The nerves are foraminate by the nervous radicals of the neural foramina columns
and can become a strain that can lead to painful numbness, weakening of the arm,
wakening and equilibrium of the leg. Neural foramina stenosis (NFS) is the restricting
of the openings of any vertebral in the spine called foramina. The risk of neural
foramina stenosis rises according to the age factor as we age discs in the spine lose
height, starting to dry excessively because of NFS, over 80% of peoples are infected.
For segmentation, HOG methods are used to detect the right area and examination,
MRI and CT imaging are used to demonstrate the stenosis for the diagnosis and
treatment manual. The segmentation system of the physician should be used because
of size, shape, variance attribute. Segmentation is the automatic boundary detection
process.
J. M. Waykule (B)
E & C Department, Sanjay Ghodawat University, Atigre, Kolhapur, Maharashtra, India
e-mail: [email protected]
V. R. Udupi
Department of E & C Engineering, Gogte Institute of Technology, Belgaum, Karnataka, India
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 309
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_31
310 J. M. Waykule and V. R. Udupi
2 Literature Review
HOG is a feature identifier used in image processing and computer vision for object
detection. A feature descriptor is an image or image patch interpretation that simpli-
fies the image by extracting useful information and discarding foreign information.
In most cases, a feature descriptor converts an image of size width * height * 3 (chan-
nels) to a feature vector/array of length n. In the case of the HOG feature descriptor,
the input image is a size 64 * 128 * 3, and the output feature vector is of length 3780.
HOG descriptor can be calculated for other sizes, but the above numbers presented
in the original paper so one can easily understand the concept with one concrete
example. The histogram of directions of gradients (oriented gradients) is used as a
feature in the HOG feature descriptor. The gradients (x and y derivatives) of a image
are useful because the magnitude of the gradients is intense around the edges and
corners (regions of rapid intensity changes) and we know that the edges and corners
provide plenty of key information about the shape of the object than the flat regions.
Gradient Calculation: Mostly in HOG feature extractor, the input image is divided
into small parts called cells, normally 8 × 8 pixels, as shown in Fig. 1. The gradient
in both vertical directions is then estimated for each pixel in the cell. Simple [−1, 0,
1] and [−1, 0, 1] T gradient filter is applied to the pixel value f(x, y) to obtain both
fx (x, y) and fy (x, y) which is defined as the gradients in both x and y directions have
been determined, the gradient Magnitude m(x, y) and the gradient direction (x, y)
could be determined as:
fy (x, y)
θ(x, y) = arctan (4)
fx (x, y)
In the case of color image, the gradients at each pixel would be calculated for all
color components separately and the gradient with the largest magnitude would be
selected to be used in next steps of feature extraction.
Histogram Formation: The first steps is to produce an orientation histogram for each
cell in the image based on the values of m(x, y) and (x, y) derived for the pixels
in the cell. The interval of [0, π] is evenly divided into the number of orientations
for the development of orientation bins to be considered in the calculation of the
histogram. The original HOG feature extraction introduced by Dalal and Triggs [1]
has a value 9 bins to be expected for a number of directions since it has shown better
result in the detection [6] for each pixel inside the cell as reference in Paper [1].
Block Normalization: The key element in the extraction of the HOG function is
the process of standardization across blocks. Once the gradient for each pixel of
the cell is determined, the neighboring cells are grouped together to create a more
spatial block. In general, blocks are characterized as overlapping numbers of adjacent
cells. Usually, each block consists of four cells, i.e. collection of 2 × 2 neighboring
cells [1]. Based on the normalization scheme adopted for this step, the four cells
vectors within a block are combined to produce a normalization factor. Considering
normalized vector function as show in paper [1], we are going to have a small constant
and L2-Hys is L2-norm followed by clipping and restricting the maximum values
and then renormalizing. Dalal and Triggs have shown that L2-Hys, L2-norm and
L1-sqrt perform equally well for normalization, while L1-norm decreases efficiency
by 5% [1, 6].
In order to measure the HOG descriptor, we first need to measure the horizontal and
vertical gradients; however, we want to calculate the gradient histogram (Fig. 2). The
gradient of the image is determined in both the “x” and “y” directions (Fig. 3c). And
in Fig. 3d. Then angle Fig. 3f and magnitude Fig. 3e of the gradients is calculates
using following two equations.
Pre-processing: Here the preprocessed step considers the conversation of Di.com
image into uint8 image by using the MATLAB function.
Spine Diseases Detection Using SVM 313
Compute Gradients with its Magnitude and Angles: The 2D filter [−1 0 1] is used
for calculation in x direction and 2D filter transpose of [−1 0 1] is used to calculation
of gradient in y-direction. The filter then calculates the horizontal difference between
the next and previous pixels and stores the result as “dx” in the current pixel; similarly
calculates the vertical difference between the next and previous pixels and stores the
result as “dy” in the current pixel.√Calculated the magnitude and angles by using the
following formula. Magnitude = dxˆ2 + dyˆ2 and Angle = tan − 1(dy/dx).
The angle and magnitude of gradients is then used for HOG calculations. The gradi-
ents of the picture are sensitive to complete illumination. If you darken the image by
dividing all pixel values by 2, the gradient magnitude will change by half, and the
histogram values will change by half.
Histogram of Gradients (HOG Descriptor): In the HOG function, the input image
is divided into small parts called cells, typically 8 × 8 pixels. The gradient in both
horizontal and vertical directions is then determined for each pixel inside the cell as
shown in the earlier point. Now the values of angles and magnitudes in each cell [64
of angles and 64 of magnitudes] are to be shaped into a histogram of 9 bins, making
9 values for each cell. The x-axis bins of the histogram are the angles of the gradients
measured above. The angles are divided into 9 values scale from zero to 180° in 20°
measures. Thus the 9 bins are [0, 20, 40, 60, 80, 100, 120, 140, 160]. Then the angles
between the bin values are divided into the left and the right parts. The left part is the
allocation to the closest bin value less than the angle and the right part is the nearest
bin value greater than the angle [180 is not a part of this bins because largest will
is similar to zero]. The magnitude of the section is the difference between the angle
and the bins. Now the gradient magnitude corresponding to the bins are inserted and
the histogram is created (Fig. 3c). The matrix is now transformed to a single array of
cell-wise histogram values and this is going to send the final HOG function vector
of matrix size M × N × 9, where M and N are the number of vertical and horizontal
cells respectively.
Boundary Representation: Image editing software (GNU Image Manipulation) is
used and the boundary of the image taken for segmentation training is traced. The
analyzed boundary then divides the image into two regions, the inside boundary
region and the outside boundary region. The area outside the boundary is filled
with Black and the area inside the boundary is filled with White. And the image is
converted to Binary image. White pixels are the border pixels, and Black pixels are
the non-boundary pixels. The generated labels are used for the formation of the SVM
classification.
Support Vector Machine (SVM Classifier): Support Vector Machine (SVM) is
supervised machine learning technique. SVM is binary classifier used for classifying
data into two classes. The SVM classifier is generally a hyper plane that divides the
data into two sections. The bins separated by the HOG are divided into two sections
Spine Diseases Detection Using SVM 315
[boundary region and non-border region]. In SVM Training the first step is to choose
a nonlinear function (kernel function) that maps the input to a higher dimensional
space. This option will also be influenced by understanding of the problem domain.
In the absence of such knowledge, one could prefer by using polynomials, Gaussians
or other basic functions. The dimensionality of the configured space may be random
high.
Classify Boundary and Non-boundary Pixels: SVM model is trained with HOG
feature descriptor and gradient magnitude characteristics are given. SVM model
offers a set of labels for border and non-border cells. SVM labels are stored in an
array. The array is “N * M” where ‘N’ is Horizontal cell numbers and ‘M’ is vertical
picture cell count. The labels array is now iterated with the length “N” segments of
“M.” if you find boundary pixels in the array then corresponding element in matrix
represent by one. The matrix is then transformed to a binary image.
Boundary on Original Input Image: The Boolean image obtained from the SVM
output label array is resized to the original image size. The pixels in the original
image equivalent to those in the boundary Binary image are labeled with red color
pixels.
Fig. 4 Spine MRI image segmentation using HOG SVM, Prewitt, Sobel and Canny resp
4 Conclusion
Fig. 5 Spine MRI image segmentation using HOG SVM, Prewitt, Sobel and Canny resp
Fig. 6 Spine MRI image segmentation using HOG SVM, Prewitt, Sobel and Canny resp
318 J. M. Waykule and V. R. Udupi
Fig. 6 (continued)
Fig. 7 Spine MRI image segmentation using HOG SVM, Prewitt, Sobel and Canny resp
References
1. Dalal N, Triggs B (2005) Histograms of oriented gradients for human detection. In: 2005 proceed-
ings of the IEEE computer society conference on computer vision and pattern recognition, vol
1, pp 886–893
2. Canny J (1986) A computational approach to edge detection. IEEE Trans Pattern Anal Mach
Intell PAMI-8(6):679–698
3. Sánchez-Fernández M (2004) Member, IEEE, SVM multiregression for nonlinear channel
estimation in multiple-input multiple-output systems, vol 52, no 8
4. Wang Z, Zhen X, Tay KY, Osman S, Romano W, Li S (2015) Regression segmentation for M3
spinal images. IEEE Trans Med Imaging 34(8)
5. Michopoulou SK, Costaridou L, Panagiotopoulos E, Speller R, Panayiotakis G, Todd-Pokropek
A (2008) Atlas-based segmentation of degenerated lumbar intervertebral discs from MR images
of the spine
6. Sakthivel K, Nallusamy R, Kavitha C (2014) Color image segmentation using SVM pixel
classification image. World Acad Sci Eng Technol Int J Comput Inf Eng 8(10)
BCI Integrated Wheelchair Controlled
via Eye Blinks and Brain Waves
Abstract People who suffer from complete or partial paralysis often require the
help of a second individual to move around from one place to another even with
automated traveling mechanisms. This dependence on another person takes a toll on
the individual. Thus, a framework is proposed wherein eye blinks and brain waves
are used to decide the decisions made by the individual. This requires no extra phys-
ical exertion. These signals will be collected by an Electroencephalogram Headset
(Neurosky Mindwave Mobile) and will then be passed to an android application
which will then transfer necessary data to the Arduino, which will use these data to
run the device. Such a framework can be executed on mechanized wheelchairs to
uphold the impaired and help in their transition of becoming independent.
1 Introduction
Wheelchairs give versatility to individuals who experience the ill effects of halfway or
complete loss of motion and have been around for an exceptionally significant time-
frame. Early Chinese models of wheeled goods utilized work carts to move both
substantial items and impaired individuals. After some time, the plan and useful-
ness of wheelchairs have improved with the development of the main lightweight,
folding, and mass-created wheelchair in 1933 by Harry Jennings and Herbert Everest
filling in as the presentation of the advanced wheelchair. Besides, the wheelchair has
developed with the headway of innovation. Notwithstanding the conventional labor
driven wheelchairs, electric and battery-controlled wheelchairs are currently like-
wise accessible. Additionally, wheelchairs can in any case be made a stride further
with the intensity of brainwaves.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 321
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_32
322 S. Prem et al.
2 Literature Survey
3 Definition of Terms
3.3 Electrooculography
3.4 Bluetooth
3.5 Microcontroller
The HC-06 Bluetooth module is one of the most common Bluetooth modules used
by hobbyists and professionals alike. It is a serial port protocol Bluetooth module
that only acts as a slave. HC-06 modules can transmit up to 10 m.
The Arduino Uno is one of the most, if not the most, popular microcontrollers around.
It is based on the ATmega238p. The Uno has fourteen digital input/output pins with
six of those doubling as PWM outputs.
The L298N Motor Driver enables the control of DC motors by amplifying the low-
current signal from the Arduino into a higher-current signal suitable for motor control.
A value based on the user’s beta brainwaves that are calculated by Neurosky’s propri-
etary algorithm. Attention is associated with the focusing of a single thought by the
user.
To understand how the Mindwave Mobile works in terms of what EEG data is
captured by its sensor and how it can identify the user’s attention and meditation
levels, a basic understanding of the different EEG Frequency Bands that the Mind-
wave Mobile can detect is needed. The five main brainwave frequencies [13] include
Gamma, Beta, Alpha, Theta, and Delta. Figure 1 shows the different EEG bands and
their associated states.
Gamma Waves. Gamma waves are associated with learning and processing new
information.
Beta Waves. Beta waves are associated with intense, logical, and analytic thinking.
They dominate one’s normal conscious state when our attention is directed towards
cognitive and real-world tasks.
Alpha Waves. Alpha waves are associated with quiet and meditative thoughts. They
aid in mental coordination, calmness, and alertness. The Alpha state of the brain is
known as its resting state.
Theta Waves. Theta waves mostly occur when one sleeps, moreover, such waves
are also present in deep meditation.
Delta Waves. Delta waves are associated with absolute relaxation. They are generated
when one is in the deepest of meditations and dreamless deep sleep.
326 S. Prem et al.
In addition to being able to output eSense meters, the Mindwave Mobile is also able
to detect blinks [13, 14]. It does this by using Electrooculography or EOG. EOG
measures the electrical potential between electrodes placed at points near the eye or
in the ocular region. Blink detection with EOG follows the concept that whenever a
person blinks, a resulting spike in the EOG data will happen.
Two of the three essential yields of the Mindwave Mobile are its eSense meter read-
ings [15]—Attention and Meditation. Attention just alludes to the user’s center level
while reflection alludes to the degree of serenity. Attention levels increment when an
individual spotlights on a solitary idea, for example, tackling a numerical question,
and diminishes when he is diverted. Contemplation levels, then again, increment
when an individual loosens up his psyche and diminishes when he gets focused.
Regarding the calculation Neurosky uses to interpret crude brainwave information
into its eSense meter readings, they, sadly, have not unveiled it. No documentation
examines the points of interest of how Neurosky does this. This might just be a
BCI Integrated Wheelchair Controlled via Eye Blinks and Brain … 327
result of protection and scholarly copyright reasons and there is no way around it.
Neurosky has, in any case, disclosed that the consideration eSense meter depends
on beta waves while the contemplation eSense meter depends on alpha waves. This
bodes well as beta waves are related to serious, legitimate, and logical reasoning while
alpha waves are related to calm and reflective considerations. Just exactly how they
measure the crude alpha and beta waves, Neurosky has not unveiled. With respect
to the precision and dependability of Neurosky’s eSense meters, they are, similar to
blink detection, accurate and reliable to a certain extent. While a portion of this can
be ascribed to the calculation Neurosky utilizes, the more plausible purpose behind
this is our powerlessness to explicitly tune into and control a specific EEG band
recurrence. There is certifiably not a person on earth that has supreme authority over
his cerebrum’s EEG frequencies. It is simply genuinely and intellectually difficult to,
state; just increase your Beta waves and quiet different frequencies. This failure will
prompt confounding and, now and again, out and out baffling outcomes. You might
be zeroing in such a great amount on attempting to tackle an analytics issue, yet
Neurosky’s eSense consideration worth will even now be low. Neurosky delivered
a disclaimer expressing how its eSense meters are intended for diversion purposes
just—confirmation that eSense meter readings are not 100% precise and solid.
5 Design Overview
The Android application will be connected by both the Mindwave Mobile and the HC-
06 Bluetooth module simultaneously. The hardware system flow is shown in Fig. 2. It
begins with the Neurosky Mindwave Mobile, worn around the user’s head, that gets
brainwave-related information and cycles it before remotely sending it through its
own inherent Bluetooth module to the Android Application. The Android application
goes about as an agent between the Mindwave Mobile and the Arduino, which is the
microcontroller that lives in the little wheelchair itself. The Android application can
likewise be considered as a security precautionary measure for the user. Since infor-
mation from the Mindwave Mobile isn’t legitimately communicated to the Arduino,
the Android application can get rid of undesirable information, guaranteeing that
only the important data is sent to the Arduino, thereby making it more efficient. The
Mindwave Mobile and the Android application is interfaced via Neurosky’s Android
application, allowing the transmission and reception of EEG and EOG data. The
subsequent information received by the Android application will at that point be
remotely sent to the HC-06 Bluetooth module, which is associated with the Arduino.
The Arduino will then pass the information to the L298N motor driver. The L298N
Motor driver has 4 DC motors associated with it and is liable for both its speed and
heading control. Depending upon the data passed on by the Arduino, the L298N
328 S. Prem et al.
Motor driver at that point sends resulting orders to the DC motors, permitting the
movement of the wheelchair.
After both the Mindwave Mobile and the Arduino’s Bluetooth module is set up in
association with the Android application, the Android application starts receiving the
signal quality worth, which can be identified as poor, medium, or good. The signal
between the temple skin and the dry sensor, and great if the dry sensor connects
properly with the temple. The Software system flow is presented in Fig. 3.
When a good signal value is obtained, the Android application commences
listening for any incoming force blink data from the Mindwave Mobile. When a
forced blink or a blink whose blink strength value is above the value of 80 is iden-
tified, the Android application begins moving between the four direction values
namely, forward, reverse, left, and right for 10 s with a 2-s interval in between the
change in direction value. This window of 10-s direction-cycle is known as command
mode. During command mode, the Android application listens for two consecutive
blinks also known as a double blink event, from the user. When a double blink even
is detected, the cycling of different directions stops and the direction at which the
cycle was at when the double blink event happened, becomes the current direction.
The blinks are considered consecutive if the time interval between two blinks is less
BCI Integrated Wheelchair Controlled via Eye Blinks and Brain … 329
than 4 s. There is a shift to the focus mode in the Android application once the direc-
tion has been chosen. When the attention values (put out by the Mindwave Mobile
once every 1 s) reaches a value of 50 or more, the Android application changes to
the running mode where it sends a command to the Arduino based on the direction
chosen earlier. Each direction has a respective Bluetooth command that will be trans-
mitted to and interpreted by the Arduino which will be present in the wheelchair.
Similar to command mode, the user exits running mode by blinking consecutively
to go back to standby mode.
From then on, the whole operation loop is repeated if the user wants to move
the wheelchair once again. The speed is kept constant all through activity when the
wheelchair is running. This is because of precision and control-gives that are natural
to the brainwave detection in the Mindwave Mobile. Along these lines, the consistent
speed can likewise be thought of as a safety measure for the user.
6 Conclusion
By using brainwaves and eye blinks to control the motion of a wheelchair, people
suffering from paralysis can easily move around from place to place without the aid
of another person and will give them a sense of independence. This is just an idea, on
the potential of implementing a Brain-Computer Interface system in an everyday item
like a wheelchair. Brain wave technology is a relatively new and upcoming field due to
which several limitations are existent. EEG and brainwave technology is effective to
a certain extent but nowhere near perfect. Blink detection isn’t 100% precise and this
issue can be solved if the blink recognition technology shows signs of improvement.
As human beings do not have complete control of their brainwaves, there might be
inconsistencies and fluctuations in the brain wave data received which makes the
brainwave detection not so precise. As far as human beings learn how to control
and manipulate individual brainwave frequencies, complete and absolute control
over brainwave is impossible leading to minor inaccuracies in the algorithm used to
obtain the required brainwave data. Mindwave Mobile has got issues like hardware
bugs that endanger the sudden stoppage of the wheelchair at times, contributing to
various problems that the researchers have no control over. Some of these issues can
be resolved by using a more advanced EEG headset like EMOTIV Epoch but is more
expensive. This is only preliminary research. More extensive researches have to be
conducted to get more conclusive results. The BCI integration applications hold great
promise in the coming years [16]. Brainwave technologies will without a doubt get
better over time and we are optimistic that soon days will come when BCI will be
successfully incorporated into everyday items.
BCI Integrated Wheelchair Controlled via Eye Blinks and Brain … 331
References
1. Carlson T, del Millan JR (2013) Brain controlled wheelchairs: a robotic architecture. IEEE
Robot Autom Mag 20(1):65–73
2. Butt A, Stanacevic M (2014) Implementation of mind control robot systems. In: Application
and technology conference (LISAT), Farmingdale, pp 1–6
3. Postolache O, Viegas V, Dias Pereira JM, Vinhas D, Silva GP, Postolache G (2014) Towards
developing a smart wheelchair for user physiological stress and physical activity monitoring. In:
IEEE international symposium on medical measurement and application (MeMeA), Portugal,
pp 1–6
4. Kumar RM, Lohit HS (2012) Design of multipurpose wheelchair for physically challenged
and elder people. SASTECH 11(1):108–117
5. Takashi G, Ann G (1998) Developing intelligent wheelchairs for the handicapped. Applied AI
System Inc. (AAI)
6. Nicolas-Alonso LF, Gomez-Gil J (2012) Brain computer interface, a review. Sensors
12(2):1211–1279
7. Keerthana et al (2010) Controlling electrical devices with human brainwaves. https://aasrc.
org/conference/wp-content/uploads/2015/04/Controlling-Electrical-Devices-with-Human-
Brainwaves.pdf
8. Manuel Adrian CA, Anne Monique BC, Dominic Even MP, Lawrence Angelo AT (2012)
PC based hands free short messaging system through facial movements using Emotiv EPOC
headset. De LaSalle University, Philippines
9. Dominguez H (2014) Bluetooth remote controlled car. University of California, Riverside,
California. https://alumni.cs.ucr.edu/~hdomi001/docs/car.pdf
10. Achkar R, Haidar GA, Dourgham H, Semaan D, Araji H (2015) Mobile controlled wheelchair.
In: IEEE European modelling symposium (EMS), Madrid, pp 429–434
11. Psychedelic information theory. https://psychedelic-information-theory.com/eeg-bands
12. Neurosky store. https://store.neurosky.com/pages/mindwave
13. Ahsan MR, Ibrahimy MI, Khalifa OO (2009) EMG signal classification for human computer
interaction: a review. Eur J Sci Res 33(3):480–501
14. EEG: Electroencephalography with Lab VIEW and Mindwave Mobile. https://cerescontrols.
com/projects/eeg-electroencephalography-with-labview-and-mindwave-mobile/
15. EEG eSense Analysis. https://store.neurosky.com/products/eeg-analyzer
16. Nicolas-Alonso LF, Gomez-Gil J (2012) Brain computer interfaces, a review. Sensors
12(2):1211–1279
Classification of Soil Nutrients Using IoT
1 Introduction
Soil is a natural body consisting of layers that are primarily composed of different
materials. Soil moisture and nutrients content is an essential factor of a crop’s growth.
Moisture must be measured. Soil nutrients content is the single most crucial factor
in determining plant growth initially. The water calculates devices are programmed
such that they detect arrangements of moisture content. No resource will be more
valuable to agriculture in the future. Soil contamination is distinguished by frozen
or liquid hazardous substances sundry with naturally occurring soil. Soil contamina-
tion can arise from several sources, which could be both naturally occurring in soil
and human made. In the Multi-dimensional flow of water, plants will uptake, and
high frequency of water sources increase the complexity in modelling soil moisture
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 333
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_33
334 S. Koli et al.
passage from trickle irrigation. We used all calculated methods to identify certain
boundary conditions to model the insinuate from water source [1].
The system will check every particle of nutrients and soil contamination a sample
is taken from the land. It refers to what will be the solution for it, and every crop
needs is different based on it. The system allows us to verify which land is suitable
for which crop which fertilizer is ideal for crop growth. Many studies have been done
about the contamination of soil due to coal mine waste, indicating particularly on
the influences on the environment, treatment of coal mine waste, but the research on
Attributes of soil heavy metal pollution around coal mine waste piles was rare [2]. In
precision agriculture and organic farming, it is important to continuously monitor the
fields as they are site-specific. Monitoring plant health is essential, which enriches
the productivity of food grains. Soil moisture is one of the primitive factors for plant
health. The water that remains in soil as a thin film aid in supplying nutrients to the
plant growth. The pollution of soils with heavy non-ferrous metals is a world problem.
Many studies have been achieving soil pollution and the portability of the most toxic
pollutants [3]. There have been many reports from several states such as Maharashtra,
Telangana, Karnataka etc. of contamination of soils with Lead, plumbum, Zinc, etc.
The lead, copper and zinc cannery are the main sources of the contamination [3].
The higher percentage of heavy non-ferrous metals in the soil, compared with the
entrance limit values are established. The children are very conscious of the lead
contamination of soil [3]. All crops have different pH preferences. Knowing the
soil’s pH can help in choosing the right crops and allow proper treatment for the
ground. Soil pH can change every year because of elements, including the type of
rain and consumption of certain nutrients. Hence it is necessary to monitor soil pH
before planting and regularly in a random area of land all over the season’s growth.
But so far, many Indian farmers have not paid attention to soil pH on their land.
The local Government Agriculture Service only takes the Soil pH testing by taking
samples of soil to be measured from a field, then measurements are carried out by lab
tests using a pH meter and using different techniques, or by the calorimeter process.
The results of dimension with conventional systems are less effective because only
the soil pH value can be known from the sample. While all models of soil pH value
of a measured land is unknown, and the measurement time needed is very long [3].
To prevent these all contamination, performing the Soil pH test is very useful for
the first progress of the system. The pH sensor can easily give value on the spot
of farmers land. Farmers can use it to know the nutrients of their land easily and
cheaply. The soil test kit can approximate the amount of nutrients like potassium,
nitrogen and phosphorus, and the soil’s pH. The system takes hazardous samples
of soil for the next process going through a colour sensor for the exact percentage
of nutrients value. The Sensor will help in providing standard recommendations to
reduce human errors. Color comparison and looking for crop recommendations was
eliminated to reduce the time analysis of soil samples. The system was collected of
one digital color sensor and an LDR-RGB LED color sensor for the color analysis
of the soil sample and it is microcontroller-based using Arduino. It also had a data
storage unit for the future use and invention. The stored data get processed using
a program made which provides the analysis and the recommendations of fertilizer
Classification of Soil Nutrients Using IoT 335
use which cannot be contained by the system. This also delivers the database of the
whole process. Standard procedures using the rapid test kit were applied to avoid
incorrect results [4]. Only after the procedures of the rapid test kit, the Sensor was
used for the colour sensing procedure. The only connection of the rapid test kit and
the device was that the device was only a replacement in the analysis of colours to
indicate the nutrient of the soil. Recommendations on the type fertilizers used were
based on the analysis of the data.
2 Required Hardware
The hardware used in the system of classification of soil is pH sensor, Color Sensor,
Arduino microcontroller, Temperature Sensor and Soil moisture Sensor.
pH scale is used to scale the acidity and basicity of a liquid. It has ranged from 1 to
14 where 1 shows acidic liquid and 14 shows basic liquid. It is expected that the pH
scale is argued to range from 0 to 14 or perhaps 1–14, but neither is correct. The pH
range does not have an upper or lower bound since, as defined above, the pH is an
indication of the concentration of H+ . For example, at a pH of zero the hydronium
ion concentration is one molar, while at pH 14 the hydroxide ion concentration is
one molar. One can go somewhat below zero and slightly above 14 in water because
the concentrations of hydronium or hydroxide ions can exceed one molar.
Arduino Uno is IoT device that can handle various sensors. It is a microcontroller
with 14 digital input/output pins, a USB connection is used to connect Arduino to
power supply and used for reprogramming. Arduino is an open-source electronics
platform based on easy-to-use hardware and software. Arduino Boards are able to
read inputs—light on a sensor, a finger on a button, or a Twitter message—and
turn it into an output—activating a motor, turning on an LED, publishing something
online. You can tell your board what to do by sending a set of instructions to the
microcontroller on the board. To do so, you use the Arduino programming language
and IDE.
The temperature sensor is used to sense the temperature in the field and is propor-
tional to the centigrade. The sensors have low output friction, the output and precise
inherent scale make interfacing to the control circuitry easy. A thermocouple or
resistance temperature detector that provides temperature measurement in a read-
able form through an electrical signal. A thermometer is the most basic form of a
temperature meter used to measure the degree of hotness and coolness. Temperature
meters are used in the geotechnical field to monitor concrete, structures, soil, water,
bridges etc. for structural changes in them due to seasonal variations. A thermocouple
(T/C) is made from two dissimilar metals that generate an electrical voltage directly
proportional to the temperature change.
3 Objectives
4 Operating of Sensor
5 Methodology
As Fig. 1 shows first we need to take soil samples from the random location of the
respected land of farmers. Each sample must contain 30–50 g of the Soil sample. The
random location of land is must needed because the contamination and the nutrients
are not same in different locations due to multiple sources of water being given to the
land borewell water is much polluted. Although the fertilizers were given in a proper
way, thus creating various salty parts of land to get average calculated contamination
we must take all random possible locations for soil samples.
After taking the random soil samples they have to mixed for average calculation
this is the way to find possibly accurate contamination and nutrients of the land. The
338 S. Koli et al.
next process is to differentiate the mixed sample’s pH values if random soil samples
were 15 then mixed samples will be divided into 3, 5 per each. Here the 3 mixed
samples get tested for pH value. Some amount of water gets added in mixed sample
for pH value test, the pH sensor’s tip is drowned in the mixed sample after the device
gets on it shows the pH value of that particular mixed sample. Performing the same
task for the remaining mixed sample. After the pH sensor test the mixed samples are
combined with the Rapid test solution for their nutrients and contamination test. The
rapid test kit contains Chemicals for the making solution of the mixed sample. First
it needs to take only 80% of water and 20% of mixed sample for a good amount of
liquid solution it will give better results. Each chemical is different for each nutrients
like phosphorus, nitrogen, zinc etc., however the chemicals get reacted with the
mixed sample it turns into various colors and each color sample is different from
another like pink for nitrogen, blue for phosphorus and orange for Potash. After
adding chemicals into the samples to make a solution it kept 5 min to get a complete
reaction and emit the colors. The level of the darkness and fitness shows the amount
of nutrients is present per cubic feet.
In the color sensing process, each sample solution is kept under the color sensor
TCS2300 unit, configured to emit light and get refracted from the solution. The
refracted light has data of how much nutrients are present and how the sample is
contaminated. For accurate results, the Sensor’s data is passed through the Arduino
Uno that verifies and gives the actual percentage of nutrients.
The dataset was first grouped assuming that there were two clusters and analyzed
for consistency. These two clusters would represent “naturally contaminated” soil
samples and “unnaturally contaminated” soil samples. The datasets were subse-
quently grouped with the assumption that three or more clusters were present. These
cluster centroids would also represent different types of contamination and poten-
tially different sources of contamination. The features were investigated to determine
a hierarchy of significance. The importance of a metal with respect to its impact on
classification was evaluated by comparing the standard deviation of the lognormal
concentration spread to the “distance” between the centroids at that feature. The
closer the concentrations at the centroids, the less “important” it was deemed to
Classification of Soil Nutrients Using IoT 339
be. The data is collected in the all the format for analysis and reporting purpose,
In analysis phase the each of mixed sample results get passed under the method of
percentage of nutrients present in the sample. If potash is more dark in orange, then
the potash amount is more than enough to present in the soil, but it won’t damage
crops. All the nutrients like Nitrogen, Phosphorus and Potash get classified for accu-
rate results. The results show much better than imagination. If the contamination
is more than land there must be a Decontamination process in which land is acidic
then use the basic fertilizer if land is basic then use limestone to maintain pH of the
land. If the nutrients are not enough to grow good crops, use fertilizers suitable for
land like nitrogen is less present in the land then use fertilizers that have nitrogen in
good amounts. Here in the solution phase, the farmers know how much their land is
contaminated and how much nutrients are present in the land.
6 Result
Efficient results are captured from the project. Tables 1 and 2 shows the recorded
knowledge at the field in many days. The hardware implementations are roaring and
reliable, and therefore, the sensors that we tend to exploit are giving idle results and
playing to the belief.
As Table 1 shows the temperature of soil and atmosphere is different for each day
also the moisture. Here we can say the atmosphere’s temperature directly affects the
moisture present in the soil.
As Table 2 shows, the pH values are different for each day because of soil moisture,
the nutrients of soil can be managed by water sources in a quality way.
Table 1 Readings of
Date and time Atmospheric Soil Soil moisture
temperature and soil moisture
(1.30 p.m.) temperature temperature (%)
25/05/2020 29 28 32
26/05/2020 28 27 30
27/05/2020 30 29 28
28/05/2020 32 31 26
A Soil Wetness Sensor has many appeal, especially in agriculture. Irrigation is a key
factor in farming. Detecting the amount of moisture and temperature in the soil and
turning on the system when the wetness level falls below a predefined value helps
keep away a lot of wastage of water and human resources. These classes of sensors
make automatics easier. A database can be formed. It can be used to direct the types
of acids, alkalis or salts present in the soil. Saltiness of soil can also be calculated
by correlating it with the output voltage. Signal carrying of the output data directly
to the user can be done using Zigbee or Bluetooth. We can get the values from a
stored database in PC so that the soil’s moisture holding capacity can be determined.
Indian farmers are facing a lot of problems, but the advancement of wireless sensor
networks in agriculture would be promising in the present scenario of water scarcity
and unpredictable weather conditions. This paper provides implementation of WSN
based soil moisture monitoring system. As WSN is battery operated to enhance the
lifetime. EWMA event detection algorithm is used which generates events only when
threshold conditions are met. Rest of the Time, the nodes are in a sleep state which
can save their energy. This work can also be further extended by considering more
than one sensor module.
8 Conclusion
References
1. Angelakis AN, Rolston DE, Kadir TN, Scott VN (1993) Soil-water distribution under trickle
source. J Irrig Drain Eng ASCE 119:484–500
2. Zhang M, Wang H (2009) Characteristics of soil heavy metals pollution around mine waste
piles. In: International conference on ES and IAT. https://doi.org/10.1109/2009.319
3. Prastika YFD, Zani NR, Arifin W, Alasiry AH (2019) Design and development soil pH mapping
portable system for crop selection using fuzzy algorithms on agricultural land. In: IEEE Asia-
Pacific conference on geoscience, electronics and remote sensing technology (AGERS), Jakarta,
pp 43–48. https://doi.org/10.1109/AGERS48446.2019.9034381
4. Regalado RG, Dela Cruz JC (2016) Soil pH and nutrient (nitrogen, phosphorus and potassium)
analyzer using colorimetry. In: IEEE region 10 conference (TENCON), Singapore, pp 2387–
2391. https://doi.org/10.1109/TENCON.2016.7848458
Non-invasive Methodological Techniques
to Determine Health of a Bone
1 Introduction
Bones are living and continuously developing tissue. Bone strength continually
changes throughout life. In the remodeling process, some bone cells dissolve and
new bone cells get developed. Bone isn’t uniformly solid, but consists of a versatile
matrix and bound the minerals. Bone matrix is collagen that gives a soft framework
to the body, while minerals add strength and hardness to the body framework [1].
M. R. Khare (B)
Department of Electronics Engineering, Walchand College of Engineering, Sangli, India
R. H. Havaldar
Department of Biomedical Engineering, Dr. M. S. Sheshgiri College of Engineering and
Technology, Belagavi, India
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 343
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_34
344 M. R. Khare and R. H. Havaldar
This mixture of collagen and minerals makes bone strong and versatile enough to
face up to different stresses.
By considering weight as a parameter, 60% of bone tissue is the inorganic
constituent, 30% organic constituents, and the remaining 10% is water. While by
considering volume as a parameter, 40% is the inorganic component, 35% is the
organic component, and the remaining 25% is water [2]. The bone may be a reser-
voir for various minerals. Bone minerals are categorized into two types like macro and
micro. Macrominerals include Calcium (Ca), Phosphorous (P), Magnesium (Mg),
Potassium (K), Sodium (Na), and Sulphur (S). Micro minerals contain Copper (Cu),
Iron (Fe), Manganese (Mg), Zinc (Zn), Cobalt (Co), etc. Both types of minerals
hold immense importance to ascertain the quality of a bone. The requirement of
Macrominerals is more as compared to micro minerals [3]. Calcium is the major
macro mineral present in the body. It plays a vital role in the formation of bones,
teeth, and as a messenger in cell signaling. The minerals Calcium and phosphorous
are present in the complex form of insoluble salt called hydroxyapatite [4]. The
variation in the proportion of Calcium to Phosphorous contents may make the bone
porous or weak. Consequently, this provides rise to osteopenia, osteoporosis, and
the probability of rupture in a bone [5]. Hence, it’s significant to determine the bone
mineral contents. Mineral contents and mineral density can be assigned with each
other through areal measurements. The areal Bone Mineral Density (aBMD) is the
ratio of bone mineral content with a particular area of the bone (ROI). Volumetric
Bone Mineral Density (vBMD) indicates the relation of bone mineral content with
the volume of the particular bone region [6].
The determination of BMD can be done either by invasive method or noninva-
sive method. The invasive method is applied to small animals like mice, rabbits,
hamsters, pigs, and dogs [7]. The method requires an actual bone sample in the form
of ash for mineral content measurement. To obtain ash of animal bones, first, they
are dried at low temperature and then burnt at high temperature so as to get bone ash.
Due to this drying and burning process, organic contents of bone vanish and only
inorganic constituents remain [7]. Scanning electron microscopy–Energy-dispersive
X-ray (SEM–EDX) [8], XRD, XRF, and LIBS [9] are some methods used to quan-
tify inorganic constituents from bone ash. These techniques have high accuracy but
require special sample preparation (in vitro) and are often slow that may take 5–
7 days to get the results [10]. So applying the same invasive technique on human
bone is not as suitable as on animal bones [11].
The noninvasive method measures the BMD, by using different scanning tech-
niques on live animals [7]. Quantification of bone density to diagnose the health of
bone is done by using quantitative imaging techniques. The purpose of Non Inva-
sive imaging techniques is to study the clinical tools for quantification of inorganic
Non-invasive Methodological Techniques to Determine Health … 345
constituents (in terms of BMD) to determine the health of a bone. The computable
imaging techniques are extensively used due to its more accuracy, reusability, and
validation on a huge number of populations.
The noninvasive clinical techniques to diagnose the health of a bone include the size of
cortical thickness, quality analysis of trabecular bone, and radiolucency from radio-
graphs. These parameters are used to quantify mineral density. Total body weight is
supported by bones like the femur, hip, forearm, etc. As a result, these bones are more
prone to fracture risk. This section describes different clinical imaging techniques
applied for the clinical analysis of the health of bone.
2.1 Radiography
X-ray source emits two types of radiations as low energy and the other high energy
[12–14]. When radiations are passed through a sample, inorganic contents present
in that sample absorb some energy. The unabsorbed energy is detected by an X-
ray detector. The difference between source energy and unabsorbed energy helps
to determine Bone Mineral Density. The value of BMD (g/cm2 ) is multiplied with
related area to give BMC (g) [14, 15]. Thus a bone mineral density measures how
much calcium and other inorganic contents are available in a particular region of the
bone.
DEXA is a standard clinical technique that gives an areal amount of bone mineral
density [12]. In DEXA scan imaging technique, bone is radiated by photon beams of
2 energy levels, high energy of greater than 70 keV and lower one is 30–50 keV [16].
DEXA has good precision, less contact to radiations, and a fast scanning time [17].
Determined BMD is used to diagnose the health of bone. The sample for DEXA scan
of the femur bone is shown in Fig. 1.
The health of a bone is categorized as normal bones, osteopenia bones, and the
one with osteoporosis based on BMD values as shown in Fig. 2. The obtained results
are helpful to determine the strength of bone. DEXA BMD outcomes are classified
from standardized T-score and Z-scores values (Table 1) [18]. It shows a variation of
measured bone density to that of a healthy person. The lesser the T score, the more
fragile bones are.
346 M. R. Khare and R. H. Havaldar
The drawback of DEXA scan is its more cost and limited availability [18, 19].
QCT gives a volumetric quantity of bone mineral density. In this imaging technique,
applied X-rays absorb by tissues of a bone and scatter the same at various directions
across the Region of Interest. To measure BMD values, samples of various bone
mineral densities are used. The received attenuation signal is used (Hounsfield units)
to convert into BMD values [12]. The obtained T-scores of QCT method are lower
than that of DEXA method. Martin-Badosa et al. have investigated the risk of fracture
in terms of volumetric BMD as shown in Table 2 [19, 20].
For QCT imaging technique WHO has given following range of BMD as shown
in Table 3 [21].
There are different types of QCT like Multidetector CT (MDCT), High-Resolution
Peripheral (HR-pQCT), Synchrotron radiation micro Tomography (SR-µCT).
MDCT offers a good spatial resolution. Bone Mineral Density and extracted
microstructures help to distinguish between healthy and unhealthy groups [22, 23].
HR-pQCT helps in getting higher accuracy volumetric images of the bone. A stan-
dardized analysis is used to quantify volumetric BMD. Due to marginal quantifica-
tion, the technique is appropriate for patients paining from spinal and hip fractures
[24, 25]. SR µCT Technique is used to obtain three-dimensional images to quantify
trabecular bone architecture [26]. This paper proposes a technique for measuring
the amount of minerals in bone samples using 3-Dimensional SR µCT images.
The captured 3 D image slices are considered to find gray levels of the images.
Variations in grey levels indicate different mineralization. Regions with darker grey
levels indicate low mineralization and newly formed bones. To calibrate the results
of mineral contents, homogeneous water solutions with different concentrations of
Dipotassium hydrogen phosphate is used. The result obtained from the chemical anal-
ysis was compared with those obtained from microradiography. SR µCT technique
provides overall mineral contents as well as the three-dimensional microarchitecture
of trabecular bone.
But Overall Limitation of any QCT method is that it has high radiation as compared
to DEXA technique.
It is a non-invasive method to assess bone structure, bone quality, and its microar-
chitecture [12, 27]. MRI is a non-ionizing Imaging technique that helps in volume-
wise checking of trabecular bone as well as cortical bone. When compared to QCT
technique MRI technique has no radiations but more expensive [28].
It is a hi-technology texture analysis method used to measure BMD [12]. The tech-
nique is about showing an association of DEXA–BMD values of forearm, femur
and spine. DXR has less possibility of human error and results are comparable with
DEXA However, the volumetric BMD measurement by this technique is not as
precise as compared with QCT technique.
3 Conclusion
The invasive and noninvasive methods are available to find BMD. The invasive
method needs actual bones to be processed to create ash of sample which is quite
a time-consuming process and requires ethical permissions. The benefit of imaging
techniques is lesser time consumption compared with invasive techniques. In this
review, different Imaging techniques like radiograph, QCT, QUS, MRI, DXR and
DEXA have been discussed. QCT and MRI are the techniques that offer volumetric
BMD but they are expensive. DEXA is a two-dimensional technique for the determi-
nation of areal BMD which is precise, more accurate but it has disadvantages like low
availability to date. For DEXA, the bones of different lengths may produce the same
BMD results by maintaining the equal ratio thereby indicating incorrect strength of
bones. Hence along with the area, we have to take into account the thickness of the
bone to predict BMD. So it’s the need in today’s situation to develop a mathematical
model to provide a better clinical analysis of bone health without additional cost and
distraction.
References
1. Boskey AL (2013) Bone composition: relationship to bone fragility and antiosteoporotic drug
effects. Int Bone Miner Soc
2. Feng X (2010) Chemical and biochemical basis of cell-bone matrix interaction in health and
disease. Curr Chem Biol
3. Costa e Silva LF, Filho SDCV, Engle T, Rotta PP (2015) Macrominerals and trace element
requirements for beef cattle. PLoS One. https://doi.org/10.1371/Journal.Pone.0144464
4. Swetha M, Sahithi K, Moorthi A, Srinivasan N, Ramasamy K, Selvamurugan N (2010) Biocom-
posites containing natural polymers and hydroxyapat ite for bone tissue engineering. Int J Biol
Macromol 47(1):1–4. https://doi.org/10.1016/j.ijbiomac.2010.03.015
5. Faibish D, Ott SM, Boskey AL (2006) Mineral changes in osteoporosis—a review. Clin Orthop
Relat Res 443:28–38
6. Humbert L, Martelli Y, Fonollà R, Steghöfer M, Di Gregorio S, Malouf J, Romera J, Del Río
Barquero LM (2017) 3D-DXA: assessing the femoral shape, the trabecular macrostructure and
the cortex in 3D from DXA images. IEEE Trans Med Imaging. https://doi.org/10.1109/Tmi.
2016.2593346
7. Kim HS, Jeong ES, Yang MH, Yang S-O (2018) Bone mineral density assessment for research
purpose using dual energy X-ray absorptiometry. Osteoporos Sarcopenia 4
8. Ellingham STD et al (2017) Scanning electron microscopy–energy-dispersive X-ray
(SEM/EDX): a rapid diagnostic tool to aid the identification of burnt bone and contested
cremains. J Forensic Sci. https://doi.org/10.1111/1556-4029.13541
9. Tariq U, Haider Z, Tufail K, Hussain R, Ali J (2016) Determination of calcium to phosphate ratio
in hydroxyapatite extracted from bovine bone using libs, vol 2, pp 48–53. eISSN 2504-8546
10. Sluiter A, Hames B, Ruiz R, Scarlata C, Sluiter J, Templeton D (2005) Determination of ash
in biomass, laboratory analytical procedure
11. Khalid M, Bora T, Al Ghaithi A, Thukral S, Dutta J (2018) Raman spectroscopy detects changes
in bone mineral quality and collagen cross-linkage in staphylococcus infected human bone.
Sci Rep 8:9417
12. Areeckal AS, Sumam David S (2019) Current and emerging diagnostic imaging-based
techniques for assessment of osteoporosis and fracture risk. IEEE Rev Biomed Eng 12:254–268
350 M. R. Khare and R. H. Havaldar
Abstract In Today’s world, security frames the most essential segment of our lives.
Face Recognition is an important part of the purpose of the security and surveillance
field. A small project which does face detection using OpenCV library on Raspberry
Pi. Face Recognition/Facial Recognition is a category of biometric software that
identifies people by their faces. The face is captured by the digital camera and the
system is trained and then it is capable of identifying the person. This paper focuses
on the implementation of a face detection system for human identification based on
the open-source computer vision library (OpenCV) with python. We also proposed a
hierarchical image processing approach to reduce the training or testing time while
improving recognition accuracy.
1 Introduction
The Concept of image processing through the python OpenCV platform has been
used for human identification through face detection. Human Identification means
to recognize particular people through their unique structures like fingerprint, palm,
and iris, and face detection. This paper is based on the implementation of the face
detection system. source library in which the source code is open and it is useful in
the visual field such as image processing. The main motto of this work is to take
and manage attendance using face recognition. The testing of this technique has
proceeded through Raspberry Pi devices [1].
Face recognition is an active and important research topic from the 1970s. Given an
input image with 1 face, face recognition systems typically first run face detection to
isolate the faces. Each face is preprocessed and then a low-dimensional representation
(or embedding) is obtained.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 351
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_35
352 P. R. Dolas et al.
2 Literature Survey
In the proposed system, we use the camera to accomplish the pictures when a motion
detects via the PIR sensor. Subsequently, we will apply a computer vision module to
the caught pictures to discover the faces. This system is extremely helpful and vital if
we want to protect the area. The application can be divided into two parts which are
motion detection and face detection. The system will not go to face detection if there
is no motion discovered. But, if a movement has been detected, then the detected
movement of the current frame will be processed by the algorithm.
A review of Face Recognition System Using Raspberry Pi in the Field of IoT Jain
et al. algorithm of face detection. To explicitly define a low-dimensional face repre-
sentation based on ratios of distances, areas, and angles. An explicitly defined face
representation is desirable for an intuitive feature space and technique. However,
in practice, explicitly defined representations are not accurate. Later work sought
to use holistic approaches stemming from statistics and Artificial Intelligence (AI)
that learn from and perform well on a dataset of face images. Statistical techniques
such as Principal Component Analysis represent faces as a combination of eigenvec-
tors. However, these deep neural network-based techniques are trained with private
datasets containing millions of social media images that are orders of magnitude
larger than available datasets for research [2, 3].
3 Methodology
Raspberry PI is a small computer that has been used to develop an embedded system
to perform a specific particular task. This electronic module has been operated with
the use of the Raspbian operating system and is based on the LINUX platform.
This module has an SD card slot, Inbuilt WIFI, and Bluetooth Connectivity, 40
GPIO pins for Input–output operations, PI Camera Port, PI Display port, Audio Port,
HDMI cable Port, 4 USB port for connecting pen drive, Mouse, Keyboard, USB
Camera, etc. It also has an Ethernet port for data sharing as well as network sharing
between computers and raspberry PI [4].
The Functionality of this system is mainly categorized in the following steps:
To enroll and detect faces using a camera connected to the ARM Cortex of
Raspberry Pi board.
To display the match status on the LCD as well as the terminal running on the
VGA (Video graphics array) monitor.
To program for the same using python language. The code imports certain modules
that enable functions such as face recognition, GPIO modules. The identification
and authentication technology operate using the following four stages shown in
Figs. 1 and 2 clear explanations in the modules are presented.
Face Detection and Recognition Using Raspberry Pi 353
Fig. 1 Block diagram of implementation of face detection and recognition system using Raspberry
Pi
Efficient human Identification through face detection has carried out relevant infor-
mation. The performance of the system is based on three steps which are datasets,
trainer, and detector python script. An algorithm that has been used for image
processing in OpenCV and especially for face detection as shown in Fig. 3. As
we talk about future modification of this project, it will be used in a high-security
system and face detection-based attendance system. Also, we can develop projects
based on image processing.
References
1. Patel T, Shah B (2017) A survey on facial feature extraction techniques for automatic face anno-
tation. In: IEEE international conference on innovative mechanisms for industry applications
(ICIMIA), Bangalore
2. Foster I, Kesselman C (1999) The grid: blueprint for a new computing infrastructure. Morgan
Kaufmann, San Francisco
3. Yang MH, Kriegman DJ, Ahuja N (2002) Detecting faces in images: a survey. IEEE Trans PAMI.
Face recognition based on elastic template. Beijing University of Technology, China
4. Mathur S, Subramanian B, Jain S, Choudhary K (2017) Human detector and counter using
Raspberry Pi microcontroller. In: IEEE international conference on innovations in power and
advanced computing technologies i-PACT2017, Vellore
Human Tracking Mechanism
for Institutions Using RFID and Facial
Recognition
Abstract The Adoption of digital content by the institutional members and students
have seen rapid growth in recent years. The students could access the content across
various devices, platform and applications, which has a direct implication on the
physical presence of the student. The institutions follow old and very traditional
base approach like a manual record of attendance to track the company of students
which consumes a lot of time and efforts from the staff members. The study looks at
the various technologies available in the market and present the implementation of
the best possible solution. The latest use of technology of Facial Recognition with
a combination of RFID will enhance the tracking process and also provide valuable
insight into student behaviour. The data collected by the system can further utilize
to improve the efficiency and effectiveness of student behaviour patterns and predict
the learning trend, which will help the institutions to make the correct decisions.
1 Introduction
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 355
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_36
356 R. Shaik and L. V. Patil
institution employees. Most of the Indian educational institution does not have the
correct framework to address these problems [1, p. 3]. Some of the common issues
faced in the traditional use of tracking the presence of student are manual errors, loss
of recorded copies which will not help the students and well as institutions. So there
is a need to have a mechanism which could automate the process of capturing the
presence of student without any manual intervention.
The existing use of technology must introduce an electronic framework that can
oversee and assist the faculty members with taking the attendance rapidly and error-
prone. The institution can quickly implement the automated framework digitally
without manual intervention. The other concern of having the attendance records in
the copy form is that the faculty member may lose it and would be difficult to retrieve
it [2, p. 2].
The digital framework is generating server related activities, which are reaching
in terabytes. The student activities and digital educational system modules receive
an enormous amount of data daily. Traditional analytical programs prolong to meet
the analysis requirements as the data requirements are exploding. To identify the
behaviour correctly, we do need a mobile base application that makes easy to access
the digital format on all the available platforms and necessary devices. In the current
days, mobile base application on various platform like Android, IOS or Windows
helps to capture correct data is provide a basis for a huge number of simultaneous
processing and analysis to get the valuable insights. The institutions for studies
are attaining valuable insights based on the processing of analytical tools to make
improvement on the education sector on the larger different culture base population
of students. While the improvement with respect to digital content are progressing
quicker, biometrics innovation are progressing at faster pace. These innovations
address the students personality through various features like fingerprints, face detec-
tion, Iris detection, retina scan samples, palm prints, voice recognition and so on. The
techniques used related to these features helps to identify and validate user authen-
tication that could be more trustworthy than other type of security system, which
involves password or identity cards. The biometric data are more trustworthy and
error prone and can contain it for more duration of time [3, p. 6].
It is atmost critical to validate the correct devices and instruments to use in logical
test cases. Common standard validation methods like radio frequency device frame-
work, Bluetooth devices, and NFC devices are few cases which can be utilise in
the process. The upfront cost of these techniques are very costly to do the imple-
mentation and has lot of restrictions to use it. In today’s generation digital related
validation process contains number of biometric related devices and systems. The
improvements in these areas and the techniques usage depends on physiological or
social features. Biometrics denotes as the “robotized distinguishing proof or confir-
mation of human personality through the estimation of repeatable physiological and
conduct attributes”. Different kinds of biometric methods are listed (Fig. 1).
Human Tracking Mechanism for Institutions Using RFID and Facial … 357
2 Related Work
The section eventually reveals some facts based on thoughtful analysis of many
authors work as follows.
Core Research recommended the utilization of a mechanized participation frame-
work, which can lessen human contribution, manual information section mistake, and
monotonous work. This framework is going to improve profitability, reduce finance
mistake, and brought down compensation swelling, decreased extra time, the retire-
ment of heritage frameworks, Elimination of paper expenses, and which can give all
the reports on request. In this framework, the workforce needs to gauge participation
physically, and just these records must be gone into the advanced framework [4, p. 2].
Another similar kind of project proposed, but as part of use case, the student should
register themselves using client-server socket program from their own devices such
as laptops. A session of the application creates an execution process for the list
of students for a specific course. These courses display when the faculty members
approve the request or start the application. The student presence is marked when
clicked to checkbox of the student name, and afterwards, click the section button to
stamp their presence.
In 2013, Bhalla et al. [5], have recommended the participation framework, which
can gauge participation in applying Bluetooth. In this task, presence marked using
teacher’s cell phone. Application installed in faculties mobile device which triggers
the notification on student mobile devices using Bluetooth and the data exchange
by PDA Media Access Control (MAC). The student appearance marks once they
acknowledge the report.
358 R. Shaik and L. V. Patil
In Ayu [6] suggested the online participation framework using NFC innovation
which named as Tochin System. In the framework, two methods of activities used
where one of them is writer mode and other is a peer to peer mode. In this framework,
each room of the institution has NFC enabled, which is a program to connect to the
faculty member’s computer.
In 2012, Josphineleela and Ramakrishnan [7, p. 4] presented a framework in
which participation has done using one of the biometric features of finger impres-
sion. This framework can apply for students and faculty members of the institu-
tion. In this method, the unique finger impression acknowledged for the presence
of the subject, and it is orchestrated into the accompanying modules Pre-preparing,
Minutiae Extraction, Reconstruction, Fingerprint Recognition and report. This new
calculation reproduces the stage picture from Miniature.
In 2010, Kadry and Smaili suggested using the iris-based framework, which is one
of the techniques of bio-metric. A remote iris base organization framework executed,
utilizing Daugman’s calculation. This biometric base framework is challenging to
implement due to the complex process of data. The implementation of the framework
is too expensive and time-consuming authentication. The use of the iris base tech-
nology is implemented in the financial sector, such as banks and financial institution,
where complex authentication is the highest priority.
3 Implemented Framework
The human tracking system is a global application framework to improve and auto-
mate the manual intervention to write down and transmit in near real-time. The proper
way of implementation consists of two options which are related to authentication
used by the RFID base device and another option is to use the biometric technique
such as facial recognition which matches with the saved images. Once the match
is successful, then presence is marked against the authenticated subject, or else it
is marked as absent. Above implementation helps to automate the subject tracking
mechanism with fraud tolerance. Figure 2 demonstrates the system workflow by
using biometric facial recognition along with the RFID authentication to track more
efficiently.
The functionality of the system, includes registration, near real-time tracking, are
as follows.
A. Student and Faculty Registration Platform
During admission of a student in an educational institution, registration form and
process is explained to register into the system. During the process, the institu-
tion will issue an RFID Tag whose unique code data set saves into the database
concerning its name. The student has to undergo through the biometric fingerprint
scanner where fingerprint information stored in the database respective to their terms
at the same time. The exact process is followed by the faculty members of the insti-
tution to register themselves with necessary information into the system. The faculty
Human Tracking Mechanism for Institutions Using RFID and Facial … 359
members can search for the specific student either by the name or ID. The system
automatically calculates the attendance percentage as per the pre-configuration set
by the authorized member of the institution. The teachers can later check the atten-
dance percentage of the student for a particular period of days. It is a mandate for
faculty members to update the courses related information into the system regularly.
Different authorization roles are created while setting up the system where the faculty
team needs to update their profile, attendance and leave base information.
Figure 3 demonstrates the data link of the process flow or system. Another critical
functionality in the system that the student who registers for the different semester
can track their daily class routine for the ongoing courses.
The administration of institution can also frequently monitor their faculty
member’s performance metrics which is calculated by the system automatically based
on the student presence corresponding to each course. In this way, institutions can
improve the quality of education for students.
B. Near Real-Time Tracking
The tracking mechanism/framework can mark the daily attendance of enrolled
students, teachers, and other institutional staff automatically without any manual
intervention. The RFID readers are installed at various places like entrance, exist
gates and hotspots of the institution. Whenever a student enters the premises, the
RFID automatically reads the unique code of RFID passive tag from their identity
card and data is transferred to the server and get saved in the database. When the
student present for an enrolled course, the RFID reader reads the unique code of the
tag and cross-checks that information present in the system database. The checks
against the attributes like semester number, subject code, classroom information and
360 R. Shaik and L. V. Patil
so on. After scanning the RFID, student photo is captured, and the image information
is transfer later to the server for processing the image against the database copy. If
the match is found, then the presence of student for that enrolled course is marked as
a success. The facial recognition device will only be active at the scheduled time the
course. The teacher who is delivering the lecture can decide on the activation time
of the system where he can set the rule to mark the attendance as absent as penalty
if the student visits the class late base on the threshold value.
The installation of the system completes at a faster pace due to less complexity.
The facial recognition unit can install on the entrance of each classroom but not at all
places, which can help to reduce the cost for the institution. The information which
is present in the database can later put in use to analyse the data to take the necessary
measure to improve the quality of education for students.
C. Mathematical Expression/Algorithm
A model can explain its effects on a different level and predict the correct implemen-
tation of the system. The following equation could show the mathematical expression
of Face validation:
X = { , E, δ, D}.
Y = Face Recognition.
= set of input values = {audio content, video content, characters}.
F = set of output symbol content = {Match concerning inform r, Match Not Found}.
Human Tracking Mechanism for Institutions Using RFID and Facial … 361
1
M
ψ= i
M i=1
where
= average no. of images, M = number of image objects,
Gi = denotes the images vector. i = Gi − .
Where, i = 1, 2, 3… M.
A = [1, 2, 3…M].
– Computation of covariance matrix: (AA ).
– Compute eigenvectors of the c covariance matrix.
– Compute Eigenfaces = No. of training image − no. of classes (total number of
people) of eigenvectors.
– Create reduced Eigenface space. The selected set of vectors are then multiplied
by the A matrix to generate the reduced result set.
– Evaluate the Eigenface of an image object determines in the question.
– Next step would be the calculation of Euclidean distance for the Eigenfaces and
the image object.
– Now the action is to search the insignificant Euclidian distance.
– The output is as follows: Image object with the less Euclidian space or image is
not recognizable.
D = {the system cannot process the voice dataset, Eigen generates the grey image
object; the key algorithms will execute till frames.}.
This section depicts the outcome of the system which intends to track human partic-
ipation which can analyze the behaviour. There are many technologies which are
present for monitoring the students like manual, Bluetooth base, infrared, Wi-Fi-
enabled application. The solution is to create feasible and affordable devices which
can remove the corns of the previous implementation. The framework uses RFID
which can read several labels and process it in parallel, which is one of the most
crucial features. The RFID works on the theory named Automatic Identification and
362 R. Shaik and L. V. Patil
percentage, which is 96% in total. After tweaking the external parameters, we see
the increase in ratio to 100 rates, also considering the time spent by the subject with
the tracking devices.
Table 2 demonstrates the elapsed time by the system concerning the previous
technology-based implementation. The average time consumed by the solution is
far less as compared to the survey performed on the other tracking technologies and
devices. The average time is 8.7 s as compared with 11.6 s, which consider as to be
the most promising results.
5 Conclusion
makes the affordable solution to the educational sector. The data captured by the
system can feed into various analytical tools to create different learning patterns as
well as a retention plan for the students. The data can also be utilized to develop
a predictive model of student behaviour which helps students and institutions to
provide better and quality education.
Acknowledgements I am incredibly thankful to my guide Dr. L V. Patil for suggesting the topic
for research and providing all assistance to complete the work. His guidance and discussion with
him are valuable in the realization of the work.
References
1. Srinidhi MB, Roy R (2015) A web-enabled secured system for attendance monitoring and real
time location tracking using biometric and radio frequency identification (RFID) technology. In:
International conference on computer communication and informatics, 978-1-4799-6805-3/15
2. Eridani D (2015) Simulation of attendance application on campus based on RFID (radio
frequency identification). In: Proceedings of 2015 2nd international conference on information
technology. 978-1-4799-986-3-0/15
3. Sundararajan M, Mahalakshmi M (2015) Tracking the student’s performance in web-based
education using scrum methodology. In: 2015 international conference on computing and
communications technologies. 978-1-4799-7623-2/15
4. Nandyal S (2015) An automatic attendance system using image processing. Int J Eng Sci. ISSN
2319 – 1813 ISSN
5. Noguchi S (2015) Student attendance management system with Bluetooth low energy beacon.
In: 2015 international conference on network based information system. https://doi.org/10.1109/
NBiS.2015.109
6. Benyo B, Doktor T (2015) Student attendance monitoring at the university using NFC. In: 2015
wireless telecommunication symposium. https://doi.org/10.1109/WTS.2015.6266137
7. Gunawan TS (2015) Design and development of portable classroom attendance system based
on Arduino and fingerprint biometric. In: The 5th international conference on emerging
technologies. https://doi.org/10.1109/ICT4M.2015.7020601
Monitoring Power Consumption
and Automation Using IOT
Riya Jain, Revati Awale, Neha Kothari, Swarali Shah, and Amit Kore
Abstract Today, the entire world is frequently facing a challenging and speedy
environment everywhere. The main problem we all are going through is the energy
crisis. Actual energy demand made on existing energy supply is Electric energy
consumption. Early India has recorded a swift growth in electricity generation since
1985 which is increasing from 179 TWh in 1985 to 1057 TWh in 2012. The rapid
increase of this came from non-conventional renewable energy resources (RES) and
coal-fired plants. In India gross utility of electricity generation was about 1384 billion
kWh in 2019–20 which represents 1.0% growth yearly as compared to 2018–2019. A
country’s electric power development is an important measure by annual electricity
consumption per capita. It reached 3084 kWh, up 42.3% from 1990. Sensor networks
have been incorporated in the management of buildings for organizations and cities.
In recent it has led to an exponential increase in the volume of data available, and
monetary savings. For this purpose, new approaches, alignments and techniques are
required to investigate and analyze information in big data environments. For this
problem, having a relevant system to monitor the power usage is the only solution.
This paper proposes an analytical and prediction model by using different energy
profiles, which will provide the user power consumption chart of a consumer over a
period of time, to perform quantitative analysis using smart meters that automatically
acquire context information. There are devices which are capable of measuring a
customer’s energy consumption, for example smart meters. In this paper, there are
two modules. The first one focuses on receiving the data from the smart meter and
also transfers it to the data analyst. The second module is the predictive module which
uses consumption data and information of the consumer in order to understand the
behavioral patterns of the consumption of electricity. These models can be used to
predict energy consumption and also identify irregularities and outliers. The customer
gets acknowledged about abnormal usage. On getting acknowledged the consumer
will get to know where and when exactly the power usage is increasing and can
control power consumption which will ultimately result in reduction of expenses.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 365
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_37
366 R. Jain et al.
1 Introduction
by these power saving modes, makes sure that whether it is not wasted and that all
is utilized by the consumers.
Various Factors influencing Electricity Prices are: Fundamental Factors, Fuel
Prices Temperature, Weather Conditions Time indices such as day of the week, month
of the year, season of the year Cost of production of electricity per unit Operational
Factors Power Transmission Congestion Power System Operating condition, Elec-
tricity production (deficit/surplus), Network Maintenance, Electricity Load, Strategic
Factors Power Purchase Agreements, Bilateral Contracts, Power Exchange, Bidding
Strategy, Market Design, Historical Factors, Price, Demand.
2 Literature Survey
3 Proposed Methodology
Figure 2 explains us the flow of our project. The Smart Meter Data is stored in the
form of an excel sheet. The data is programmed to be received by the database on
every minute basis. The data sending procedure is called the auto generator in the
diagram. The data is then received by the Mysql database and stored. As the location
to store data is the database, no extra technique or method is used to do the process
[2]. The data is then validated and cleaned. Based on this data, the prediction on
future consumptions is done. In this system, one can get predicted data within a span
of a week. The predicted data directly shows the overall expected consumption in
units and in price. A user can store much wise data upto a month in the system. So
that the user can easily compare the electricity bill he paid and the bill on the app.
A report is generated of electricity consumed allowing the user the choice to select
the timespan of the report. The entire data is easily accessible and visible on the app.
The app, made in android studio, help the user to view data in graphical form and the
price of consumption [3]. The app is the interface to view the graphical form of the
content stored in the database. The Smart meter data used here is from a pre-installed
meter which reports reading on a regular basis of one minute. This data is in excel
sheet format. The sheet involves many parameters of electricity, like, voltage, current,
kvarh, kWh, pf and many more. This data with its every minute detail is being stored
in our database of application. The database is stored in the system. The database is
accessed using MySQL. There are 2 tables in the database, one of the users details,
second of the every minute detail of electricity. These use simple and few complex
queries of MySQL. The app also provides a feature of viewing prediction to the
user. The prediction is available for all parameters as well as for total consumpy and
bill. The restriction for prediction is upto 5 days. Prediction is basically a method
of forecasting based on data analytics which requires a trained model. The model of
prediction needs to be trained on the data set. The data set is then divided as a test
data set to check how appropriate it predicts. Various techniques are used and the
one which gives the maximum percentage of approximate value is usually chosen.
As prediction is only the forecast based on the past data, it shouldn’t be considered
that this is the exact value one will get in future as it could be influenced with many
unknown and unexpected factors. The interface of the app shows many pages which
can give a good idea of data. The main page has graphical bar representation of the
current data. The graph is the price of the consumed electricity bill along with the
units consumed. Next page has the historical data. It shows all the past consumption
a user has done right from the data has started to store. It shows data in parameters.
Each parameter is shown with its consumption till date. Next page is the prediction
page where the user can watch his prediction of the selected date. Each parameter
prediction is visible to the user. Next page in the app is report generation. This
feature helps the user to compare or see the record of his consumed electricity. The
generation of reports requires starting and ending dates. The total consumption unit
wise and price wise is visible to the user as it is in the electricity bill paper. This
will help the user to compare the bill received from MahaVitaran and the one the
app shows. The entire functioning of the system as a prototype is done using xampp.
When the xampp MySQL servers are switched on, the device like mobile or laptop
can access it using the localhost. For localhost connection all devices need to be in
the same network/router. Only then the devices will be able to access the database on
the server. A virtual network interface is created from which the user can request data
using the port 80. The virtual network interface revives the request and then sends
the request to the local web server services. The request is answered back in the same
flow, i.e. the data requested is sent in reverse flow. The tool used for communication
is the IP address. In this prototype, the above said method is used as developers use
the same technique for testing applications which work over the internet.
Monitoring Power Consumption and Automation Using IOT 371
4 Result
Then the user selects the time scale and finally a graph according to his requested
data is plotted. For a clear understanding we can refer to the pictorial representations
shown. The system also enables the user to keep a check on the past consumptions.
Users can also predict their consumption for the upcoming days and can get a detailed
report of any day including prices of their consumption.
Here, in the screenshots, Fig. 3 is the signin/signup page where the user has to enter
his information to login in into his account. The user has a choice to select the
page that should be visible. Figures 4, 5 and 6 show the data stored in the database
according to the day. Figure 7 shows the historical data stored in the database in bar
chart form. Also Fig. 8 shows the facility of comparing 2 parameters of electricity
as a time. Figure 9 shows the prediction of the selected date to the user, which shows
prediction of all parameters. Figure 10 shows the report generation page where the
user can fetch his data of electricity consumption selecting from date and last date.
The generated report is in PDF format.
374 R. Jain et al.
6 Conclusion
Smart meters are used to produce considerable volume of data presenting the oppor-
tunity to enhance utility end customer service, improve energy efficiency and lower
the cost and to reduce the bill and save energy for consumers. The system aims
to analyze the data, understand the energy profiles and appliances that affect the
consumption profiles to group similar consumers together. Setting standards for effi-
cient energy consumption is aided by the analysis and it also identifies the outliers
and irregularities in the consumption patterns. The system also calculates the energy
consumption of devices and even makes the energy unit reading to be handy. Hence
it also reduces the wastage of energy and brings awareness among them. Even it can
reduce manual intervention. The proposed system will overall help all the industries
to lower down their electricity bills and will give a detailed analysis of the electricity
consumed by appliances during the particular duration of time.
376 R. Jain et al.
References
1. Nezhad AJ, Wijaya TK, Vasirani M, Aberer K. SmartD: smart meter data analytics dashboard
2. Medina CC, Joe Mel U. Monitoring and control of electricity consumption using Rasberry-Pi
through IoT
3. Borle P, Saswadhar A, Hiwarkar D, Kali RS (2013) Automatic meter reading for power
consumed. Int J Adv Res Electr Instrum Eng 2(3):982–987
Light Fidelity (Li-Fi): Future 5G Wireless
Connectivity to Empower Rural India
1 Introduction
The congested radio frequency spectrum is the main restraint on the increasing
demand for connectivity with high speed. The latest Cisco Visual Networking Index
predicts projects global IP traffic also called as web traffic to approximately triple
from 2017 to 2022. The expected growth of overall IP traffic is 396 EB per month
up to 2022. In 2017, overall IP traffic observed was 122 EB per month. It is esti-
mated to increase with a Compound annual growth rate (CAGR) of 26% [1]. This
increase in CAGR is due to the growing share of mobile traffic over the total IP traffic.
P. A. Satarkar (B)
Computer Science and Engineering Department, Swami Ramanand Teerth Marathwada
University, Nanded, MH, India
e-mail: [email protected]
G. V. Chowdhary
School of Computational Sciences, Swami Ramanand Teerth Marathwada University, Nanded,
MH, India
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 377
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_38
378 P. A. Satarkar and G. V. Chowdhary
This rapid increase in the number of mobile devices and growth in social services
like Facebook, Video on demand and Twitter, etc. are the primary reason for this
tremendous growth in mobile traffic. Along with the limitation in spectrum band-
width, interference due to electromagnetic wireless devices and power inefficiency
and security are other important issues in RF wireless communication. To overcome
these drawbacks, Visible Light Communication (VLC), also called as Free Space
Optics (FSO) networks could become a promising candidate.
The visible spectrum has a wide range for wavelengths from 380 to 750 nm
and supports a wide frequency range of 430–790 THz [2]. The advantages of Li-Fi
network over RF are outlined below [3]:
• License-free unused spectrum
• Immune to electromagnetic interference
• Support broadband, peer to peer communication
• High bandwidth
• Low power consumption
• Inexpensive optical components
• Secure as light cannot penetrate through walls.
VLC has a wide range of applications which include indoor Li-Fi network to
provide high-speed internet up to 10 Gbps, vehicle to vehicle communication to
avoid road accidents, in hospitals and in aircraft to avoid interference with RF signals
of other machines, underwater communication for submarine or remotely operated
vehicles and signboards in the areas like railway station, traffic, airports.
With all the above advantages and applications of VLC, some challenges are
observed during the actual implementation of VLC. They can be described as:
(a) Interference with normal light sources
(b) Data loss in case of shadowing
(c) Fluctuation and dimming of light
(d) Integration with other available technologies.
In this paper, we start with a survey of the potential of Li-Fi, working principle,
system overview, indoor applications, and challenges are discussed. The remaining
paper is arranged as follows. We start by providing system components and system
overview in Sect. 2. In Sect. 3 hybrid Li-Fi network is discussed with designing
issues which are followed by Applications of Li-Fi in Sect. 4. Section 5 contains
Challenges and concluding remarks.
Light Fidelity (Li-Fi): Future 5G Wireless Connectivity … 379
VLC uses intensity modulation and direct detection for information transmission.
Li-Fi uses the physical layer of VLC for communication. This section is composed
of an overview of the VLC system with its two primary components: transmitter and
receiver.
1. VLC Transmitter
In VLC, the transmitter consists of single or multiple LED lamps, a driver circuit,
modulation techniques, and other components as shown in Fig. 1 [4]. For example,
during transmission, simple On–Off Keying modulation uses two separate levels of
light intensity representing the data bit “0” and “1”. LED serves two purposes illu-
mination and communication. While designing care has to be taken that illumination
should not get affected by communication. The performance of the VLC system is
dependent on the design of LEDs. White light can be generated with three different
LEDs: The dichromatic, tri-chromatic, and Tetra-chromatic modes [2]. RGB combi-
nation is preferred for communication as it can use Color Shift Keying to modulate
the data with three LEDs using different color wavelengths.
2. VLC Receiver
The VLC receiver has 3 components: optical amplifier circuit, optical concentrators,
and filter as shown in Fig. 1 [4]. The optical concentrator is used to compensate for
the beam divergence of LEDs which results in attenuation in the case of a large area.
VLC receiver is responsible for light detection and conversion into photocurrent. In
VLC we can use silicon photodiode, PIN diode, or avalanche photodiode as receiver
[5]. After a comparison between these three types, it is observed that cost and gain
are higher for avalanche photodiode than a PIN photodiode. To resolve the inter-
ference effect in VLC from sunlight and other light sources, optical filters should
be introduced in the design. Two types of receivers to receive the signal from the
transmitter:
(1) A Photodetector: It is also referred to as a photodiode or Non-imaging receiver.
This device converts the received light signal into an electric current. It is better
suited for stationary receivers.
(2) Imaging Sensor: It is also called a camera sensor used to receive the transmitted
light signals and has a larger Field Of View (FOV) in the case of mobile
applications. They are slow and energy expensive.
Based on the type of application we can choose the proper type of VLC receiver.
Depending on the transmitters and receivers location and network range supported,
these networks can be categorized into:
(i) Satellite Networks (SNs)
(ii) Outdoor Networks (ONs)
(iii) Indoor Networks (INs).
Indoor networks are applicable for wireless communications inside offices, build-
ings, and houses. Usually, a base station is connected with multiple users through
optical wireless links i.e. Light Emitting Diode (LED). For indoor communication
OWC links are divided into two types based on the way of propagation:
1. Line Of Sight (LOS) links: An LOS link requires both the transmitter and
receiver should be inline. It is again subdivided into three types: (a) Both the
transmitter and receiver face towards each other as shown in Fig. 2a [6]. This
method improves power efficiency and resists to distortion caused by ambient
and artificial light sources. (b) The transmitter and receiver are not facing towards
each other as shown in Fig. 2b [6]. Wide beam transmitters and wide FOV
receivers are required for the transmission. It needs high power levels to cope up
with the high optical loss and the multipath-induced distortions. Here multipath
fading issue is resolved. (c) Hybrid design method where the transmitter and
receiver can have different levels of directionalities, such as a narrow beam
Light Fidelity (Li-Fi): Future 5G Wireless Connectivity … 381
transmitter which is not aligned with a wide FOV receiver as shown in Fig. 2c
[6]. LOS links achieve higher capacity because of a better power budget and the
absence of multipath propagation effects.
2. Non-LOS (diffused) links: In the Non-LOS links, signals transmitted from the
source do not directly arrive at the receiver. They are reflected from different
surfaces like walls, ceiling, floor, or other objects and arrive at time intervals to
the receiver. This causes multipath distortions and makes the path loss estimation
more difficult. The Non-LOS architecture with non directed transmitter–receiver
pair as shown in Fig. 2f [6] is called diffused systems which is the most robust
system and easy to implement for mobile communication systems. Non-LOS
links are more robust when coming across obstacles in the light path.
Allocation (RA) are major concerns as Wi-Fi gets overloaded whereas Li-Fi near it
remains underloaded. Many researchers have focused on these issues and came up
with different solutions. APS and RA are resolved using 3 approaches: fuzzy logic,
optimization, and machine learning. Aspects like handover and mobility are consid-
ered in [7] to improve system throughput. Further to optimize the throughput, [8]
used load balancing, and power allocation is considered in [9] to improve energy effi-
ciency. Demir et al. [10] used particle swarm optimization but all of these optimization
techniques introduced high computational complexity.
Fuzzy logic techniques are implemented in [11] to improve data rates with a
dynamic load balancing approach. In [12] APS is carried out in 2 steps, the first
user gets associated with Wi-Fi and the remaining users to Li-Fi. It is observed that
similar results are obtained with quite less computational overhead.
In [13] Reinforcement learning with knowledge transfer is implemented to handle
network selection. The decision probability distribution is used in [14] for APS. In
[15], the Reinforcement learning load balancing approach is used by considering the
following parameters: throughput, user satisfaction, and outage probability along
with computational complexity. Trust Region Policy Optimization (TRPO) is used
in this paper to find an optimal policy at each step.
The comparative study for average network throughput and computational
complexity is done for all different methods for APS and RA which is presented
in Table 1.
Here the worst case performance in terms of computational complexity (Big-O
notation) is compared for four types of methods, namely, SSS, Iterative, Reinforce-
ment learning and Exhaustive search. NAP represents number of access points, Nu
represents number of users, I represent number of iterations for iterative method.
4 Applications of Li-Fi
Li-Fi applications are vast as they are dependent on the key characteristics, such as
EMI free, energy efficiency, security, improved data rate, availability, license-free,
and integrated networking capability. We can use Li-Fi for satellite communication,
dense urban environment, aircraft and petroleum stations, Outdoor communication
like vehicle and underwater communication, indoor communication for localization,
augmented reality, and cellular communication.
Based on the literature review we will discuss the important challenges that can
be open issues for the researchers. The solution to these issues will help in the
deployment of Li-Fi practically in near future.
A. LOS Alignment: To maximize the channel response and achieve high data rates,
the transmitter and receiver should align their FOV. However, practically due
to mobile user position and orientation of the receiver may get changed. So
the receiver’s FOV will not be aligned with the transmitter. It causes a drop in
received optical power. Designing the techniques to handle the LOS problem
is extremely challenging and is an important direction for future research.
B. Shadowing: When any object or mobile user blocks the LOS, an immediate
reduction in the observed optical power is witnessed; this causes data rate
reduction and data loss. So a large scope is there to understand and work on the
phenomenon of shadowing in VLC. As the shadow effect is for a small duration
of time we have to utilize the reflected optical power promptly.
C. Design and Energy-Related Issues of Receiver: For receiving a light signal in
current VLC we either use a photodiode or an imaging sensor. For the stationary
users FOV can be fixed using photodiode but for mobile users FOV of the
imaging sensor is sufficiently large for communication. Due to a large number
of photodiodes imaging sensors can become slow and consume a large amount
of energy. Hence the achievable data rate will get reduced. Thus, it is challenging
to design a receiver that can give a high data rate while moving and should work
with low energy consumption.
D. LED to Internet Connectivity: LEDs are required to connect to the internet
for broadband applications. Large numbers of LEDs are used for illumination
and connecting all of them to the internet via Ethernet is a costly process. The
advantage of reusing the LED infrastructure for communication is no more
useful. A large number of LEDs cause interference, which hampers the achiev-
able data rate. Power line communication is suggested [16] which provides
available power lines for communication. It only introduces the requirement of
Ethernet to power the modem and power to the VLC modem. The performance
and coverage issue [17] is resolved in this method.
384 P. A. Satarkar and G. V. Chowdhary
E. Dimming Control: During the daytime, if LEDs are switched off then no
communication is possible. LED should be “on” to use the communication
link. Hence similar to RF communications, power consumption for data trans-
mission is not free. The solution can make the LED brightness level low so that
LED is considered as “off”. Further work is required to solve this problem.
F. Inter-Cell Interference: If within a small cell area, LEDs are deployed densely
then it certainly leads to inter-cell interference. This reduces performance due to
low SINR. The solution to this problem can be network MIMO or rearrangement
of LEDs. In the network, MIMO LEDs synchronize the transmissions to ensure
high SINR. In the rearrangement method, LEDs are arranged to get minimum
interference. Research is required for design and analysis to get an optimized
solution.
G. Uplink Issues: More focus is given on the downlink performance problem but
less attention is given to the uplink problem. Mobile users with a smartphone
can use imaging sensors to downlink effectively. But for uplink, if a flashlight
or notification indicator is used then it will cause battery consumption with
user visual irritation. Also, the alignment and orientation of the smartphone are
important for the uplink connection. The solution is, to use RF for uplink [18].
It creates the issues of hybrid VLC with Wi-Fi.
H. User Mobility: For a mobile user the coverage of the VLC network is required
at every point to provide a high data rate and avoid interruption in indoor
communication. The link-layer techniques are designed to facilitate data rate
adaptation, frame aggregation, etc. If RF is used along with VLC then handover
related issues will arise.
I. Security: Light cannot penetrate through walls but the light leaking from
windows, doors, etc. will become a concern for security issues.
J. Interference from Sunlight: For outdoor applications, this problem is associated
with transmission beams which cause low SNR. The solution can be optical
filters but more research is still going on to solve this issue completely.
This survey paper has provided an introduction to visible light communica-
tion, its merits over the current RF system, and the need for VLC to facilitate the
increasing demand for data rate. The shadowing effect followed by the data loss
can be handled effectively using different strategies for APS and RA in a hybrid
network. The comparative study shows SSS gives the weakest performance while
Exhaustive search gives the best performance with complexity overhead. RL gives
better performance with moderate complexity.
Light Fidelity (Li-Fi): Future 5G Wireless Connectivity … 385
References
1. Cisco (2019) Cisco visual networking index: forecast and methodology, 2017–2022, 27 Feb
2019
2. Khan LU (2017) Visible light communication: applications, architecture, standardization and
research challenges. Digit Commun Netw 3:78–88
3. Son IK, Mao S (2017) A survey of free space optical networks. Digit Commun Netw 3:67–77
4. Lee CG (2011) Visible light communication, advanced trends in wireless communications.
In: Khatib M (ed) InTech. ISBN: 978-953-307-183-1. Available from https://www.intechopen.
com/books/advanced-trends-in-wireless-communications/visible-light-communication
5. Kharraz O, Forsyth D (2013) Performance comparisons between PIN and APD photodetectors
for use in optical communication systems. Opt Int J Light Electron Opt 124(13):1493–1498
6. Pathak PH, Feng X, Hu P, Mohapatra P (2015) Visible light communication, networking, and
sensing: a survey, potential, and challenges. IEEE Commun Surv Tutor 17(4):2047–2077
7. Wang Y, Haas H (2015) Dynamic load balancing with handover in hybrid Li-Fi and Wi-Fi
networks. J Lightw Technol 33(22):4671–4682
8. Wu X, Safari M, Haas H (2017) Joint optimization of load balancing and handover for hybrid
LiFi and WiFi networks. In: Proceedings of IEEE wireless communications and networking
conference (WCNC), Mar 2017, pp 1–5
9. Kashef M, Ismail M, Abdallah M, Qaraqe KA, Serpedin E (2016) Energy efficient resource
allocation for mixed RF/VLC heterogeneous wireless networks. IEEE J Sel Areas Commun
34(4):883–893
10. Demir MS, Sait SM, Uysal M (2018) Unified resource allocation and mobility management
technique using particle swarm optimization for VLC networks. IEEE Photon J 10(6):1–9
11. Wang Y, Wu X, Haas H (2016) Fuzzy logic based dynamic handover scheme for indoor Li-Fi
and RF hybrid network. In: Proceedings of IEEE international conference on communications
(ICC), May 2016, pp 1–6
12. Wu X, Safari M, Haas H (2017) Access point selection for hybrid Li-Fi and Wi-Fi networks.
IEEE Trans Commun 65(12):5375–5385
13. Wang C, Wu G, Du Z (2018) Reinforcement learning based network selection for hybrid VLC
and RF systems. In: Proceedings of MATEC web of conferences, vol 173. EDP Sciences, Les
Ulis, p 03014
14. Wang J, Jiang C, Zhang H, Zhang X, Leung VCM, Hanzo L (2018) Learning-aided network
association for hybrid indoor LiFi-WiFi systems. IEEE Trans Veh Technol 67(4):3561–3574
15. Ahmad R, Soltani MD, Safari M, Srivastava A, Das A (2020) Reinforcement learning based
load balancing for hybrid LiFi WiFi networks. IEEE Access 8:132273–132284
16. Ma H, Lampe L, Hranilovic S (2013) Integration of indoor visible light and power line
communication systems. In: Proceedings of 17th IEEE ISPLC, Mar 2013, pp 291–296
17. Yousuf M, El-Shafei M (2007) Power line communications: an overview—part I. In:
Proceedings of 4th international conference IIT, Nov 2007, pp 218–222
18. Rahaim M, Vegni A, Little T (2011) A hybrid radio frequency and broadcast visible light
communication system. In: Proceedings of IEEE GC Wkshps, Dec 2011, pp 792–796
Optimized Dynamic Feature Matching
for Face Recognition
Abstract Since the last three decades, face detection and recognition have become
very active and a huge part of image processing research. In real-time applications
like video surveillance, front views cannot be guaranteed as input. Hence the failure
rates can degrade the performance of the face recognition system. The proposal aims
to introduce a novel PFR method termed as DFM that combine Sparse Representation
Classification (SRC) and FCN for resolving the partial face recognition issues. As
the major contribution, this proposal aims to tune the sparse coefficient of DFM in
an optimal manner, such that the reconstruction error should be minimal. Moreover,
this proposal introduces Jaccard Similarity Index measure to calculate the similarity
scores among the gallery sub feature map and probe feature map. For optimization
purpose, this work deploys a hybrid algorithm that hybrids both the concepts of Grey
Wolf Optimization (GWO) and Sea Lion Optimization (SLnO) algorithm.
1 Introduction
“Face recognition is the process of identifying the faces from the images during
human computer verification” [1]. Under the controlled condition, the face recog-
nition algorithm [2, 3] obtains more optimal results, whereas, under uncontrolled
condition, various issues are faced during the recognition process. Mostly, the face
recognition techniques [4, 5] are classified into two types based on the occlusion
problem: they are part-based and holistic. The human face [6] is presented in the
captured image and it focuses on noisy environment as non-frontal pose, shading,
facial association and overexposing, thus it leads to a problem known as partial face
G. G. Patil (B)
Department of Computer Science and Engineering, SVERI’s College of Engineering, Pandharpur,
India
R. K. Banyal
Department of Computer Science and Engineering, Rajasthan Technical University, Kota, India
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 387
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_39
388 G. G. Patil and R. K. Banyal
recognition [7]. Traditional face recognition systems usually fail to address the prob-
lems of partial face detection. Moreover, the partial face recognition system does
not provide proper run time effectiveness. Introducing a realistic face recognition
system [8] can handle partial faces directly without any synchronization, and it will
be resilient to occlusions, and overcomes the issues regarding pose and illumination.
Partial faces can be determined from the images and accordingly, detection can
be effectively performed from various facial segments in to a complete face [9]. The
important concern is to develop the advanced face detectors for the identification of
occluded and partially noticeable faces. The similarity [10] among the probe partial
faces and gallery faces can identify the improved partial faces and hence the face part
is occluded. To identify the facial landmarks [11], the alignment method is provided
and then occluded.
The conventional face recognition methods use the deep learning methods by
the Convolutional Neural Network (CNNs). The benefits of deep learning method
require trained datasets to identify the better characteristics of data representation
[12]. The collection of large-scale datasets of faces including real-world variation
provides the faces-in-the wild on the web. The trained datasets in CNN based on face
recognition methods obtains better accuracy so that one can know the features [13,
14] whether they are robust in the real-world conditions. The exploitation of deep
learning approaches based on computer vision assists the facial recognition task in
an effective manner [15]. However, certain machine learning approaches like Fully
Convolutional Networks (FCNs) does not contain a permanent input as in the spatial
feature maps similar with the input size of images [16]. Therefore, it turns out to be
essential to develop a more advanced PFR system for detecting the partial faces in
an optimal manner [17].
2 Related Works
(RPSM) and Multi-Key point Descriptor with Gabor Ternary Pattern (MKD-GTP)”.
In 2016, Weng et al. [22] have implemented a new PFR method for finding the
concentration of an individual from their partial faces. “Robust Point Set Matching
(RPSM)” method was presented to match its geometrical and textural information
of the facial characteristics. In 2019, He et al. [23] have proposed Dynamic Feature
Matching (DFM) with Sparse Representation Classification (SRC) and FCNs for
resolving the issues persisting in partial face detection. Here, the DFM does not
necessitate the previous information regarding the position of partial faces. More-
over, the feature maps were computed from the whole image and it resulted in a
speedy process. Finally, the experimental outcomes have shown better effectiveness
and advantages over the traditional models. In 2016, Lei et al. [24] have presented
efficient 3D face recognition methods for identifying all the challenges in PFR. The
proposed method introduced the “Two-Phase Weighted collaborative Representation
Classification (TPWCRC)” framework for rectifying the issues found in the samples
images. The proposed model solved the following parameters like lost data, defor-
mations, occlusions and accessibility of limited training samples and thus it offered
higher efficient results. In 2019, Aminu and Ahmad [25] have established a “Locality
Preserving Partial Least Square Discriminant Analysis (LPPLSDA)” with conven-
tional “Partial Least Squares Discriminant Analysis (PLS-DA)” for face recognition.
The LPPLS-DA model hold better capability of capturing local structures and thus it
was found to be well suited for detecting the facial images. Finally, the outcomes of
proposed method were provided with improved consistency and reduced computa-
tional complexity. In 2019, Patil and Banyal [26] have presented techniques of deep
learning for image recognition. In 2020, Patil and Banyal [27] have presented uncon-
strained feature matching technique for face recognition. Table 1 show the reviews
on the partial face recognition systems.
Table 1 (continued)
Author Adopted methods Features Challenges
He et al. [23] DFM • Improved efficiency • No consideration on
• Better accuracy partial RE-ID and
iLIDS databases
Lei et al. [24] TPWCRC • Minimal false rates • Requires more
• High-quality facial deliberation on
data computational
efficiency
Aminu and Ahmad [25] LPPLSDA • Enhanced specificity • Have to solve the
• Increased accuracy sample size issues
3 Methodology
PFR is the process of verifying or finding the facial emotions of human faces from
the video frame, video source, digital images or still images. Multiple methods are
involved in these systems in which the features of selected facial images are compared
within the images in database. This proposal aims to introduce a novel PFR method
termed as DFM that combine Sparse Representation Classification (SRC) and FCN
for resolving the partial face recognition issues. As the major contribution, this
proposal aims to tune the sparse coefficient of DFM in an optimal manner, such
that the reconstruction error should be minimal. Moreover, this proposal introduces
Jaccard Similarity Index measure to calculate the similarity scores among the gallery
sub feature map and probe feature map. For optimization purpose, this work deploys a
hybrid algorithm that hybrids both the concepts of Grey Wolf Optimization (GWO)
and Sea Lion Optimization (SLnO) algorithm. “The GWO algorithm mimics the
leadership hierarchy and hunting mechanism of grey wolves in nature [28]. The
SLnO algorithm imitates the hunting behaviour of sea lions in nature. Moreover, it
is inspired by sea lions whiskers that are used in order to detect the prey [29]”. Thus,
the partial faces will be detected in an effective manner. The overall architecture of
the implemented work is given by Fig. 1.
4 Conclusion
References
10. He M, Zhang J, Shan S, Kan M, Chen X (2020) Deformable face net for pose invariant face
recognition. Pattern Recogn 100:Article 107113
11. Meinhardt-Injac B, Kurbel D, Meinhardt G (2020) The coupling between face and emotion
recognition from early adolescence to young adulthood. Cogn Dev 53:Article 100851
12. Trigueros DS, Meng L, Hartnett M (2018) Enhancing convolutional neural networks for face
recognition with occlusion maps and batch triplet loss. Image Vis Comput 79:99–108
13. Grati N, Ben-Hamadou A, Hammami M (2020) Learning local representations for scalable
RGB-D face recognition. Expert Syst Appl (in press)
14. García E, Escamilla E, Nakano M, Pérez H (2017) Face recognition with occlusion using a
wireframe model and support vector machine. IEEE Lat Am Trans 15(10):1960–1966
15. Young SG, Tracy RE, Wilson JP, Rydell RJ, Hugenberg K (2019) The temporal dynamics of
the link between configural face processing and dehumanization. J Exp Soc Psychol 85:Article
103883
16. Kim H, Kim G, Lee S-H (2019) Effects of individuation and categorization on face
representations in the visual cortex. Neurosci Lett 70824:Article 134344
17. Iranmanesh SM, Riggan B, Hu S, Nasrabadi NM (2020) Coupled generative adversarial
network for heterogeneous face recognition. Image Vis Comput 94:Article 103861
18. Lahasan B, Lutfi SL, Venkat I, Al-Betar MA, San-Segundo R (2018) Optimized symmetric
partial facegraphs for face recognition in adverse conditions. Inf Sci 429:194–214
19. Mahbub U, Sarkar S, Chellappa R (2019) Partial face detection in the mobile domain. Image
Vis Comput 82:1–17
20. Greening SG, Mitchell DGV, Smith FW (2018) Spatially generalizable representations of facial
expressions: decoding across partial face samples. Cortex 101:31–43
21. Duan Y, Lu J, Feng J, Zhou J (2018) Topology preserving structural matching for automatic
partial face recognition. IEEE Trans Inf Forensics Secur 13(7):1823–1837
22. Weng R, Lu J, Tan Y (2016) Robust point set matching for partial face recognition. IEEE Trans
Image Process 25(3):1163–1176
23. He L, Li H, Zhang Q, Sun Z (2019) Dynamic feature matching for partial face recognition.
IEEE Trans Image Process 28(2):791–802
24. Lei Y, Guo Y, Hayat M, Bennamoun M, Zhou X (2016) A two-phase weighted collaborative
representation for 3D partial face recognition with single sample. Pattern Recogn 52:218–237
25. Aminu M, Ahmad NA (2019) Locality preserving partial least squares discriminant analysis
for face recognition. J King Saud Univ Comput Inf Sci (in press)
26. Patil GG, Banyal RK (2019) Techniques of deep learning for image recognition. In: 2019 IEEE
5th international conference for convergence in technology (I2CT), Bombay, pp 1–5. https://
doi.org/10.1109/I2CT45611.2019.9033628
27. Patil GG, Banyal RK (2020) A dynamic unconstrained feature matching algorithm for face
recognition. J Adv Inf Technol 11(2):103–108. https://doi.org/10.12720/jait.11.2.103-108
28. Mirjalili S, Mirjalili SM, Lewis A (2014) Grey wolf optimizer. Adv Eng Softw 69:46–61
29. Masadeh R, Mahafzah B, Sharieh A (2019) Sea lion optimization algorithm. Int J Adv Comput
Sci Appl 10:388–395
Novel Secure Routing Protocol
for Detecting and Presenting Sybil Attack
1 Introduction
Vehicular Ad-hoc Network is the technology [1] which can forms a secure network
between vehicles, i.e. Vehicles communicate to every alternative and transfer infor-
mation to another vehicle. VANET provides safety to driver of vehicles by exchanging
messages between vehicles. VANET is not secure because many types of attacks can
be appeared in it so it can leads to insecurity of drivers of vehicles. Vehicular Ad Hoc
Network (VANET) is an emerging area for research. Vehicular Ad-Hoc Network is
a challenging topic because of its mobility and link disruption. Many researchers [2]
have been working on specific issues of VANET like routing, broadcasting, Quality of
Service, security, architectures, applications, protocols, etc. Security is a main issue
in VANET because malicious drivers in the network disrupt the system performance.
Sybil attack creates multiple identities which lead to subvert the computer System
S. M. Sawant (B)
CSE Department, SVERI’s College of Engineering, Pandharpur, India
S. M. Shinde · J. S. Shinde
SVERI’s College of Engineering, Pandharpur, India
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 393
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_40
394 S. M. Sawant et al.
[3]. Novel secure routing protocol detects and prevents the Sybil attack particular on
Vehicular ad- hoc network. The proposed Secure routing protocol is based on ad-
hoc on demand (AODV) distance vector secure routing (DVSR) for Mobile ad-hoc
network (MANET) or other wireless ad-hoc network (WAN). This protocol main-
tains routing information and route discovery that detects and prevents Sevier Sybil
attack and each node have unique identity and entry in route table [2]. Inter vehicle
communication is passing and receiving the information to increase traffic efficiency,
detection of road conditions, avoid collisions, detect emergency situations and overall
increase of the efficiency of network. In a MANET or VANET, mobile or vehicular
nodes are making huge impact on the performance of routing protocols because of its
varying mobility characteristics. So many of the scholars and researchers [4] devel-
oped Ad hoc On-demand (AODV), Dynamic Source Routing (DSR), Destination
Sequence Distance Vector (DSDV) routing protocols for MANET but they cannot
directly used in VANET. Because of its apparently, widely varying mobility charac-
teristics of mobile or vehicular nodes are expected to have a significant impact on the
performance of routing protocols. Therefore even though researchers [4] have devel-
oped routing protocols like Ad hoc On-demand Vector (AODV), Dynamic Source
Routing (DSR), Destination Sequence Distance Vector (DSDV) etc. for MANET,
these protocols cannot be directly adopted in VANETs, efficiently, because of the
rapid variation in link connectivity, high speed and extremely varied density of vehic-
ular nodes in VANET. Researchers have developed special routing protocols for
VANET [3], and these are aimed to adapt rapidly. However, these recent protocols
are not fully secure and able to prevent attacks on VANET. These attacks are fraud
information, denial of service, black hole, alternative attack and Sybil attack. Sybil
attacks become a serious threat which can affect the functionality of VANETs for
the benefit of the attacker. The Sybil attack is the case where a single faulty entity,
called a malicious node, can present/create multiple identities known as Sybil nodes
or fake nodes. Sybil attack has been found in mostly peer-to-peer network where a
node in the network can operate as multiple identities at the same time it can gains
the authority and power in reputed systems. The main purpose behind this attack
is to gain the majority of influence in the network actions in the system. In a Sybil
attack a single entity that is computer has the capability to create and operate multiple
identities such as user accounts or IP address based accounts. For the other observers
these multiple attacks look like a real unique identity.
2 Literature
(i) Correct—entities that follow the protocols and rules setup in the network
honestly (whose honesty is verified).
Novel Secure Routing Protocol for Detecting and Presenting … 395
(ii) Faulty—entities whose behavior are arbitrary and can’t be predicted. They
don’t honestly follow the protocols and rules in the network.
(iii) A communication cloud: A very general cloud through which messages
between different entities travel.
(iv) Pipe: to connect an entity with the communication cloud.
3 Methodology
The proposed novel Secure Routing protocol for Ad hoc Network (SRAN) is a routing
protocol for detecting and preventing Sybil attack. It is based on AODV and does
not allow Sybil node into Route discovery by eliminating the node from the route
table. Important views of SRAN protocol are route request packet format (RREQ),
route reply packet format (RREP) and route error packet format (RERR).
Route Request Packet Format
In SRAN routing protocol, if source wants to send message to destination then it first
broadcasts the route request (RREQ) to its neighbors. Neighboring node receives
route request packet format, if receiving node is not destination and does not have
396 S. M. Sawant et al.
route to the destination then it rebroadcast the route request packet format and same
time backward route is created to the source. If the receiving node is destination node
or it has current route to the destination then Route Reply (RREP) is generated.
1. RREQ ID: A sequence number uniquely identifying the particular RREQ when
taken in association with the source node’s IP address.
2. Source IP Address: The IP address of the Source.
3. Source Sequence Number: The Sequence number of Source.
4. Source Unique ID: The Unique Identification of Source.
5. Destination IP Address: The IP address of the destination for which a route is
selected.
6. Destination Sequence Number: The latest sequence number received in the
past by the source for any route towards the destination.
7. Destination Unique ID: The Unique Identification of Destination.
8. Hop Count: Number of hops needed to reach destination.
Route Maintenance
Once route is defined then route maintenance is also required. It is to provide infor-
mation about link of the route as well as route to be modified due to movement of one
Novel Secure Routing Protocol for Detecting and Presenting … 397
Fig. 2 Difference of
performance of AODV
protocol and SRAN protocol
or more nodes in the route. Every time route is used to send packet then its expiry
time is updated by adding current time and Active Route Timeout (ART). ART is a
constant value that defines how long new route is kept into routing table of node after
last transmission done. ART defines both source and intermediate node. If route is
not used in the predefined period then node can’t be sure that route is still valid or not
and then this route is removed from routing table. It ensures that no any unnecessary
packet loss.
4 Result
Secure Routing protocol (SRAN) can improve the performance of routing in secure
communication. It means that this protocol will detect and prevent Sybil attack.
In SRAN total numbers of packets sent by source are received successfully at
destination. Table 1 shows the result and depicted in Fig. 2.
5 Conclusion
especially in the performance that successfully detects and prevents Sybil attack.
SRAN routing protocol maintains unique identity of the node in order to establish
higher secure communication between vehicles to vehicle. SRAN routing protocol
is the best solution for the secure communication.
References
1. Haque MM, Misic J, Misic V et al (2013) Vehicular network security. In: encyclopedia of
wireless and mobile communications, 2nd edn
2. Gadkari MY, Sambre NB (2012) VANET: routing protocols, security issues and simulation tools.
J IOSR J Comput Eng (IOSRJCE)
3. Fonseca E, Festag A (2006) A survey of existing approaches for secure ad hoc routing and their
applicability to VANETS. J NEC Netw Lab, Version 11:1–28
4. Bai R, Singhal M (2006) DOA: DSR over AODV routing for mobile ad hoc networks. IEEE
IoT Model for Heart Disease Detection
Using Machine Learning (ML)
Techniques
Abstract From the last decade, a tremendous spotlight is on the giving quality
medical services because of the exponentially growing of life threatening illnesses
of the patients. There are numerous components that influence the health condi-
tion of each person and a few illnesses are more dangerous and cause death of the
patient. And in the present age, the most common reason of death is heart disease.
This research work presents the IoT based system for heart disease detection using
the Machine Learning (ML) technique. It consists a novel preprocessing stage that
provides more accurate classification of the ECG signal. Also, this novel prepro-
cessing method removes the noise effectively from the raw ECG data. The classifi-
cation performance was evaluated using the various classifiers such as KNN, Naïve
Bayes and Decision tree that detects the normal and abnormal heart-beat rhythms.
With the obtained results, we have observed that the preprocessing has improved
the classification performance. This technique further proves that the decision tree
has good performance over the KNN and Naïve Bayes with respect to the accuracy,
sensitivity and precision.
1 Introduction
The health care industry has been adopting new technologies for providing better
and smart healthcare facilities [1]. With the IoT, remote and real-time monitoring of
patients is made possible and this unleashes the potential to continuously monitor
the health and helps the physicians to give suggestions or treatment in a timely
manner. As a larger community of people are suffering from heart disease, it is
vital to carry out diagnosis at the early stage to save lives and help to support a
healthy lifestyle of people. The health care monitoring has improved tremendously
due to the development of different IoT capabilities and instruments to track patient’s
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 399
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_41
400 M. K. Gawali and C. Rambabu
health conditions regularly [2]. The patients can also interact with the doctor more
easily which gives the satisfaction of treatment and it also reduces the hospital stay
and healthcare expenses. The main focus of employing IoT in healthcare system
is to set up a fully automated environment for patient monitoring and providing
assistance and care to patients in real-time. There is a rise in the need for a portable
system that can be used at home by the patient for measuring their ECG profiles and
diagnose their disorder in real-time. So in this paper, an extensive review is carried
out to find the existing technologies that are available for monitoring heart related
diseases. It is understood from the analysis, that the collected raw data contains noise
and irrelevant contents. These are irrelevant and incorrect data that are not useful for
diagnosis. This noise and huge variation in data leads to reduction in the classification
accuracy, sensitivity and precision. Therefore, in this paper a novel pre-processing
approach is used to remove noise and unrelated data from ECG signals. Relevant
attributes are identified using correlation technique to enhance data efficiency. The
machine learning classifier algorithm such as KNN, Naïve Bayes and Decision tree
are used for classifying the ECG signals based on waveforms [3]. The classifier that
obtains better performance metrics can be used for diagnosing the variation in the
ECG waveform and identify the type of abnormality and disorders.
2 Literature Review
Most of the time people go to the hospital only when after they suffer from cardiac
disease. In the traditional ECG setup, the medical instruments are housed in the
hospital, patients need to visit the hospital to check their heart disorders and study
their physiology of heart. During this process, the patient’s activities are limited.
Frequent visit to hospital increases the medical expenses and puts a burden on hospital
authorities. Early intervention is essential for the survival of patients, there has been
a lot of focus and attention on building an automated system for the detection of
abnormalities of heart signals. In [3] an IoT based wearable architecture was proposed
to measure the ECG signals. This system provides a portable platform where a non-
intrusive wearable sensor is used to collect the patient’s ECG signals and send them
to IoT cloud via the smart phone enabled Bluetooth or ZigBee technologies. The data
stored in the cloud can be retrieved by the specialist for further processing using data
analytics to find the disease. The data analytics procedure of data cleaning, storage,
analysis and generation of warning alerts to the concerned specialist in a real-time
manner can be performed by having access to the remote server. To facilitate the
early diagnosis of heart disease, machine learning techniques are employed. From
the health dataset, the investigation were performed to study the abnormal functions
of the heart. To classify the signals, the amplitude and interval periods of the cardiac
waves were analyzed using machine learning classification algorithms such as SVM,
Adaboost, ANN and Naïve Bayes [4]. Identifying accurate classifiers will assist the
physician in making quality decisions on diagnosis and timely treatment. There are
different types of arrhythmia diseases that are related to cardiac rhythm disorders. To
IoT Model for Heart Disease Detection Using Machine Learning … 401
ensure proper diagnosis, statistical and dynamic features extraction of ECG signals
is necessary [5]. So in this paper, heart rate variability is computed to generate alerts
when the patient is affected by arrhythmia disease.
In [6], to reduce the time consuming process of manually checking the ECG data,
a new classifier was proposed to distinguish normal and abnormal heartbeat rhythm.
This classifier removes noise and extracts ECG features. This classifier provided
better performance when compared with other machine learning classification algo-
rithms. The time computation is comparatively reduced and helps in identifying
arrhythmia disease. Early detection of abnormal pulse rates is also crucial for the
survival of the patient. So, to improve survival, a mechanism for the automatic detec-
tion of cardiac arrest was proposed. The ECG based pulse detection system uses the
random forest classifier (RF) [7]. The ECG data were processed to remove noise and
extract the features. The features were fed to random forest classifier and compared
with other existing classifiers. The RF classifier resulted in improved performance
helps the practitioners in making quick decisions for providing appropriate treat-
ment. The pre-processing are widely used in various fields for data cleaning, data
transformation, data integration and data reduction [8]. The identification of missing
values, noisy data and detecting outliers are performed on the data to perform data
cleaning. This data cleaning process provide significant improvement in the perfor-
mance of the classifier. Different pre-processing techniques are available that can be
applied to the dataset for improving the performance metrics. From [9], it can also
be understood that the preprocessing helps in better performance of the classifier. In
[10], a data driven approach used the outlier based alert system for identifying the
anomaly data of patients to reduce the measurement errors. When trained dataset
was tested in real-time system, the system proved to be effective. In the following
section, the novel pre-processing is proposed for classifying ECG data.
3 Proposed Methodology
The portable IoT system is designed to work with sensors and microcontroller. The
components that are used for setting up the portable system are:
• LM35 Temperature Sensor
• Pulse Sensor
• AD8232 ECG Sensor
• Arduino Uno.
These 3 sensors are connected to the Arduino Uno microcontroller to collect
the body temperature, heartbeat rate and ECG signals. The different reading of the
patient’s vital signs are gathered and send for testing by the classifier model which
are using the dataset for detecting the abnormalities. The prototype is shown in Fig. 1.
402 M. K. Gawali and C. Rambabu
The fundamental concept behind the proposed methodology is to enhance the pre-
processing of ECG data. The proposed model has two main steps: Pre-processing
and classification of heart disease data. The block diagram of the proposed system
is shown in Fig. 2.
3.2 Dataset
The historical health data of the patient was taken from the heart disease health dataset
for training the classifier. The dataset used for training the classifier for testing the
accuracy, sensitivity and precision of classification is the heart disease dataset which
was created by V.A. Medical Center, Long Beach and Cleveland Clinic Foundation.
This dataset is available as an open source from the UCI machine learning repository
[11]. In this dataset, 75 sets of attributes are available but 14 of these attributes are
considered for the prediction of heart disease. The dataset comprises records of 303
patients.
3.3 Correlation
To train the classifier model, heart disease dataset is used. Correlation is used to
identify the relationship between two continuous, quantitative variables. The iden-
tification of relevant attributes is implemented using the correlation technique [12].
All the irrelevant attributes are not considered for training the classifier model. The
correlation coefficient is computed to determine the relationship between the dataset
attributes. This improves the performance of the classifier algorithm by removing
weakly correlated attributes. This improves the performance of the classifier algo-
rithm by removing weakly correlated attributes. To better understand the correlation
between the attributes, the correlation graph is plotted is shown in Fig. 3.
3.4 Preprocessing
This helps to identify the association between the attributes. Correlation values range
from −0.25 to +1. Positive correlation represents that the column attributes either
increases or decrease together. A negative correlation indicates that one attribute
will increase and other one decreases or vice versa. The correlation graph obtained
is shown in Fig. 3 and is depicted in different colors. The dark maroon states that the
attributes are weakly correlated with another and orange color represents a strong
correlated with one another.
3.5 Processing
In the pre-processing phase, missing values in the dataset are identified before using
them for classification. The collected data in the dataset might contain erroneous
entries with noise, missing values, null values and incorrect values. The handling of
404 M. K. Gawali and C. Rambabu
IQR = Q3 − Q1
where Q3 represents the third quartile and Q1 represents the first quartile. Therefore,
it is essential to eliminate them so that errors of classification algorithm outcome is
reduced.
IoT Model for Heart Disease Detection Using Machine Learning … 405
3.6 Classification
The classification algorithm used are performing the pre-processing and classification
of dataset values are Naïve Bayes, K nearest neighbour (KNN) and Decision tree
[13, 14]. The KNN uses the Euclidean distance for computing the nearest neighbour
attribute.
D(P, Q) = ( pi − qi )2
pi represents the training set of attributes with a given class qi attributes. To determine
the class from nearest neighbour list, the majority vote of class labels among the K
nearest neighbours are computed using the formula
y = arg max Wi ∗ I (v = yi )
( pi ,qi Dz )
where the distance of z is the set of k closest training examples to z and v represents
the class labels.
Naive Bayes performs classification by using the formula
E
H p p(H )
p = H
E p(E)
n
q = arg max P(q) P( pi /q)
i=1
The dataset is divided into two subsets, one for training and another for testing.
The training and testing dataset size are varied to identify a best classifier model.
The Naïve Bayes works well for large datasets when compared with other algo-
rithms. Also the decision trees algorithm is used because it can handle categorical
and numerical data. The classifier identifies normal and abnormality from the heart
disease dataset. The Naïve Bayes algorithm considers each of the attributes indepen-
dently to the probability that the patient is suffering from heart disease. It identifies
the maximum likelihood of the normal or abnormality from the patient’s health data.
The accuracy is computed by comparing the test value and predicted values. The
performance of models is evaluated in terms of accuracy with different test sizes.
The decision tree accuracy is high when compared with the Naïve Bayes.
4 Results
The real-time ECG signals of patients were collected from the sensors are fed for
testing via the microcontroller. The retrieval of ECG data from the experimental
setup is shown in Fig. 4. The classifier model is tested using Python.
The classifiers used are Naïve Bayes, KNN and decision trees. The classifier
is trained and validated using the dataset from UCI database. This pre-processing
procedure improves the performance of the classifier algorithm. Figure 5 shows the
Outliers computed using the IQR score. Those values that have large variations are
identified and removed.
Attributes
Fig. 6 Comparison of
accuracy with and without
pre-processing
Fig. 7 Comparison of
sensitivity with and without
pre-processing
408 M. K. Gawali and C. Rambabu
Fig. 8 Comparison of
precision with and without
pre-processing
test was conducted by changing the training the dataset size. When more data was
used for training the classifier, the classification was improved. From the results, it is
found that the best performance was achieved by Decision tree classifier algorithm
when the preprocessing was performed to the data in terms of accuracy, sensitivity and
precision. So from the analysis, it is found that pre-processing improves classification
performance and this can be used for testing of ECG signals for abnormality and
disorders.
5 Conclusion
In this research work, we have discussed the major breakthrough that can be brought
about in the healthcare industry by the IoT platform. The benefits and challenges faced
by the healthcare system are presented. In this paper, a novel pre-processing technique
was proposed to improve the classification of ECG data using machine learning
algorithms. The classification algorithm used for training the system are KNN, Naïve
Bayes and decision trees. With the introduction of pre-processing technique, the
performance of decision tree outperformed other classification algorithms in terms
of accuracy, sensitivity and precision. The accuracy of the predicted is evaluated.
This method can be used for identification of normality and abnormality of ECG
signals and assist in making early and accurate diagnosis. This technique further
proves that the decision tree has good performance over the KNN and Naïve Bayes
with respect to the accuracy, sensitivity and precision.
References
4. Kalaivani RLDV (2019) Machine learning and IoT-based cardiac arrhythmia diagnosis using
statistical and dynamic features of ECG. J Supercomput 0123456789
5. Hammad M, Maher A, Wang K, Jiang F, Amrani M (2017) Detection of abnormal heart
conditions based on characteristics of ECG signals. J Meas 125:634–644
6. Elola A, Aramendi E, Irusta U, Del Ser J, Alonso E, Daya M (2018) ECG-based pulse detection
during cardiac arrest using random forest classifier
7. Bhaya WS (2017) Review of data preprocessing techniques in data mining
8. Zhu C (2016) Influence of data preprocessing, vol 10, no 2, pp 51–57
9. Hauskrecht M, Batal I, Valko M, Visweswaran S, Cooper GF, Clermont G (2013) Outlier
detection for patient monitoring and alerting. J Biomed Inform 46(1):47–55
10. https://archive.ics.uci.edu/ml/datasets/Heart+Disease
11. Rawashdeh M, Al Zamil MG, Hossain MS, Samarah S, Amin SU, Muhammad G (2018)
Reliable service delivery in tele-health care systems. J Netw Comput Appl 115:86–93
12. Alsheikh MA, Lin S, Niyato D, Tan H (2015) Machine learning in wireless sensor networks.
J Algor Strateg Appl 1–23
13. Qi J, Yang P, Min G, Amft O, Dong F, Xu L (2017) Advanced internet of things for personalised
healthcare systems: a survey. J Pervas Mob Comput 41:132–149
14. Gelogo YE, Hwang HJ, Kim H-K (2015) Internet of things (IoT) framework for u-healthcare
system. Int J Smart Home 9(11):323–330
Security Threats and Their Mitigations
in IoT Devices
Abstract Internet of things (IoT) is another worldview converging with the social
networks, permitting data sharing between the individuals and electronic gadgets.
Likewise, it is expected for omnipresent connectivity among different entities or
things using Internet. Anyway, security and privacy issues are the major concerns
for IoT. The heterogeneous technological advancements, inherent vulnerabilities of
IoT devices, poor design IoT standard invites the cyber attack. This research work
mainly aims to address the security threats and issues on different layers of IoT archi-
tectures and their possible mitigations. Also, it provides a taxonomic representation
of the major 3-layers of IoT architecture with their protocol stack. Finally, we have
highlighted the most challenging security threats and their mitigations with some
future research work proposals.
1 Introduction
Internet of Things (IoT) is one of the key components of digital and transformation of
digital world along with Social, Mobile, Analytics and Cloud (SMAC). It is otherwise
called as Internet of Everything or Industrial IoT. IoT, Big Data and SMAC can
helps as a numerous possibilities that were unheard earlier. It takes the absolute
center stage for the product vendors, system integrators, software companies and
IT sector companies. Today’s Industry analysts says, there can be around 26 billion
devices on the IoT (Cisco estimate 50 billion) by the end of 2020 and the data
exchange will be 40 Zettabytes over the networks [1]. According to the McKinsey
Global Inst., IoT market will have the major impact of $3.9–$11.1 trillion per year by
2025 over various applications i.e. smart cities, smart industries, home, offices, retail
environment, worksite, logistic and navigation, smart vehicles, etc. [2]. The abstract
level of IoT model contains various physical devices, or sensors i.e. controllable
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 411
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_42
412 S. Gupta and N. Lingareddy
sensors, RFID (Radio Frequency Identifications), IoT getaways, web servers [3] as
depicted in Fig. 1.
The term “things” from IoT comprises both the physical and cyber worlds [4].
IoT brings up many challenges and holds much promise i.e. the data generated,
stored or transmitted through IoT devices, so many security issues and privacy of
the users can have serious consequences. Every challenge to the IoT system must
be secured, controllable and privacy to the smart users, only when the IoT systems
are built up with security. The general architecture of an IoT system is divided with
three layers i.e. (1) Physical layer, (2) Network Layer and (3) Application Layer.
The three layer architecture of IoT system is depicted in Fig. 1. The deployment
process involves various technologies i.e. WSNs, RFID, Bluetooth, NFC [5], IP,
EPC (electronic product code), Wi-Fi and various actuators [6]. The intelligent and
smart applications of IoT interconnected devices helps personal as well as economic
benefits to the society [7].
In IoT many sensor devices and smart peoples have the connectivity through Internet
to provide services at anytime, anywhere, and any types of services. IoT also provide
services at any business, anybody, any one, any context, any device, anything, any
path and any Network. Owing to the wide range of impact on daily life, all the sensor
devices are connected through Internet and all are also defenseless to all privacy and
security issues. Security in the field of Information Technology (IT) [8–10] considers
three features: confidentiality, integrity, and availability as the prime objectives and
are called as CIA triad [11]. Security is defined as a process by which unauthorized
access to the system state is prevented and thus the privacy is not compromised.
Confidentiality refers to the secrecy of data, whereas integrity confirms that the data
is not changed in transmission [12, 13]. Further, availability provides the smooth
transmission of data whenever it is required. In a secure network some of the required
capabilities are:
Security Threats and Their Mitigations in IoT Devices 413
i. Data Authentication: The sensed data and related information collected from
secured, authenticated devices must be followed some technical mechanism
and allow to transmit.
ii. Resilience to Attacks: During data transmission if the system crashes, it should
be automatically recover itself as same data uses in different domain. A cloud
server must be protect smartly and intelligently himself from an intruders or
eavesdropper.
iii. Client Privacy: At client side, the used data and information must be secure and
safe. The private data should be protected i.e. no other types of clients can’t be
access the private data from the client.
iv. Access Control: Only authenticated and authorized person can access the
control. The general user can access the system by providing user name and
password and their access rights, which will be controlled by the system
administrator. Different user can access the specific portion of the database
or programs to smoothen running the system.
Security issues are divided into different sub-categories, viz., data confidentiality,
monitoring and tracing of activities, avoidance of malicious insiders, hijacking
of services or processes including phishing, lack of transparency into providers’
service provisions and procedure environments, fraudulent activities and exploita-
tion, management of multi- instance in multi-tenancy etc. [14, 15]. Moreover, prob-
ability of attack through side channel, escaping the sandboxed environment, can
access the virtual machine and hence unauthorized or spoofed access to the host are
also a possibility [16–18]. Encryption techniques are also the most important tool in
providing multidimensional security services for IoT [3].
In IoT mostly, applications data are personal related or industrial. These applications
data must be provided with security and confidentiality against any kind of unautho-
rised access [19]. The biggest challenging factor in IoT is security. Information is to be
transmitted over the communication channel with more security in the network. The
IoT improve the communication between devices but still there are so many issues for
time (response time), scalability, and availability. In IoT though machine to machine
technology is the first phase, but it enables new applications and to bridge diverse
technology by connecting physical objects together in support of intelligent deci-
sions. Open Web Application Security Project (OWASP) has defined ten top security
issues associated with IoT devices i.e. (1) Insecure network services, (2) Insecure
web interface, (3) Insufficient authentication or authorization, (4) Non availability
of transport encryption, (5) Privacy concern, (6) Lack of secured cloud interface, (7)
Lack of secured mobile interface, (8) In-sufficient security configuration, (9) Lack
of secure software/firmware and (10) Less physical security [20].
414 S. Gupta and N. Lingareddy
The reference model of IoT can be represented by seven levels. The levels are:
i. Physical devices and controllers
ii. Connectivity
iii. Edge computing
iv. Data accumulation
v. Data abstraction
vi. Application
vii. Collaboration and processes.
Broadly different levels of an IoT system can be described as: Level 1: IoT system
consists of single device that performs sensing/actuation, performs analysis storing
data and host the applications. These levels are suitable for modelling less complexity
and less cost solutions where the data is medium and analysis requirements are
computationally extensive. Mostly these are used in home automation. A level 2:
IoT system has same as level 1 IoT system but it included local analysis. Information
storage is at the clouds and applications are generally cloud-based. These are more
appropriate for solution where the data usage is more and the requirement is same as
level 1 IoT. Mostly these are used in smart irrigation. But in Level 3 IoT system has
same as level 1 and 2 but data can be stored and processed in clouds and cloud-based
applications.
These system levels are involved in big data analysis and computationally inten-
sive requirements as depicted in Fig. 2. Level 4 IoT system consists various nodes
that performs local processing, i.e. cloud-based applications where data’s are stored
in the cloud and observer nodes (not performing any control functions). The observer
node can subscribe to and receive information collected in the cloud from the IoT
devices and also process information. These are more applicable for big data, and used
Security Threats and Their Mitigations in IoT Devices 415
In IoT system model different layers of IoT are facing various attacks including active
and passive attacks. The attacks can be of two types depending on their behavior in
network i.e. active and passive attacks [21]. Network behaviors are
i. Active Attack: The attacker disturbs the performance of the network by stealing
the data at the time of communication in active attack.
ii. Passive Attack: The communications channels are observed and form its usage
history the passive attacker steal the information [21].
As discussed earlier, IoT framework model can be spoken to by three principle
layers for example Physical, system and application layer. Every one of these layers,
has its own innovations and some security shortcomings. The security problems
of each layers are discussed with possible threats and next part of this section, we
discussed the feasible solutions to that threats.
Attacks in Physical/Perception Layer
The physical layer incorporates sensors and actuators to perform estimation of temp,
speeding up, mugginess and functionalities like questioning area [22]. The funda-
mental security dangers in physical layer are because of constrained hub assets and
appropriated composed structure. The main threats are:
i. Tampering: These attacks generally focused on hardware components and an
attacker is to be physically present into the IoT systems and continue its process
Security Threats and Their Mitigations in IoT Devices 417
to make system busy. A few models are hub altering and vindictive code infu-
sion. In hub altering assailant can harm the sensor hub by electronically denies
to get to, alter the sensitive information or physically replace the part of its
hardware or entire node. But in malicious code injection, the attacker injects
its malicious code physically on to a node and access to that node from that
IoT system.
ii. Impression: Authorization/Validation in the disseminated systems are so trou-
blesome. So enabling malevolent hub to make a phony distinguishing proof for
noxious assaults.
iii. Denial of Service: An attackers can adopt the finite processing abilities of each
node, to make the system unavailable.
iv. Routing Attackers (WSN, RSN): In the Data collection and forwarding
processes, intermediate malicious node modify the routing path and make the
system busy.
v. Data Transit Attack (DTA): Various attacks like sniffing, Man in the middle
attacks on the integrity and confidentiality during data transit.
Sometimes active attacks/vulnerable attacks can prevent the IoT devices smartly.
Prevention of IoT devices from the vulnerable attacks can be done by deploying
some security constraints [25, 26]. According to the behavior, different categorized
attacks are:
(a) Low level attack—When network is attacked by intruders and that and it’s
attack is not secure)
(b) Medium Level attack—When intruders are listening to the medium while
changing the data integrity.
(c) High level attack—When intruders is entered into the network and it can alter
the intensity as well as modify the original data.
Extremely High level attack—When Intruders attack on the network with the
adoption of unauthorized access and doing the illicit operations leading to suspended
or unavailability of networks or congestion of the network (Tables 1 and 2).
Solving the Security Challenges in Device Level
In designing phase of products the security aspect should be incorporated. The secu-
rity aspect can be introduced in operating system level and it should be extended
through the device stack to the implemented applications and having hardware secu-
rity capabilities. Generally IoT devices are having 70% security threat and 25%
security aspect concerns per device as depicted in Fig. 4. As most of IoT devices
may not be designed with security concern and it leads to susceptibility and config-
uration management problem. Figure 5 depicts about the organization’s controls for
controlling IoT devices.
Security Threats and Their Mitigations in IoT Devices 419
6 Conclusion
Now a day’s IoT seems to be unbeatable and the overwhelming use of smart devices
cannot be reversible. Unless and until the security issues are addressed, organiza-
tions require to be attentive, putting appropriate controls, to ensure security risks
against the applications and focus on IoT devices those are performing well and who
are connected to their networks. The main idea behind this paper is to highlight the
security issues and their challenges to the different layers of IoT and deliberating
the security concern in various protocols and their possible corrective measures. IoT
devices became soft targets as they are deprived of security mechanisms. Security
mechanisms should be incorporated to all IoT related devices along with the commu-
nication networks. To protect from introducers or threats, we should have used default
password and for first time user, install all the security enabled requirements for all
the smart devices.
Table 2 Security threat in IoT automation and the mitigations
420
References
1. Van Der Meulen R (2015) Gartner says 6.4 billion connected ‘things’ will be in use in 2016,
up 30 percent from 2015. Stamford
2. Manyika J, Chui M, Bisson P, Woetzel J, Dobbs R, Bughin J, Aharon D (2015) Unlocking the
potential of the internet of things. McKinsey Global Institute
3. Lamba A, Singh S, Singh B, Sai S, Muni R, Islands C (2018) Quantum computing technology
(QCT)—a data security threat. Int J Emerg Technol Innov Res 5(4): 801–806
4. Ning H (2013) Unit and ubiquitous internet of things. CRC Press, Inc., Boca Raton
5. Sarangi M, Singh D, Khuntia M (2019) A potential solution for man-in-middle security issues
through near field communication (NFC). Int J Eng Adv Technol (IJEAT) 8(4):492–498
6. Al-Fuqaha A, Guizani M, Mohammadi M, Aledhari M, Ayyash M (2015) Internet of things:
a survey on enabling technologies, protocols, and applications. IEEE Commun Surv Tutor
17:2347–2376
7. Khan R, Khan SU, Zaheer R, Khan S (2012) Future internet: the internet of things architecture,
possible applications and key challenges. In: Proceedings of the 10th international conference
on frontiers of information technology, Islamabad, 17–19 Dec 2012, pp 257–260
8. Hashizume K, Rosado DG, Fernández-Medina E, Fernandez EB (2013) J Internet Serv Appl
4(5)
9. Ochani A, Dongre N (2017) Security issues in cloud computing. In: International conference
on I-SMAC (IoT in social, mobile, analytics and cloud) (I-SMAC)
Security Threats and Their Mitigations in IoT Devices 423
10. Tianfield H (2012) Security issues in cloud computing. In: IEEE international conference on
systems, man, and cybernetics
11. Sammy F, Vigila S (2016) A survey on CIA triad for cloud storage services, vol 9, pp 6701–6709
12. Paruchuri VL (2016) Data confidentiality in cloud using encryption algorithms. Int J Cloud
Comput Super Comput 3(2):7–18
13. Lamba A (2019) SR-MLC: machine learning classifiers in cyber security—an optimal
approach. Int J Res Inf Sci Appl Tech (IJRISAT) I(3)
14. Chandrika Sai Priya A (2016) Integrated framework for multi-user encrypted query operations
on cloud database services. Int J Cloud Comput Super Comput 3(2):1–6
15. Aich A, Sen A (2015) Study on cloud security risk and remedy. Int J Grid Distrib Comput 8(2)
16. Bhattacharyya D (2018) Space and security issues in cloud computing: a review. Int J Secur
Appl 12(6):37–46
17. Thirupathi Rao N, Sravani A, Bhattacharyya D, Kim T (2018) Security and assurance aspects
to be observed in cloud computing based data centers: a study. Int J Secur Appl 12(4):1–14
18. Thirupathi Rao N, Bhattacharyya D (2019) Security aspects to be considered in cloud
computing based data centers: a tutorial. Int J Database Theory Appl 12(1):27–42
19. Sicaria S, Rizzardia A, Griecob LA, Coen-Porisinia A (2015) Security, privacy and trust in
internet of things: the road ahead. Comput Netw 146–164
20. Frustaci M, Pace P, Aloi G, Fortino G (2017) Evaluating critical security issues of the IoT
world: present and future challenges. IEEE Internet Things J
21. Kocakulak M, Butun I (2017) An overview of wireless sensor networks towards internet of
things. In: IEEE 7th annual computing and communication workshop and conference (CCWC),
pp 1–6
22. Bhardwaj I, Kumar A, Bansal M (2017) A review on lightweight cryptography algorithms
for data security and authentication in IoTs. In: 2017 4th international conference on signal
processing, computing and control (ISPCC). IEEE, pp 504–509
23. Puthal D, Nepal S, Ranjan R, Chen J (2016) Threats to networking cloud and edge datacenters
in the internet of things. IEEE Cloud Comput 3(3):64–71
24. Pongle P, Chavan G (2015) A survey: attacks on RPL and 6LoWPAN in IoT. In: 2015
international conference on pervasive computing (ICPC). IEEE, pp 1–6
25. Hennebert C, Santos JD (2014) Security protocols and privacy issues into 6LoWPAN stack: a
synthesis. IEEE Internet Things J 1(5):384–398
26. Raza S, Wallgren L, Voigt T (2013) SVELTE: real-time intrusion detection in the internet of
things. Ad Hoc Netw 11:2661–2674
27. Adnan AH et al (2015) A comparative study of WLAN security protocols: WPA, WPA2. In:
Proceedings of international conference on advances in electrical engineering (ICAEE), Dhaka,
pp 165–169
28. Sulaiman AG, Al Shaikhli IF (2014) Comparative study on 4G/LTE cryptographic algorithms
based on different factors. IJCST 5
29. Gelogo YE, Caytiles RD, Park B (2011) Threats and security analysis for enhanced secure
neighbor discovery protocol (SEND) of IPv6 NDP security. Int J Control Autom 4(4):179–184
30. Neisse R, Steri G, Fovino IN, Baldini G (2015) SecKit: a model-based security toolkit for the
internet of things. Comput Secur 54:60–76
31. Raza S, Shafagh H, Hewage K, Hummen R, Voigt T (2013) Lithe: lightweight secure CoAP
for the internet of things. IEEE Sens J 13(10):3711–3720
32. Brendel J (2003) World-wide-web server that finds optimal path by sending multiple syn +
ack packets to a single client. U.S. Patent 6,587,438, issued 1 July
33. Mishra N, Swagatika S, Singh D (2020) An intelligent framework for analysing terrorism
actions using cloud. In: New paradigm in decision science and management. Springer,
Singapore, pp 225–235
424 S. Gupta and N. Lingareddy
34. Lu C, Blum BM, Abdelzaher TF, Stankovic JA, He T (2002) RAP: a real-time communication
architecture for large-scale wireless sensor networks. University of Virginia: Department of
Computer Science, Charlottesville
Age and Gender Classification Based
on Deep Learning
Abstract Now a days, age and gender classification became relevant to growing
applications such as human-computer interaction (HCI), biometrics, video surveil-
lance, electronic customer relationship management, forensic, and many more. The
social networks have a huge amount of data available, but often, people do not
provide some of their data, such as age, gender, and other demographics. Age and
gender of a person from face image is one such significant demographic attribute.
Face recognition is an extremely difficult issue in the field of image analysis and
computer vision on the grounds. Although Face recognition had been done over a
couple of decades with traditional image processing features, recently Deep Convo-
lutional Neural Networks (CNNs) has been proved that it is the descent method for
the classification task, especially of uncontrolled, real-world face images. In this
research, we propose a deep learning CNN method to extract discriminative features
from the uncontrolled, real-world dataset and classify those images into gender and
age group. In this paper, learning representation is shown through the use of the deep
learning-based convolutional neural network (CNN), essential growth in the perfor-
mance can be achieved on these works. To this end, we would like to mention that
our CNN model can be used even when the number of learning features is restricted.
We evaluate our technique on two prominent datasets. Accuracy up to 80% obtained
for the Adience datasets and 95% for the Gender Classification dataset from Kaggle.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 425
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_43
426 T. Agarwal et al.
1 Introduction
Recently, the age estimation from face image has drawn out large attention in
computer vision due to its significant application in video surveillance, demography,
internet access control and security [1]. Age and gender which are the two facial
key attributes, play very important roles in social interactions, making gender and
age estimation from a one face image is a prime task in intelligent applications [2].
In contrast, the facial age estimation is a difficult process and get easily affected
by many elements such as identity, gender, and extrinsic factors, including artificial
position, environment, human lifestyle and facial expressions. Datasets of the human
facial image have different demographical attributes [3], ex. Regarding the same age
of two persons, one can be a male with a face that appears old, while the other can
be a female with a face that appears young. Hence, enabling a computer system
to categorize the face images based on gender and age of the person is yet a diffi-
cult task. In the last two decades, the facial images classification based on age and
gender has received in many of the research. Past proposals to estimate facial image
attributes have depended on differences in facial feature aspect. Many approaches
have established classification strategies designed specifically for gender or age clas-
sification tasks [4]. Still from a given facial image, accurate physical age estimation
is a difficult problem.
In this research, we approach Deep Convolutional Neural Network, which has
novel face recognition and age, gender classification capabilities. A Deep Convolu-
tional Neural Network CNN is widely used for image recognition and processing.
Convolutional Neural Network has input layers, output layers, and multiple hidden
layers, many of which are convolutional. To check the efficiency of the CNN model,
we approach Adience dataset [5]. This dataset has all categories of images for training
as well as testing purposes.
The rest of the paper is organized as follows. In the next section, we reviewed
prominent techniques associated with age and gender classification based on deep
learning. Our proposal for classification using CNN is detailed in Sect. 3. The dataset
used for the implementation and its associated results can be found in Sect. 4. We
conclude the paper in Sect. 5.
2 Associated Works
Before describing the proposed method, we would give a broad survey of significant
work-related to age and gender classification, which were based on CNN.
In recent years there has been increasing attention, and many methods have been
put forth for the difficulty with automatically extracting age features from the face
images—a detailed survey of many techniques found in [6]. Previous age estimation
methods were based on computing ratios between different measurements of the
facial attributes [7]. The requirement of the input image in methods that represent
Age and Gender Classification Based on Deep Learning 427
the aging process is to be frontage, and well smoothen and thus, current observed
results only on constrained datasets having frontage images [8]. Few have addressed
the difficulties that arise from real-world variations in picture quality, while some
previously have shown fairly good results on constrained input images such as near-
ideal lighting, angle, and visibility [9].
A detailed associated work of gender classification methods with automatically
detected aligned faces can be found in [10]. One of the previous methods for SVM
gender classifiers were used by [11]. The appearance, intensity, quality and texture
features were used with related information in [12]. Deep CNN has additionally been
successfully applied to applications, including facial keypoint detection [13].
A human face contains a large variation in appearance that captured from the
real world conditions. Thus, some of the approaches having difficulty under such
uncontrolled conditions and they unable to provide an ethical way to handle the
outliers. Thus, [14] together proposed a new methodology called Robust Cascaded
Pose Regression to reduce the risks to the outliers by detecting obstruction precisely
and used robust shape-indexed features.
Detail discussion about gender classification and age estimation using CNN found
in [15]. CNN model was used for the gender classification and adience dataset is used
for testing and training in [16].
3 Proposed Methodology
Figure 1 shows a general block diagram of gender and age classification. Initially,
the gender is classified and then it is placed in the appropriate age group. A CNN
based deep learning model is employed. A general understanding of this can be seen
in Fig. 2. The parameter tuning for this CNN model is tabulated in Sect. 4. However,
we explain the general processing inside a CNN model.
Deep Learning has one of the algorithms named Convolutional Neural Networks
(CNN). This algorithm takes the input as an image, and several attributes of the image
are learnt through the kernels. Important objects present in the image are learnt by
CNN, which allows them to detect one image from the other. The convolutional
neural network learns the specific attribute of a person to identify gender easily. One
dominant feature of Convolutional Network is its ability to pre-process the data by
itself as other Machine Learning algorithms need to provide the features manually.
Thus, they may not spend a resource in data pre-processing. They are also able to
learn the features and develop filters of their own.
Convolution
Convolution layer performs the extraction of features from the face image and reduces
the size of images for the faster computations, as shown in Fig. 3. The convolutional
neural network learns the patterns within the face image and will recognize it in the
picture.
There are three convolution layers in the network, followed by a pooling layer.
Convolution layer- Parameter tuning
1. Convolution Layer1- 96 filters with size 3 × 7 × 7 are convolved with stride
4 and padding 0, providing an output size of 56 × 56 with 96 kernels. This is
followed with Non-linearity ReLU, maximum pooling which reduces the size
to 28 × 28 with 96 kernels.
2. Convolution Layer2- 256 filters with size 96 × 5 × 5 are convolved with stride
1 and padding 2, providing an output size of 28 × 28 with 256 kernels. This is
also followed with Non-Linearity ReLU, maximum pooling, LRN and reduces
the size to 256 × 14 × 14 as output.
3. Convolution Layer3- 256 filters with size 256 × 3 × 3 are convolved with stride
1 and padding 1, and the layer followed with ReLU and maximum pooling,
provide an output size of 256 × 7 × 7.
covered by the filter. Maximum pooling returns the maximum value in the image
portion covered by the kernel. Maximum pooling is used in the proposed work to
reduce the dimension of the input image, as shown in Fig. 5.
Fully Connected Layer
A Fully connected layer is used to compute the class score. 512 neurons are fully
connected to the 256 × 7 × 7 output of Convolution layer 3 and followed by a
Non-Linearity ReLU layer and dropout layer. At the end, there is a softmax layer
connected on top of the fully connected layer to provide the loss and final class score
and maximum probability considered as a predicted class, as shown in Fig. 6.
We test our CNN design on the Adience benchmark dataset [5] and the Gender
Classification dataset [17]. The Adience benchmark dataset is prominently signified
for the gender and age classification. The automatically uploaded images on Flickr
from smart-phone devices has made the Adience dataset. These images without prior
Age and Gender Classification Based on Deep Learning 431
filtering uploaded on Flickr, as is normally the case on social platform viewing situ-
ations of these images are highly unconstrained, considering real-world challenges
of face aspect in Internet images. The entire Adience dataset includes 2284 subjects
with 26,000 images roughly. These images are used in the techniques to achieve the
performance gain rather than better processing attributed to the network architecture.
Another dataset for testing and training is the Gender Classification dataset down-
loaded from Kaggle [17]. The dataset is of cropped images of males and females. It
is split into training and the validation directory. Training contains 2500 images of
each class, and a validation directory contains 100 images of each class.
Below Table 1 shows dataset distribution of Adience faces benchmark into age
and gender classes [5].
Below Table 2 shows the distribution of gender classification dataset into different
gender classes [17].
Table 2 Gender
Gender No. of images
classification dataset
Male 2600
Female 2600
Both 5200
432 T. Agarwal et al.
The parameters are tuned in the algorithm to control the behaviour of parame-
ters. Normally the algorithm creates a different model for each value of the tuning
parameter. The parameters chosen for tuning in the CNN algorithm are shown in
Table 3 for age and gender classification.
Results are observed by using the system having the specifications, as shown in Table
4.
The methodology used of deep learning based on CNN is for face detection and
recognition, which requires certain limiting factors for the orderly design of the
system. Based on the various parameters, changes like by providing one hidden layer
then two, we have observed the changes in the output and changes in the accuracy.
We have also observed the changes by providing initially one convolution and
pooling layer and then increasing the convolution and pooling layer. There is a change
in the result after every parameter change, and based on the observations, we have
taken the most effectual parameter for face recognition with average accuracy. As
we do the preprocessing of images, it will suppress unwilling distortions or enhance
some important for further processing that reduces the complexity and increases the
accuracy of CNN.
We train our CNN design on 80% of the Gender classification dataset and the
remaining 20% used for testing of CNN model. We achieved more accuracy on
Gender classification.
Confusion Matrix:
The performance of classification algorithms is summarized in a confusion matrix.
Classification accuracy alone can be misleading if we have an unequal number of
observations in each class or if we have more than two classes in our dataset. As we
have 2 class for gender and 8 age group, so the confusion matrix is 2 × 2 and 8 ×
8. Confusion matrix for the age group on Adience dataset and for Gender class on
Gender Classification dataset as shown in Fig. 7a and b.
We have implemented the CNN model on the dataset as mentioned in Table 1,
and sample results are mentioned in figure from Figs. 8a and 9.
Image was taken of a child and detected as male gender within the age group of
4–6 as positive misclassification.
Image taken of a girl and truly detected as female gender within the age group of
25–32.
Image taken of an old woman and truly detected as Female gender within the age
group of 48–53.
Image was taken of a boy and detected as male gender within the age group of
25–32 as positive misclassification.
Image was taken of a sofa and no face detected.
Fig. 7 a Confusion matrix- adience dataset, b. Confusion matrix- gender classification dataset
434 T. Agarwal et al.
Fig. 8 a Gender Male with Age 4–6, b. Gender Female with Age 25–32, c. Gender Female with
Age 48–53, d. Gender Male with Age 25–32
5 Conclusion
This paper presented the age and gender classification using CNN method to process
real world unconstrained database. The test result verified the effectiveness of the
proposed method using a deep convolutional neural network.
Taking an example of the related problems of face recognition, we analyze how
good deep Convolutional Neural Network algorithm performs on these tasks using
unconstrained data. Hereby it can be concluded that pre-processed images datasets
give good accuracy. Also, some parameters such as kernel size, number of convolution
layers and hidden layers affect the accuracy.
Considering the smaller size of unconstrained image datasets categorized for age
and gender, CNN provides an improved result for age and gender classification.
Adience dataset and Gender Classification dataset were used for the testing and
training of this CNN model. As a result of this experimental studies, gender classifica-
tion on adience was done with 80% accuracy, and gender classification on the Gender
Classification dataset was done with 95% accuracy. As we increase the number of
layers in the CNN model, it provides improved accuracy.
Acknowledgements We wish to record our deep sense of gratitude and profound thanks to our
project teacher Prof. Prashant Aher for helping us on drafting this paper. He educated us on the
ethics of paper writing and to bring this paper into fruition.
References
1. Liu X, Li S, Kan M, Zhang J, Wu S, Liu W, Han H, Shan S, Chen X (2015) AgeNet: Deeply
learned regressor and classifier for robust apparent age estimation. In: Proceedings of the IEEE
international conference computer vision workshop, pp 258–266
2. Rothe R, Timofte R, Van Gool L (2016) Deep expectation of real and apparent age from a
single image without facial landmarks. Int J Comput Vis
3. Duan M, Li K (2018) An ensemble CNN2ELM for age estimation. IEEE Trans Inf Forensics
Secur 13(3)
4. Deep Kaur K, Rai P (2017) An analysis on gender classification and age estimation approaches.
Int J Comput Appl (0975–8887), 171(10)
5. Kaggle, https://www.kaggle.com/ttungl/adience-benchmark-gender-and-age-classification
6. Fu Y, Guo G, Huang TS (2010) Age synthesis and estimation via faces, A survey. IEEE Trans
Pattern Anal Mach Intell 32(11):1955–1976
7. Kwon YH, da Vitoria Lobo N (1999) Age classification from facial images. In: Proceeding
conference computer vision and image understanding, vol 74. IEEE, pp 1–21
8. Geng X, Zhou ZH, Smith-Miles K (2007) Automatic age estimation based on facial aging
patterns. IEEE Trans Pattern Anal Mach Intell 29(12):2234–2240
9. Eidinger E, Enbar R, Hassner T (2014) Age and gender estimation of unfiltered faces. IEEE
Trans Inf Forensics Secur 9(12):2170–2179
10. Makinen E, Raisamo R (2008) Evaluation of gender classification methods with automatically
detected and aligned faces. IEEE Trans Pattern Anal Mach Intell 30(3):541–547
11. Moghaddam B, Yang MH (2002) Learning gender with support faces. IEEE Trans Pattern Anal
Mach Intell 24(5)
12. Perez C, Tapia J, Estevez P, Held C (2012) Gender classification face images using mutual
information and feature fusion. Int J Optomechatronics 6(1):92–119
13. Sun Y, Wang X, Tang X (2013) Deep convolutional network cascade for facial point detection.
In: IEEE conference on computer vision and pattern recognition. Portland, OR, pp 3476–3483
14. Burgos-Artizzu XP, Perona P, Dollar P (2013) Robust face landmark estimation under occlusion.
In: IEEE international conference on computer vision (ICCV). Sydney, Australia, pp 1513–
1520
15. Trivedi G, Pise NN (2020) Gender classification and age estimation using neural networks: a
survey. Int J Comput Appl (0975–8887) 176(23)
Age and Gender Classification Based on Deep Learning 437
16. Inik O, Uyar K, Ulker E (2018) Gender classification with a novel convolutional neural network
(CNN) model and comparison with other machine learning and deep learning CNN models. J
Ind Eng Res 4(4):57–63. ISSN-2077–4559
17. Kaggle, https://www.kaggle.com/cashutosh/gender-classification-dataset
Handling of Auxiliaries in Kannada
Morphology
Bhuvaneshwari C. Melinamath
1 Introduction
B. C. Melinamath (B)
Department of Computer and Engineering, SVERI College of Engineering, Pandharapur,
Maharastra, India
e-mail: [email protected]
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 439
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_44
440 B. C. Melinamath
study. The comparative study would yield rich information regarding the semantic
structure of languages, apart from being useful for translation systems, etc. Auxil-
iaries are the helping verbs. Auxiliary verbs are used in conjunction with main verbs.
Auxiliary verbs usually accompany the main verb. The main verb provides the main
semantic content of the clause. An example is the verb have in the sentence I have
finished my dinner. Here, the main verb is finished, and the auxiliary “have” helps
to express the perfect aspect. To know state of the art, we carried out the literature
survey.
2 Literature Survey
The syntax of auxiliaries has given rise to much discussion in the generative literature
[1] discussed finite-state transducers based system for Hindi, [2] have discussed the
auxiliaries from a generative perspective [3, 4] focus on syntactic structures. With
respect to auxiliary [5], have focussed on serial verb construction using auxiliaries
[6]. Focussed on the syntax of valuation in auxiliary-participle constructions [7] has
focussed on verb phrase use of ellipsis, phases and the syntax of morphology.) [8] has
discussed issues conserving the Clitics, morphological merger, and the mapping to
phonological structure, [9] has discussed the Kannada morphology aspect concerning
syntax and semantic of Kannada Language.
English auxiliary verbs or helping verbs such as will, shall, may, might, can, could,
must, ought to, should, would, used to, need are used in conjunction with main verbs
to express shades of time and mood. The combination of helping verbs with main
verbs creates what are called verb phrases in English. In English, shall is used to
express simple future. The different forms of the has. have used to express tenses
present perfect and past perfect. There is also a separate section on the Modal auxiliary
such as can, could, may, might, must, ought to, shall, should, will, and would, do not
change the form of different subjects. For example “I can write". in this example, the
modal auxiliary express various meanings of necessity, advice, ability, expectation,
permission, possibility, etc.
3 Proposed Methodology
The auxiliaries occupy more than 75% of corpus file, hence analysis and handling
of Kannada auxiliary is crucial for translation purpose, the methodology is shown
in Fig. 1. The input module reads the input and tokenizes the input sentence in to
Handling of Auxiliaries in Kannada Morphology 441
words and each word is converted into transliteration module using ir.pl program,
each transliterated word is searched in the dictionary, if word is found then tag in
the dictionary is assigned to it, otherwise the word is passed to next module to check
for the morphological inflections. The morph module analysis the input and gives
output with part of speech tags.
The general occurrence of aspect auxiliary is past participle form of the verb followed
by aspect auxiliary. Aspect auxiliary always preceded by past participle form. (past
verbal participle + Aspect auxiliary verb). Another essential distinction between
Kannada auxiliary and English auxiliary is the verbs which are acting as aspect
markers have used as the main verb. Also, this is not true with English auxiliaries
like has, shall etc. The example illustration is given in Table 1.
Kannada has a set of verbs that may be added to verbal participle to give certain
semantic nuances to the meaning of the sentence. Aspect markers are very similar to
the main verbs in their morphology and syntax. In fact, they are derived from certain
main verbs. But semantically they do not express the lexical meaning like their main
verbs express in their auxiliary aspect usage. The aspectual biDu ‘completive’ does
not mean the same as main verb biDu ‘leave’. Consider an example of verb formation
by adding a set of the auxiliary, as shown in Table 1. In Kannada thousands of
such verb forms can be generated. That is why the study of auxiliaries in Dravidian
languages is a challenging task. In Kannada language, the verbs which are acting as
442 B. C. Melinamath
aspect auxiliary are actually the main verb. Apart from their usage as the main verb,
these verbs are acting as auxiliaries also. This is not true for the English language.
Consider an example below.
Handling of Auxiliaries in Kannada Morphology 443
Here bareduko Du is a though looks like a simple single verb. But it is the complex
verb, which is formed by the addition of aspect auxiliary ‘koDu’ to the baerdu form,
which is a non-finite (past verbal participle form) of verb ‘bare’ (write). This kind
of formation of verb leads to lakhs of complex verbs. But in reality, there are only
around 2000 basic verb roots. Another thing is the complex verbs formed by this
process follow the TAM (tense, aspect and modality) inflections to the auxiliary verb
and not to the first verb. But however, the meaning of such formation is inferred from
the first verb itself. In the below example in Table 1 “tiMdubiDu”. The meaning here
is eaten itself; it is not like ate and then left. This process of complex verb formation
is productive and regular in Kannada. It is not wise to store such complex verbs in the
dictionary; instead, we have handled the formation of such complex verbs through
our morphology. In currently existing morphological systems these verbs are stored
as basic verbs and kept in the dictionary. Another speciality of Kannada is more than
one aspect marker can be attached; this is not the case with English. Consider an
example below.
In this example, 4 aspect markers are added; one can observe the complexity in
word formation with respect to Kannada, the meaning of this word is inferred from
first verb maaDu ‘do’. TAM inflections, PNG (person, gender, number) inflections
follow last verb biDu (Leave). From the CIIL (central institute of Indian language)
3 million corpus, different auxiliaries are explored and are given in Table 1.
444 B. C. Melinamath
The modal auxiliary in example 7 and 8, aagu and paDu denote passive
constructions.
A Kannada Sample file Account4.aci.out the file from DoE CILL Corpus is selected
for analysis, and it is observed that many words in the corpus are formed with
auxiliaries and also modal auxiliaries. The analysis varies with the type of corpus.
The occurrence of various auxiliaries in the corpus is shown in Fig. 2.
In the above corpus the auxiliary haaku ‘put’ has occurred maximum followed by
nooDu and koDu, auxiliaries biDu and hoogu have not occurred at all. The occur-
rences various with the type of corpus used for testing. Similarly, the occurrences of
modal auxiliaries in the file is also tested and shown in Fig. 3.
Randomly five files F1, F2, F3, F4, F5 of size around 1000 words is taken for
testing, and the result of auxiliary verbs is shown in Fig. 4. It shows that auxiliary
beeku ‘must” have occurred maximum times followed by aagu, paDu andkuuDadu
has not occurred. This variation depends on the corpus.
The morphology of Kannada becomes complex due occurrence of auxiliaries as
bound morpheme with the root or derivative stem. In contrast, they occur as free
morphemes in English. We observe that 5 to 6 levels of aspect marker can be added
to the derivate stem to form a complex verb root in Kannada. Auxiliary verbs help
the main verb to denote the actions of the subject. They help in making compound
words and passive voice statements. Auxiliaries are useful in serial verb construction.
40
35
30
25
20
15
10
5
0
aaDu biDu hoogu koDu nooDu haaku
Aspect Auxiliaries
446 B. C. Melinamath
Modal Auxilaries
60
No Of Modal Auxilaries
50
40
30
20
10
0
Morphology of auxiliary verbs like aspect auxiliaries and modal auxiliaries play
a very important role as far as a Dravidian language like Kannada is considered.
The morphological richness of any language is the presence of auxiliaries as bound
morphemes.
References
1. Singh DM, Seema S (2012) FST based morphological analyzer for Hindi language. J Comput
Sci Issues 9:349–353
2. Akmajian A, Steele SM, Wasow T (1979) The category AUX in universal grammar. Linguis Inq
10:1–64
3. Chomsky N (2001) Derivation by phase. In: Kenstowicz M (ed) Ken hale: a life in language.
MIT Press, Cambridge, MA, pp 1–52
4. Emonds J (1978) The verbal complex V’-V In French Linguistic Inquiry 21
5. Heine K (2001) The serial verb constructions and other complex predicates
Handling of Auxiliaries in Kannada Morphology 447
Abstract Patients suffering from dysarthria have trouble controlling their muscles
involved in speaking, thereby leading to spoken speech that is indiscernible. There
have been a number of studies that have addressed speech impairments; however addi-
tional research is required in terms of considering speakers with the same impairment
though with variable condition of the impairment. The type of impairment and the
level of severity will help in assessing the progression of the dysarthria and will also
help in planning the therapy.This paper proposes the use of Convolutional Neural
Network based model for identifying whether a person is suffering from dysarthria.
Early diagnosis is a step towards better management of the impairment. The proposed
model makes use of several speech features viz. zero crossing rates, MFCCs, spectral
centroids, spectral roll off for analysis of the speech signals. TORGO speech signal
database is used for the training and testing of the proposed model. CNN shows
promising results for early diagnosis of dysarthric speech with an accuracy score of
93.87%.
1 Introduction
P. Dumane · S. Chavan
Don Bosco Institute of Technology, Kurla (W), Mumbai 400070, MS, India
B. Hungund (B)
Narsee Monjee Institute of Management Studies, Vile Parle, Mumbai 400056, MS, India
e-mail: [email protected]
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 449
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_45
450 P. Dumane et al.
dysarthria types. Signs and indications of dysarthria differ based upon the cause and
the type of dysarthria. These disorders may lead to symptoms like slurred speech, slow
speech, inability to speak louder than a whisper or speaking too loudly, rapid speech
that is difficult to understand, nasal, raspy or strained voice, uneven or abnormal
speech rhythm, difficulty in moving tongue or facial muscles.
Dysarthria can be caused by damage to the nervous system, and because of the
intertwined intricate relationship between the different functioning elements, it is
difficult to consider any of the functions like respiration, phonation, articulation inde-
pendent of each other. Respiration is controlled by abdominal muscles, by thoracic
musculature and by the diaphragm. Phonation is controlled by the laryngeal struc-
tures. Resonance is determined by pharyngeal and oral musculature, by the soft
palate, by the tongue and by the nasal cavities. Articulation is under the control of
tongue, jaws and lips [1].
The differing degrees of defects are an outcome of the problems at one or more of
these control points because of which velocity, range and amplitude of the movement
are altered. Based on the corresponding site of neural impairment, dysarthria can be
classified into the types viz. flaccid dysarthria, spastic dysarthria, ataxic dysarthria,
hypo kinetic dysarthria, hyperkinetic dysarthria, mixed dysarthria [2].
Production of speech sounds involves a series of internal and external events
that occur at different levels in the auditory system. Decision support for planning
the therapy and also improving the automatic recognition of dysarthric speech is
possible with the assessment of the level of severity of dysarthria [3]. Convolutional
Neural Network (CNN) is a promising tool as presented in this paper that can be
used effectively for detection of dysarthria as compared to the traditional methods.
The manuscript is organized as: Sect. 2 discusses the dysarthria related previous
work. Section 3 covers the proposed methodology used for developing the classi-
fication model using CNN. Results are discussed in Sect. 4. Section 5 gives the
conclusion.
2 Literature Review
Literature suggests that the dysarthric speech severity classification according to the
types of dysarthria can be improved if the automatic assessment of dysarthric speech
can be done. The first step towards this is to identify whether the speech data is
that of a dysarthric or a non-dysarthric person. In [3], 31 audio descriptors and 10
feature sets are used for classification of the severity level of dysarthria. Spectral and
harmonic features were estimated using multi-taper based spectral estimation. Balaji
and Sadashivappa [4] have compared the speech waveforms of persons by mapping
the unintelligible speech data with the intelligible data. Genetic algorithms have been
used for selecting speech disorder specific prosody features in [5] and support vector
machines were used for diagnosis and assessment of dysarthric speech. Spangler
et al. [6] used the XGBoost classification algorithm for a fully automatic process for
detection of dysarthria. Kim et al. [7] utilized the KL-HMM framework composed
Dysarthria Detection Using Convolutional Neural Network 451
3 Proposed Methodology
The research was segregated into four sections—data collection and preparation,
data analysis, data preprocessing and data modeling as shown in Fig. 1.
Step 1: Data Collection and Preparation. Universal Access dysarthric speech
corpus and the TORGO database [12, 13] was used for training and testing the model
developed using CNN. This database consists of dysarthric articulation of aligned
acoustics and measured 3D articulatory features from Female and Male speakers. The
data consists of recordings for both dysarthric and non-dysarthric persons mainly to
identify the hidden articulatory parameters using statistical pattern recognition.The
data consists of.wav files of both dysarthric females and males, and non-dysarthric
females and males. The sample data files with their details are given in Fig. 2.
Step 2: Data Analysis. In addition to auditory speech evaluation, some other
techniques of analysis of the speech signals like detecting the changes in the
vowel formants, f0, changes in duration of the speech signal, amplitude variations,
prolonged vowel duration, prolonged voice onset time, variations in tempo, etc.
Reduced values of f1 and f2 are also seen to be associated with patients of dysarthria
as a result of reduction in tongue movements [14]. In this paper, the features used for
assessment are short time Fourier transform, spectral centroids, spectral bandwidth,
spectral roll off, zero crossing rate, Mel Frequency Cepstral Coefficients (MFCC).
The plots of these features for a female with dysarthria are presented in Fig. 3 as a
sample. The features extracted from males and females suffering as well as those not
suffering from dysarthria were studied.
Step 3: Data Pre-processing. As preprocessing of the data, the unstructured data
was converted into numeric features considering 128 sample rates. The audio signals
can be converted into a Mel Spectrogram. Furthermore, to train the model, the target
variable which was whether the person is suffering from dysarthria or not is converted
into categorical variable followed by one-hot encoding. The input with 8216 audios
with 128 sample rate features (columns) and 2-Labelled target variables was prepared
to model the data and train the convolution model. These data are further split into
80% data for training and 20% data for testing the model.
Step 4: Data Modeling. The pre-processed data is provided as an input to the
first CNN layer of the model [9–11]. For each audio file, the data was separated
with the shape (16, 8, 1), thus the first CNN with 64 filters had an input of (64 ×
16 × 8) with the kernel (3 × 3). This layer generated 640 parameters and shrunk
the height and width dimensions and was used to identify the dysarthria or non-
dysarthria patterns from the audio. These layers generate several patterns which will
further pool the features using max-pool method. In max-pool of two-dimensional
method, the important unique features are identified and are further fed to the second
convolution layer with 128 filters of 8 × 4 shrink image. The second convolution
layer and max-pool layer has a 3 × 3 and 2 × 2 kernel respectively to shrink the
height and width of image, the input to this layer was (128 × 8 × 4). At the end
of the convolution layers, a dropout layer was fitted to overcome the problem of
over-fitting and under-fitting, with the frequency rate of 0.1. To complete the model,
a dense layer was fitted to make a classification problem. Before that, a flatten layer
was added which converts the 3D tensor into 1D tensor for dense layer. In the model,
the second convolution and max-pool layer generates an output shape of 128 × 8 ×
4. Hence to flatten the 3D tensor, these shapes generated 1024 parameters as an input
to first dense layer of 1024 neurons. At the end of the model, 2 neurons of dense layer
were added as an output layer of the model with the softmax activation function. The
presented model in Fig. 4 was built using TensorFlow library. The summary of the
model is given in Table 1. To compile the model, various hyper-parameters were
Dysarthria Detection Using Convolutional Neural Network 453
Fig. 3 Speech analysis (Type 1: Female with Dysarthria) a Original speech b Short term fourier
transform (STFT) c Spectral centroids d Spectral bandwidth (p = 1, p = 2, p = 3) e Spectral roll-off
( Zero crossing rate g MFCC spectrogram[Sentence: You wish to knowall about my grandfather]
454 P. Dumane et al.
Fig. 3 (continued)
tuned and Adam optimizer was built, with a learning rate of 0.0005. The parameters
of the model are described in Table 2.
It can be observed that after 10 epochs, the training loss starts decreasing and holds
a linear loss, though there were some fluctuations observed in between 11 and 20
epochs. Similarly, it can be observed that testing losses are fluctuating for each epoch.
This may be due to over fitting or under fitting of the model. The call-back function
was implemented to stop training the model as the testing set achieves the lowest
loss. Furthermore, it can be observed that after 30 epochs of training of the model,
it gives training accuracy of 96.59% and testing accuracy of 93.87%. Thus it can be
considered that CNN model does identify the unique features among the voice data
to classify the given data into that of dysarthria or non-dysarthria patients.
456 P. Dumane et al.
The model which was built to classify the given audio into a dysarthria or non-
dysarthria patient achieved a standard accuracy on a testing set. These sets were
unseen by the model. It can be observed that out of the 2054 random samples of
audios, 1928 audios are correctly predicted and 126 are not correctly predicted,
this can be inferred from the confusion matrix given in Table 3. The errors for
not predicting the labels correctly infers that 75 audios which were actually non-
dysarthria persons are predicted as suffering from dysarthria, similarly for 51
dysarthria patients are predicted as non-dysarthria persons. The reason for the errors
in the model could be because there were certain sentences which were very difficult
to classify, for example some sentences are in the form of only one word—‘sleep’.
The precision of predicting the correct labels is 95.83%. The evaluation metrics for
the model predicted on the test set are presented in Table 4.
The model assures that the area under the curve of non-dysarthria samples guar-
antees 98.46% to correctly identify the actual labels. This can be observed from
the confusion matrix that out of total 1247 non-dysarthria voices 1172 are correctly
predicted signifying the highest ratio. The AUC graph is displayed in Fig. 4.
5 Conclusion
With advancements in the field of artificial intelligence, deep learning models are
useful for decision support systems in the diagnosis of patients suffering from
dysarthria. Disorders like dysarthria not only create trouble for the patient in
conveying their thoughts but also bring in social embarrassment for the patient which
adds to the mental trauma. CNN based model definitely holds promise for helping
in the diagnosis of severity of dysarthria. Analysis helps in planning the therapy and
can also help in recognition of dysarthric speech. In this paper, the accuracy score
of 93.87% has been obtained in the classification of dysarthric speech using CNN
model. Future work involves considering more number of features and correlating
them for classifying the severity of dysarthria. In spite of some clear evidences, there
are some misclassifications which need further investigation.
Dysarthria Detection Using Convolutional Neural Network 457
References
1. Edwards M (2012) Disorders of articulation: aspects of dysarthria and verbal dyspraxia, vol 7.
Springer Science & Business Media
2. Hodge M (2013) Developmental dysarthria (2013). Cambridge Handbooks in Language and
Linguistics. Cambridge University Press. https://doi.org/https://doi.org/10.1017/CBO978113
9108683.004
3. Bhat C, Vachhani B, Kopparapu SK (2017) Automatic assessment of dysarthria severity level
using audio descriptors. In: IEEE International conference on acoustics, speech and signal
processing (ICASSP). IEEE, pp 5070–5074
4. Balaji V, Sadashivappa G (2019) Waveform analysis and feature extraction from speech data of
dysarthric persons. In: 2019 6th international conference on signal processing and integrated
networks (SPIN). Noida, India, pp 955–960. https://doi.org/10.1109/SPIN.2019.8711768.
5. Vyas G, Dutta MK, Prinosil J, Harár P (2016) An automatic diagnosis and assessment of
dysarthric speech using speech disorder specific prosodic features. In: 2016 39th international
conference on telecommunications and signal processing (TSP). Vienna, pp 515–518. https://
doi.org/10.1109/TSP.2016.7760933.
6. Spangler T, Vinodchandran NV, Samal A, Green JR (2017) Fractal features for automatic
detection of dysarthria. In: IEEE EMBS international conference on biomedical & health
informatics (BHI). Orlando, FL, pp 437–440. https://doi.org/10.1109/BHI.2017.7897299.
7. Kim M, Kim Y, Yoo J, Wang J, Kim H (2017) Regularized speaker adaptation of KL-HMM for
dysarthric speech recognition. IEEE Trans Neural Syst Rehabil Eng 25(9):1581–1591. https://
doi.org/10.1109/TNSRE.,.2681691
8. Millet J, Zeghidour N (2019) Learning to detect dysarthria from raw speech. In: 2019 IEEE inter-
national conference on acoustics, speech and signal processing (ICASSP). Brighton, United
Kingdom, pp 5831–5835. https://doi.org/10.1109/ICASSP.2019.8682324A.
9. Khamparia A, Gupta D, Nguyen NG, Khanna A, Pandey B, Tiwari P (2019) Sound classification
using convolutional neural network and tensor deep stacking network. IEEE Access 7:7717–
7727
10. Wang Z, Chen L, Wang L, Diao G (2020) Recognition of audio depression based on
convolutional neural network and generative antagonism network model. IEEE Access
8:101181–101191
11. Sasmaz E, Tek FB (2018) Animal sound classification using a convolutional neural network.
In: 2018 3rd international conference on computer science and engineering (UBMK). IEEE,
pp 625–629
12. The TORGO Database (2020) Acoustic and articulatory speech from speakers with dysarthria,
https://www.cs.toronto.edu/compling-web/data/TORGO/torgo.html
13. Rudzicz F, Namasivayam AK, Wolff T (2012) The TORGO database of acoustic and
articulatory speech from speakers with dysarthria. Lang Resour Eval 46(4):523–541
14. Refai MS, Aziz AA, Osman DM, El-Shafee SF et al (2012) Assessing speech intelligibility in
a group of egyptiandysarthric patients. Egypt J Otolaryngol 28(1):49
AR for Maintenance Training During
COVID-19 Pandemic
Abstract COVID-19 situations are getting worse day by day. Buying behaviors
and economic priorities for purchasing new goods have changed. In these days,
maintenance of existing equipment has gained more importance. For the purpose of
training the maintenance workers or end users in this pandemic situation, augmented
reality is like a blessing today. The maintenance procedures can be communicated
remotely on demand with the help of web based AR and android apps. This paper
describes advantages of using AR based instruction over paper based instruction to
train maintenance operators. The authors have developed web based AR and android
apps for augmenting 3D models. Many benefits observed by using this technology
are reduced human errors reduced execution time, reduced downtime of equipment,
reduced cost, and increased productivity. Paper mainly focuses on ways of delivering
maintenance instructions remotely.
1 Introduction
The COVID-19 pandemic has changed the planet forever, and consequently
the impacts from this are far-reaching. Globally, companies are pressured to under-
take strict work from home policies whenever possible, and outright cessation of
operations while social distancing and remote work weren’t possible. Uniquely, this
become real in each marketplace in each nation. Today, a few markets and areas are
returning to normalcy, at the same time as others are persevering with to conform
to the brand new normal of labour. Interestingly, COVID-19 did not present many
J. Pawar (B)
SVERI’s College of Engineering (Poly.), Ranjani Road, Pandharpur, Maharashtra 413304, India
e-mail: [email protected]
T. Bansode
COER, Fabtech Technical Campus, Sangola, Maharashtra 413307, India
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 459
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_46
460 J. Pawar and T. Bansode
2 Related Works
Training maintenance professionals is a crucial part for any industry. Many authors
have worked to study the use of Augmented reality in maintenance training. A review
paper [3] has collected some publications in this regard out of which 18% have
used air for the Aeronautics industry. They have reported that 34% of papers have
described the use of air for assembling the equipment. In paper [4] a pilot study of
Aeronautical engineers mentioning interview based research has explained the use
of wearable Technology. The study reports that some maintenance tasks require less
time if you use AR based applications. In paper [5] the industrial aspect of using
AR is reported with 91% improvement in quality of maintenance work and 32%
improvement it is achieved in required time to complete a manufacturing operation.
These studies have focused on air for maintenance and proved that are significantly
reduces time and improves performance. Some researchers have also proven the
usability of AR in training and education.
For conveying instructions of routine maintenance of Desktop machines or laptops
regarding replacement of Ram chip or hard disc drive AR based solution is much
beneficial. AR has wide acceptance as educational technology in many educa-
tional settings including school going kids for undergraduate students. As a part
of maintenance training this smart Technology has proved its ability.
AR for Maintenance Training During COVID-19 Pandemic 461
3 AR as Immersive Technology
History. The study of augmented reality (AR) can be traced back to Ivan Suther-
land’s invention of the head-mounted display (HMD) in the 1960s. But it gained
a lot of attention after 1990, when Tom Caudell coined the term. Furthermore, for
many years, AR was constricted to academic research and entertainment through
sci-fi films. Early AR systems were mostly experimental and developed focusing on
specific tasks, such as maintenance and repair tasks. More recently, AR has started
to successfully step away from laboratories and appear in a wide range of applica-
tions, which include advertising, education, for advanced driver-assistance systems,
advertising, defense, manufacturing, networking, tourism, medicine, smart cities, for
social media apps that use face detection [6].
According to Gartner research, AR is poised to become an important workplace
tool. Many companies are increasingly applying AR tools to create effective training
programs to increase employee engagement, reduced efforts for actual training and
to achieve success in the highly competitive global marketplace.
AR for remote assistance: Remote assistance has been the leading use case in AR
for years. Quick and reliable ROI, through both direct travel cost reduction and
increased efficiency through instant knowledge access (and, secondarily, reduced
downtime). Adding value to a process that, at best, was a video call on a mobile
device or, at worst, a traditional telephone call or an asynchronous support ticket,
has been straightforward for AR on both head-worn and mobile devices. Annotation,
spatial awareness, automatic data capture and creation, visualization integration, and
step-by- step instruction all have proven value on both device types.
Remote work is unique in its universal applicability. Internal usage for the enter-
prise sector is relatively straightforward, but customer-facing use cases are growing.
Post- sales support for devices or services has been an increasingly valuable use case,
with AR providing a level of visualization and easy-to-follow guidance not possible
with existing means. Telehealth and telemedicine fit in a similar way.
AR for training: Related to remote work and assistance, training in AR and
VR has been a powerful value driver for companies. Realistic simulation has had
a home in VR, while more real-world training and knowledge transfer fits well
in AR. The capability for in-situ training, where training and workflow accomplish-
ment happen simultaneously, thanks to AR, allows instantaneous employee ramp-up.
With a renewed focus on off-site capabilities, training content can be delivered and
consumed wherever a user is located, with proper visualization and simulation of the
target workflow environment available through the device.
462 J. Pawar and T. Bansode
Data capture: Again tying into remote work and training, data capture in AR
has not seen as much direct activity as training and remote assistance, but it is being
positioned as a promising longer-term approach to content capture, creation, and
distribution. Today, there is an increased emphasis on the quick and easy capture of
data and, more broadly, workflows to maintain efficiency with a reduced workforce
and enable more seamless knowledge capture and transfer. Capturing and logging
data through AR serves as a connection point between data-driven Internet of Things
(IoT) platforms and human workers: data from workers, who without that connection
point would be lost, can now be fed into the system and fully leveraged for analytics
and other uses.
Hands free operation: With head-worn devices, hands-free data access and inter-
action is possible. Input methods like voice, gesture, and gaze can completely elim-
inate hands-on interaction for both the AR/VR device and anything the device can
access. Prior to the pandemic, hands-free was a valuable assurance of an uninter-
rupted workflow and employee safety. Today, there is an added layer of sanitation
potential anywhere head-worn devices are leveraged.
In AR, the virtual objects are superimposed to the real word so that by moving the
point of view of the camera, the symbol positions with respect to the external reference
system does not change. This is obtained by computing the position of the camera
with respect to the external environment. AR based applications are of two types-
one that used markers and other that use objects or specific locations to superimpose
the virtual content. The marker based AR experiences have the facility to run over
any equipment like laptop or handheld device but they have a limitation that the user
should have the printed marker available for generating augmented content in the
camera view. AR.JS is the library which can provide marker based augmentation in
the web browser. Another type of AR experiences uses Android apps specifically
designed to use camera Input and location information to decide the content for
augmentation. Some recent libraries like AR core from Google identify the plane
surface and place the augmented model on that plane surface. In some cases such
AR experiences need extra hardware facilities. The authors have developed market-
based AR experience for augmentation of 3D model and video for web application
based interface. For handheld device based interface, they have developed an Android
application which can augment 3D model of a measurement scale and maintenance
device called Hand held electric saw.
Fig: 1 shows Snapshot of web based application using model-viewer component.
Using model-viewer allows us to augment virtual content without marker. Fig: 2 is
Snapshot of a 3D model of a cube and polygon over a marker in live webcam feed.
Fig: 3 Snapshot of augmentation of 3D model over a marker provided by Google
Poly recorded using browser in a handheld device.The libraries used for development
of web applications using markers shown in figure are AR.js and Aframe.js.
Fig: 4 Snapshot of three dimensional augmented keyboard in real time using a
mobile app. Fig: 5 is the screenshot for Mobile Augmented reality using android
app. It displays a virtual desktop computer placed near a wall. Fig: 6 snapshots of 3D
model of handheld electric saw augmented in real space with mobile app. This app
can be further developed to augment the series of instructions for using that electric
saw for the purpose of maintenance. These instructions will provide the operator
directions to use and operate the saw in real time. The operator is able to see the
464 J. Pawar and T. Bansode
3d model and the instructions are provided in the same window where the model
appears.
In this paper we have discussed the use cases for AR and their impact during COVID-
19 situations. The system described here will be very much useful and it will surely
prove its capacitance while training maintenance workers.
AR for Maintenance Training During COVID-19 Pandemic 465
Fig. 3 Snapshot of
Marker-based AR using 3d
models available on Google
Poly
Fig. 4 Snapshot of three-dimensional virtual keyboard augmented in real-time using a mobile app
Fig. 6 Snapshots of 3D
model of handheld electric
saw augmented in real space
with mobile app
the enterprise market come down to device cost and ease of use; these issues are
the same in the consumer market. Many of the first wave of consumer AR devices
have prioritized keeping prices down, while delivering on promised value. Google
Glass was technically launched as a consumer product, but the price and lack of
clarity around value quickly pushed the device to the enterprise market where it
found purchasers.
Looking ahead, things become both more clear-cut and more complicated. Proven
value in work-from-home and remote collaboration extends beyond the COVID-19
impact. A consistent increase in remote work and tele presence usage overall naturally
includes AR/VR, and as companies and customers return to normal, much of this
AR for Maintenance Training During COVID-19 Pandemic 467
adaptation will persist. The pandemic forced pilots for remote assistance and other
AR use cases for many previously hesitant to adopt immersive technology. These
forced programs will encourage future investment, serving as a kick start for the
market that had been occurring slowly but surely over time.
References
1. Gattullo M, Uva AE, Fiorentino M, Gabbard JL (2015) Legibility in industrial AR: Text style,
color coding, and illuminance. IEEE Comput Graph Appl 35(2)
2. Azuma RT (1997) A survey of augmented reality. Presence: Teleoperators Virtual Environ 6:355–
385
3. Palmarini R, Erkoyuncua JA, Roy R, Torabmostaedi H (2018) A systematic review in augmented
reality applications in maintenance. Rob Comput Integr Manuf 49:215–228
4. Robertson T, Bischof J, Geyman M, Ilse E (2017) Reducing maintenance error with wearable
technology. IEEE
5. The Area website (2020) https://thearea.org/augmented-reality-can-increase-productivity/.
[Accessed: June 2020]
6. The PTC website (2020) https://www.ptc.com/en/service-software-blog/augmented-reality-mai
ntenance-and-repair. [Accessed: June 2020]
7. Airbus Company website (2020) https://airbus-xo.com/real-benefits-virtual-reality/ [Accessed:
June 2020]
Face Recognition Based Attendance
System Using Cv2
Abstract Face recognition based attendance based system will be used in the near
future in classrooms instead of the traditional system; it may replace even biometric
attendance systems. The purpose of the present work is to devise a novel attendance
system using cv2. Facebook also uses face recognition technology as it tags the names
of faces as soon as you upload photos which have been tagged by you previously. The
algorithm identifies the unique features of the faces in the database and encodes them
into pattern image. Python modules are used to Then the machine learning algorithm
called classifier is used to find the name of the person. Image capture, facial features,
face recognition and attendance system, are the stages of the procedure.
1 Introduction
In this paper, face recognition is done with the help of Python programming. All the
faces are initially taken in image form and folder is prepared. The Names are also
given to each image. Facial features are found so that the database is ready.
Once the database is ready, The camera is used to take attendance, and its output
is used along with a database for face detection and face recognition.
After face detection and face recognition, the program prints the result immedi-
ately after the program is run. Also, another program loop can express the output in
separate excel sheet with names, present, absent, date and, time details.
V. Khairnar
Ramdeobaba College of Engineering, Nagpur, Maharastra, India
C. M. Khairnar (B)
Government College of Engineering, Nagpur, Maharastra, India
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 469
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_47
470 V. Khairnar and C. M. Khairnar
Algorithm of Viola and Jones, Principles analysis are used [1]. The first step used is
to detect the face and its recognition, after this the classifier program has compared
the faces with the database of students faces.
Optical flow field [2] method is used here in this study of the face detection; it
uses objects, comparatively a new method for recognition and detection.
The portable device [3] is proposed and developed here for the student attendance
system. The device is to be designed and optimized for camera, server and other
hardware is also to be customized.
A filtering system [4] based on Euclidean distances calculated by three face recog-
nition techniques, namely Eigenfaces, Fisherfaces and Local binary pattern has been
developed for face recognition.
Automated attendance management system [5] along with the techniques used to
handle the threats like spoofing.
The new type of touchless attendance system is studied [6] for the performance,
efficiency and accuracy. The efficiency depends on the hardware as well as the
database quality collected.
Touchless attendance system for institute or classroom is developed and studied
[7].
The present system [8] studied here is about face detection and used for the class-
room attendance using Discrete Wavelet Transforms (DWT) and Discrete Cosine
Transform (DCT), the facial feature extraction is easy by this method.
Deep Learning algorithm helps in detecting the faces in the classroom when
captured by high definition camera and then subjected to the classifier program [9].
The research [10, 11], includes for Face detection Students and the system is
based on CNN perspectives and algorithms.
3 Methodology
In all the biometric attendance process, the biometric features of the person are
compared with the database templates previously prepared at the time of enrollment
of all the persons. The methodology used in our research can broadly be divided into
the following procedural stages.
Camera for the classroom having an optimum resolution. SSD memory for storage of
image database. Computer with a powerful graphics card. The cost of this attendance
Face Recognition Based Attendance System Using Cv2 471
system is mainly the cost of hardware like computer and camera. Software required
is windows/ Linux for execution. For development Face recognition library.
“OpenCV” Python library is used to capture the images generated by the camera.
Nowadays, the camera of the Smartphone also has the face detection feature if noted
carefully, as that feature is used to, focus on the faces and capture.
Another example of face detection is that Facebook also uses this feature. In case
we upload the group photo on Facebook, then it will automatically tag the faces in the
photos, if and only if you have previously tagged those faces with names. Facebook
remembers those tag names using machine learning and can detect them.
Feature base technique is used in this library for face recognition. Facial features
such as the nose, eyebrows, lips are used. The shape of these facial features is different
for different persons. Hence detections can be possible with the help of these deep
learning techniques.
The unique feature in the face, such as long/short nose, which distinguishes it
from other person faces is specially used to recognize and detect the face of a person.
Histogram of an oriented gradient, pattern image is used in this type of machine
learning. The algorithm matches the image captured by the camera with these pattern
images in databases, which highlight the special features of the person, and detects
the face of the person. This image ensures the detection of the face even if it is slightly
tilted/ oriented or in bad light conditions.
Then basic machine classification algorithms can easily find the names of the
person.
In our program “Face recognition” Python library built using deep learning, is
used for face detection and face recognition.
4 Algorithm
import datetime
from datetime import datetime
import cv2
import face_recognition
import numpy as np
472 V. Khairnar and C. M. Khairnar
import openpyxl
def addInExcel(d):
fp = “Attendance.xlsx”
wb = openpyxl.load_workbook(fp)
sheet = wb.get_active_sheet()
max_row = sheet.max_row
max_column = sheet.max_column
now = datetime.now()
# print(“Data Acquired :”, d)
sheet.cell(row=1, column=max_column + 1).value =
now.strftime(“%m/%d/%Y, %H:%M:%S")
for index in range(1, max_row):
sheet.cell(row=index + 1, column=max_column + 1).value =
d[sheet.cell(row=index + 1, column=2).value]
wb.save(fp
video_capture = cv2.VideoCapture(0)
obama_image = face_recognition.load_image_file("images/obama.jpg")
obama_face_encoding = face_recognition.face_encodings(obama_image)[0]
gates_image = face_recognition.load_image_file("images/gates.jpg")
gates_face_encoding = face_recognition.face_encodings(gates_image)[0]
known_face_encodings = [
obama_face_encoding,
gates_face_encoding
]
known_face_names = [
"Barack Obama",
"Bill Gates"
]
attendance_list = {"Barack Obama": "Absent", "Bill Gates": "Absent", " ": " "}
face_locations = []
face_encodings = []
face_names = []
process_this_frame = True
while True:
ret, frame = video_capture.read()
small_frame = cv2.resize(frame, (0, 0), fx=0.25, fy=0.25)
rgb_small_frame = small_frame[:, :, ::-1]
if process_this_frame:
attendance_list[name] = "Present"
process_this_frame = not process_this_frame
Face Recognition Based Attendance System Using Cv2 473
cv2.imshow(’Video’, frame)
if attendance_list[i] == "Present":
present += 1
elif attendance_list[i] == "Absent":
absent += 1
print("Total Present :", present)
print("Total Absent :", absent)
print("------------------------------------------")
attendance_list[’Total Present’] = present
attendance_list[’Total Absent’] = absent
# print(attendance_list)
addInExcel(attendance_list)
Break
cv2.destroyAllWindows()
video_capture.release()
"""
------------------------------------------
Attendance as per 2020-05-06 09:48:11.096916
Name Absent/Present
Barack Obama : Present
Bill Gates : Absent
:
Total Present : 1
Total Absent : 1
474 V. Khairnar and C. M. Khairnar
The library used for getting the output in excel sheet is “Openpyxl” It is a Python
library for expressing the data in excel sheet.
The sample attendance for one person is taken. Database for two people is prepared
one with name Barak Obama and another with name Bill Gate, Fig. 1 shows the
screenshot capture for one person in front of the camera.
The output is shown in excel sheet generated using a Python program is shown
in Fig. 2. The excel sheet screenshot shows the present, absent, date and time of
the attendance taken. The efficiency of the program is 90%, hence can be used for
practical purposes.
6 Conclusion
Deep Learning algorithm helps in detecting the faces in the classroom when captured
by the high-definition camera and then subjected to the classifier program.
The future scope for the work can be refining the algorithm so that proxy
attendance can be detected. Also, Cloud Integration can be done.
Face Recognition Based Attendance System Using Cv2 475
References
1. Wagh P, Thakare R, Chaudhari J, Patil S (2015) Attendance system based on face recognition
using eigen face and PCA algorithms. In: International conference on green computing and
internet of things (ICGCIoT). Noida, pp 303–308. https://doi.org/10.1109/ICGCIoT.2015.738
0478
2. Bao W, Li H, Li N, Jiang W (2009) A liveness detection method for face recognition based
on optical flow field. In: International conference on image analysis and signal processing.
Taizhou, pp 233–236. https://doi.org/10.1109/IASP.2009.5054589
3. Bhattacharya S, Nainala GS, Das P, Routray A (2018) Smart attendance monitoring system
(SAMS): a face recognition based attendance system for classroom environment. In: IEEE 18th
international conference on advanced learning technologies (ICALT). Mumbai, pp 358–360.
https://doi.org/10.1109/ICALT.2018.00090
4. Samet R, Tanriverdi M (2017) Face recognition-based mobile automatic classroom attendance
management system. In: 2017 international conference on cyberworlds (CW). Chester, pp
253–256. https://doi.org/10.1109/CW.2017.34
5. Chintalapati S, Raghunadh MV (2013) Automated attendance management system based on
face recognition algorithms. In: IEEE international conference on computational intelligence
and computing research. Enathi, pp 1–5. https://doi.org/10.1109/ICCIC.2013.6724266
6. Surekha B, Nazare KJ, Viswanadha RS, Dey N (2017) Attendance recording system using
partial face recognition algorithm. In: Dey N, Santhi V (eds) Intelligent techniques in signal
processing for multimedia security. Studies in computational intelligence, vol 660. Springer,
Cham
7. Kar N, Debbarma MK, Saha A, Pal D (2012) Study of implementing automated attendance
system using face recognition technique. Int J Comput Commun Eng 1(2)
8. Lukas SM, Desanti AR, Krisnadi DRI (2016) Student attendance system in classroom using
face recognition technique. In: International conference on information and communication
technology convergence (ICTC). Jeju, pp 1032–1035. https://doi.org/10.1109/ICTC.2016.776
3360
476 V. Khairnar and C. M. Khairnar
1 Introduction
It is not easy to distribute rationing in most populated countries like India. Second
highest populated countries like India there are some reserved castes and the people
under poverty line have been benefited with the livelihood commodities by the Public
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 477
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_48
478 S. K. Rudrawar et al.
distribution system. Ration card holders of various types like BPL,APL and Antyod-
hana Anna Yojana have been benefited with the low price purchase of goods and the
food items like wheat, rice, sugar, cereal, Kerosene, edible oils etc. with the viable
amount by government schemes [1, 2].
Aadhar card is considered as the symbolic representation of citizenship of Indians.
There are lots of schemes that have been provided with this as it is linked at different
places such as banks, hospitals, rationing shops which contain all the details of
customers like identification marks, photo, name, economic status, religious status
[3, 4].
This current ration cards system can be replaced by card less system proposed in
the paper by proper implementation of hardware structure and creating the database
of all customers containing their personal/family details, ration card details, biometric
details and by linking it to Aadhar card which will be stored in cloud for maintaining
transparency [5]. The automated ration distribution system can be installed at the
ration shop which is having two layers of security. First is fingerprint verification
and the other is OTP.
Hardware structure has the fingerprint scanner; Load cell, Motors, speaker
and touch screen display as a user interface are interfaced to Raspberry Pi. The
System will be linked with the mobile phone and the government database. If some
unauthorized user tries to access an account, the authorized person will get alerted.
2 Literature Survey
In the present Public Distribution System (PDS), paper based ration cards are used
by the customers which are categories into three parts based on the economic status
of the family. This ration card needs to be renewed every year, which is a very time
consuming process. Nowadays it is replaced with the rationing machine known as
EPoS machine which is used to authenticate the user and generate a bill based on
the total material user has been allotted. Then the ration shop owner gives him the
utilities [6].
Dr. M. Pallikonda Rajesekan in his research paper “Automatic Smart Ration Distri-
bution System for Prevention of Civil Supplies Hoarding in India” proposed the
system used for distribution of ration with the automated system. Which takes the
manual input from the user for registration of ration cards and then for distribution
of material. He proposed the system which uses the GSM based communication and
AUTORAT system [7].
Padmashree S. in his research paper “Automatic Ration Material Dispensing
System” also worked on the similar kind of system. Here he has proposed a system
which works on the RFID technology for the authentication of users. RFID cards use
the radio frequency to detect the card of a particular user; each card has its own unique
identification number [8]. Madhu Kumar N. and his team used the combination of
both RFID and the GSM based technology for the authentication and communica-
tion with users [9]. Brendon Desouza states in his research work “Smart Ration Card
IoT Enabled Secured Card Less Ration Distribution System 479
Automation System” that the GSM based technology can be used to achieve the high
level of security in Public Distribution services. He developed the system by using
the real time embedded system by using ARM based microcontroller [5].
The system we are proposing here is totally card less system and Google firebase
is used to store the customer’s information as well as online updating the material
allotted to the customer once taken which makes the system more transparent and
avoids malfunctioning. Customers from any region can collect their allotted grains
from any other region location where this hardware is placed or fixed. So here we
addressed and provided a solution for One Nation One Ration scheme.
3 Block Diagram
The Fig. 1 shows the block diagram of the system and components required to run
the entire system. Each block work in hand with others to perform tasks.
The microcontroller unit incorporates the weight measurement system which
controls the motion of valves by providing signals to servo motors. Raspberry Pi
is the most important part of this system which acts like the heart of a human and
is used to perform all computational activities. By providing particular delay to the
opening and closing of the valve we can withdraw the required amount of material.
• Raspberry Pi3 is a 32-bit processor that has two serial communication ports. One
of these ports is used for FINGERPRINT MODULE and GPIO pins are used to
connect different peripherals like Load cell, speaker and servo motors. 3.5 inch
capacitive touch screen interface is provided which shows the information on the
screen also used to take actions.
• A DC servo motor is used to control the outlet valve to dispense the rice and
wheat.
• After successful verification and entering OTP with the help of the keypad shown
on the screen, the transaction will be done.
Proposed system operates on 230 V input and 5 V, 1 A output to power the
raspberry pi board. The Raspberry Pi board is having support for the Linux operating
system on which we have different modules such as eSpeak (Google text to speech
for Linux), finger print module and python 3.7 to write the source code for the
application. Load cell not coming alone it requires HX711 is an SOC having 24- bit
ADC which is designed for weighing measurement and also used in industries for
control applications to the load cell which is used for weight measurement.
Load cells convert the forces applied into an electrical signal which is then
measured and converted into standard form. It will work on the principle of change in
resistance results in change in voltage [10]. Upon application of external stress it will
produce changing output voltage. Available in different shapes, sizes and numbers of
wires (either 4 or 6 depending upon distance between load cell and microcontroller).
The load cell is nothing but the four strain gauge which is configured in Wheatstone
bridge configuration with four different resistors. An excitation voltage up to 10 V
is applied to one end and the voltage is measured between the other two ends. It
is at equilibrium if no force is applied then the output is zero volts or near to zero
when all the resistors are having the same resistance and also known as balanced
bridge circuit. By the application of strain there is the change in resistance resulting
in the variable output voltage. Then this is noted and digitized then amplified the
small millivolts level signals into high amplitude signals. Load cell has four wires
(Two are excitation wires and other two are for output).positive or negative sign in
the output voltage of load cell is dependent upon location of strain applied. Whereas
HX711 has also four wires as VCC, SCK, DT and GND. When output is not fetching
any data then the digital output pin DT is at high. Serial input clock SCK must be
low.
Whenever DT is at 0, data sent by the application 25–27 positive pulses at the
SCK pin, data is shifted out from DT output pin. Each SCK pulse shifts from one
bit, starting with the MSB bit first, until each 24 bits is shifted out. After the 25th
pulse DT pin will pull back to high [11]. To avoid communication error, SCK pulse
should not be less than 25 and not more than 27 within one conversion period.
Fingerprint scanners are a light sensitivity transducer to produce digital images.
And they are available in different variations, many of them have external four pin
connection interface By way of serial interface, fingerprint sensor can communicate
with microcontroller runs on either 5 V or 3.3 V power supply. TX/RD pin connects
IoT Enabled Secured Card Less Ration Distribution System 481
with TXD (TX-OUT pin of the microcontroller). We have used optical fingerprint
scanners, capture high contrast images.
Details are captured with the high number of diodes per inch. They work by
reflecting light over their fingers and taking photos. The ridges of fingerprint help to
produce the digital image.
Figure 2 shows the verification of the user with the help of Fingerprint sensor.
Then this sensor verifies the user by looking into Aadhar Data stored.
We have used the Raspberry Pi 3 module which is supporting the operating system
like Linux. To make our system more user friendly GUI (Graphic User Interface) has
been developed. It can be made with Python programming. We can create GUI with
third party modules like tkinter. For that we have to import all supporting modules.
We can also import the Google text to speech for Linux named as eSpeak by which
we will read out all the instructions.
Source code is the heart of all the hardware which is interfaced to the micro-
controller board. In Python IDE we can write, compile and run the source code.
Google Firebase is used to create and store a real-time database of all the customers.
The data like name, address, Aadhar number, allotted ration, bank account number,
mobile number is stored in the tabular format like we store in excel sheet. Database
is managed and stored on the cloud services provided by Google cloud and Amazon.
Fig. 2 Fingerprint
verification
482 S. K. Rudrawar et al.
4 Methodology
This system proposes the advanced Ration Distribution system, named as “IoT
enabled secured card less Ration Distribution System”. The proposed system is
similar to the ATM and proposed to work for 24/7. It must have a biometric input
to get started. It incorporates an embedded and IOT based automatic ration shop.
Figure 3 shows the front and back end of the proposed system. Steps involved in
implementation of systems are
• Data Collection: We are collecting the data of the user from the Aadhar database
as well as by linking the ration cards of the user with the Aadhar card ID. It will
hold all the data such as Aadhar number, ration card number, bank account details
and allocated goods.
• Network Configuration: Now from the collected data in the system we will
configure the settings which are mainly required for allocating goods to the user
and also to verify the user.
• Storage: We require the storage area to store the grains like rice and wheat as well
as some commodities like sugar.
• Provision: Providing the allocated goods and the information after verifying the
authorized user.
• Warning: If someone tries to use your account then the system will give alert to
both administration as well as particular users. And also it conveys the message
of deposition of material into their account and withdrawal of money from their
bank account.
• Alert: The customer should be alerted whenever there is use of his identity with
the system. This is done by sending an OTP and the emails with the associated
account.
5 Working
6 Prototype Testing
Figure 4 shows the hardware present inside the system, Where two conveyor belts
platforms are used to deliver the goods after successful verification, validation and
measurement. The Hoffer present above is used to deliver the grains by opening and
closing the valve into the container placed on the platform. It takes almost a couple of
minutes and then it measures the delivered grains with the help of a load cell placed
below the conveyor belt. Then to deliver the goods to the customers those belts are
run until it reaches the customers. And the next transaction restarts only after placing
the container. GUI makes the process smooth and user friendly is shown below in
the Fig. 5 which uses touch screen to interact with the user.
484 S. K. Rudrawar et al.
We have used the Raspberry Pi 3 module which is supporting the operating system
like Linux. To make our system more user friendly GUI can be more user friendly.
GUI is an abbreviation for Graphic User Interface. It can be made with Python
programming. We can create GUI with third party modules like tinker CAD. For that
we have to import all supporting modules. We can also import the Google text to
speech for Linux named as eSpeak which will read out all the instructions. Source
code is the heart of all the hardware which interfaced to the microcontroller board.
In Python IDE we can write, compile and run the source code.
Figure 6 shows customers database. Google firebase is used to create and store.
Google firebase is used to create and store real-time database of all the customers.
The data like name, address, Aadhar number, allotted ration, bank account number,
IoT Enabled Secured Card Less Ration Distribution System 485
mobile number is stored in the tabular format like we store in the Excel sheet. After
successful verification of biometrics, the user receives OTP on the registered mobile
number as shown in Fig. 7. The user has to enter the OTP. The user has three chances
to enter the 4 digit OTP if the password does not match an alarm is triggered. After
successful verification, the user has full access to the system where he can select
the ration material and the quantity of material required. Once transaction has been
completed the database will get updated with amount of grains allocated as per recent
transaction. As shown in Fig. 8. materials will get dispatched through the conveyor
belt inside the system. Hence, the experimental results show that the proposed system
is easy to access and prevent the ration from theft activity.
In a populated country like India it is very difficult to manage the records of the public
distribution system. The proposed system not only helped to manage the records but
also it made the distribution of goods among all in a very systematic and transparent
way. It minimized the chances of malpractices done by the shop owners in order to
gain the profit. Also reduced the human intervention which may result in incorrect
measurement and the reduced the chances of infection to others in the situation like
COVID-19.
Proposed system has achieved accuracy and transparency by verification of the
correct user and maintains the record of the goods. It also saved the time consuming
process of distribution of goods at ration shops. With the help of biometric scanning
it reduced the chances of malpractices and the adulteration as there is no human inter-
vention while distribution of goods. This can be a part of an E-commerce platform
IoT Enabled Secured Card Less Ration Distribution System 487
which promotes the cashless transaction. The proposed system will not only aid the
government agencies but will also help to digitize the system and in term help to
deploy resources efficiently to the citizens.
One limitation of proposed work is to get the solution for high volume of goods
storage.
In this year we have faced the rapid spread of COVID-19 which spread with
contacts of people. It’s our future work to make contactless with the help of face
scanning and with the voice commands which can be achieved with Image processing
as well as some signal processing techniques.
References
1. https://instrumentationtools.com
2. https://mahafood.gov.in/website/english/PDS.aspx
3. https://www.indiaenvironmentportal.org.in
4. https://www.tnpd.gov.in
5. Bagul G, Desouza B, Gaikwad T, Panghanti A (2017) Smart ration card automation system.
Int Res J Eng Technol (IRJET) 04(05)
6. Padmavathi R, Mohammed Azeezulla KM, Venkatesh P, Mahato KK, Nitin G (2017) Digi-
talized aadhar enabled ration distribution using smart card. In: International conference on
recent trends in electronics information and communication technology (RTEICT), vol 2(2)
May 19–20
7. Pallikonda Rajesekara M, Balaji D, Arthi R, Daniel P (2017) Automatic smart ration distribution
system for prevention of civil supplies hoarding in India. In: International Conference on
Advanced Computing and Communication Systems (ICACCS -2017), Coimbatore, India. 978-
1-5090-4559-4. 06–07 Jan 2017
8. Aishwarya M, Nayaka AK, Chandana BS, Divyashree N, Padmashree S (2017) Automatic
ration material dispensing system. In: International conference on trends in electronics and
informatics ICEI 2017. 978-1-5090-4257-9
9. Balasubramani A, Sunil K, Madhu KN (2018) Cashless automatic rationing system by using
GSM and RFID technology. In: Proceedings of the Second International conference on I-
SMAC. ISBN: 978-1-5386-1442-6.
10. Fitzgerald DW, Murphy FE, Wright WMD, Whelan PM, Popovici EM (2015) Design and
development of a smart weighing scale for beehive monitoring. In: 2015 26th irish signals and
systems conference (ISSC). Carlow, pp. 1–6. https://doi.org/10.1109/ISSC.2015.7163763
11. B. Buvansari, G. Ramya, R. Shivapriyaa (2015) GSM based home security system. Int J Eng
Tech Res (IJETR). 3(2), ISSN: 2321–0869
Voice Assisted Bots for Automobile
Applications
1 Introduction
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 489
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_49
490 S. K. Rudrawar et al.
list. The camera is used as distance sensor to view obstacle. User will able to see a
view of the left-side and the right-side of your car using camera capture the image
during accident in order to identify who made mistakes.
2 Literature Survey
3 Block Diagram
As shown in Fig. 1, the Voice Controlled Infotainment System consists input devices
such as microphone, OBD II and output devices such as 7” HDMI TFT LCD display,
camera, speaker. The microprocessor works on car battery only through the charging
point for mobile.
Microphone
It takes input from user and convert it into electrical signal.i.e. it acts as transducer.
Signal coming from Mic is analog in nature. The speech input from microphone is
given to the raspberry pi and there input speech is compared with command configure
Voice Assisted Bots for Automobile Applications 491
line in raspberry pi. This process is done by Google. Only one command is effective
at a time.
Raspberry Pi 3
The Raspberry Pi 3 board model B + has a processor of 1.2 GHz 64-bit quad-core
ARMv8 CPU and 1 GB RAM which almost acts like a mini computer. Raspberry
pi 3 Board has 802.11n wireless LAN and Bluetooth 4.1 for this install Raspbian
Jessey in the memory card used for the board. Raspberry Pi 3 has a LINUX based
operating system call Raspbian. There are also 40 GPIO pins which can be used
as both digital input, digital output and to control and interface with various other
devices in the real world, four USB ports, one HDMI port, one Ethernet port, one
3.5 mm Audio jack, micro USB power supply. This board also has serial connections
for connecting a camera (CSI) and a display (DSI).The Camera Module is a great
accessory for the Raspberry Pi, allowing users to take still pictures and record video
in full HD. In our system we use Pi camera to capture image if human appearance
is detected. This captured image is sent by email to owner of the car. Before using
mic, first check whether microphone records properly. First, check if microphone is
listed using the command “lsusb”. Check if mic comes up on the list. Next, in order
to set the mic recording volume to high. To do this, enter the command \alsamixer”
in the terminal. On the graphical interface that shows up, press the up/down arrow
keys to set the volume. Press F6 (all) and then select the Mic from the list. Again,
use the up arrow key to set the recording volume to high.
Steps to connect Headphones with raspberry pi:
• Connect headphones to Rpi
• Type command in terminal raspi-con g
• Go to advance setting Audio
• Select headphone and click ok
492 S. K. Rudrawar et al.
Raspbian OS
Raspbian OS is a Debian-based operating system built for Raspberry Pi boards. It is
the official operating system of the Raspberry Pi single board computers. It provides
a modified XLDE desktop environment. It includes thousands of packages. Raspbian
OS has been used in this project to provide a user-friendly desktop environment for
running the applications on the Raspberry pi. It provides a python command terminal
which is used to execute any python statement. The python code for the sensor,
camera and motors were done in the python IDLE window by default available in
the Raspbian OS package [6].
Python command shell
It is the window where we write the python commands for running and testing the
sensor or camera or the working of the motors. We initialize the.py files here and run
them to get the desired output. The image captured by the pi camera is processed
and worked upon by executing the python commands for the same. To connect the
server to the raspberry pi, we have to enter the command and URL of the domain to
stream the camera [7].
Linux Commands:
• Man: It is used to show the manual of the inputted command. Just like a lm on
the nature of lm, the man command is the meta command of the Linux CLI.
• ls: It show all of the major directories led under a given le system show all the
folders stored in the overall applications folder. The ls command is used for
viewing folders and directories.
• cd: It will allow the user to change between le directories.
• mv: It allows a user to move a le to another folder or directory.
Voice Assisted Bots for Automobile Applications 493
4 Methodology
The Pi camera placed on the bot, takes image recognized by bot is shown in Fig. 2.
The bot moves according to voice command received, bot hardware is shown Fig. 3.
The voice commands given to the system were actuated, actions like move left,
right, forward, backward and also the infotainment system, just by asking “Jarvis
Play music” or Jarvis tell me something about … The system greets the user and
asks how could I help! Thus the system is human interactive.
The concept is programmed by playing the wav files through a speaker if any
voice command is received. The wav files from the SD card module and given as
input to the inverting terminal of the amplifier of LM358 whenever a constant supply
from voltage driver is given to non-inverting terminal. Table 1 is showing different
recorded sounds of voice recording of the robot for different voice commands.
Here an application has been built through which we could monitor and access
the motions of the bots. We need to press the voice button icon, just as we do for OK
GOOFLE!. and once the Google-API is on we need to give commands such as turn
forward, right etc.
Voice Assisted Bots for Automobile Applications 495
Figures 4 and 5 shows that once voice command like forward is given to bot,
the Google API application displays that command on screen, once it’s clear that
the same command which was made is displayed on screen we can proceed ahead
(Fig. 6).
The desired outputs were built successfully using Python and its libraries. We
imported various modules such as speech recognition, pyttx—Python text to speech
and many other required for converting speech to text and vice versa. We used to give
commands which can be seen in the above image which states that it first initializes
Jarvis and then it greets the user as per the time and then listens to users command
and recognize, if the audio command given by user is recognized then the following
action is executed, you can see I said open Youtube in the above program and it
executes the same. In this way the smart assistant works over your voice.
6 Conclusion
References
1. Gaoar, Kouchak SM (2017) \Minimalist design, 2017, “An optimized solution for intelligent
interactive infotainment systems”. Intell Syst Conf 553–557
2. Mi-JinKim, Wook Jang J (2010) A study on in-vehicle diagnosis system using OBD-II with
nevigation. Int J Comput Sci Netw Secur 10(9):136–140
3. Isabella, Retna E (2012) \Study paper on test case generation for GUI based testing. Int J Softw
Eng Appl 3(1):139–147
4. Jarvis R, Ho N, Byrne J (2007) Autonomous robot navigation in cyber and real worlds. In:
2007 international conference on cyber worlds (CW’07). Hannover, 2007, pp 66–73. https://
doi.org/10.1109/CW.2007.31
5. Obispo SL (2017) M3 Pi: Raspberry Pi OBD-II touch screen car computer, 1–30
6. Sonnenberg J (2012) Service and user interface transfer from nomadic devices to car info-
tainment systems. In: International conference on automotive user inter-face and interactive
vehicular applications, pp 162–165
7. Prabha SS, Antony AJP, Meena MJ, Pandian SR (2014) Smart cloud robot using raspberry
Pi. In: International conference on recent trends in information technology, Chennai, pp 1–5.
https://doi.org/10.1109/ICRTIT.2014.6996193
8. Menciassi JH, Park S, Lee S, Dario GP, Jong-Oh P (2002) Robotic solutions and mechanisms
for a semi-autonomous endoscope. In: IEEE/RSJ international conference on intelligent robots
and systems, vol 2. Lausanne, Switzerland, pp 1379–1384. https://doi.org/10.1109/IRDS.2002.
1043947
9. Denker, İşeri MC (2017) Design and implementation of a semi-autonomous mobile search
and rescue robot: SALVOR. In: 2017 international artificial intelligence and data processing
symposium (IDAP), Malatya, pp 1–6. https://doi.org/10.1109/IDAP.2017.8090184
10. Bensalem S, Gallien M, Ingrand F, Kahloul I, Thanh-Hung N (2009) Designing autonomous
robots. IEEE Robot Autom Mag 16(1):67–77. https://doi.org/10.1109/MRA.2008.931631
Content-Based Image Retrieval Using
Color Histogram and Bit Pattern
Features
Nandkumar S. Admile, Akshay A. Jadhav, Swagat M. Karve,
and Anil A. Kasture
Abstract To Identify the Particular image from a huge dataset is a key problem in
Image Processing. For image retrieval, a block truncation coding based method is
used. The image features derived from colour quantizer as well as a Bitmap image.
Using colour quantizer Color Histogram Features (CHF) is obtained, and using
Bitmap image Bit Pattern Feature (BPF) is obtained. The various distance measures
are employed to match the similarity between images. Simulated result shows better
in term of Average Recall Rate (ARR) and Average Precision rate (APR).
1 Introduction
In recent years image retrieval is a key innovative topic. In earlier days annotated,
the based method is used for image retrieval, but in this method, we have to manually
provide the description of each image. Hence in order to solve this issue instead of
this method content-based image retrieval invented. It is the method to identify the
same image available in the dataset—a number of techniques employed for image
retrieval. In 1979 Delp and Mitchell developed the Block Truncation Coding (BTC)
which is normally used for image compression [1]. Block truncation coding method
split the image into noncontiguous image block. From each block, high and low mean
values are calculated by performing Thresholding low mean values bitmap image is
derived. In [5–7] author obtains better accuracy using block truncation coding based
image retrieval. In this paper, the author uses RGB colour space to extract several
of image features. In [8] instead of RGB colour space, the author uses a different
colour space to extract several features of the image. The author obtains the improved
results. In [3, 4] and [9] BPH and BCCM derived to obtain the sameness of contents.
In [10], the author uses greyscale colour, which proves it achieves better image quality
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 499
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_50
500 N. S. Admile et al.
and efficiency as compared to ODBTC [11] and EDBTC [12]. This paper describes
Dot Diffusion Block Truncation Coding (DDBTC) [13, 14] for retrieval of images.
Colour Histogram Features and Bit Pattern Features are extracted.
In the DDBTC technique (Fig. 1), a colour image is divided into a number of non-
overlapping image blocks. The DDBTC encoder generates two quantizers (minimum
and maximum) from the RGB colour space (12). Consider f (i, j) be the image block
at position (i, j). Where i = 1, 2 … M/m and j = 1, 2 … N/n. Let fR (x, y), fG (x, y)
and fB (x, y) are the RGB pixel values in the image block (i, j) where x = 1, 2 … m
and y = 1, 2 … n.
Qmin = {min fR (x, y), min fG (x, y), min fB (x, y)} (1)
Qmax = {maxfR (x, y), maxfG (x, y) and max fB (x, y)} (2)
In the Eqs. (1) and (2) Red, Green, Blue colour is represented by R And B.
The gray scale image is obtained using Eq. (3) as:
The feature descriptor of the image is extracted using DDBTC colour quantizers as
well as Bitmap Image. Using minimum and maximum quantizers, Color Histogram
Feature is removed and using Bitmap Image Bit Pattern Features is extracted.
1. Color Histogram Feature (CHF)
Using minimum and maximum quantizers shown in Eqs. (5) and (6) Color Histogram
Feature is extracted.
Content-Based Image Retrieval Using Color Histogram and Bit … 501
i = 1, 2 … M/m; j = 1, 2 …. N/n.
CHFmin and CHFmax are calculated using Eqs. (7) and (8) respectively:
M N
CHFmin (k) = Pr imin (i, j) = k/i = 1, 2.... ; j = 1, 2.... (7)
m n
where k = 1, 2, Nmin.
M N
CHFmax (k) = Pr imax (i, j) = k/i = 1, 2.... ; j = 1, 2.... (8)
m n
k = 1, 2, Nmax.
For performing image retrieval first, dataset is created, which is called as training
image dataset. Forth similarity measurement CHF and BPF features are used. The
L1 distance represented in Eq. (9), L2 distance represented in Eq. (9), and Modified
Canberra distance are used to calculate the distance between images.
The Various distance metrics are formulated as:
1. L1 distance
N
min
query target
δ(query, target) = α1 ./CHFmin − CHFmin (k)/
k=1
N
max
query target
+ α2 ./CHFmin (k) ∗ CHFmin (k)/
k=1
Nb
target
+ α3 ./BPF.query (k) − BPFmin (k)/ (9)
k=1
Content-Based Image Retrieval Using Color Histogram and Bit … 503
2. L2 distance
N
min
query target
δ(query, target) = α1 ./CHFmin − CHFmin (k)/ ∧ 2
k=1
N
max
query target
+ α2 ./CHFmin (k) ∗ CHFmin (k)/ ∧ 2
k=1
Nb
target
+ α3 ./BPF.query (k) − BPFmin (k)/1 /2 (10)
k=1
Performance Measures:
The APR and ARR are measured by the following Eqs. (12) and (13)
1
Nt
APR = Nq(L) (12)
Nt L q=1
1
Nt
ARR = Nq(Nr) (13)
Nt Nr q=1
The Corel Dataset is employed [2] to perform the image retrieval task. A sample
image is given in “Fig. 2”.Image dataset contains only natural images.
For Experimental result, one image is randomly selected mentioned in “Fig. 3”.
Minimum and Maximum quantized image obtained by performing the quantization
process mentioned in “Fig. 4”. From Grey colour space Bitmap image is derived
mentioned in “Fig. 5”.
By performing the experiment distance between the query image and target image
is measured shown in “Fig. 6”.
Table 2 Calculations of
Class Modified Can. Dist Modified Can. Dist.
modified of Canberra
Distance [13] APR ARR
African 0.8750 90.19
Beach 0.8750 90.19
Building 0.8750 90.19
Bus 0.9375 95.19
Dinosaur 0.9375 95.19
Elephant 0.9375 90.19
Flower 0.9375 95.19
Food 0.9375 95.19
Horse 0.8750 90.19
Mountain 0.9375 95.19
6 Conclusion
Dot diffusion block truncation coding used for performing the image retrieval. The
image brightness and colour distribution represent by colour histogram feature (CHF)
accurately. The image contents are characterized by bit pattern feature (BPF). The
DDBTC provide the highest average precision rate and average retrieval rate. In
future work, the same technique can also be applied for video retrieval. In which
an image sequence DDBTC technique can be applied directly. Use another colour
space can also be possible by this technique.
References
1. Delp EJ, Mitchell OR (1979) Image coding using block truncation coding. IEEE Trans Comm
COM-27 9:1335–1342
2. Corel Photo Collection Color Image Database. [Online]. Available: https://wang.ist.psu.edu/
docs/realted/. Accessed 2001
508 N. S. Admile et al.
3. Qiu G (2003) Color image indexing using BTC, (2003). IEEE Trans Image Process 12(1):93–
101
4. Gahroudi MR, Sarshar MR (2007) Image retrieval based on texture and colour method in
BTC-VQ compressed domain. In: Proceeding of the international symposium signal process
Its applications, pp 1–4
5. Silakari S, Motwani M, Maheshwari M (2009) Color image clustering using block truncation
algorithm. Int J Comput Sci Issues 4(2):31–35
6. Lai CC, Chen Y-C (2011) A user-oriented image retrieval system based on interactive genetic
algorithm. IEEE Trans Instrum Meas 60(10):3318–3325
7. Yu FX, Luo H, Lu ZM (2011) Colour image retrieval using pattern co-occurrence matrices
based on BTC and VQ. Electron Lett 47(2)
8. Yu F-X, Luo H, Lu Z-M (2011) Colour image retrieval using pattern co-occurrence matrices
based on BTC and VQ. Electron Lett 47(2):100–101
9. Xingyuan W, Zongyu W (2013) A novel method for image retrieval based on structure element’s
descriptor. J Vis Commun Image Represent 24(1):63–74
10. Guo JM, Liu YF (2014) Improved block truncation coding using optimized dot diffusion. IEEE
Trans Image Proc 23(3):1269–1275
11. Guo JM, Prasetyo H (2014) Content-based image retrieval using feature extracted from
halftoning based BTC. IEEE Trans 24(3):1010-1024
12. Guo JM, Prasetyo H (2014) Effective image retrieval using feature extracted using dot diffused
BTC. IEEE Trans 24(3):1010–1024
13. Admile NS (2018) Image retrieval based on block truncation coding. In: IEEE international
conference on communication and electronics systems (ICCES-2018). Coimbatore, India
14. Admile NS, Kanase A, Khiste R (2019) User-oriented image retrieval using feature extracted
by dot diffusion block truncation coding. In: IEEE international conference for convergence in
technology (I2CT-2019). Pune, India, pp 29–31
KNN and Linear SVM Based Object
Classification Using Global Feature
of Image
Abstract Machine learning plays a vital role in Object classification due to its
various applications viz autonomous vehicle, driverless cars. In our research work we
have considered machine learning algorithm, linear support vector machine (SVM)
and K-Nearest Neighborhood (KNN) for classification of object like car and truck
which are essential for Autonomous vehicle applications. We have performed RGB
to gray conversion followed by histogram of gradient (HOG) for feature extraction
before applying to KNN and SVM for classification. The dataset required for the
experimentations for training and testing are utilized from kaggle website and the
performance of SVM and KNN have been evaluated on these testing data. Results
show that SVM outperforms the KNN providing accuracy of 71.3%.
1 Introduction
Machine learning is very important topic now days. According to global survey
growth in machine learning is 48.3%. We can implement so many applications with
machine learning algorithms such as in medical field, robotics, and weather fore-
casting. Driverless car is one of the famous applications of machine learning. With the
help of machine learning algorithms we can design system with minimum program.
Computer program and its development is prime work of machine learning algorithm
and that can access data and use it learns from them. Feature selection is very impor-
tant task in image processing and computer vision applications [1, 2]. Processing
on huge amount of data makes your algorithm inefficient so feature selection is
important [3–5]. Features are nothing but numerical values extracted from images
[6]. Feature or descriptors selection algorithms [7–9] broadly categorize into local
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 509
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_51
510 M. M. Bhosales et al.
feature and global. The entire image describe by Global feature and region of interest
is describe by local features like speeded up robust features (SURF) [10], scale-
invariant feature transform (SIFT) [11], Local binary patterns (LBP) [12]. Global
descriptors are generally used in image classification, object detection. Texture in
an image patch is represented by local descriptor. Invariant moment, histogram of
oriented gradient (HOG), Co-HOG, shape matrices, hue, haralic texture are some
examples of global descriptors. Nowadays, deep learning algorithms used for high
level semantic features. Convolution neural network(CNN) based approaches are
also useful for feature selection of large number of dataset [13] So overall this area
of feature improvement with traditional algorithm is still remains an open research
area [14, 15]. This work is preliminary step of real time object detection. This work
mainly focused on comparison of different machine learning algorithms, based on
performance of the trained model selection of best classifier is possible. Here we
have used HOG feature that is global feature of images. We used SVM and KNN
algorithms and compared the performance of these classifiers and then we have
recognized that SVM works better for this particular set of parameters application.
We have applied this on database in future we can apply it on real time with more
features and with some different algorithms. The paper is organized in 5 sections.
Section 2 elaborates Literature Review. Methodology discussed in Sect. 3, in which
we have explained block diagram of the system then about HOG feature and different
machine learning algorithms like KNN and SVM. Results have been discussed in
Sect. 4 wherein we have compared SVM and KNN with different parameters which
is followed by conclusions in Sect. 5 and future scope in Sect. 6.
2 Literature Review
There is a huge research has been going on object detection as specified in [16–
18]. Solved the problem of grouping and extraction of image features from complex
images using hierarchical based approach like voting and clustering studied in [19].
Fused local codebook less model feature and local binary pattern features for high
resolution scene classification elaborated in [20]. In [21], authors work on two prob-
lems; the methods defining directly texture in color space but more emphasis on
color. Combining different features with their good and bad parameters may worsen
the performance of system so they proposed new algorithm which will combined
features by using Color Intensity Based Local Difference Pattern (CILDP) and bag
of visual words (BOW) and the average precision improved by 12.53% and 6.68%.
Work focused on object detection on [22] tracking system based on SIFT and SURF
and their comparative analysis. SURF gives more accuracy as compared to SIFT. In
[23] paper Combined global feature with local feature for texture classification, with
the combination global matching and Local Binary Pattern Variance (LBPV) giving
drastic improvement in the more than 10% accuracy in sometimes.
KNN and Linear SVM Based Object Classification Using Global … 511
3 Methodology
The proposed methodology, overall algorithm has been splinted into two parts one
is training and other is testing.
(1) Training Model: Overall training model has been splinted into 3 steps. As
shown in Fig. 1a. First step is preprocessing then feature extraction then train
data using different machine learning algorithms.
(a) Gray scale conversion is utilized for preprocessing: Let v(x) = input
image as and converter output image = z(x) as specified in [24].
v(x) Is uniformly increasing step wedges and z(x) will convert that
uniformly increasing step wedges into exponentially increasing wedges.
(b) Feature Extraction using HOG: This gives information about appearance
about local object and shape, using edge direction distribution or inten-
sity gradient distribution [25]. Image is segmented into small connected
regions and then calculates histogram of those regions. Overall final
descriptor is addition of all these histogram [26]. For enhancing the
Fig. 1 a Block diagram of training phase of system, b Block diagram testing phase of system
512 M. M. Bhosales et al.
of y(j) = −1. Distance between hyperplanes and nearest q(j) point should be
maximum. Equation of the hyperplane is,
s.
q −d = 0 (5)
−−→
where s = normal vector to the hyperplane, d/||q|| = offset of hyperplanes from
origin along s. Figure 2 shows graph of SVM.
B. K-Nearest Neighborhood (KNN): it is useful for both classification and
regression problem. Figure 3. shows flowchart of KNN.
(2) Second part of algorithm is testing. In this we do same operation on testing
data which we already performed on training data. And provide that data as an
input to trained model and predict the output. As shown in Fig. 1b.
– d= -1
4 Results
For experimentation work we have used open source database of kaggle [27] and
python 3.6 for simulation. Database having two classes, truck and other is car. First
we have divided dataset into two parts, 70% for training and 30% for testing using k-
fold algorithm then resized trained images into 128*128 images after that converted
into gray scale then extracted total 16,384 features of 782 images using HOG as
shown in sample image of truck and car in Figs. 5 and 6 respectively.
After feature extraction we have trained our model using KNN algorithm as shown
in Fig. 7 with K (nearest data point) = 5, k- fold algorithm for cross validation with
k (cross validation parameter) = 10. This provides model training accuracy 65.45%
and model testing mean accuracy 59.84% and standard accuracy 58.60%.
Further we have trained our model with SVM as shown in Fig. 8. For this we have
selected linear kernel and with random state parameter is 9 with k-fold algorithm for
cross validation with k = 10. This gives model training accuracy 63.63% and model
testing mean accuracy as 71.31% and standard accuracy is 74.03.comparative study
between SVM and KNN as shown in Table.1.
5 Conclusion
In this paper, we have discussed various object detection papers in the literature
review. We presented comparative study between KNN and SVM algorithm for
object classification for two class truck and car. From our work we have observed
that for SVM with linear kernel and random state = 9, k (cross validation = 10) our
model accuracy is 71.31% and for KNN with K (nearest data point) = 5, k (cross
validation) = 10 our model accuracy is 59.84%. So SVM outperforms KNN.
516 M. M. Bhosales et al.
6 Future Scope
In future scope focus need to be given on local features or combination of local and
global descriptors for better results for same database which are considered in this
paper and for improvement in results you can improve database count and you can
modify SVM and KNN model parameters such as kernel, C, gamma, random state
etc. Also you can compare with different other classifiers such as logistic regression,
random forest, decision tree etc. you can also add more classes for same applications
such as motorcycles, humans, traffic signals etc. Based on our current results and
future experimentations, results will be helpful in the area of driverless.
KNN and Linear SVM Based Object Classification Using Global … 517
Acknowledgements We would like to show our gratitude to JSPM’s Rajarshi Shahu College of
Engineering, for sharing their pearls of wisdom with us during the course of the proposed work.
References
1. Chen J, Lei B, Song Q, Ying H, Chen DZ, Wu J (2020 Proceedings of the IEEE/CVF. In:
Conference on computer vision and pattern recognition (CVPR), pp 392–401
2. Chen S, Tan X, Wang B, Lu Hu X, Fu Y (2020) Reverse attention based residual network for
salient object detection. IEEE TIP 29:3763–3776
3. Fan DP, Lin Z, Ji Z, Fu H, Cheng M (2020) Taking a deeper look at co-salient object detection.
In: CVPR, pp 2919–2929
4. Banerji S, Verma A, Liu C (2012) Cross disciplinary biometric systems. LBP and color
descriptors for image classification. Springer, Berlin, pp 205–225
5. Yu J, Qin Z, Wan T, Zhang X (2013) Feature integration analysis of bag of features model for
image retrieval. Neurocomputing 120:355–364
6. Miao J, Niu L (2016) A survey on feature selection. Procedia Comput Sci 91:919–926
7. Lowe DG (2004) Distinctive image features from scale-invariant keypoints. Int J Comput Vision
60(2):91–110
8. Yongsheng D (2017) Multi-scale counting and difference representation for texture classifica-
tion. Vis Comput. https://doi.org/10.1007/s00371-017-1415-4.
9. Ledoux A, Losson O, Macaire L (2016) Color local binary patterns:compact descriptors for
texture classification. J Electron Imaging 25(6):061404
10. Bay H (2008) Speeded-up robust features (SURF). Comput Vis Image Underst 110(3):346–359
11. Ojala T, Pietikäinen M, Harwood D (1996) A comparative study of texturemeasures with
classification based on featured distributions. Pattern Recognit 29(1):51–59
12. Shuang B, Zhaohong L, Jianjun H (2017) Learning two-pathway convolutional neural networks
for categorizing scene images. Multimed Tools Appl 76(15):16145–16162
13. Felzenszwalb P, Girshick R, McAllester D, Ramanan D (2010) Object detection with
disciminatively trained part-based models. IEEE Trans Pattern Anal Mach Intell 32:1627–1645
14. Lee H, Grosse R, Ranganath R, Ng A (2009) Convolutional deep belief networks for scalable
unsupervised learning of hierarchical representations. In: Proceedings annual international
conference on machine learning, pp 609–616
15. Mohan A, Papageorgiou C, Poggio T (2001) Example-based object detection in images by
components. PAMI 23(4):349–361
16. Papageorgiou C, Poggio T (2000) A trainable system for object detection. IJCV 38(1):15–33
17. Viola P, Jones MJ, Snow D (2003) Detecting pedestrians using patterns of motion and
appearance. In: The 9th ICCV, Nice, France, vol 1, pp 734–741
18. Schneiderman H, Kanade T (2004) Object detection using the statistics of parts. IJCV
56(3):151–177.
19. Gavrila DM (1999) The visual analysis of human movement: A survey. CVIU 73(1):82–98
20. Foresti GL, Regazzoni C (2000) A hierarchical approach to feature extraction and grouping.
IEEE Trans Image Proc 9(6):1056–1074
21. Bian X, Chen C, Tian L, Du Q (2017) Fusing local and global features for high-resolution
scene classification. IEEE J Sel Top Appl Earth Observations Remote Sens 10(6):2889–2901
22. Li L, Feng L, Wu J, Sun M, Liu S (2018) Exploiting global and local features for image retrieval.
J Central South Univ 25(2):259–276
23. Sakai Y, Oda T, Ikeda M, Barolli L (2015) An object tracking system based on SIFT and SURF
feature extraction methods. In: 18th International conference on network-based information
systems
518 M. M. Bhosales et al.
24. Guo Z, Zhang L, Zhang D (2010) Rotation invariant texture classification using LBP variance
(LBPV) with global matching. Pattern Recognition 43(3):706–719
25. Gavrila DM, Giebel J, Munder S (2004) Vision-based pedestrian detection: the protector+
system. Proceeding of the IEEE Intelligent Vehicles Symposium, Parma, Italy
26. Dalal N, Triggs B (2005) Histograms of oriented gradients for human detection. In: Proceedings
of the 2005 IEEE computer society conference on computer vision and pattern recognition
(CVPR’05), vol 05, pp. 1063–6919
27. https://www.kaggle.com/enesumcu/car-and-truck
MURA: Bone Fracture Segmentation
Using a U-net Deep Learning in X-ray
Images
1 Introduction
Medical image processing is the technique for visualization of interior parts and
posterior parts of the body for clinical analysis. Medical image processing helps to
represent interior structures of the bone, muscle and is useful for diagnosis.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 519
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_52
520 K. Ghoti et al.
This study gives details about the different methods for fracture or crack detection
based on X-ray images. For that purpose, we have been studied frequently used
techniques of previous papers. For this paper, we used the MURA (musculoskeletal
radiographs) dataset, which contains seven different classes such as hand, wrist,
finger, shoulder, hummus, elbow, forearm. These images are classified as normal
and abnormal in the classification task [5].
The final goal of the paper is to segment out the lesion from the remaining bone
tissue that the temporal changes can be examined. The use of a deep learning approach
based on the U-Net model in if to avoid the human interventions and it can be seen as
automatic and accurate combine bone and fracture segmentation. This could permit
doctors to find a similar lesion in the bone X-ray images.
In this paper, MURA dataset is used only upper extremity part of the human body in
shown Fig. 1.
In MURA dataset, seven different upper extremity part of the human body are
present such as elbow, humerus, shoulder, finger, forearm, hand and wrist showed in
Fig. 2. The shoulder is an upper part of the back and arms. Long bone in the arm that
runs from the shoulder and elbows is known as the humerus. Elbow is the visible
joint between the upper and lower part of the arm. The forearm is part of the arm
between the elbow and wrist. The wrist is the joint connecting the hand with the
forearm. The finger is a basic part of the hand. In the above study, different views
are present in Fig. 3.
Fig. 2 Upper extremity part included shoulder, wrist, elbow, forearm, humerus, hand [7]
Fig. 3 Different view of shoulder, elbow, wrist, hand and finger [7]
3 Mura
4 Related Work
crack and input radiograph are first preprocessed by rotating, cropping, and applied
an aspect ratio to yield a fixed resolution of 1024 × 512. DCNN is an extension of
the U-Net architecture and achieved the fracture probability of 0.98.
Dong et al. [13] developed a brain tumor segmentation using U-Net based deep
convolutional networks. Proposed Model is trained using Multimodal Brain Tumor
Image Segmentation (BRATS 2015) dataset, which consist of 220 high-grade brain
tumor and 54 low-grade tumor cases. Cross-validation has shown that this method
can obtain promising segmentation efficiently.
Ng et al. [14], proposed a method for the classification using DenseNet169 archi-
tecture based deep convolutional network. MURA dataset is used Hight classification
accuracy of 0.93% including seven different classes.
5 Deep Learning
6 Methodology
The while images which are required for this project are collected from Stanford
radiologists the images are in the size of 512 × 512 after resizing. we have used
2000 images in the MURA database including elbow, forearm, hand, finger, humerus,
wrist, shoulder.
A. Input Image: Input image is taken from the Database. The size of the image is
512 × 512 in ‘.png’ format (Fig. 6).
B. Conversion: In this stage, the input image is converted into the Gray in shown
in Fig. 7.
C. Creating Mask: For the segmentation stage, the mask of the input image is
needed. Mask is nothing but the region of interest. We can create the region of
interest of various shapes like circle, polygon, rectangle, and ellipses, but here
we used freehand shape because bones do not have any fixed shape.
Mask for bone fracture segmentation:
In the deep learning approach, we need ground truth data to compare network
generated masks with the annotated masks. Some datasets are provided with
their ground truth masks. But this dataset does not have annotated masks with
them. So, the masks are manually created, and they binarized to get the masks
as shown in the Fig. 6. The ground truth marked on MURE data is verified by
the expert radiologist.
D. Data Augmentation: Augmenting the data is nothing but increasing the number
of images in the database to get better generalization in the Model used where
more amount of the data, as well as variation, is needed so, for getting enough
variation, we must generate more data from the dataset with limited images.
This is what data augmentation is. The augmentation process increases the
training images based on the operations which are used when a neural network
is developed, which improves the overall network performance whenever the
Model Creating
Mask
Result
7 Results
2(G ∩ S)
DICE = (1)
|G| + |S|
8 Dependencies
For the deep learning approach, there are some libraries, and high computation
systems are required. Following some dependencies used for the training model,
Keras with a TensorFlow backend Next to this We used scikit-learn, simpleitk,
beautiful soup, OpenCV.
9 Conclusion
The paper presented bone fracture Segmentation Using Deep learning. This method
required to test and train images that contains the original input image and mask
images, respectively. The system has been tested on images obtained from Miles
Research. The proposed method gives approximately 95.95% for training accuracy
and 90.29% for testing accuracy, including bone fracture segmentation. The proposed
system is very effective and can be implemented in real-time applications. We can
also classify the bone X-ray images, whether it is either normal or abnormal. The main
advantage of the approach, which is presented in this paper is its uniform nature and
can be applied to different medical image segmentation tasks. The proposed Model
achieved an acceptable performance with the MURA dataset images.
References
12. Lindsey R, Daluiski A, Chopra S, Chapelle ALA et al (2018) Deep neural network improves
fracture detection by clinicians. In: PNAS
13. Dong H et al (2017) Automatic brain tumor detection and segmentation using U-net based
fully convolutional networks. In: MIUA 2017, 069, v3: ’Automatic Brain Tumor Detection
and Segmentation Using U-Net
14. Rajpurkar P, Irvin J, Bagul A, Ding D, Duan T, Mehta H, Yang B, Zhu K, Laird D, Ball RL,
Langlotz C, Shpanskaya K, Lungren MP, Ng A et al (2018) MURA, 22 May 2018, “Large
dataset for abnormality detection in musculoskeletal radiograph.” arXiv preprint arXiv:1712.
06957v4.
15. https://images.app.goo.gl/RdQGf9XyrJrcwXh66
16. Ronneberger O, Fischer P, Brox T (2015) 18 May 2015, U-Net: Convolutional networks for
biomedical image segmentation. arXiv:1505.04597v1
Effective Usage of Oversampling
with SMOTE for Performance
Improvement in Classification Over
Twitter Data
Abstract This paper highlights an attempt for addressing the issue of imbal-
anced classification resulted due to deployment of machine learning algorithms
over an imbalanced dataset. It has used Synthetic Minority Oversampling Tech-
nique (SMOTE). This type of augmentation of the dataset is extremely necessary
as it leads to poor performance in the minority class. Four machine learning algo-
rithms were deployed on the Twitter dataset using the Python platform. Standard
data preprocessing including data cleaning, data integration, data transformations,
and data reduction was carried out first as the most necessary arrangement before
experimentations.
1 Introduction
D. Patil (B)
Department of Computer Science, Smt. K.R.P. Kanya Mahavidyalaya, Islampur, Sangli, MS, India
P. Katyare
MCA Department, PCCOE, Pune, India
P. Bhalchandra
School of Computational Sciences S.R.T.M. University, Nanded, MS, India
A. Muley
School of Mathematical Sciences, S.R.T.M. University, Nanded, MS, India
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 533
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_53
534 D. Patil et al.
under sampling and oversampling can also work depending upon the situation and
distribution of the sample. The SMOTE (Synthetic Minority Over-sampling Tech-
nique) [2, 4] is an example of oversampling approach where we do not duplicate data
points or examples in the minority class rather we synthesize new examples from the
existing examples so as to add new information in the model. This is done by selecting
minority groups and making changes in the existing minority instances. This is also
known as amplification. To demonstrate the working of SMOTE for performance
improvement of selected classification algorithms, we have used a Twitter dataset
[5] that has 973 networks with 81,306 nodes and 1,768,149 edges. The dataset has
been anonymized to preserve confidentiality. An Ego network is drawn to represent
this database using positive and negative edges. To demonstrate the actual working,
we have identified two classes for our study and this created class imbalance problem.
Due to imbalance distribution of class in the considered dataset we witnessed down-
graded performance of the selected machine learning algorithms. This is addressed
by re-sampling of dataset [3, 4]. Accordingly SMOTE method was implemented.
Numbers of experiments are done to find best accurate classification model for our
study according to the objectives.
2 Experimental Set Up
Standard data preprocessing including data cleaning, data integration, data trans-
formations, and data reduction [1, 6] was carried out first as the most necessary
arrangement before experimentations. Data reconstruction [6, 7] was done later
to precisely realize under sampling as well as over sampling strategies. We have
assumed that the imbalance is there by default. Hence one class was labeled as
majority and other as minority [7]. Imbalance distribution always hampers training
mode and perhaps make entire training biased by totally ignoring minority class.
Hence random sampling was done and the dataset was randomly re-sampled. This is
done either of the way like, delete examples from majority class (under sampling) or
duplicate examples if the minority class(over sampling). We work on the assumption
that the predictions on minority class are important for improved performance. For
these issues, we have used SMOTE (Synthetic Minority Oversampling Technique)
for oversampling of dataset [5]. Its standard algorithm [2, 3] is presented in Fig. 1.
Table shows an example of calculation of random synthetic samples. The amount
of over-sampling is a parameter of the system, and a series of ROC curves can be
generated for different populations and ROC analysis performed.
During the course of experimentations, we have found that our Twitter dataset
ego network graph shows approximately 21,476 number of positive sample edges in
the majority class and 10,738 number of negative sample edges in the minority class.
This was, however, an observation for the training set used in cross-validation. The
SMOTE algorithm in Fig. 1 selected minority class instance first at a random and
then found its k nearest minority class neighbors. Creation of synthetic instances then
started by choosing randomly, one of the k nearest neighbors and then connecting
Effective Usage of Oversampling with SMOTE for Performance … 535
these both to form a line segment in the feature space. The synthetic instances
are generated as a convex combination of the two chosen instances. This mecha-
nism of creating synthetic creation can be repeated as many times as needed until a
targeted balanced percentage is reached. However, this is great deal of disadvantage
to SMOTE approach to take into consideration new synthesized instances without
taking into consideration of the majority class [2, 8]. Figure 2 shows positive and
negative edges of the ego network during the course of observations.
Now, we have applied selected machine learning algorithms for classification
including K-Nearest Neighbors Algorithm (KNN), Logistic Regression Classifier
Algorithm (LRC), Artificial Neural Network Algorithm (ANN) and Support Vector
Machine Algorithm (SVM) for comparison of the performance improvement.
The KNN [1, 9] is a widely used classification algorithms which makes prediction
based on the neighborhood distance with an assumption that the similar things exist
in proximity [10].
The LRC [1, 9] is typically recommended for categorical based classification. It
is typically used for binary classification problems. Decisions in LRC are made with
the help of a logistic curve limited to values between 0 and 1. A sigmoid function is
used for testing of positive and negative values [11].
The ANN [1] works mimicking of human brains. We have used a multilayer
perceptron as a common artificial neural network model. It has an input layer, one
or more hidden layers, and an output layer. Neurons have an activation function that
operates upon the value received from the input layer. The outputs from the first
hidden layer neurons are multiplied with the weights of the second hidden layer; the
results are summed together and passed to the neurons of the proceeding layers. This
process continues until the outer layer is reached.
The SVM [9, 10] is extremely recommended for two-class learning classification
problems. It uses a hyper plane for separation of these two classes.
All these considered algorithms are implemented on the dataset and their perfor-
mance is evaluated using a confusion matrix. This matrix has rows and columns
where rows are the actual class and the columns show predicted class. In order to
tabulate results, we assumed following terminology,
Effective Usage of Oversampling with SMOTE for Performance … 537
1. Use of Confusion matrix for describing performance of the model. It has some
attributes like False Positive (FP), False Negative (FN), True Positives (TP) and
True Negatives (TN).
2. Total sample count is sum of all FP, FN, TP and TN.
3. Use of Accuracy parameter to record correctness of the model, which is TP +
TN/Total sample count.
4. Precision is taken as TP + FP/Total sample count.
5. Recall is taken as TP + FN/Total sample count.
6. F1 Score is harmonic mean of Precision and Recall as they both deal with
relevance. It is best value at 1 and worst at value 0.
7. Area Under the Curve (AUC) shows ability of the model to distinguish between
FN and FP classes and used as a synopsis of receiver operator characteristic
(ROC) curve that narrates about performance of binary classification algorithm
[11].
Above considerations used to measure the performance of the model are shown
in the Table 1. Graphical representations of the results are shown in Figs. 3 and 4.
85
80 77.7
75.89
75
70
65
KNN LR ANN SVM
538 D. Patil et al.
100
50 F1-score
Recall
0 Precision
KNN LR ANN SVM
Precision Recall F1-score
3 Conclusion
This paper demonstrates data augmentation for the imbalanced dataset related to
minority class using SMOTE approach. This was mainly done to demonstrate the
performance improvement of selected classification algorithms over imbalanced
Twitter dataset. Experiments were done on the python platform. It is witnessed that
the KNN showed maximum accuracy whereas Logistic regression as the lowest
accuracy for the SMOTE based performance enhancements.
References
1. Jan S, Ruby R, Najeeb PT, Muttoo MA (2017) Social network analysis and data mining, by ©
2017, IJCSMC Saima Jan et al. Int J Comput Sci Mob Comput 6(6):401–404
2. Chawla NV, Bowyer KW, Hall LO, Kegelmeyer WP (2002) SMOTE: synthetic minority over-
sampling technique. J. Artif Intell Res 16:321–357
3. Mohammed AJ, Hassan MM, Kadir DH (2020) Improving classification performance for a
novel imbalanced medical dataset using SMOTE method. Int J Adv Trends Comput Sci Eng.
9(3). ISSN 2278–3091
4. Papagelis M, Das G (2010) Sampling online social networks. IEEE Trans Knowl Data Eng
X(X):Month 20xx
5. Web resource at https://www.kaggle.com
6. Boyd D, Ellison N (2009) Social network sites. J Comput-Mediated Commun from: https://
jcmc.indiana.edu/vol2013/issue2001/boyd.htl
7. Wasserman S, Galaskiewicz J (1994) Advances in social network analysis: research in the
social and behavioral sciences. Sage Publications, Thousand Oaks
8. Zheng Z, Cai Y, Li Ye (2015) Oversampling method for imbalanced classification. Comput Inf
34:1017–1037
9. Jiawei H, Kamber M (2001) Data mining: concepts and techniques San Francisco. California,
Morgan Kauffmann
10. Muthuselvi et al (2012) Information retrieval from social network, IJCA. In: Proceedings on
E-governance and cloud computing services vol 4
11. Erdogan SZ et al (2006) A data mining application. A student database. J Aeronaut Space
Technol 2(2)
Multi-Classification of Breast
Histopathological Image Using Xception:
Deep Learning with Depthwise Separable
Convolutions Model
Abstract One of the best methods for Breast cancer diagnosis is histopathological
images from the visual analysis of histology, but pathologist requires lots of expe-
rience and training to an accurate diagnosis. Therefore, computer-aided diagnosis
(CAD) is an automated and more precise method. Recent developments in computer
vision and deep learning (DL), DL based models are popular in analyzing the hema-
toxylin–eosin (H&E) stained breast cancer digital slides. This paper proposed a
deep learning-based framework, called multi-classification of breast histopatholog-
ical image using Xception: Deep Learning with Depth wise Separable Convolutions
model (MCBHIX). Xception based on depthwise separable convolution layer. We
trained this network from scratch for binary classes and for multi-classes BreakHis
dataset. The accuracy achieved by MCBHIX- 99.01% for binary type and 96.57%
for Multiclass.
1 Introduction
Breast cancer (BC) is the most common type of cancer, especially in women, 14%
of cancers in Indian women. For every four minutes, an Indian woman is diagnosed
with breast cancer. A 2018 report of Breast Cancer statistics recorded 1, 62,468 new
registered cases and 87,090 reported deaths. The survival rate Cancer is 60% for the
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 539
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_54
540 S. D. Pujari et al.
Indian women as compared to 80% in the U.S [1]. In 2020 BC goes as high as two
million [2].
More accurate and early detection and diagnosis of BC reduce the mortality
rate, there are many tests available including physical examination, Mammography,
tomosynthesis, magnetic resonance imaging(MARI), ultrasound and biopsy [3]. Out
of them, biopsy serves as the gold standard for BC detection. Pathologists need profes-
sional background, rich experience for manual classification of breast hematoxylin
and eosin stained slides is costly. Recent improvements in biomedical image anal-
ysis using computers, to improve clinicians accuracy the computer-Aided Diagnosis
(CAD) systems have developed that can help pathologists to be more productive,
objective and consistent in diagnosis.
For classification BC, many handcrafted feature-based algorithms like geometric
means of symmetric positive definite matrices (mSPD) [4, 5], morphological [6]. But
it is hard to classify H & E stained images using handcrafted features, due to complex
structure, huge variation in inter-class and intra-class, the similarity in micro-
scopic images these are challenges. Supervised Machine learning algorithms-support
vector machine(SVM) [2, 7], K-Nearest Neighbor, Decision Tree [8] Recently,
Deep learning is continuously fueling medical image processing, bilinear convo-
lutional neural network (BCNN) [9], Deep layer CNN(DCNN) [10], VGG16 [11],
ResNet(MuDeRN) [12] these classified in only two classes benign and malignant.
Malignancy tumour detection is still a challenging problem. In this work, we use
the openly available BreaskHis Dataset, which contains eight types. We designed a
model which does a binary classification as well as multi-classification images into
eight classes’. There are treatment options available for BC patients, and the BC
subtype could help to predict the patient’s response to therapy; for example, invasive
lobular cancer gains a clear benefit from systemic treatment when compared to inva-
sive ductal cancer. The correct recognition of benign lesion type is also important
because the patient’s risk of developing subsequent BC varies among different types
of benign lesions [12]. Xception Model which is entirely dependent on depth wise
convolution layers [13] which reduces the computational cost of the training model.
Table 1 The BreakHis database distributed into four magnification levels and main classes, their
subcategories
Class Sub-class Magnification level Total
40X 100X 200X 400X
Benign Adenosis 114 113 111 106 444
Fibroadenoma 253 260 264 237 1014
Tubular adenoma 109 121 108 115 453
Phyllodes tumor 149 140 150 130 569
Malignant Ductal carcinoma 864 903 896 788 3451
Lobular carcinoma 156 170 163 137 626
Mucinous carcinoma 205 222 196 169 792
Papillary carcinoma 145 142 135 138 560
Total 1995 2081 2013 1820 7909
The BreakHis dataset mainly consists of two main classes: Benign and malignant,
from 24 patients have 2480 benign images, and from 58 patients have 5429 malignant
images [14, 15].
The benign and Malignant Histopathological images again divided into
four subclasses each. Benign classes-Adenosis(A), Fibroadenoma(F), Tubular
Adenoma(TA), Phyllodes Tumor(PT) and Malignant Classes-Ductal Carci-
noma(DC), Lobular Carcinoma(LC), Mucinous Carcinoma(MC), Papillary Carci-
noma(PC).The statistics of all data given in Table 1.
2.2 Methodology
The input image is from a database with 229 × 229 pixel resolution having
morphological features took, which is then augmented using Keras ImageDataGen-
erator function. Augmentation means creating more images adding some operation
like rotation, zooming, shift, and flip with slight modification without changing the
original structure.
Classifier: The classifier for classification we used the DCNN based xception
model is a low-cost, low computation model, as shown in Fig. 4. It is entirely based
on the Depth wise separable convolution layers. It has 36 convolution layers of
kernel size 3 × 3 to extract features which structured into 14 modules, all have a
linear residual connection between them except the first one [13].
Residual called identity/skip links solve vanishing gradient problems. A convolu-
tion layer used to extract the features, learn features from frames in CNN. The depth
wise Convolution shown in Fig. 3. Reduce computational cost shown in (Eqs. 2, 3)
which reduce the over fitting and computational cost.
Fig. 4 Xception: Deep Learning with Depth wise Separable Convolutions model
(arXiv:1610.02357v3)
N × D f 2 × Dk 2 × M (1)
(1/N + 1/Dk2) times lesser multiplications for the depth-wise Convolution than
the standard Convolution (Eq. 3).
The proposed model uses Adam optimizer, categorical cross-entropy and relu as
an activation function.
Adam Optimizer: Adam is an adaptive moment based algorithm used to update
the network weights during the training of data. Adam works. It is memory efficient,
computationally efficient and requires less hyperparameter tuning [17].
Loss: Categorical cross-entropy is log loss and used for multi-class classification
shown in (Eq. 4) [18].
Categorical cross − entropy = − [( y log ŷ) + (1 − y)log 1 − ŷ ] (4)
ReLu: Rectified Linear Unit is the activation function, gives the output if the input
is positive otherwise zero. Many networks use it as a default activation function.
3 Result Discussion
The experimental results of the proposed MCBHIX model have compared with
existing VGG16. MCBHIX has trained on GPU Tesla P100, for 100 epochs, in
Google Colab Notebook is a free cloud service and supports free GPU. MCCHIX
performance has better than VGG16 in terms of accuracy, Loss and computation
shown in Table 2. We achieved accuracy- 96.57%, which is 23.08% higher, Loss −
8.6%, which is 11.88% lesser than VGG16s accuracy and Loss has shown in Table
2.
Table 2 Accuracy (%), Loss (%) and many other parameters of MCBHIX and VGG16
Classification CNN Image Accuracy Loss CNN Computational
architecture size (%) (%) layers Parameters (million)
Multi-class VGG16 224 × 73.49 20.48 16 14.9
24x3
MCCBHIX 229 × 96.57 8.6 36 22.9
229 x 3
Binary VGG16 224 × 92.58 19.56 16 14.7
224x 3
MCCBHIX 229 × 99.01 1.68 36 22.5
229x3
Multi-Classification of Breast Histopathological Image Using Xception: Deep … 545
4 Conclusion
In this paper, we proposed the MCBHIX model to classify H&E stained breast
histopathological microscopy images into (i) Binary classification and (ii) Multi-
class Benign and Malignant classes into eight categories A, F, TA, PT, DC, LC, MC,
PC. This model has trained on the BreakHis dataset. We achieved an accuracy of
99.01% for binary class and 96.57% for Multi-Class classification.
References
1. Rangarajan B et al (2016) Breast cancer: an overview of published Indian data. South Asian J
Cancer 5(3):86
2. Reddy A, Soni B, S Reddy (2020) Breast cancer detection by leveraging machine learning.
ICT Express
3. Guo Y, Shang X, Li Z (2019) Identification of cancer subtypes by integrating multiple types
of transcriptomics data with deep learning in breast cancer. Neurocomputing 324:20–30
4. Khan AM, Sirinukunwattana K, Rajpoot N (2014) Geodesic geometric mean of regional covari-
ance descriptors as an image-level descriptor for nuclear atypia grading in breast histology
images. In: International workshop on machine learning in medical imaging. Springer, Berlin.
5. Janowczyk A, Chandran S, Madabhushi A (2013) Quantifying local heterogeneity via
morphologic scale: distinguishing tumoral from stromal regions. J pathol informa. 4(Suppl)
6. Das A, Nair MS, Peter DS (2020) Batch mode active learning on the riemannian manifold for
automated scoring of nuclear pleomorphism in breast cancer. Artif Intell Med 103:101805
7. Chekkoury A et al (2012) Automated malignancy detection in breast histopathological images.
In: Medical imaging: computer-aided diagnosis. 2012. International society for optics and
photonics
8. Gupta P, Garg S (2020) Breast cancer prediction using varying parameters of machine learning
models. Procedia Comput Sci 171:593–601
9. Wang C et al (2017) Histopathological image classification with bilinear convolutional neural
networks. In 2017 39th annual international conference of the ieee engineering in medicine
and biology society (EMBC). IEEE
10. Zainudin Z, Shamsuddin SM, Hasan S (2019) Deep layer CNN architecture for breast cancer
histopathology image detection. In: International conference on advanced machine learning
technologies and applications. Springer, Berlin
11. Toğaçar M et al (2020) BreastNet: a novel convolutional neural network model through
histopathological images for the diagnosis of breast cancer. Phys A 545:123592
12. Gandomkar Z, Brennan PC, Mello-Thoms C (2018) MuDeRN: multi-category classification
of breast histopathological image using deep residual networks. Artif Intell Med 88:14–24
13. Chollet F ( 2017) Xception: deep learning with depthwise separable convolutions. In:
Proceedings of the IEEE conference on computer vision and pattern recognition
14. Yosinski J et al (2014) How transferable are features in deep neural networks? In: Advances in
neural information processing systems
15. Benhammou Y et al (2020) BreakHis based breast cancer automatic diagnosis using deep
learning: taxonomy, survey and insights. Neurocomputing 375:9–24
546 S. D. Pujari et al.
16. Cireşan DC et al (2013) Mitosis detection in breast cancer histology images with deep neural
networks. In: International conference on medical image computing and computer-assisted
intervention. Springer, Berlin
17. Kingma DP, Ba J (2014) Adam: a method for stochastic optimization. arXiv preprint arXiv:
1412.6980
18. Ho Y, Wookey S (2019) The real-world-weight cross-entropy loss function: modeling the costs
of mislabeling. IEEE Access 8:4806–4813
Dense Haze Removal Using Convolution
Neural Network
1 Introduction
Visibility of the outdoor images decreases due to the presence of fog within the
atmosphere. Thus, within the presence of the haze or fog particles, a computer vision
algorithm faces difficulty to realize the specified output as generally they anticipate
an information picture without a quality debasement. In this manner, the presence
of the cloudiness or haze particles within the atmosphere degrades the performance
of computer vision algorithms like object detection [1], moving object segmentation
[2] etc. Therefore, to enhance the performance of vision algorithms within the hazy
environment, image de-hazing may be a required pre-processing task. Research in
the field of image de-hazing is roughly divided into learning based methods [3–5]
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 547
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_55
548 M. Dongare and J. Kendule
Fig. 1 Haze-free image recovered by the proposed method. Left column: Input hazy image Right
column: Haze-free image recovered using the proposed method
and prior based methods [6–13]. Among these, prior based methods believe the haze
relevant prior and extract haze relevant features.
Figure 1 shows the outdoor hazy image and the haze-free image recovered by the
proposed net- work. The proposed network is built using basic principles of inception
and dense modules. Thus, named as inception-dense network (ID-Net). The proposed
ID-Net bypasses the estimation of intermediate feature maps and directly recovers
the haze-free image.
The key features of this work are listed below:
1. End-to-give up conditional generative adverse community named as ID-Net is
proposed for picture de-hazing.
2. A novel generator community is proposed that’s de- signed the usage of a
aggregate of each inception and dense module.
3. Experimental evaluation has been done on current benchmark datasets.
2 Literature Survey
Haze impact is without delay proportional to the intensity of an item from the
digital digicam device. To apprehend this non- linearity, numerous methods had
been proposed consisting of polarized filters [14, 15], use of more than one pix of
equal scenery [16, 17], earlier primarily based totally hazy models [6–13] etc. At the
beginning, with inside the vicinity of photograph de-hazing, Schechner et al. [14,
15] proposed the polarized filters. Their method works with more than one pix of the
equal scene however differs in polarization angle. This method fails because of its
multi-photograph dependency. Nayer et al. [17] overcame the hardware complexity.
By utilizing 3D geometrical model Cozman et al. [16] resolved multi-image depen-
dency this technique is based upon the depth information of the hazy images. In the
last decade because of the convincing assumptions regarding the haze spread or haze
Dense Haze Removal Using Convolution Neural Network 549
density, image de-hazing has made remarkable progress. Tan et al. [11] Proposed
assessment enhancement of the hazy scene. By maximizing the nearby assessment
of the hazy photo they eliminated haze. However, whilst there may be a intensity
discontinuity within side the hazy photo this approach fails and create blocking off
artifacts. He et al. [8, 18] proposed darkish channel prior (DChP) to repair the visi-
bility within side the hazy scene. It accommodates of darkish pixels i.e. pixels which
can be having very low depth amongst one of the colour channels for a given hazy-
unfastened scene. DChP fails in complex edgy systems and additionally undergoes
the halo effect [3]. To estimate the sturdy trans- undertaking map of the hazy scene,
researchers observe post- processing strategies which include guided filtering [19],
median filtering [9, 20] etc. Lai et al. [21] proposed priors to estimate the most
excellent transmission map. Wang et al. [22] applied multi- scale retinex set of rules
to estimate the brightness components. Further, with the help of a bodily model,
they recovered the haze-unfastened photo. Zhu et al. [13] proposed a color atten-
uation prior (CAP) which having a HSV color space to extract the haze-relelated
features. To avail the benefits of multiple haze priors, Tang et al. [12] proposed
regression framework for image de- hazing. They have proposed the extraction of
varied haze relevant features using existing haze relevant priors and learned the
integrated features to estimate the robust scene transmission map. However, it will
increase the mistakes upstretched to the hired priors. Thus, to reduce the cascading
error, researchers employ convolutional neural networks (CNN). Existing learning-
primarily based totally approaches [3–5, 23, 24] estimate the scene transmission
map the usage of CNN. Further, a international airlight estimation accompanied via
way of means of atmospheric scattering version restores the haze-loose scene. All
above strategies offers the identical perception that so one can get better a haze-loose
photo, estimation of a correct scene transmission map is essential. The atmospheric
mild is calculated one by one and the easy photo is recovered the usage of the atmo-
spheric scattering version. Still any such manner does now no longer at once degree
or reduces the reconstruction distortions. As a end result it’ll supply upward thrust
to the sub-ideal photo recovery quality. The mistakes in every separate estimation
step will amplify the general error. In this aspect, Li et al. [25] proposed an end-to-
end architecture called as AOD-Net. They analyzed the internal relationship between
traditional atmospheric model and the end-to-end de-hazing network. Further, Swami
et al. [26] discussed an end-to-end network based on conditional GAN for image clar-
ification. Recently, researchers [27–29] make use of unpaired training approach for
haze removal [27, 28, 30] utilized unpaired training approach for image de-hazing
whereas [29] found its use for moving object segmentation. In the next Section, we
have discussed the proposed method for single image haze removal.
550 M. Dongare and J. Kendule
The proposed generator network architecture is divided into three parts namely:
(1) Encoder block (2) Inception block, and (3) Decoder block. Encoder/Decoder
block consists of simple convolution/ deconvolution layer followed by non-linear
activation function (ReLU). We use instance normalization [31] to normalize the
network feature maps. Figure 2 show the encoder blocks. We have designed four
encoder blocks among which initial two blocks down-samples the input feature maps.
Purpose of the down-sampling operation is to increase the receptive field in the
network. Another advantage of the down-sampling the feature maps are to reduce
the computations. It is well known fact that the number of computations in CNN
are directly proportional to the spatial size of the feature maps those which are
processed through the network. Down-sampling operation is purposefully designed
so as to reduce the number of computations and to increase the receptive field. We
design four encoder and two decoder blocks with filters having a spatial size of 3 × 3.
Table 1 Quantitative
SOTS SSIM PSNR CIEDE2000
analysis of image De-hazing
on SOTS Database [37] TPAMI-11 [8] 0.8179 16.6215 9.9419
TIP-14 [13] 0.8364 19.0524 8.3102
TIP-16 [3] 0.8472 21.1412 6.2645
ECCV-16 [5] 0.8102 17.5731 10.7991
ICCV-17 [25] 0.8512 19.0868 8.2716
NIPS-18 [35] 0.8299 18.3767 11.5527
WACV-18 [4] 0.8152 20.1186 8.3060
CVPR-18 [40] 0.8800 22.3113 –
CVPR-18 [41] 0.8378 20.8100 –
ECCV-18 [42] 0.8640 20.3125 7.1152
WACV-19 [27] 0.8919 22.9207 8.9262
TIP-19 [43] 0.8261 21.3400 –
IJCV-19 [44] 0.8600 21.5600 –
TIP-19 [45, 46] 0.8716 21.4375 –
CVPRW-19 [46] 0.8500 19.8280 8.2993
CVPRW-19 [47] 0.9221 24.0200 –
Proposed method 0.9383 27.5070 5.2428
D-Hazy [23] is a standard dataset used to evaluate the performance of various algo-
rithms for image dehazing. It comprises of pair of 1449 indoor hazy and respective
haze-free scenes. We utilized the entire database i.e. 1449 images for quantitative
analysis of the proposed network for image de-hazing. The performance of the pro-
posed network is compared with the existing state-of-the- art methods on D-Hazy
[23] database as shown in Table 1.
4 Experimental Results
scene. Additionally, there is a huge improvement in the PSNR and CIEDE2000 of the
proposed RI-GAN as contrasted and the current cutting edge strategies. Proposed
network builds SSIM by practically 9%. SOTS information base [37] is created
from set of 50 pictures and their individual profundity maps from NYU-profundity
information base [38, 39]. From each fog free picture and its profundity map, 10
foggy pictures are created with various estimations of β and airlight utilizing climatic
dispersing model. Hence, despite the fact that there is a cover of some scene in D-
Hazy and SOTS information base, distinctive airlight and thickness of murkiness
have a huge effect between them. In this manner, we assessed the exhibition of
proposed RI-GAN on SOTS information base. We thought about each of the 500
murky pictures for the examination. Table 1 depicts the result of the proposed and
existing methods on SOTS database. We can observe that proposed Dense- Inception
network outperforms the other existing state-of- the-art methods in terms of PSNR
and appears very close to AODNet [25].
Because of the inaccessibility of pair of this present reality murky and murkiness
free scenes, it is hard to convey quantitative examination of picture de-preliminaries
calculations for true cloudy scenes. In this way, we convey just subjective examina-
tion for this present reality dim scenes. Five as often as possible utilized true foggy
scenes are used here for investigation. Result correlation of proposed and existing
methodologies on these pictures is appeared in Figs. 3 and 4. We can plainly see
that the proposed RI-GAN produces the proper scene data simultaneously protects
the primary subtleties in recuperated dimness free scene. We think about the conse-
quences of existing earlier based hand-created and learning approaches [3, 4, 8] and
start to finish dehazing approach [25]. Qualitative examination shows that proposed
Fig. 3 Visual after effect of proposed and existing strategies on genuine foggy pictures. a Input
murky picture. After effects of b [8], c [5], d [4], e [25], f [40], g [42], h [48]
Dense Haze Removal Using Convolution Neural Network 553
Fig. 4 Visual consequence of proposed and existing strategies on true murky pictures. a Input
foggy picture. Consequences of b [8], c [5], d [4], e [25], f [40], g [42], h [48]
Network beats the other existing methodologies and produces an outwardly charming
dimness free scene.
5 Conclusion
References
1. Dudhane A, Murala S (2019) Cardinal color fusion network for single image haze removal.
Mach Vis Appl 30(2):231–242
2. Patil PW, Murala S (2018) Msfgnet: a novel compact end-to- end deep network for moving
object detection. IEEE Trans actions Intell Transp Syst
3. Cai B, Xu X, Jia K, Qing C, Tao D (2016) Dehazenet: an end-to-end system for single image
haze removal. IEEE Trans Image Proces 25(11):5187–5198
4. Dudhane A, Murala S (2018) IEEE, 2018, Cˆ 2msnet: A novel approach for single image haze
removal. In: 2018 IEEE winter conference on applications of computer vision (WACV), pp
1397–1404
5. Ren W, Liu S, Zhang H, Pan J, Cao X, Yang M-H (2016) Single image dehazing via multi-scale
convolutional neural networks. In: European conference on computer vision. Springer, Berlin,
pp 154–169
6. Fattal R (2008) Single image dehazing. ACM Trans Graphics (TOG), 27(3):72
7. Gibson KB, Vo DT, Nguyen TQ (2012) An investigation of dehazing effects on image and
video coding. IEEE Trans Image Proces 21(2):662–673
8. He K, Sun J, Tang X (2011) Single image haze removal using dark channel prior. IEEE Trans
Pattern Anal Mach Intell 33(12):2341–2353
9. Huang S-C, Chen B-H, Wang W-J (2014) Visibility restoration of single hazy images captured
in real-world weather conditions. IEEE Trans Circuits Syst Video Technol 24(10):1814–1824
10. Ancuti CO, Ancuti C, Hermans C, Bekaert P (2010) A fast semi-inverse approach to detect
and remove the haze from a single image. In: Asian conference on computer vision. Springer,
Berlin, pp 501–514
11. Tan RT (2008) Visibility in bad weather from a single image. In: IEEE conference on computer
vision and pattern recognition, 2008. CVPR 2008. IEEE pp 1–8
12. Tang K, Yang J, Wang J (2014) Investigating haze-relevant features in a learning framework
for image dehazing. In: Proceedings of the IEEE conference on computer vision and pattern
recognition, pp 2995–3000
13. Zhu Q, Mai J, Shao L (2014) Single image dehazing using color attenuation prior. In: 25th
British machine vision conference, BMVC
14. Schechner YY, Narasimhan SG, Nayar SK (2001) Instant dehazing of images using polariza-
tion. In: Proceedings of the 2001 IEEE computer society conference on computer vision and
pattern recognition, CVPR 2001. vol 1, pp I–I
15. Shwartz S, Namer E, Schechner YY (2006) Blind haze separation. In: 2006 IEEE computer
society conference on computer vision and pattern recognition (CVPR’06), vol 2, pp 1984–1991
16. Cozman F, Krotkov E (1997) Depth from scattering, Jun 1997. In: Proceedings of IEEE
computer society conference on computer vision and pattern recognition, pp 801–806
17. Nayar SK, Narasimhan SG (1999) Vision in bad weather. In: The proceedings of the Seventh
IEEE international conference on computer vision, vol 2. IEEE, pp 820–827
18. He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In:
Proceedings of the IEEE conference on computer vision and pattern recognition, pp 770–778
19. He K, Sun J, Tang X (2010) Guided image filtering. In: European conference on computer
vision, Springer, Berlin, pp 1–14
20. Yu J, Xiao C, Li D (2010) Physics-based fast single image fog removal. In: 2010 IEEE 10th
international conference on signal processing (ICSP), IEEE, pp 1048–1052
21. Lai Y, Chen Y, Chiou C, Hsu C (2015) Single-image dehazing via optimal transmission map
under scene priors. IEEE Trans Circuits Syst Video Technol 25(1):1–14
22. Wang J, Lu K, Xue J, He N, Shao L (2018) Single image dehazing based on the physical model
and msrcr algorithm. IEEE Trans Circ Syst Video Technol, pp 1–1
23. Ancuti C, Ancuti CO, De Vleeschouwer C (2016) IEEE 2016, D-hazy: A dataset to eval-
uate quantitatively dehazing algorithms. In: 2016 IEEE international conference on image
processing (ICIP), pp 2226–2230
Dense Haze Removal Using Convolution Neural Network 555
24. He K, Sun J (2015) Convolutional neural networks at con- strained time cost. In: Proceedings
of the IEEE conference on computer vision and pattern recognition, pp 5353–5360
25. Li B, Peng X, Wang Z, Xu J, Feng D (2017) Aod-net: All- in-one dehazing network. In:
Proceedings of the IEEE international conference on computer vision. pp 4770–4778
26. Swami K, Das SK (2018) Candy: conditional adversarial networks based end-to-end system
for single image haze removal. In: 2018 24th international conference on pattern recognition
(ICPR). IEEE, pp 3061–3067
27. Dudhane A, Murala S (2019) IEEE, 2019, Cdnet: Single image de-hazing using unpaired
adversarial training. In: 2019 IEEE Winter Conference on Applications of Computer Vision
(WACV), pp 1147–1155
28. Engin D, Genc A, Kemal Ekenel H (2018) Cycle-dehaze: Enhanced cyclegan for single image
dehazing. In: Proceedings of the IEEE conference on computer vision and pattern recognition
workshops, pp 825–833
29. Patil P, Murala S (2019) Fggan: A cascaded unpaired learning for background estimation and
foreground segmentation. In: 2019 IEEE winter conference on applications of computer vision
(WACV), IEEE, pp 1770–1778
30. Szegedy C, Liu W, Jia Y, Sermanet P, Reed S, Anguelov D, Erhan D, Vanhoucke V, Rabinovich
A (2015) Going deeper with convolutions. In: Proceedings of the IEEE conference on computer
vision and pattern recognition, pp 1–9
31. Ulyanov D, Vedaldi A, Lempitsky V (2016) Instance normalization: the missing ingredient for
fast stylization. arXiv preprint arXiv:1607.08022
32. Isola P, Zhu J-Y, Zhou T, Efros AA (2017) Image-to-image translation with conditional adver-
sarial networks. In: 2017 IEEE conference on computer vision and pattern recognition (CVPR).
IEEE, pp 5967–5976
33. Wang Z, Bovik AC, Sheikh HR, Simoncelli EP (2004) Image quality assessment: from error
visibility to struc tural similarity. IEEE Trans Image Proces 13(4):600–612
34. Sharma G, Wu W, Dalal EN (2005) The ciede2000 color- difference formula: Implementation
notes, supplementary test data, and mathematical observations. In: Color research & appli-
cation: endorsed by inter-society color council, the colour group (Great Britain), Canadian
society for color, color science association of Japan, Dutch society for the study of color, The
Swedish colour centre foundation, colour society of Australia, Centre Franc¸ais de la Couleur
30(1):21–30
35. Yang X, Xu Z, Luo J (2018) Towards perceptual image dehazing by physics-based disentan-
glement and adversarial training. In: In Thirty third-second AAAI conference on Artificial
Intelligence (AAAI-18)
36. Zhu J-Y, Park T, Isola P, Efros AA (2017) Unpaired image- to-image translation using cycle-
consistent adversarial net- works. In: 2017 IEEE international conference on computer vision
(ICCV). IEEE, pp 2242–2251
37. Li B, Ren W, Fu D, Tao D, Feng D, Zeng W, Wang Z (2019) Benchmarking single-image
dehazing and be- yond. IEEE Trans Image Proces 28(1):492–505
38. Silberman N, Hoiem D, Kohli P, Fergus R (2012) Indoor segmentation and support inference
from rgbd images. In: European conference on computer vision. Springer, pp 746–760
39. Srivastava RK, Greff K, Schmidhuber J (2015) Highway networks. arXiv preprint arXiv:1505.
00387
40. Ren W, Ma L, Zhang J, Pan J, Cao X, Liu W, Yang M-H (2018) Gated fusion network for
single image dehazing. In: The IEEE conference on computer vision and pattern recognition
(CVPR)
41. Zhang H, Patel VM (2018) Densely connected pyramid de-hazing network. In: Proceedings of
the IEEE conference on computer vision and pattern recognition, pp 3194–3203
42. Yang D, Sun J (2018) Proximal dehaze-net: a prior learning- based deep network for single
image dehazing. In: Proceedings of the European conference on computer vision (ECCV), pp
702–717
43. Hu H-M, Guo Q, Zheng J, Wang H, Li B (2019) Single image defogging based on illumination
decomposition for visual maritime surveillance. IEEE Trans Image Process
556 M. Dongare and J. Kendule
44. Ren W, Pan J, Zhang H, Cao X, Yang M-H (2019) Single image dehazing via multi-scale
convolutional neural net- works with holistic edges. Int J Comput Vision, pp 1–20
45. Dudhane A, Murala S (2019) Ryf-net: Deep fusion network for single image haze removal.
IEEE Trans Image Proces 29:628–640
46. Dudhane A, Singh Aulakh H, Murala S (2019) Ri-gan: An end-to-end network for single
image haze removal. In: Proceedings of the IEEE conference on computer vision and pattern
recognition workshops, pp 0–0
47. Chen S, Chen Y, Qu Y, Huang J, Hong M (2019) Multi-scale adaptive dehazing network. In:
Proceedings of the IEEE conference on computer vision and pattern recognition work- shops,
pp 0–0
48. Guo T, Li X, Cherukuri V, Monga V (2019) Dense scene information estimation network for
dehazing. In: Proceedings of the IEEE Conference on computer vision and pattern recognition
workshops, pp 0–0
ICT Based Societal Technologies
Diabetic Retinopathy Detection
with Optimal Feature Selection:
An Algorithmic Analysis
Abstract This work aims to establish a new automated Diabetic Retinopathy (DR)
recognition scheme, which involves phases such as “Preprocessing, Blood Vessel
Segmentation, Feature Extraction, and Classification”. Initially, Contrast Limited
Adaptive Histogram Equalization (CLAHE) and median filter aids in pre-processing
the image. For blood vessels segmentation, Fuzzy C Mean (FCM) thresholding is
deployed that offers improved threshold values. As the next process, feature extrac-
tion is performed, where local, morphological transformation oriented features and
Gray-Level Run-Length Matrix (GLRM) is based on extracted features. Further, the
optimal features are selected using a new FireFly Migration Operator-based Monarch
Butterfly Optimization (FM-MBO) model. Finally, Convolutional Neural Network
(CNN) is deployed for classification purposes. Moreover, to attain better accuracy,
the count of convolutional neurons of CNN is optimally elected using the proposed
FM-MBO algorithm.
1 Introduction
S. S. Basha
Y.S.R. Engineering College of Yogi Vemana University, Proddatur 516360, India
S. J. Badashah (B)
Sreenidhi Institute of Science and Technology (Autonomous), Hyderabad, Telangana State
501301, India
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 559
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_56
560 S. S. Basha and S. J. Badashah
that offers automated diagnosis in a precise way. It diagnoses DR by mining the optic
disc by concerning the above-said issues in the existing systems. Also, it aids the
experts in taking appropriate decisions, thus leading to quicker and more accurate
diagnostic decisions with high reliability. The major contribution is to introduce a DR
detection system with optimization assisted CNN that intakes the optimally selected
features as input. Moreover, an algorithmic analysis that proves the betterment of
proposed work with respect to varied parameters. The paper is arranged as follows.
The reviews are presented in Sect. 2. Section 3 describes the proposed DR detec-
tion: pre-processing, segmentation and feature extraction. Section 4 explains the
optimal feature selection and optimization-based classification. Section 5 discusses
the outcomes and the paper is concluded by Sect. 6.
2 Literature Review
In 2020, Kumar et al. [6] have established a better method for detecting hemorrhages
and MA that contributed on the overall enhancement in the earlier identification of
DR. In 2020, Zago et al. [7] have sestablished a novel methodology for recognizing
DR by means of a deep patch-based model. In 2019, Liu et al. [8] have implemented a
novel technique called WP-CNN that made use of numerous weighted paths of CNN.
In 2018, Wan et al. [9] have implemented an automated technique for classifying a
specified set of fundus images.
Pxi σ
β= 1+ Mss (1)
drr 100
Equation (1) portrays the clip point estimation, in which β represents the clip
point, Pxi indicates pixel count in every section,drr represents the dynamic range in
particular block, Mss denotes maximum slope, σ refers to clip factor. Thus, Bclahe is
achieved from the pre-processing stage by means of CLAHE.
Diabetic Retinopathy Detection with Optimal Feature Selection … 561
Median Filter [11]: The Bclahe processing in median filtering is given by Eq. (2),
in which, gr(xe, ye) and Bclahe (xe, ye) indicates the output image and original image
in order and Hm denotes the 2-D mask. Thus, Bmedian is achieved from the median
filter.
Now, the image Bmedian is subjected under blood vessel segmentation. Usually, the
FCM [12] model partitions Me components into Xe classes. The entire pixels in
an image indicated as Me defined as Me = Mea × Meb and Xe = 3 that denotes
3-class FCM clustering. Here the optimized objective function is given by Eq. (3),
in which hele ∈ He, He = {he1 , he2 , he3 , . . . , heXe } indicates the cluster centre and
wene ∈ W e, W e = {we1 , we2 , we3 , . . . , weMe } specifies the measured data. The
criterion in ∗ indicates similarityamongstthe cluster center and measured data
ge ∈ [1, ∞] denotes a real integer.
Xe
Me
Ob = (eln )ge wene − hele 2 (3)
le=1 ne=1
The segmented blood vessel Bfcm is then exposed to the feature extraction process,
anywhere local, morphological transformation and GLRM √ features are
extracted.
√
Local Features [13]: Here, the 4 scales denoted by 2, 2, 2 2, 4 are exploited
for image filtering. In addition, a 2-D Gaussian filter is deployed for highlighting the
vessel area. The 2-D Gaussian kernel is formulated by Eq. (4).
1 xe2 − ye2
Geke (xe, ye) = exp − (4)
2π σ 2 2σ 2
∈ {3, 7, 11, 15, 19, 23} and θ represents angular revolution θ ∈ MeB , MeB =
ie
xe0 ≤ xe ≤ π, xe mod π 12 = 0 .
Seie = Seieθ (6)
θ∈MeB
Retinal
Fundus image
Pre-Processing
Blood Vessel
Segmentation
CLAHE
FCM Thresholding
Median Filtering
Local features
Convolution
al Neurons Optimal Morphological
feature transformation
selection based features
GLRM features
CNN Proposed
FM-MBO
Affected Healthy
∗
From the extracted features, the precise or optimal features Fefinal are selected via
the optimization algorithm, and they are subjected tothe classification process. This
precise selection enhances the detection accuracy, thereby increases the system
performance.
∗
The optimal feature set Fefinal is then subjected to classification using CNN frame-
work. Usually, CNN [16] includes input and output layer and convolution layer i.e.
multiple hidden layers. Multiple layers are exploited to accomplish the overall feature
map as given by Eq. (7).
le
Ca,b,c = wicleTi ua,b
le
+ bicle (7)
le
In Eq. (17), Ca,b,c refers to convolutional feature, wicleTi specifies weight vector,
le
and bic indicates bias term of cth filter in le layer, ua,b
le
specifies the input patch at
the position (a, b) in le layer. The nonlinearities in CNN are determined using an
th
le
activation function which is given in Eq. (8), where Bia,b,c indicates activation value.
le
Bia,b,c = Bi Ca,b,c
le
(8)
The resolution of feature maps is reduced by the pooling layer denoted by pool(·)
and it is defined in Eq. (9).
le
Pia,b,c = pool Bii.j,c
le
, ∀(i, j) ∈ Ria,b (9)
1
Nn
Li = li θ ; vi(ne) , tl (ne) (10)
Nn ne=1
In the above equation, Ria,b denotes local neighborhood at the position (a, b) and
each variable in CNN is denoted by θ . The optimal parameters are portrayed by
minimizing the specific loss function as revealed in Eq. (10), where, Li is the loss of
CNN, Nn is the desired input–output relations ui(ne) , vi(ne) ; ne ∈ [1, 2, . . . , Nn] ,
ui(ne) is the neth input feature, vi(ne) specifies target label, and tl (ne) indicates output.
Two convolutional layers are exploited in CNN, where O1 and O2 denotes the count
of hidden neurons that needsto be optimized using the proposed FM-MBO algorithm.
The minimum and maximum bounding limits of O1 and O2 are 1 and 50.
564 S. S. Basha and S. J. Badashah
OB = max(C) (11)
The solution encoding of this work is illustrated in Fig. 2. This includes a selection
∗
of optimal features Fefinal from the extracted features Fefinal . Among the 244 extracted
features, only 128 feature subsets are optimally selected. Further, the count of hidden
neurons O1 and O2 are optimally tuned using the algorithm.
Even though various optimization issues can be resolved by the existing MBO,
it suffers from poor performance and pre-mature convergence [17]. Therefore, to
improve its efficiency, FireFly (FF) algorithm is incorporated with the conventional
MBO model. In general, monarch butterflies are found at certain places that are
considered as land1 and land2. The migration operator is deployed for producing
new ones and after giving birth to new ones, the parent butterfly leaves the pack.
Migration Operator: Consider MPe as the total population, ce represents the
proportion of butterflies in land1 (subpopn1) and land2 (subpopn2) correspondingly.
me+1
The migration operation is specified by Eq. (12), in which, Jie,ye refers to yeth
componentof Jie and Jse1 represents the novel location of the monarch butterfly se1 .
me+1
Jie,ye = Jse1,ye
me
(12)
The component ye for the new monarch butterfly is instigated depending on the FF
updateas given in Eq. (14) [18], where, initial parameter signifies the present position
Diabetic Retinopathy Detection with Optimal Feature Selection … 565
of Jbest that denotes the best monarch butterfly in land1 and land2.
me+1
Jxe,ye = Jbest,ye
me
(15)
In addition, when ce > rnm, the memory gets updated as specified in Eq. (16), in
which, Jseme3 ,ye refers to yeth component of Jse3 that is chosenfrom land2 in a random
manner. At this point, se3 ∈ {1, 2, . . . , MPe2 }
me+1
Jxe,ye = Jseme3 ,ye (16)
me+1
Jxe,ye = Jxe,ye
me+1
+ λ × fnsye − 0.5 (17)
Equation (17) reveals the memory update depending upon the constraint,rnm >
arbmv, here, arbmv refers to balancing the value of butterfly, and fns denotes monarch
butterfly xe walk steps. It is evaluated by deploying control flight as indicated in
Eqs. (18) and (19).
The weighting element λ is computed as per Eq. (19), where Hedim denotes the
value of a single move.
566 S. S. Basha and S. J. Badashah
The performance analysis of the applied DR detection model is exposed in Figs. 3 and
4 for optimistic measures and undesirable measures respectively. From the study, the
FM-MBO system has achieved optimal performance for all the trials when matched
to the other recent schemes. Precisely, from Fig. 3a, the values of accuracy attained
by the presented approach when α = 0.2 and α = 0.5 is 3.53% and 2.35% superior
than the values achieved when α = 0.1 and α = 0.3. On considering precision, the
value of α = 0.3 is found to be higher, which is 46.77%, 3.23%, 20.97%, and 6.45%
better than the values of α at 0.1, 0.2, 0.4 and 0.5 respectively. Thus, the improvement
of the presented methodology has been confirmed via algorithmic analysis.
Fig. 3 Algorithmic analysis of the proposed model by varying the parameters α and γ of FF with
respect to measures like a Accuracy b Sensitivity c Specificity d Precision e NPV f F1-score g MCC
Diabetic Retinopathy Detection with Optimal Feature Selection … 567
Fig. 4 Algorithmic analysis of the proposed model by varying the parameters α and γ of FF with
respect to measures like a FDR b FPR c FNR
6 Conclusion
This paper introduced a advanced DR diagnostic model that includes phases like
“preprocessing, blood vessel segmentation, feature extraction, and classification”.
Initially, pre-processing is done for which CLAHE and median filter were deployed.
Next to pre-processing, blood vessel segmentation was performed using FCM thresh-
olding model. Subsequently, certain features were extracted from the segmented
images. In addition, optimal feature selection was carried out by means of suggested
FM-MBO algorithm, and more over, classification was done by optimizing the convo-
lutional neurons in CNN using the same algorithm.Lastly, the algorithmic study was
held for verifying the superiority of the adopted model. Particularly, on observing the
analysis on precision, the value of α = 0.3 is found to be higher, which is 46.77%,
3.23%, 20.97% and 6.45% better than the values of α at 0.1, 0.2, 0.4 and 0.5 respec-
tively. Thus, the enhancement of the adopted FM-MBO algorithm was validated
effectively.
References
1. Wan S, Liang Y, Zhang Y (2018) Deep convolutional neural networks for diabetic retinopathy
detection by image classification. Comput Electr Eng 72:274–282
2. Leontidis G (2017) A new unified framework for the early detection of the progression to
diabeticretinopathy from fundus images. Comput Biol Med 90:98–115
3. Arias MEG, Marin D, Ponte B, Alvarez F, Bravo JM (2017) A tool for automated diabetic
retinopathy pre-screening based on retinal image computer analysis. Comput Biol Med 88:100–
109
4. Romero RR, Carballido JM, Capistrán JH, Valencia LJU (2015) A method to assist in the diag-
nosis of early diabetic retinopathy: Image processing applied to detection of microaneurysms
in fundus images. Comput Med Imaging Graph 44:41–53
5. Gupta G, Kulasekaran S, Ram K, Joshi N, Gandhi R (2017) Local characterization of neovas-
cularization and identification of proliferative diabetic retinopathy in retinal fundus images”.
Comput Med Imaging Graph 55:124–132
6. Shanthi T, Sabeenian RS (2019) Modified Alexnet architecture for classification of diabetic
retinopathy images. Comput Electr Eng 76:56–64
568 S. S. Basha and S. J. Badashah
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 569
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_57
570 M. Rahamat and B. Lavanya
1 Introduction
2 Review of Literature
3 Research Methodology
The current study has been carried out with the following objectives:
1. To examine the students perception towards Virtual Learning.
2. To explore the satisfaction levels of contemporary students over the use of
Virtual Learning system.
3. To understand the challenges faced by the students, propose a model for
enhancing virtual class room learning effective.
This study is an empirical and exploratory in nature. Primary data has been
collected with the help of a self—designed structured questionnaire. Secondary data
is collected from various journals, articles, websites and online resources. Conve-
nient sampling method was used to collect data. The Virtual Learning variables
(Perception, Satisfaction, Drawbacks) were measured using 5-point Likert scale.
The data collected were subjected to reliability analysis, descriptive Statistics, T-Test
and ANOVA and Correlation has been applied using SPSS Version 26.
Table 4 Descriptive Statistics for the statement about the drawbacks of virtual learning
Statements N Mean Std. deviation
I think I get socially isolated while using virtual learning 142 3.47 1.22
Virtual class is difficult to handle and therefore frustrated to use 142 3.39 1.22
Virtual learning can damage our health as we use laptop for long 142 3.65 1.31
time
Internet connectivity is the major problem I face using virtual 142 3.83 1.24
learning
I do face technical problem when I use virtual learning to learn 142 3.79 1.17
Do you think virtual class provide the right amount of theoretical 142 2.57 1.16
and practical experience
I prefer to learn from books rather than from the virtual 142 3.59 1.20
learning/course learning
I do not have computer and therefore I find it difficult to learn 142 3.04 1.29
through virtually
574 M. Rahamat and B. Lavanya
From Table 4, it is seen that most of the students neutrally agree virtual learning
is a drawback because they feel isolated, due to the technical problems like internet
connectivity, sitting foe long hours in front of laptop make them difficult to handle
and feel that virtual leaning is a disadvantage as it provides low quality experience
than the practical one.
From Table 5, it can be said that students agree to all statements and strongly
believe that drawing attention to core content, attracting and gaining the students
attention and taking continuous feedback makes virtual learning effective.
Table 6 shows the correlation between the perception and satisfaction of students
towards virtual leaning. It can be seen that Pearson Correlation value 0.827 which
indicates that there exist significant and positive relationship between students
perception and satisfaction with regard to virtual learning. This stipulates that the
student’s are satisfied with the way they perceive in connection with virtual learning.
From Table 7, assuming equal variance the significance (2 tailed) value is greater
than 0.05. Hence we reject H0 and accept H1. Therefore there is a significance
difference male and female PG student towards satisfaction of virtual learning.
Table 5 Descriptive statistics of strategies used for making virtual learning effective
Statement N Mean Std. deviation
Developing a detailed outline and research core content 142 4.21 1.04
beforehand
Focus on friendly content delivery 142 4.38 1.14
Ask engaging and thought-provoking questions that will drive the 142 4.28 1.13
discussion
Draw attention to important content and offer periodic recaps 142 4.36 1.21
Keep it short and simple to avoid information overload 142 4.35 1.16
Monitor the students progress about the understanding of subjects 142 4.27 1.13
through virtual learning
Ask for continuous feedback to improve virtual learning 142 4.37 1.16
variances
assumed
Equal 0.064 139.188 0.949 0.00992 0.15533 −0.29719 0.31703
variances not
assumed
H0: There is a no significance difference in satisfaction between male and female students towards virtual learning
H1: There is significance difference in satisfaction between male and female students towards virtual learning
575
576 M. Rahamat and B. Lavanya
From Table 8, the significance value (0.036) is less than 0.05. Hence we accept
H0 and reject H1. Therefore there is no significance difference in perception between
different age groups towards virtual learning.
6 Findings of Study
From the study it was found that most of the students have positive perception and are
not completely satisfied with virtual learning as it is new way of learning due to this
pandemic situation, from the analysis it was also found that students neutrally agree
that virtual learning makes them isolated from their friends and technical issue is
the major disadvantage they face while using virtual learning. The analysis disclose
that there exist positive correlation between perception and satisfaction of students
towards virtual learning From the study it was also found that there is difference
in the way male and female students perceive virtual learning but irrespective of
different age groups there is no difference in perception of students with regard to
virtual learning.
Based on the strategies and suggestions of the students a model is proposed for
making virtual learning effective (Fig. 1).
Institution:
• Institution should frame flexible rules and regulations for virtual classes.
• Institution should either develop or choose good technology platform for
education.
• Should develop user friendly to LMS (easily understandable).
Students Perception and Satisfaction Towards ICT … 577
•INSTITUTION •FACULTY
Upgrading
Flexible Rules skills that
and virtual classes
regulations demand
Expediting
Active
the ward’s
Participation
learning
•STUDENT •PARENT
• Institution should provide access to all the forms of resources to students for
learning.
Faculty:
• Faculty should be trained to conduct virtual learning.
• Special/extra efforts need to be put by the faculty for creating a real class room
experience and for problematic/Quantitative subjects.
• Faculty should provide clear and concise instructions and exercises to follow.
• Faculty should monitor discussions to clarify students’ postings, highlight good or
interesting comments and ideas, provide insight, and ensure every voice is heard.
• Faculty should provide the necessary components of successful interaction:
Lesson and Session plans, Learning material, animated PPTs, Content related
videos, demonstration, practice sessions, feedback, and assessment.
Students:
• Students should be active learner as well as active listener.
• Begin with a flexible schedule.
• They should be participative in all activities.
• Should be responsible in completing their task.
• Students should be motivated and committed to learn and work on their own.
• Students should give constructive feedback for effective improvement in learning.
Teachers should consider the feedback given by the students and take the corrective
actions for making the virtual class in an effective manner. In the same way faculty
should also suggest the institution in using new tools to make the students familiar
to work with virtual learning.
578 M. Rahamat and B. Lavanya
Parents:
• Should monitor their wards participation in the virtual classes.
• Should stay in touch with the faculty and to know the progress.
• Should support their ward and help them to stay physically, mentally and
emotionally strong.
8 Conclusion
Shutdown of the institutions due to covid-19 pandemic has resulted in the greater
use of ICT. Online education is not so easy as speaking into the microphone at one
end, and connecting a laptop and listening on the other; there are challenges faced
at both ends of the spectrum. It is true that ICT is playing a vital role in teaching
and learning but at same time there are many challenges that are to be addressed
for smooth functioning of various services provided through ICT. So, in the current
research an attempt has been made to study the perception and satisfaction levels of
students towards virtual learning and found that there exist a positive relationship
between the two perception and satisfaction. During this pandemic situation, it is very
important that students have positive perception with respect to institution, faculty,
and parents. It is the responsibility of teacher to engage the students during the class
to make it interesting as well as student themselves to have a cooperative attitude to
learn from virtual platform. Positive the perception higher the satisfaction level of
students towards virtual learning. Strengthening teacher-student interaction, utilize
student’s feedback, making more creative and useful learning makes virtual learning
effective. It can be concluded that, during this hard and tough time ICT stood as a
boon to students all around the world.
References
Apeksha M. Gopale
1 Introduction
A. M. Gopale (B)
SNDT Women’s University, Mumbai, India
AIKTC, Mumbai University, Mumbai, India
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 579
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_58
580 A. M. Gopale
small spaces thus increases chances of Covid-19 spread. During pandemic the main-
tenance of personal vehicle became need of an hour due to increased usage.This is
the experience of countless salon-goers and automobile owners.
Thus a system based on android application is being proposed in this paper to
manage appointment slots in order to manage spread through social gathering. The
proposed system focuses on proper scheduling of appointments such that minimum
number of customers who are under service will be present in the service centers and
will give time to service providers to sanitize the area before the next set of customers
visit.
The app connects clients to service providers, putting all the information 24 * 7,
in one convenient place and on demand as well. The app eases the management of
staff, business growth and track of customers for service providers.
2 Literature Review
Qaffas and Barker [2] proposed a web based appointment management system for
students to take appointment of their advisers. This is web based method developed
using ASP.NET 2008. Nimbekar et al. [3] proposed an idea of an android application
for salon appointment but there work do not include any results.
Sherly et al. proposed web based appointment scheduling for health-care [4]
by taking a case study. Majority of the appointment management applications are
web based and in the field of healthcare [6, 7]. Malik et al. developed an android
application based appointment schdeular for patients [5].
Chai and Wen proposed salon management system based on PHP and MySQL
[8]. Xandaro Scheduling Software [9] focuses on the appointment function. This
software is easy to handle. It’s user interface is simple and direct which is helpful for
naive users. Unique Salon Software [10] is complex software to use and one need to
read the user manual. This homepage is unorganized. It uses search function in all
the modules. It simplifies the retrieval of specific information. So, users can obtain
the information faster and easily. Advantage Salon software [11] is organized and
neat which helps the user to use it easily.
Sale et al. proposed [16] web based solutions for automobile service management.
Considering the need of today more than a web based solution [12–15, 17–19], an
app based solution is any time preferable by users. Thus in the proposed system the
idea is considered keeping needs of users in mind.
3 Proposed System
The proposed system is very efficient and easy to handle for owner and client both.The
speed and accuracy will be maintained in proper way. It has features such as facility
to register staff and regular clients along with their personal information, appoint-
ment handling, view non-working days through an event calendar, handling various
An Appointment Scheduler: A Solution to Control Covid-19 Spread 581
services along with their respective prices, time required etc., providing customer
payment handling option, system generated invoices, generating reports for analysis,
maintaining database, reminders through emails, giving staff rating facility to know
the most efficient staff members.
This application is built in “Android” as a front end and “Firebase” as a back end.
The required system specification are as follows:
Android Smartphone: version 4.0 (or Above), RAM: 512 MB, Device Memory:
50 MB (or above), Internet Connectivity.
Figure 1 shows the system architecture for shop owner. Owners can register them-
selves by providing personal and shop information along with GPS enabled photo
which can help to verify genuine service providers. The service owner will provide
details of the various services along with their charges and offers available. Shop
adder module can fetch the reviews available online about the shop or staff. Appoint-
ment manager keeps track of available and booked slots, generates report on confir-
mation of a booked slot. Notification engine gives confirmation of booked slot to the
owner of the shop.
Figure 2 shows the system architecture for customer. Clients can create their profile
by providing personal information and genuine customers are verified through OTP.
Clients can check previous ratings of shop or staff available, can check list of services
and charges of specific shop and can book the available slot. On confirm booking they
will get notification though email. All the details of shop, owner, staff, appointments
and client are saved in database.
Account Registration: Registration of user is done which makes it easier for the
user to book appointment. As they don’t need to fill personal details from scratch
every time, they want to make an appointment.
Get Location: User will know where they are located, showing the location of
the shop of their locality.
Customized Account: User can change it personal details according to his
convenience.
An Appointment Scheduler: A Solution to Control Covid-19 Spread 583
Email/Push Notification: This module makes use of email and enable shop owner
to update their customer. The shop owner can let their customer know about any
promotional offer run by them.
Marketing and Promotion: Marketing and other means to promote your products
and services is very convenient and easy to manage.
User Record: Every point of fail transaction is track and tied to each user service
history, providing quick rebooking. No restrictions on uses and user unlimited access
to any number user, staff and devices, to use if to the fullest.
Billing: This module is responsible for the generation of customers billing and
invoice.
Feedback: Collect real-time user feedback which inside to improve overall
experience.
4 Results
The proposed system is implemented as an android app for a Saloon Client and the
screenshots of various activities are as follows.
Main screen shows two options for the user either to register as an Owner or a
Customer for very first time (Fig. 3a).
If user selects Owner as an option he/she need to register as an Owner with personal
and shop information including Geo tagged photo of the shop then on verification
the registration will be confirmed (Fig. 3b–d). On successful verification the Owner
will get log in screen in the next attempt. Owner can add staff details, rates of various
services, can see booked appointments, can change profile photo (Fig. 3e–h).
If user selects Customer as an option on main screen then the user is directed to user
registration and it will be verified through OTP. On Completion of verification user
can log in.User can see various shops providing the same services, can see calender,
time slots, services provided by shops, book or cancel appointments (Fig. 4a–h).
5 Conclusion
Fig. 3 a Main screen, b Owner registration, c Details for registration, d Message after registration,
e Owner log in, f Staff details, g Services with charges, h Profile and cover photo settings
6 Future Scope
Fig. 4 a Customer registration, b Customer log in, c List of service providers, d Selection of date,
e Selection of slot, f Staff selection and services with charges, g Summary of booking, h Booking
confirmation message
References
4. Sherly IS, Mahalakshmi A, Menaka D, Sujatha R (2016) Online appointment reservation and
scheduling for healthcare-a detailed study. J Int J Innov Res Comput Commun Eng
5. Malik S, Bibi N, Khan S, Sultana R, Rauf SA (2017) Mr. Doc: a doctor appointment application
system
6. Symey Y, Sankaranarayanan S, binti Sait SN (2013) Application of smart technologies for
mobile patient appointment system. J Int J Adv Trends Comput Sci Eng
7. Choudhari SB, Kusurkar C, Sonje R, Mahajan P, Vaz J (2014) Android application for doctor’s
appointment. Int J Innov Res Comput Commun Eng
8. Chai A, Wen CC (2017) SHEARS Inc. Salon management system. J Int J Inform Vis
9. Xandaro: Free Salon Software (2017) URL: https://xandaro.com/
10. Unique Salon Software (2017) URL: https://uniquesalonsoftware.com
11. Advantage Salon Software. URL: https://www.aknaf.com/salon
12. B. Plan:Pretty (2011) Quick business plan beauty booking made simple. Confidential
Proprietary Bus Plan 1–52 29
13. Salon Today (2018) https://www.salontoday.com/article/81571/tech-pilots-voga-salon-and-
salon-clouds
14. M. S. International, The Bussiness Side of Beauty, Millennium Systems International. https://
www.millenniumsi.com/blog/65-salon-marketing-ideas (2018)
15. SBDCNe, Small Business Development Center: SBDCNet News Magazine Theme built.
https://www.sbdcnet.org/small-business-research-reports/beauty-salon-2014.
16. Sale HB, Bari D, Dalvi T, Pandey Y (2018) Online management system for automobile services.
J Int J Eng Sci Comput
17. Chavan S (2014) Automobile service center management system. J Int J Sci Res Publ 4(3).
ISSN 22503153
18. Shivasankaran N, Senthilkumar P, Scheduling of mechanics in automobile repair shops. J Indian
J Comput Sci Eng (IJCSE)
19. Selokar N, Masne V, Pimpalkar R, Puranik S, Bhoyar N, 24*7 vehicle management systems
for automobile industry. e-ISSN: 2395-0056
Bandobast Allocation and Attendance
System
Abstract This System (Website and App) is aimed at developing a Bandobast Allo-
cation and Attendance System (BAAS) that is of importance to Pandharpur Police
for various Bandobast like VIP Bandobast, Election Bandobast, and mostly for Wari
Bandobast. This system is used for the management of any Bandobast in Pandharpur
(or in any City). This system is being developed for CO’s to allocate duties to all
police, and further heads can take attendance of their subordinates. Later the control
room can print the entire report in just one click. During any Bandobast it is very
hard to allocate duties manually (informing each one about it) and also it may not
be possible to take attendance manually of more than 5000 police staff. Besides, it
is very hard to change the duty allocation or to know whether everyone is at their
allocated point.
1 Introduction
The title of the project is “Bandobast Allocation and Attendance System” is a Hybrid
application (Website + Android APP) that aims at providing easier management of
any type Bandobast like VIP Bandobast or Wari Bandobast. Surely, many times there
are various types of Police Bandobast (like Wari in Pandharpur) and City. The police
station has to manage it properly and has to submit a report of it. But how can the
Police Station manage the Bandobast [1] with less manual work if they are still using
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 587
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_59
588 P. S. Bhandare et al.
the traditional (manual) method for this? In the traditional method, before two days
of Bandobast City Police Station used to know about list of details of Police staff
coming for the bandobast. Then, they were allocating the police to the Points (All
things were typed manually). On a working day, all Police members used to come to
the Police station to know their duty details. There was no facility for attendance or
checking the allocation. The most difficult task was to replace or remove or to change
a complete point of a member during bandobast. Thus, there were more chances of
human errors, and sometimes due to lack of attendance, system members may leave
their point before the end of their shift. Nowadays, technology has changed many
aspects of life, and people’s daily life is becoming indivisible from the network due
to the development of the Internet. With our hybrid system, the process gets much
faster and more efficient than the traditional way. During any Bandobast it is very
hard to allocate duties manually (informing each one about it) and also it may not
be possible to take attendance manually of more than 5000 police staff. Besides, it
is very hard to change the duty allocation or to know whether everyone is at their
allocated point. After all these CO’s were typing the entire report (minimum 400
pages) of bandobast, which was the most tiresome and time-consuming task. Our
system makes this easier.
Even though the Pandharpur police tried to make it online, but they can’t because
the only single system is not enough to solve the issue and they are facing several
problems during process management. There is also a problem when any police
staff doesn’t know his/her point of duty they cannot check their duty details online.
Police stations provide proper Smart I-cards and instruction on a paper the paper
or Smart I-card may be lost. BAAS system helps any police member to know their
duty details at any given time, and CO’s can easily administrate the bandobast,
record maintenance becomes easy, report generation is easy, I-card generation is
possible. This system also provides a feature of attendance along with the real-time
and location of the user, which keeps all police at their point, and any critical situation
can be managed easily. With all this, there are more advantages like the attendance
without any intervention can even be taken without having an internet connection and
automatically get uploaded to the server when the connection is available. The initial
task of CO’s to create and allocate a bandobast becomes much easier and efficient
with the system. Earlier this was taking more than four days to manage the simplest
bandobast, but now our system reduced this time to only one day or half-day and
also this can be done with less manpower.
The BAAS System is developed to implement the whole process in the City Police
Station (Pandharpur) starting with the registration of all police staff coming for the
bandobast. The registration includes details like Buckle Number, Full name, Name of
posting, Rank, Aadhar number, Duty type. Then COs can allocate duties to them. The
allocated person can check their duty details, team, head name, instruction, Smart
I-card in the App. Then all print heads can take attendance of respective subordinates
by scanning QR codes on their Smart Ecards. Bandobast heads have the authority to
take attendance of anyone who is part of the bandobast. During bandobast, COs can
change the allocation of any point or person.
Bandobast Allocation and Attendance System 589
The most important and advanced concept used in the system is attendance [2].
It is because we have used QR code for attendance (never used before for any Police
Station or any organization in India, only some Maleshian school uses this concept to
take attendance of students) [3]. Every police have provided the Smart I-card having
a QR Code which uniquely identifies the police officer [4]. Whenever the respective
head scan that QR code with ‘BAAS App’ then and then only he is marked as present.
Besides increasing the accuracy in the presence of a member at the required point,
we also record his/her location [5] during scanning. If any member is absent, then
the respective head marks him/her as absent with some remark. One more important
part is, the attendance can be marked even though the user is offline, and whenever
the connection is available, the data will be sent to the server. One more important
part is, the attendance can be marked even though the user is offline, and whenever
the connection is available, the data will be sent to the server. Earlier Control Room
has to decide for the shift of duties like A, B, A and on next day, and it will be like B,
A, B, and so on. But our system itself generates the chart for this by just providing
the start and end date of bandobast [6]. Depending on the same, there is a Bandobast
report of a minimum of 500 pages (typed manually in an earlier system), it was also
automatically generated with more detailed information. Thus, all repeated typing
effort is reduced. At the end of the day, the control Room has to prepare the absentee
report, and the system provides the same in one click only. To use the app, every user
has to login with a valid username and password. The app provides different features
and rights to each user as per the role of the user like sector head, control Room,
point head, etc. Hence lower-level users can’t administrate the bandobast, and hence
security and integrity of critical data are maintained [7]. As the system include both
Website and App, the control room can easily access the necessary details of each
police anywhere with ease. Hence this becomes an advantage of this system.
2 Related Work
In this system, BAAS is very user friendly and dynamic. In this control room can
easily use the Website and App with all facilities. All other special officers like SP,
DYSP, PI have some administration rights. So as per the previous system, all work
can be done with ease as well as some new features are also added to improvise the
system [8]. Any head can know his team instantly after allocation; he/she can easily
scan the QR code to take attendance of the team. General Police user will easily
know their duty details and instructions for the point. By just entering the start and
end date control room will have the report of the entire bandobast. The absentee can
be printed even on a daily basis. By just providing an Aadhar card number, I-card is
generated needs not to fill manually. Any Police person will understand this system
in a few minutes [9]. This system is being developed for CO’s to Allocate duties to
all police, and the further officer can take attendance of their subordinates. Later the
control room can print the entire report in just one click. For this, all police staff has
to register with detailed information so that duties can be allocated to them. Once
590 P. S. Bhandare et al.
all police staff is registered to the system, CO’s can easily allocate them through the
Website to any required point. Then each point’s Head Officer can take attendance of
his/her subordinates using the Android App. Every user can see his/her duty details
in the app. CO’s can change allocation in a few seconds using the app. The app helps
to generate I-Cards for all in a few seconds. The app provides different features
to each user as per the role of the user; hence lower-level users can’t administrate
the bandobast, and hence security and integrity of critical data are maintained. As
the system include both Website and App, the Control Room can easily access the
necessary details of every police anywhere; hence this becomes an advantage of this
system. Now any Bandobast can be managed with ease in less time and less manual
work. Now instead of caring about allocation, attendance and report, the Control
room can control the bandobast properly [10]. The works like bandobast creation,
allocation, and report generation are easy to handle with the desktop, so we have
developed a Website “https://www.pandharpurpolice.in”. Other works like checking
duty details, taking attendance (even offline), I-card generation, allocation changing
and checking, instruction showing can be done easily with a mobile device. That’s
why the “BAAS” App was also developed.
3 System Architecture
The proposed system is tested by running the system by using it in a VIP Bandobast
and Magh Wari Bandobast with the online host server. We also checked the ability
of the system by registering more than two thousand users. We also created some
testing bandobast. To check the offline feature of the app, we turned o an internet
connection and marked the attendance of some members. All types of results are
properly tested against earlier manual reports. This continuing education course is
intended for any general bandobast like Election bandobast, event Bandobast, etc.
This course introduces a variety of functions and concepts that facilitate the transition
from manual handling of a bandobast to a computerized process of allocation and
attendance system and strengthen the practice for years to come. The SP of Solapur
District Mr Manoj Patil is very satisfied with the overall system as they have got the
result as they expected. Now it is very easy for them to handle any Bandobast.
5 Conclusion
The project entitled “Bandobast Allocation and Attendance System” is the Hybrid
System (APP + Website) that deals with the issues related to all types of Police
Bandobast in any Town. This project is successfully implemented two times in real
situations (in VIP Bandobast for Uddhav Thakare and Magh Wari) with all the
features mentioned in the Application requirements specification. The application
provides appropriate information to police according to the chosen service. The
project is designed to keep a view on all the issues faced during any type of Police
Bandobast. Deployment of our application will certainly help the City Police Station
(specially Pandharpur Police) to reduce unnecessary wastage of time, paperwork, etc.
Now instead of caring about other issues, the Control room can handle the bandobast
properly without having any stress. The system has helped the COs to do their work
in one or two days despite 5 or 6 days. Bandobast is a delicate event that can be
managed now efficiently and with less work. So this serves the right purpose in
achieving the desire requirements of both the police. It will be quite difficult to get
adopted with the system, but not much time it will take.
6 Future Work
D. We are trying to make it more user-friendly so that old people in the police can
understand it.
E. In the future, we will have NIC certificates for more security of the system.
F. Now we have only some activities in offline mode; we try to get all activities
in offline mode.
G. We will make the system capable of handling bandobast for a crowd of more
than 25 lakh people (like Kumbh Mela).
References
1. Zaharchuk TM, Atcheson RM, Palmar J (1981) Study of police management. ISBN 0-662-
11643-7)
2. Mohammad MH (201l) Attendance using QR code at University of Sulaimaniyan. J Math
Comput Sci 4(3). https://www.unifoglive.com/QRcode
3. Pandya KH, Galiywal HJ (2014) A survey on QR codes: in context of research and application.
Int J Emerg Technol Adv Eng 4(3)
4. Jivasoftt CORP (2010) On duty police scheduling software. www.osocer.com
5. Kasinathan V, Mustapha A, Chandran R (2018) Employee tracking using GPS. MATEC Web
Conf 150:05015
6. Ernst AT, Jiang H, Sier D (2004) StaVT scheduling and roistering. Euro J Oper Res. https://
www.elsevier.com/locate/dsw
7. Dezezhyts Y (2013) Thesis on android application development, Published by HAGAHELIA.
The University of Applied Science
8. Wei X, Manori A, Pasai N (2017) Smart attendance system using QR code. Int J Smart Bus
Technol 5(1). EZSCAN attendance system from SAMMY by IEEEIMS
9. Ghuman SS (2014) Software testing techniques. Int J Comput Sci Mobile Comput 3(10). ISSN:
2320-088X)
Fire Fighting Un-manned Air Vehicle
for Remote Areas
Abstract Fire accident results in catastrophic injuries and devastating damage. The
death rate in India due to fire accidents was almost 2.5 times more than in other
parts of the world. Fire fighting is a highly difficult and challenging task for human
beings to access the remote target areas. By using unmanned aerial vehicle, quick
response to the fire affected area can be achieved and also firefighters will get the
visual information of the fire accident. This work focuses on the implementation
of Un-manned Air Vehicles (UAV)s that can extinguish the fire. The proposed fire
fighting UAV system consists of Hexacopter as a platform. Hexacopter is a UAV that
works with six motors to achieve stable flight and better lift loading capability. The
goal of this work is achieved with stable and robust hexacopter along with dropping
mechanism which is used to drop the fire extinguishing ball on fire-affected area
and camera interface for live video footage. The description of the proposed work is
briefly described as well as determines the principle functionality.
1 Introduction
The research on Unmanned Aerial Vehicles has always been an interest of engineers
and hobbyists due to its wide range of applications such as rescue mission, mapping,
aerial photography, border patrol, etc. These vehicles do not require any pilot on-
board and can be easily controlled from the base station. These vehicles are equipped
with GPS and hence they can autonomously fly to the desired location when a path
is defined. Hexacopters are equipped with six motors to provide thrust to the vehicle.
With the proper arrangement of these motors, stability can be obtained [1]. Depending
N. S. Bhat · K. S. Shashidhara
Department of ECE, Nitte Meenakshi Institute of Technology, Bangalore, India
e-mail: [email protected]
V. Dakulagi (B)
Department of ECE, Guru Nanak Dev Engineering College, Bidar, Karnataka, India
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 593
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_60
594 N. S. Bhat et al.
2 Related Work
The Design and implementation method of “fire fighting UAV” and methods to
extinguish the fire has been explored [6, 7]. It also contains the control structure
of UAV and mission control architecture. The water spraying technique is used on
the fire-affected area which is complicated, instead of spraying technique using fire
extinguishing balls makes the scenario less complex and more efficient [6].
Two separate drones one forgetting the visual information and another for drop-
ping the fire extinguishing ball [8]. The improvisation in our project is that we are
achieving both of these objectives using single UAV. This includes battery time,
flying capacities, and flight velocity with respect to battery voltages which helps to
build a drone [9]. The suitable materials and component used to make the drones has
been discussed [10] with respect to applications.
3 Proposed System
3.1 Objective
The objective of this project is to give a quick response to the fire-affected area and
to acquire better information and a better view of the fire-affected area. This project
Fire Fighting Un-manned Air Vehicle for Remote Areas 595
also helps to extinguish the fire which is not easily accessible by human firefighters.
It also helps to reduce the risk of the life of human firefighters.
When the fire station gets to know about the fire accident, they plan on the resources
to be taken to extinguish the fire. Fire Extinguisher vans are deployed to reach the
fire accident area. There is a delay in giving a faster response to the accident. More
the delay implies more is the spreading of fire and hence more casualties. Once they
reach the accident area they manually analyze the situation and try to get valuable
information from that. Due to many restrictions such as the height of buildings, dense
forests, etc., it becomes difficult to get information from the accident place. In these
situations, the firefighters need to risk their lives in order to get clear information
about the accident and extinguish the fire.
In order to overcome the above-mentioned drawback, we have implemented a
system which contains a Drone along with dropping mechanism. The Drone helps
in providing quick assistance to the fire accident and the dropping mechanism helps
to drop the fire extinguisher ball on the fire. The camera interfaced with the Drone is
used to take live visuals of the accident, thus acquiring more information about the
incident. The Drone and the dropping mechanism is controlled by a remote in the
base station. The sensors such as accelerometer, compass, GPS present in the Drone
provides stability in order to achieve different flight modes.
Fire fighting UAV system consists of two main units Drone Unit, and Dropping
Mechanism as shown in Fig. 1, they co-ordinate with each other and respond to the
accident immediately. It helps in gathering more information from the fire accident
area and reduces the risk of firefighters.
A Drone uses six motors along with propellers to produce a thrust which enables it to
lift itself from the ground. The flight controller board forms the heart of the hexacopter
which processes different instructions and performs desired actions. Electronic speed
controllers are used to control and regulate the speed of an electric motor. Electronic
Speed Controller (ESC) also provides a reversing of the motor rotating direction and
dynamic braking. It changes DC battery power into 3-phase AC for rotating brushless
596 N. S. Bhat et al.
motors. The battery is used to provide power to all the circuitry and motors. Higher
the battery power more is the performance of the drone in-terms of flight time.
The drone consists of six motors of 1400 kV. Three motors rotate in a clockwise
direction and the other three in an anti-clockwise direction. Each motor provides
a thrust of 1000 g. The speed of the motors is controlled by ESCs. With the help
of ESCs speed controlling feature, the direction of the drone can be changed. The
flight controller which is positioned at the center monitors the firmware within the
ESC and controls the rotating rate of the motors. The flight controller takes the
input from the receiver and adjusts the motor speed accordingly via the electronic
speed controller. The flight controller is also equipped with many sensors such as a
gyroscope accelerometer, barometer, and GPS. External GPS is installed in the drone
to provide co-ordinate information and altitude information of the drone. Pixhawk
flight controller is used in this project, since it has additional features when compared
to other flight controllers. An Anti-vibrator kit is installed with the flight controller
in order to get a precise reading from the sensors. The camera along with transmitter
and receiver assists in acquiring video footage of the fire accident.
Fire Fighting Un-manned Air Vehicle for Remote Areas 597
Mission Planner is the firmware used to calibrate all the sensors present in the flight
controller. It is used to program different flight modes for the drone, such as Loiter,
Alt Hold, Stabilize, and Return to Launch, etc. This firmware is also used to program
different switches of the transmitter to perform specific actions. It is used for battery
monitoring to enable special functions like, failsafe.
This unit is used to hold and drop the fire extinguisher ball on the desired location.
It consists of a container with two doors along with a locking mechanism. The fire
extinguisher ball is placed inside the container and locked. When the drone reaches
the desired location where the ball is to be dropped, with the help of the remote
the lock is made open. When the lock is opened the ball drops and the fire gets
extinguished.
5 Experimental Results
The Drone unit and the Dropping Mechanism unit have been developed and the
system is successfully coordinated. Initially the drone without payload as shown in
Fig. 2 has been tested successfully.
Then the drone was made to fly with a payload of around 1.5 kg (Overall weight
of the drone is around 4 kg) as shown in Fig. 3. During this testing drone flew for
around 2 min, suddenly it crashed from 40 ft height. On this crash, a drone wing,
motor, flight controller, and propellers got damaged.
After installation and proper calibration of GPS, both loiter and RTL mode worked
properly as per our requirements as shown in Fig. 4. Drone flight time testing was
carried out and drone flight time was recorded to be 20 min.
The dropping mechanism system is integrated with the Hexacopter to obtain Fire
Fighting UAV as shown in Fig. 5.
As depicted in Fig. 6, Drone fleet successfully and same has been captured.
6 Conclusion
In this paper, we built a UAV that can carry a payload (fire extinguisher ball) and
deploy it on the fire. The camera integrated on the UAV provides visuals to the UAV
controller. The UAV is controlled remotely with the help of RF Transceiver. The drone
is designed to extinguish the fire on buildings and forests. It also consists of a wireless
camera to get the live video footage of the particular locations in fire-affected places.
The objective of this work is accomplished with a steady and robust hexacopter along
with a dropping mechanism which is used to drop the fire extinguishing ball on fire-
affected area and camera interface for live video footage. The proposed drones could
Fire Fighting Un-manned Air Vehicle for Remote Areas 599
be exploited in the emergencies which incidentally happen, and the fire station needs
sophisticated technologies for the rescue missions. The demand for such drones is
indeed very high in aircraft and fire rescue operations.
References
1. Zakaria AH, Mustafah YM, Mudzakkir M, Hatta M, Azlan MNN (2015) Development of load
carrying and releasing system of Hexacopter. In: Control conference (ASCC), pp 1–6
2. Chen R, Cao H, Cheng H, Xie J (2019) Study on urban emergency firefighting flying robots
based on UAV. In: 2019 IEEE 4th advanced information technology, electronic and automation
control conference (IAEAC), 20–22 Dec. ISSN: 2381–0947. https://doi.org/10.1109/IAEAC4
7372.2019.8997723
3. Chen Y, Zhang Y, Xin J, Wang G, Mu L, Yi Y, Liu H, Liu D (2019) UAV image-based forest
fire detection approach using convolutional neural network. In: 2019 14th IEEE conference
on industrial electronics and applications (ICIEA), 19–21 June 2019. https://doi.org/10.1109/
ICIEA.2019.8833958
4. Aydin B, Selvi E, Tao J, Starek MJ, Use of fire extinguishing balls for a conceptual system of
drone-assisted wildfire fighting. https://doi.org/10.3390/drones3010017
5. Radu VT (2019) Use of drones for firefighting operations. Master thesis report MSc. Risk and
Safety Management, 10/01/2019, Aalborg University, Esbjerg, Denmark
Fire Fighting Un-manned Air Vehicle for Remote Areas 601
6. Qin H, Cui JQ, Li J, Bi Y, Lan M, Shan M, Liu W, Wang K, Lin F, Zhang YF, Chen BM
(2016) Design and implementation of an unmanned aerial vehicle For autonomous firefighting
missions. In: 12th IEEE international conference on control & automation (ICCA) Kathmandu,
Nepal, June 1–3
7. Alagirisamy M (2020) Efficient coherent direction-of-arrival estimation and realization using
digital signal processor, IEEE Trans Antennas Propag 68(9): 6675–6682
8. Suprapto BY, Heryanto MA, Suprijono H, Muliadi J, Kusumoputro B (2017) Design and devel-
opment of heavy-lift hexacopter. In: 2017 international seminar on application for technology
of information and communication
9. Hwang M-H, Cha H-R, Jung SY (2018) Practical endurance estimation for minimizing energy
consumption of multirotor unmanned aerial vehicles
10. Baranek R, Solc F (2012) Modelling and control of a Hexa-copter. In: 2012 13th international
Carpathian control conference
Human Age Classification
and Estimation Based on Positional
Ternary Pattern Features Using Ann
Abstract In this paper, Positional Ternary Pattern features based Human Age classi-
fication using Artificial Neural Network for Forensic science application. The classifi-
cation of human age from facial pictures plays an important role in pc vision, scientific
discipline, and forensic Science. The various machine and mathematical models, for
classifying facial age together with Principal Component Analysis (PCA), Positional
Ternary Pattern (PTP) are planned yields higher performance. This paper proposes a
completely unique technique of classifying the human age group exploitation Artifi-
cial Neural Network. This is often done by preprocessing the face image initially and
so extracting the face options exploitation PCA. Then the classification of human age
is finished exploitation Artificial Neural Network (ANN). The method of combining
PCA and ANN performs higher rather than victimization separately.
1 Introduction
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 603
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_61
604 S. V. Jagzap et al.
wrinkles with age growth. Therefore, representing said facial changes is imperative
in age recognition.
Human faces, as necessary visual cues, convey a major quantity of nonverbal infor-
mation to facilitate the real-world human-to-human communication. As a result, the
trendy intelligent systems are expected to own the aptitude to accurately acknowledge
and interpret human faces in real-time. Facial attributes, like identity, age, gender,
expression, and ethnic origin, play an important role in real facial image analysis
applications as well as transmission communication, human–computer interaction
(HCI), and security. In such applications, numerous attributes will be calculable from
a captured face image to infer the additional system reactions. For instance, if the
user’s age is calculable by a pc, associate age-specific human–computer interaction
(ASHCI) system could also be developed for secure network/system access manage-
ment. The ASHCI system ensures young children don’t have any access to web pages
with adult materials. A slot machine, secured by the ASHCI system, will refuse to
sell alcohol or cigarettes to the underage individuals. In image and video retrieval,
users might retrieve their pictures or videos by specifying a needed age vary. Ad-
agency will verify what reasonably scroll advertisements will attract the passengers
(potential customers) in what age ranges employing a latent laptop vision system.
Although automatic image-based age estimation is a very important technique
concerned in several real-world applications, it’s still a difficult drawback to estimate
human ages from face pictures. Face pictures of two people with totally different ages.
Since totally different people age quite; otherwise, the ageing method is decided by
not only the person’s factor however conjointly several external factors, like health,
living style, living location, and atmospheric condition. Males and females can also
age otherwise thanks to the various extent in victimization makeups and accessories.
The way to extract general discriminative ageing options, whereas reducing the
negative influence of individual variations still remains an associate degree open
problem.
There exits some work on age synthesis and rendering within the last many
decades, but there are only a few publications on age estimation owing to the
complexity of ageing patterns. The age progression displayed on faces is uncon-
trollable and personalized. Such special characteristics of ageing variation cannot
be captured accurately owing to the prolific and heterogeneous information sent by
human faces. On the opposite hand, the age estimation downside is totally different
from the matter of face recognition with age variation, wherever the goal is to estimate
facial identities whereas no ages are calculable from the input faces. The research
effort on age estimation might facilitate recognizing faces containing age variations.
In this paper, we propose a PTP and PCA feature extraction to detect faces in face
images. The feature extraction first detects the faces using the RGB colour model
and divides the face region into blocks of equal size then extracting the face features
using PCA.
Human Age Classification and Estimation Based … 605
2 Review of Literature
This paper concerns the estimation of facial attributes, particularly, age and gender
from images of faces acquired in difficult, within the wild conditions. This downside
has received way less attention than the connected downside of face recognition,
and especially, has not enjoyed an equivalent dramatic improvement in capabilities
demonstrated by modern face recognition systems. Here, we tend to address this
downside by creating subsequent contributions. First, in answer to at least one of the
key issues of age estimation analysis absence information we provide a singular data
set of face pictures, tagged for age and gender, non-inheritable by smart-phones
and alternative mobile devices, and uploaded without manual filtering to online
image repositories. We tend to show the pictures in our assortment to be tougher
than those offered by alternative face-photo benchmarks [1]. Second, we tend to
describe the dropout-support vector machine approach utilized by our system for
face attribute estimation, so as to avoid over fitting. This technique, impressed by
the dropout learning techniques currently popular deep belief networks, is applied
here for coaching support vector machines, to the most effective of our data, for
the primary time. Finally, we tend to present a robust face alignment technique that
expressly considers the uncertainties of facial feature detectors. We tend to report
in-depth tests analyzing each the issue levels of up to date benchmarks likewise
because of the capabilities of our own system. These show our technique to outstrip
progressive by a good margin [2].
We think about the matter of automatic age estimation from face pictures. Age esti-
mation is sometimes developed as a regression downside relating the facial expression
and therefore the age variable, and one regression model is learnt for all ages [3]. We
have a tendency to propose a hierarchical approach, wherever we have a tendency to
initial divide the face pictures into numerous age teams, so learn a separate regression
model for every cluster. Given a take a look at the image, we have a tendency to initial
classify the image into one amongst the age teams, so use the regression model for
that specific cluster. To enhance our classification result, we have a tendency to use
many alternative classifiers and fuse them mistreatment the bulk rule. Experiments
show that our approach outperforms several states of the art regression ways for age
estimation [4].
Human age will give important demographic data. During this paper, we have
a tendency to tackle the estimation older in face pictures with possibilities. The
look of the planned methodology is predicated on the relative order older labels
within the information. The age estimation drawback is reworked into a series of
binary classifications achieved by convolution neural network. Every classifier is
employed to evaluate whether or not the age of input image is larger than a precise
age, and therefore the calculable age is obtained by adding likelihood values of those
classification issues. The planned method: Deep possibilities (DP) of facial age show
enhancements over direct regression and multi-classification strategies [5]. Design
Sparse Features like wrinkles, dark spots, moles etc. were used to analyze the image
and estimate the age [6].
606 S. V. Jagzap et al.
3 Block Diagram
4 System Analysis
The block diagram for the proposed algorithm is as shown in Fig. 1. System analysis is
done by steps Input image acquisition, Database creation, Pre-processing, Positional
Ternary Pattern, Principal Component Analysis, GLCM Feature Extraction, Artificial
Neural Network, Result.
4.1 Preprocessing
Figure 2 shows the Positional Ternary Pattern (PTP). Which assigns eight-bit
computer code to every element of a picture? Initially, Kirsch compass masks
compute the sting response of eight neighborhood pixels.
This process using creates a GLCM, use the gray co-matrix function. Particular
function create grey level Co-occurrence matrix GLCM by calculation how often a
particular pixel with the intensity value I occur in a specific spatial relationship to a
pixel with the value j. This process number of grey levels in the image determines
the size of the GLCM..In GLCM can also derive stoical measures, and the derive
statistics from GLCM and more information from plot correlation shown in Fig. 3.
How to grey co-matrix calculate the 1st three value in given GLCM.
The performance of the artificial neural network was evaluated in terms of coaching
performance and classification accuracies. Artificial Neural Network offers quick and
correct classification and could be a promising tool for the classification of the result.
The ANN with FF is trained with reference options set and desired output exploitation
‘new’ and ‘train’ command. Here, target one for dataset1, two for dataset2 and
dataset3 area unit taken as the desired output. Once the coaching, updated weight
issue and biases with alternative network parameters area unit hold on to simulate
with input options. At the classification stage, take a look at image options area unit
utilized to simulate with trained network model exploitation ‘sim’ command. Finally,
it returns the classified price as one, two or three supported that the choice is going
to be taken as our age classification.
Human Age Classification and Estimation Based … 609
In this work, our own database in spite of the existence of other databases has been
created. In this technique, we are using Gabor filter we can calculate the individual
face and mark the key points after that we are applying LDA feature extraction we
can get the output. The output result should become 75% of accuracy only. But in our
proposed system we are introducing the detecting human face then classifying age
using PTP, PCA and Artificial Neural Network classification we can get the output.
The output result should become 92% of accuracy. In Figs. 4 and 5 shows the result
of the child and young age group. In Figs. 6 and 7 shows the result of the middle and
old age group.
6 Conclusion
In this paper, By using MATLAB 2014, we have a tendency to propose a way for
human age classification using a positional ternary pattern, PCA and ANN classifica-
tion and used this prediction in our automatic system for detecting human behavior.
This work could be a continuous study from previous analysis on the employment
of heterogeneous knowledge for autonomous detection of human behavior. Face
610 S. V. Jagzap et al.
pictures are a sort of knowledge which might improve the proof in work a suspect or
an occurrence. This technique can cut back the scope in distinctive suspicious persons
within the forensic investigation space. The primary step within the methodology
involves extracting the options from face pictures employing a positional ternary
pattern and PCA. Then, we have a tendency to applied ANN to classify the ages.
The transfer learning was then accustomed with success utilize the already learn
data for a replacement task with the restricted dataset, despite the fact that, once
the dataset is little, the results area unit far better than mistreatment traditional hand-
crafted options like Haar. The planned technique was tested on the Face datasets, and
therefore the results are unit terribly encouraging the odds of accuracy for age clas-
sifications are 92.33% and 80.17% respectively, that demonstrates that the answer is
possible. Within the future work, the results are going to be combined with another
dataset. Knowledge fusion techniques are going to be applied to integrate all the
knowledge to provide an improved overall call which might assist investigators to
find suspicious persons.
612 S. V. Jagzap et al.
References
1. Eidinger E, Enbar R, Hassner T (2014) Age and gender estimation of unfiltered faces. IEEE
Trans Inf Forensics Secur 9(12):2170–2179
2. Choi SE, Lee YJ, Lee SJ, Park KR, Kim J (2011) Age estimation using a hierarchical classifier
based on global and local facial features. Pattern Recogn 44(6):1262–1281
3. Guo G, Fu Y, Dyer CR, Huang TS (2008) Image-based human age estimation by manifold
learning and locally adjusted robust regression. IEEE Trans Image Process 17(7):1178–1188
4. Mahalingam G, Kambhamettu C (2011) Can discriminative cues aid face recognition across
age? In: Proceedings of IEEE international conference on automatic face gesture recognition
workshops, Mar 2011, pp 206–212
5. Zheng T, Deng W, Hu J (2017) Deep probabilities for age estimation. In: 2017 IEEE visual
communications and image processing (VCIP)
6. Suo J, Wu T, Zhu SC, Shan S, Chen X, Gao W (2008) Design sparse features for age estimation
using hierarchical face model. In: Proceedings of 8th 2008 international conference on automatic
face and gesture recognition
Object Recognition Using Fuzzy
Classifier
Abstract In this paper, object recognition is proposed using combined DRLBP and
SIFT features for high efficient signal transfer system applications. The aim of this
research is to develop a non-real-life application of a security lock system employing
object recognition methodology. DRLBP is chosen for the object recognition algo-
rithmic program. Arduino microcontroller is employed to represent the response to
object identification. USB serial communication is employed to inter-object between
the MATLAB and Arduino UNO Microcontroller. First, the image of the individual
is captured then the captured image is then transferred to the information developed
in MATLAB during this stage, the captured image compares to the training image
within the database to see the individual standing. If the system acknowledges the
individual as an authentication person or un-authentication person, the result is sent
to the Arduino UNO microcontroller.
1 Introduction
Object identification identifies the external body part in recordings and computerized
photos will be viewed as associate degree example order issue and therefore, the
underlying amount of any object acknowledgement framework. The not insignificant
summation of utilization zones, for instance, Human—laptop Inter objects, Security
Systems, closed-circuit television, Content primarily based Image Retrieval, and then
forth demonstrates the importance of object identification and acknowledgement
calculations. As a rule, an object recognition framework can gain an image and
acknowledges the confronts freelance of stance, scale or outward appearances. The
target of our work is to differentiate and acknowledge covering confronts, one object
lies or reaches out over and covers some portion of another face, during an image.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 613
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_62
614 S. A. Atole et al.
2 Review of Literature
During this paper, we tend to plan a non-contact technique to classify the object form
by victimization Support Vector Machine (SVM) technique [1]. This algorithmic rule
consists of three steps: head segmentation, object plane identification, and object form
classification. The accuracy rate is sixty eight 68% [2]. A compact and reliable room
attending system victimization RFID and object verification is presented during this
paper. The RFID system identifies the student victimization the RFID card, and any
identification of the student is applied victimization object recognition technique [3].
The performance of the system is tested for frontal object verification, head create
varied object verification, and detection of proxy attending is applied. It’s found that
the projected theme verifies the identity of the scholar properly of concerning 98 for
frontal object and to makes an attempt on poses varied object verification [4, 5]. The
proxy attending detection applied for frontal object resulted in Associate in nursing
potency of 73.28% Associate in Nursing for various poses resulted in the potency
of 79.29% [6]. Object recognition is that the core application within the biometric
technology space. It’s wide employed in the advanced application of artificial intelli-
gence and laptop vision. Raising business of object recognition and low social control
makes a call for participation of it will increase within the last decade—the new tech-
nique used as feature extractor addition to ancient PCA [7]. Gathering options by
victimization slope technique and PCA is to search out the optimum objects vectors
because of the inputs to the classifier (NNMLP neural network) [8]. Results have
unconcealed acceptable correct classification. As a knowledge check set, we have a
tendency to used BIO-ID knowledge base within the projected system [9].
3 Block Diagram
Fig. 1 Block diagram of object recognition using combined DRLBP and SIFT features
4 System Analysis
1. Input image
2. Database creation
3. Preprocessing
4. DRLBP and Histogram
5. SIFT Features
6. GLCM Feature Extraction
7. Fuzzy Logic classifier
8. Result.
4.1 Preprocessing
Image pre-processing is the term for operations on images like changing the RGB
image to a grey one by adjusting the resolution of the image as needed. These opera-
tions don’t increase image information content; however, they decrease it if entropy
is associate degree metric. The aim of pre-processing is associate degree improve-
ment of the image information that suppresses unwanted distortions or enhances
some image options relevant for more process and analysis task. Plane Separation
on Red/inexperienced/Blue happens.
616 S. A. Atole et al.
4.2 DRLBP
In Fig. 2 DRLBP process shows the descriptor local binary pattern is used to compare
all the pixels, including the center pixel with the neighboring pixels in the kernel
to improve the robustness against the illumination variation. An LBP code for a
neighborhood was produced by multiplying the threshold values with weights given
to the corresponding pixels and summing up the result. LBP codes are weighed using
gradient vector to generate the histogram of robust LBP, and discriminative features
are determined from the robust local binary pattern codes. DRLBP is represented
in terms of the set of normalized histogram bins as local texture features. It is used
to discriminate the local edge texture of object invariant to changes of contrast and
shape.
The SIFT algorithm takes a picture and transforms it into a set of native feature
vectors. Every of those feature vectors is meant to be distinctive and invariant to
any scaling, rotation or translation of the image. Within the original implementation,
these options are often accustomed to notice distinctive objects in different pictures,
and therefore the rework is often extended to match objects in pictures. This report
describes our own implementation of the SIFT algorithmic program and highlights
potential direction for future analysis. The SIFT options represented in our imple-
mentation are computed at the sides and that they are invariant to image scaling,
rotation, the addition of noise. They’re helpfully attributable to their distinctiveness,
which allows the right match for key points between objects. These are achieved by
victimization our Gradient-Based Edge Detector, and therefore the native descriptors
bestowed round the key points. Edges are poorly outlined and typically exhausting
to observe; however, there are still giant numbers of key points can be extracted
from typical pictures. Thus we are able to still perform the feature matching even
the objects are small. Generally, the photographs are too swish to search out that a
lot of options for an identical, and in that case tiny low object may well be unrec-
ognized from the coaching pictures. In the next step, we are going to attempt to
perform some object identification, and that we select the closest neighbor or second-
closest neighbor algorithmic program that could be a smart technique to try to the
key points matching. There’s another helpful technique to acknowledge objects by
learning an applied mathematics model. During this technique, a symbolic logic
model is employed to acknowledge the objects. Associate in Nursing Expectation–
Maximization (EM) algorithmic program is employed to find out the parameters in
an exceedingly most chance framework. Hopefully, we are able to limit the model
to a tiny low quantity of components that is economical for matching objects.
This process using creates a GLCM, use the grey co-matrix function. Particular
function create grey level Co-occurrence matrix GLCM by calculation how often a
particular pixel with the intensity value I occur in a specific spatial relationship to a
pixel with the value j. This process number of grey levels in the image determines the
size of the GLCM. By default, grey co-matrix uses scaling to reduce the number of
intensity values in an image to eight, and you can use the Num Levels and the Gray
618 S. A. Atole et al.
Limits parameters to control this scaling of grey levels. GLCM has revealed certain
properties about the spatial distribution of the grey levels in the texture signals.
In the section, Fuzzy rule-based mostly system that is employed for facial features
recognition from object expression. Fuzzy is one helpful approach for fuzzy classi-
fication, which may verify the intrinsic division in an exceedingly set of untagged
knowledge and notice representatives for solid teams. Fuzzy integrals were accus-
tomed to describe the uncertainty of facial features. Facial features house might
be created mechanically and compared for expression classification. The core of
our system could be a Fuzzy Rule-based mostly system that is employed for facial
features recognition from object expression. Fuzzy Logic is often won’t too kind
linguistic models and comes with a solid qualitative base. Fuzzy systems are utilized
in several classification and management issues as well as facial features recognition.
In the existing system, we are using two databases one is input database, and another
one is storage database. In this technique, we are using Gabor filter we can calculate
the individual object and mark the key points after that we are applying LBA feature
extraction we can get the output. The output result should become 75% of accuracy
only. But in our proposed system, we are introducing the detecting object using
DRLBP, SHIFT and Fuzzy Logic classification we can get the output. The output
result should become 95% of accuracy.
In Figs. 3 and 4 shows the input image and detection of the face after that in
Fig. 5 shows the dominant rated local binary pattern and then shows the histogram
of dominant rotated local binary pattern in Fig. 6. After completion of histogram of
DRLBP process then shows the R-Eye, L-Eye, Nose and Mouth in Fig. 7. In Figs. 8
and 9 shows the authentication result. In Fig. 10 shows the Performance Analysis of
the input image.
6 Conclusion
In our purport technique to agnize an object that is relegated into a dyad of categories.
The strategy introduced during this article was tested on the JAFFE information
which incorporates ten persons. The information consists of ten pictures. 70% of
the information was used for training, and therefore the remaining 30% for testing
pictures. The accuracy of the planned system is compared with those of different
Object Recognition Using Fuzzy Classifier 619
ways. It might be seen that the popularity of the planned technique performs higher
than the opposite technique. In the future scope of my project is we are working
video to video processing for surveillance and biometric application.
620 S. A. Atole et al.
Fig. 5 DRLBP
Fig. 6 Histogram of
DRLBP
Object Recognition Using Fuzzy Classifier 621
Fig. 9 Result
622 S. A. Atole et al.
References
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 623
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_63
624 V. M. Sale et al.
1 Introduction
Basically, for any software development task, a developer must know the project
background, particularly the architecture of the system, design, working, and the rela-
tions between the various components using any available documentation. Program
conception occurs in a bottom-up method, a top-down approach or some combination
of both [1, 2]. The traceability is an essential factor for any software project, and if
we use it, it could be valuable from different viewpoints for the development. When
we develop source code for any system that source code can be traced and become
identical with the requirement and analysis because we develop a source code as per
the requirements [3, 4]. A traceability link is a connection between the source code
and requirement.
Requirement traceability helps software engineers to trace the requirement from
its emergence to its fulfilment [3, 5]. Traceability may not help us to know how
different components of systems are interlinked and dependent on each other in the
same system. We may also fail to find the impact of change on the software and system
[6]. A major goal of traceability is a link of, in the lack of originality. Requirements
and other artifacts traceability links [4]. Therefore, we should look at traceability
from all the aspects of traceability regarding scope and coverage. Requirements
traceability has proved much important over the past decade in the scientific literature.
It is defined as “the ability to illustrate and go after the life of a requirement, in both
the onward and toward the back direction” [4, 5, 7]. Traceability links among of a
system and its source code help us in reducing comprehension system attempt.
The developers can add, remove, or modify features as per the users’ demand while
updating the software. At the same time, software maintenance and development,
requirement traceability links become marginal because no developer can devote
effort to update it. However, to recover traceability links later is a very painful and
tedious task; also it is costly for developers too [2]. A developer usually does not
update requirement-traceability links with source code. Requirements and source
codes are different from each other, which decrease the textual similarity [3, 6].
It is most significant to verify that the requirements are properly executed in the
design. This is done with requirements traceability which is usually referred to as
[2, 5, 8] “the capability to demonstrate and go after the life of a requirement, in both
onward and toward the back direction” [9]. Requirements traceability confines the
relationships between the requirements and source code. The traceability is one of
the needs of different stakeholders—project sponsors, project managers, analysts,
designers, maintainers, and end-users, because of their need, priority, and goal [1,
2, 4]. The requirements traceability is a point of a system in which the requirements
are linked to their sources and the objects formed during the system development
An Effective Approach for Accuracy of Requirement … 625
life cycle based on these requirements [7]. The important part of requirements engi-
neering and elicitation phase is that the underlying principle and sources to the
requirements are captured to understand requirements evolution and verification [3].
Modifications in design appear. During design phase requirements, traceability
helps keep track of when the change request is implemented before a system is
redesigned. Traceability can also give information about the validation, significant-
conclusion, and postulation behind requirements [3, 7]. After the delivery of the
system, [1] modifications occur due to various reasons (e.g. to a changing envi-
ronment). The traceability helps us complete, more accurate cost and schedule of
change(s) can be resolute, instead of depending on the engineer or programmer who
is expert [7].
Traceability information allows answering:
1. What is the effect when the requirement is changed?
2. Where is a requirement applied?
3. Are all requirements assigned?
4. Which requirement is deal with by a requirement?
5. Is this requirement essential?
6. What design decisions affect the implementation of a requirement?
7. What are the benefits of this technique and what were the further options?
8. Is the implementation compliant with the requirements?
9. Is this design element necessary?
10. How do I interpret this requirement?
Benefits of Traceability
1. Prevents losing of knowledge
2. Helps for the verification process
3. Impact analysis
4. Change control
5. Process monitoring
6. Better software quality
7. Reengineering
8. Reusability
9. Reduction of Risk.
This section presents an environment on the I.R. technique and a review of the related
work. Traceability approach can be separated into three main categories, i.e. dynamic,
static, and hybrid. The dynamic approach gathers and examines execution traces [10]
to recognize the technique that a software link has belen carrying out in the particular
scenario. However, it couldn’t help to differ in overlapping circumstances, because
there are some limitations to a single method [2]. The legacy system may not be
applicable, due to bugs and/or some other issues. Thus, to collect execution traces is
not possible.
Static traceability approaches [11, 12] use source code structure and/or textual
information for recovering traceability associations among high-level and low-level
software artifacts. The combination of static and dynamic information is hybrid
traceability. The study shows that a combination of dynamic and static information
can perform better than the single I.R. technique [5].
Information Retrieval (I.R.); refers to a method that would calculate textual simi-
larities of different documents. The textual similarity is calculated using the terms
that occurred in the documents. If two documents have a number of general terms,
those documents are measured to be similar. The analysis of different I.R. methods
can be in three steps [13]. First, after preprocessing such as stop word removal and
stemming, a corpus is made from the documents.
Second, each document is represented as access in an index. The term-by-
document matrix is a common index, where the document as rows and each term as a
column. The incidence of the term arising in the document is the values in the matrix.
Third, by using a cosine similarity formula, the similarity among the index entries is
calculated [14]. The presentation of the key entries and the formula for calculating
the similarity varies depends on the I.R. method. We use the VSM IR method in this
paper and briefly describe it in the following paragraph.
In the Vector Space Model (VSM) [13], each document represents the vector of
terms. In the term-by-document matrix, each row can be measured as one document’s
vector in the space of terms that occur in all documents. The calculation of similarity
of two documents is based on the cosine angle between vectors of each document.
In general, the cosine angle between vectors of the two documents will reduce as the
different documents share more terms. Hence, the higher similarity of the documents
will occur.
An Effective Approach for Accuracy of Requirement … 627
4 System Architecture
For any software evolution, the essential task is, a developer must understand the
project background [2, 8], in particular, the system planning, propose, how to imple-
ment, and the relations among the different artifacts using any available documen-
tation. Program understanding occurs in a bottom-up way, a top-down way, or some
mixture thereof. Different types of data, ranging from domain-specific knowledge
to general programming knowledge can be used throughout program conception
[3]. Traceability links between source code and part of the documentation, e.g.,
requirements, abet both top-down and bottom-up conception.
Requirement traceability is defined as, “the capability to demonstrate and go
after the life of a requirement, in both onward and toward the back direction” [9].
Traceability links are also necessary to make sure that source code is reliable with
its requirements and that all and only the specified requirements have been imple-
mented by developers. Traceability links are useful in decreasing understanding
effort between the requirements of a system and its source code [4–6]. The trace-
ability information is also useful for software maintenance and development tasks.
For instance, once a developer has traceability links, a user can easily trace what
software artifacts must be changed for the development of a new requirement.
Even with the importance of traceability links, in software maintenance and devel-
opment, as developer update features, requirement traceability links become outdated
because developers do not dedicate effort to update them later [2, 6, 8]. This lacking
traceability information is one of the main issues that contribute to project failure
and difficult to sustain. Unsatisfactory traceability information results in the need
for costly and painstaking tasks of manual recovery and maintenance of traceability
links [9]. These manual tasks may be frequently required depending on how normally
software systems evolve or are maintained.
As a result, the literature proposed methods, techniques, and tools to improve
traceability links automatically [2, 8, 9, 15]. Researchers used information retrieval
(I.R.); techniques, to recover traceability links between high-level documents, e.g.,
requirements, instruction booklet pages, and plan documents, and low-level docu-
ments, e.g., source code and UML diagrams. I.R. techniques compute the textual
similarity between each two software artifacts, e.g., the source code of a class and
a requirement [2–5, 8]. A high textual similarity means that the two artifacts most
likely share numerous concepts and that; therefore, they are likely linked to one
another.
628 V. M. Sale et al.
The proposed work is based on the IR-based RTAs process is typically divided into
three main steps. Figure 1 shows the IR-based RT links revival process.
First, every the textual information with the requirements and source code is taken
out and preprocessed by splitting terms [3–5], removing stop words and remaining
words are then steamed to its grammatical origin. Second, all the stemmed terms
are weighted using a term weighting system. Last, an I.R. technique calculates the
similarity between requirements and source code documents [6]. Lastly, it creates
a ranked list of probable traceability links. An elevated comparison between two
documents shows a probable semantic connection between them.
5.1 Preprocessing
To generate traceability links, we remove all the identifiers from source code and
terms from requirements. In this, I.R. techniques are used as an engine to create links
between requirements and source code. I.R. techniques imagine that all documents
are in the textual format [8, 16]. To remove source code identifiers, a source code
parser is used. The parser throw-outs extra information, e.g., primary data types and
keywords, from the source code and gives only identifier names. The identifiers and
terms are removed by filtering, stopper, and stemmer process [10, 17].
The primary step is term splitting [2–4]. A text normalization step renovates
all upper-case letters into lower-case letters. This step eliminates non-textual, i.e.,
some numbers, mathematical symbols, brackets, etc., information and extra white
spaces, from the documents. Some identifiers/terms could be united with some special
characters, e.g., underscore, and/or Camel Case naming reunion. Therefore, divide
all the united terms to make them separate. For example, Hello India and hello India
are split into the terms “hello India” [2].
The following step is the stop word removal [2, 8]. The input for this step is the
normalized text that could include some general words, e.g., articles, punctuation,
etc. These general words are measured as noise in the text because it does not be a
symbol of the semantics of a document. Hence, in this step, a stop word list is used
to eliminate all the stop words.
The next step is stemming. An English steamer, for example, would recognize the
terms “excel,” “excellence,” and/or “excellent” as based on the root “excel” [2, 5].
An I.R. technique calculates the similarity between two documents based on similar
terms in both documents. Still, due to different postfix, I.R. techniques would judge
them, e.g., add and addition are like two different documents, and the result would
be a low similarity between two documents. Thus, it becomes important to perform
the morphological investigation to exchange plural into the singular and to take back
infected forms to their morphemes [3, 18].
Following two main factors are considered important [2]:
1. Term frequency (T.F.): T.F. is often called home frequency. If a term appears
several times in a document, then it would be allocated higher T.F. than the
others.
2. Global frequency (G.F.): If a term appears in various documents, then the term
is considered global. It is also known as inverse document frequency (IDF).
An I.R. technique (IRT) changes all the documents into vectors to calculate their
similarities [12, 19]. To change documents terms into vectors, each term is allocated a
weight. A variety of schemes for weighting terms have been proposed in the literature.
Widely used weighting schemes are differentiated as probabilistic. In the following,
the term identifiers to refer to all source code entities, i.e., class name, method name,
variable name, and comments [5].
If a term comes out multiple times in single or multiple documents, then IRT
would propose that document as a relevant document to a query [5, 8, 11]. However,
multiple amounts of a term do not show that it is an important term.
5.3 IR Techniques
To create sets of traceability links, various I.R. techniques are used, to identify
concepts in the source code, carry out experiments using different I.R. techniques to
recover traceability links [6, 8].
630 V. M. Sale et al.
6 Proposed Algorithm
The data preprocessing is prepared to eliminate needless content from the text and
to find out the original form of the words [3, 5]. The preprocessing of the data is
completed by a valid method such as Stop word removal and Stemming to the data
composed of the customer.
For work out, stop words are words that are filtered out preceding to, or following,
processing of text [3, 4]. Stop words are ordinary words that take less significant
meaning than keyword. These stop words are a few of the most common, short func-
tion words, such as the, a, an, is, at, which, that, and on, etc. Stop-word elimination is
the method of eliminating these words. To find out the words from a text, all needless
content must be removed, so it is needed to remove the stop words from the text put
into an array [5, 12].
Algorithm:
The following is an algorithm for stop word removal
1. Acquire the input
2. Establish the glossary of stop words
3. Divide factors into words
4. Assign new word list to store words
5. Collect outcome in the String Builder
6. Loop during the entire terms
7. Come again, string with words detached.
6.3 Stemming
Words get from the input of the data are created to be too sparse to be useful as
features for categorization as they do not simplify well. The presence of a large
number of inflexions of the same word, this is the common reason for stemming.
Hence, the origin form of the word is to be taking out as a feature [2, 3, 5].
Stemming is the method, for decreasing derived words to their origin form. Stem-
ming program is commonly known as stemming algorithms or stemmers [3]. Even
as writing the sentence for a grammatical basis, it contains various forms of a word,
for example, collect, collection, collecting and/or collected. In many circumstances,
it would be helpful for a finding for one of these words to revisit the word in the
An Effective Approach for Accuracy of Requirement … 631
set to take away the required content from a given sentence [3, 5, 20]. The goal of
stemming is to decrease variation form and sometimes derivationally related forms
of a word to a common base form [2, 5, 8].
Stemming algorithm:
The stemming algorithm consists of different steps of stemming applied sequentially.
Within each stage, there are various principles to select rules, such as choosing the rule
from every rule group that applies to the longest suffix. The algorithm of stemming
works as follows:
Rules Illustrations
S → cats → cat
EED → E.E. agreed → agree
(∗v∗)E.D. → plastered → plaster
(∗v∗)ING → cutting → cut
There are three main reasons for stemming algorithm, or stemmer, as follows.
The first reason of a stemmer is to cluster the words according to their theme.
Many words are the root from the same stem, and we can consider that they belong
to the same concept (e.g., act, actor, action) [3, 4, 8]. The different forms are created
by attaching affixes (prefixes, infixes, and/or suffixes) but, in English considering
only suffixes, as normally prefixes and infixes change the meaning of the word, and
a bit of them would lead to errors of bad topic resolve [18, 21].
The next reason of a stemmer is openly associated to the [3–5, 11] I.R. process,
as containing the stems of the words agree to some point of the I.R. process to be
better, among which we can stress the ability to index the documents according to
their theme, as their terms are clustered by stems or the extension of a query to obtain
to a greater extent accurate results.
The extension of the query permits it, for refining by replacing the terms, it covers
the related topics, which are also there in the collection [4, 8, 22]. This alteration
can be done routinely and obviously to users, or the system can propose one or more
superior method of the query.
Finally, the conflation of the words allocation the same stem leads to a decrease
of the vocabulary to be taken into the process, as the entire terms contained in the
natural input collection of documents can be decreased to a set of topics [3, 5, 6].
This directs to a decrease of the space needed to store the formation used by an I.R.
system and after that also lightens the computational weight of the system.
632 V. M. Sale et al.
9 Experimental Results
Our goal is to develop software which should be able to trace the source code
according to requirement by applying steaming algorithm. Traceability links are
useful in decreasing understanding effort linking the requirements of a system and
its source code. The traceability information is also useful to the software mainte-
nance and development tasks. For instance, with traceability links, a user is able to
simply trace what software parts must be changed for the development of a new
requirement. As, show in Table 1, comparison for different files with similarity IR
accuracy, time complexity.
The experimental outcomes of the work are given below. Here we have given
the result of the proposed work, which implements the data preprocessing with a
stemming algorithm to identify the similarity between requirement and source code
(Figs. 2 and 3).
The proposed system can be evaluated based on traceability links such as I.R.,
and VSM. The analysis of the proposed system results and discusses observations
from the experiential evaluation of the Trustrace.
The experiential evaluation of the proposed system supports the Trustrace united
with IRT is helpful in increasing the precision and recall principles of some baseline
requirements traceability links.
Achieving Target
Program conception arises in a bottom to top way, a top to bottom way, or mixture
them. Developers use knowledge throughout program conception, from particular
domain knowledge to common programming knowledge. The linkage of traceability
among source codes with requirements of different stakeholders associated with
goals, help both top-down and bottom-up conception, which reduces the under-
standing exertion. Requirements Traceability is related to requests and different
improvement factors. The possible objectives are:
1. The best excellence of the system which is in the development phase;
2. To classify the product which are in the development phase with different
artifacts’; and
3. The facility to deal with change.
The requests and correlation with all the different parts related to it, such as
different modules of the system, investigation of outcomes, execution of the system
with different test cases, test actions, test outcomes and requirements of the system
all kinds should be traced. For any software development, the essential task is, a
An Effective Approach for Accuracy of Requirement … 635
designer has to know the development background, such as the development plan-
ning, purpose, how to implement, and the associations among the different parts
using any available documentation. Structures of system understanding take place in
a start to end way, an end to start way, or grouping of both. Different types of data,
ranging from particular knowledge to broad programming knowledge can be used
throughout program conception.
Requirement traceability helps software engineers to trace the requirement from
its emergence to its fulfilment. Traceability links are useful in decreasing developing
effort between the source code of a system and its requirements.
The traceability of requirements is a point of a technique. The requirements and
their sources are obviously correlated to each other and the objects formed in the
life cycle of system development depend on the requirements. During design phase
traceability of requirement helps to keep track of when the modification of request is
executed before a system is redesigned. Traceability is also able to give notification
about the validation, significant conclusion, and postulation of requirements.
11 Future Work
In future work, a broader survey, a more diverse set of stakeholders for multiple
domains. Different repositories and traceability recovery activities will be used for
replicating the study of traceability. The overall goal is to enhance the state-of-the-art
by addressing the shortcomings in a current research that uses I.R. technique to mine
vague software repositories and to persuade future researches to follow the lead.
Furthermore, a number of methods to get better the accuracy of the proposed trace-
ability improvement technique. An appropriate cost-beneficial model for traceability
needs further study as regards the costs of sustaining traceability and regarding the
benefits of traceability for other tasks in the process of software development. This
should be just the focal point of future work.
12 Conclusion
The traceability is the most important factor and precious from different points of
view for any software project development. We have developed a source code for
the system that source code can be traced and become identical with the requirement
analysis because we develop a source code as per the requirements.
From the results, it shows that our approach is performing better than other
techniques. It provides better consistency compared to existing systems. Thus,
this technique is useful for to minimize the execution time during the software
development.
For development of any software, requirement traceability plays a vital role in the
maintenance of software. Creating traceability links manually is one of the costly
636 V. M. Sale et al.
and lengthy works. The major finding of evaluation that is the proposed system
generates accurate traceability links during development with high precision and
recall. Requirements specification for requirements traceability is formed alongside
all the investigations, which drives both their direction and focus.
References
1. Shinde SK, Sale MVM (2015) A survey on accuracy of requirement traceability links
during software development. In: International conference on innovations & technological
developments in computer, electronics and mechanical engineering (ICITDCEME), 2015
2. Ali N, Guéhéneuc Y-G, Antoniol G (2011) Trust-based requirements traceability. In: Sim SE,
Ricca F (eds), Proceedings of 19th IEEE international conference program comprehension, pp
111–120, June 2011
3. Divya KS, Subha R, Palaniswami S (2014) Similar words identification using naive and TF-IDF
method. Int J Inf Technol Comput Sci 11:42–47. Published Online Oct 2014 in MECS https://
www.mecs-press.org/. https://doi.org/10.5815/ijitcs.2014.11.06. Copyright © 2014 MECS
4. Khetade PN, Nayyar VV (2014) Establishing a traceability links between the source code
and requirement analysis. A survey on traceability. In: International conference on advances in
engineering and technology (ICAET-2014) 66, Page (IOSR-JCE), e-ISSN: 2278-0661, p-ISSN:
2278-8727, pp 66–70
5. Ali N, Guéhéneuc Y-G, Antoniol G (2011) Factors impacting the inputs of traceability recovery
approaches. In: Zisman A, Cleland-Huang J, Gotel O (eds) Springer, Berlin
6. Muthamizharasi S, Selvakumar J, Rajaram M (2014) Advanced matching technique for
trustrace to improve the accuracy of requirement. Int Innov Res Sci Eng Tech (ICETS’14)
3(1)
7. Ramesh B, Jarke M (2001) Toward reference models for requirements traceability. IEEE Trans
Softw Eng 27(1):58–93
8. Ali N, Guéhéneuc Y-G, Antoniol G (2013) Trustrace: mining software repositories to improve
the accuracy of requirement traceability links. IEEE Trans Softw Eng 39(5):725–741
9. Ramesh B, Edwards M (1993) Issues in the development of a requirements traceability model.
In: Proceedings of the IEEE international symposium on requirements engineering, pp 256–259
10. Poshyvanyk D, Guéhéneuc Y-G, Marcus A, Antoniol G, Rajlich V (2007) Feature location
using probabilistic ranking of methods based on execution scenarios and information retrieval.
IEEE Trans Softw Eng 33(6):420–432, June 2007
11. Hayes JH, Antoniol G, Guéhéneuc Y-G (2008) PREREQIR: recovering pre-requirements via
cluster analysis. In: Proceedings of 15th working conference on reverse engineering, pp 165–
174
12. Antoniol G, Canfora G, Casazza G, Lucia AD, Merlo E (2002) Recovering traceability links
between code and documentation. IEEE Trans Softw Eng 28(10):970–983
13. Marcus A, Maletic JI (2003) Recovering documentation-to source-code traceability links
using latent semantic indexing. In: Proceedings of 25th international conference on software
engineering, pp 125–135
14. Panichella A, McMillan C, Moritz E, Palmieri D (1993) When and how using structural
information to improve IR-based traceability recovery
15. Gotel O, Finkelstein (1994) An analysis of the requirements traceability problem. In:
Proceedings of the first international conference on requirements engineering, pp 94–101
16. Schwarz H, Ebert J, Winter A (2009),Graph-based traceability: a comprehensive approach.
Softw Syst Model
17. Heindl M, Biffl S (2005) A case study on value-based requirements tracing. In: Proceedings
of the 10th European software engineering conference. Lisbon, Portugal
An Effective Approach for Accuracy of Requirement … 637
18. Lindvall M, Sandahl K (1998) How well do experienced software developers predict software
change? J Syst Softw 43(1998):19–27
19. Cleland-Huang J, Chang CK, Ge Y (2002) Supporting event-based traceability through
high-level recognition of change events. In: Proceedings of the 26th annual international
computer software and applications conference on prolonging software life: development and
redevelopment. Oxford, England, pp 595–602
20. Winkler S, Pilgrim J (2010) (2010), A survey of traceability in requirements engineering and
model-driven development. Softw Syst Model 9(4):529–565
21. Clarke S et al (1999) Subject oriented design: towards improved alignment of requirements,
design, and code. In: Proceedings of the 1999 ACM SIGPLAN conference on object-oriented
programming, systems, languages, and applications, Dallas, TX, pp 325–329
22. Von Knethen A (2002) Change-oriented requirements traceability. Support for evolution of
embedded systems. In: Proceedings of international conference on software maintenance, pp
482–485
23. Wieringa R (1995) An introduction to requirements traceability. Technical Report IR-389,
Faculty of Mathematics and Computer Science
24. Lehman M, Ramil J (2003) Software evolution—background, theory, practice. Inf Process Lett
88(1–2):33–44
Clustering of Fruits Image Based
on Color and Shape Using K-Means
Algorithm
1 Introduction
Machine learning is a branch of statistics and computer science in which the machine
predicts analyzing the data provided to it. In machine learning, there are various
techniques and algorithms: Supervised learning, Unsupervised learning, Clustering,
Classification, etc. [1]. The clustering algorithm is an unsupervised machine learning
technique [2]. Unsupervised machine learning well defined unknown patterns in data.
Here, we study the object clustering problem by using the K-means algorithm. K-
means algorithm is the elemental and most popular unsupervised machine learning
clustering algorithm. In clustering algorithm K variable represented by data set of
cluster group. In all cluster, centroid is define. The centroid is a data point present
at the center in all clusters. The tricks to define the centroids far away from each
other so that variation is less. After all this process, each and every data point in
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 639
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_64
640 V. Maskar et al.
the cluster is assigned to the nearest centroid. The K-means clustering algorithm is
very expedient in grouping new data [3]. Clustering analysis is the organization of
patterns into clusters based on similarity. Intuitively, patterns within a valid cluster are
more similar to each other than they are pattern belonging to a different cluster. It is
important to understand the difference between clustering and discriminant analysis.
Clustering is unsupervised classification and discriminant is supervised classifica-
tion [4]. In supervised classification we provided with collection of labeled data and
the unsupervised learning is provided data into unknown pattern [5]. A learner can
cash in of examples (data) to capture characteristics of interest of their unknown
underlying probability distribution. Data are often seen as examples that illustrate
relations between observed variables. A major focus of machine learning research is
to automatically learn to acknowledge complex patterns and make intelligent deci-
sions supported data. The ability of a program to find out from experience that’s, to
switch its execution on the idea of newly acquired information.
2 Literature Review
This proposed paper help to understand the clustering techniques which is a defined
by Dr. P. Tamilselvi and Dr. K. Ashok Kumar. In this paper, the plant leaves are
grouped based on the colors in the leaves. Totally, three categories are specified to
represent the leaf with more green, leaf with yellowish shades and leaf with reddish
shades. The task is performed using image processing. The leaf images are processed
in the sequence such as image pre-processing, segmentation, feature extraction and
clustering [1].
The Paper present an automated system for classification of fruits. Ruaa Adeeb
Abdulmunem Al-falluji have been proposed an automated system for classification
Clustering of Fruits Image Based on Color … 641
of fruits was constructed using an ordinary camera. The fruits classify based on
this color, Shape, and Texture. Gray Level Co-occurrence Matrix (GLCM) is used
to calculate texture and Best accuracy was achieved by using the Support Vector
Machine (SVM). In this paper all the fruits were analyzed on the basis of their color,
shape and texture and then classified using different classifiers to find the group that
gives the best accuracy [2].
Chithra and Henila have been applied image processing system for fruits classifi-
cation. Machine vision system for fruit identification and disease identification of
fruits or vegetables under the research in the agriculture industry. Machine vision
system for fruit identification and disease identification of fruit or vegetable is one
among the current topic that is under research in the agriculture industry. As a part of
this current research area, this fruits classification using image processing techniques
was developed [6].
Won have been proposed different types of the clustering algorithm, clustering
paradigms used in the machine learning system. They have been applied to a wide
range of domains; for instance, bioinformatics, speech recognition, and financial
analysis [3].
Muhammad and Yan have been applied the basic idea of machine learning and its
various techniques. The algorithms of machine learning are useful in areas where
deploying explicitly written in high- speed performance. Machine Learning can be
considered as a subset of Artificial Intelligence. Machine learning has many more
algorithm; those algorithms can be seen as building blocks to make computers learn
to behave more intelligently by somehow generalizing rather than just storing and
retrieving data items from a database system and other applications [4].
642 V. Maskar et al.
Gianey and Choudhary have been proposed the idea about machine learning and in
that paper also learn machine learning techniques and tools. In machine learning
empowers system with the ability to learn automatically and get better with expe-
rience without being explicitly programmed. In machine there are many algorithms
are useful in areas where deploying explicitly designing application with high speed
performance are unfeasible [5].
Obulesu et al. have been proposed the current world lots of electronic data is generated
in every field using machine learning algorithms and techniques. In that using the
different techniques like as the decision tree, Support Vector Machine, Naive Bayes
Classification, Linear Regression, etc. In the current world many data is generated
in each and every field. These data contains useful information to predict the future
[7].
Mei have been proposed Object clustering is a very challenging unsupervised learning
problem in machine learning and pattern recognition. In this paper defined the object
clustering technique it is an unsupervised machine learning technique. The experi-
ment conducting on a small clustering dataset shows that the k-means with binary
sketch templates for object clustering is very promising and the learned mixture
of templates is also meaningful for understanding the results. Number of clustering
algorithms which continue to appear in the literatures; each new clustering algorithm
performs slightly better than the existing ones on a specific distribution of patterns.
Data mining techniques, Machine learning algorithms, and Image Processing tech-
niques are applied to clustering the object. Different authors have proposed different
methods for the classification of the object [8].
Clustering of Fruits Image Based on Color … 643
3 Methodology
Methodology is the flow the content of the project in that we define the actual process
of the project. In order to classify the procedure of work. In that chapter we define
the system architecture, proposed algorithm, Flowchart and simple example of the
clustering technique.
In Fig. 1 shows that the how the system are work. Take picture of fruit image using
the camera. Images are there then extract the feature of that image as like color and
shape. After extracting feature we divide that image either shape or color wise. Using
the K-means clustering algorithm for distinguish fruits image. When we apply this
technique we got result in color or shape wise.
The goal of this algorithm is to find groups in data, with the number of groups
represented by the variable K. The algorithm works step by step to assign each data
point to one of K groups based on the features that are provided. Here, Data points
are based on feature similarity.
Algorithm
(1) K points are placed into the object data space representing the initial group of
centroids. At beginning decide the value of k.
(2) Each object or data point is assigned into the closest k. Randomly put initially
partition that classifies the data into k cluster. Assign value of k randomly.
(i) Take the first k training sample value which is single-element of clusters.
(ii) Assign each of the remaining training sample value to the cluster with
the nearest centroid. After each assignment, recomputed the centroid of
the gaining cluster.
(3) After all objects are assigned; the positions of the k centroids are recalculated.
Take each sample value in sequence and compute its distance from the centroid
of each of the clusters. If a sample is not currently in the cluster with the closest
centroid, switch this sample to that cluster and update the centroid of the cluster
gaining the new sample and the cluster losing the sample.
(4) Steps 2 and 3 are repeated until the positions of the centroids no longer move. K
centroids are created randomly (based on the predefined value of K). K-Means
allocates every data point in the dataset to the nearest centroid (minimizing
Euclidean distances between them), meaning that a data point is if the decided
data point is closet to centroid than it is considered. Then K-means recalculates
the centroids by taking the mean of all data points assigned to that centroid’s
cluster, hence reducing the total intra- cluster variance in relation to the previous
step. In the K-means refers to averaging the data and finding the new centroid.
The algorithm step by step process between steps 2 and 3 until some criteria is
met (e.g. the sum of distances between the data points and their corresponding
centroid is minimized, a maximum number of iterations is reached, no changes
in centroids value or no data points change clusters).
3.3 Flowchart
Initially, the K-means algorithm is works based on a data point. In Fig. 2 shows how
the K-Means clustering algorithm works.
In that first decide the number of cluster means K = 1, K = 2 or K = 3 etc.
Choose the centroid of an object and select the distance of that particular object. If
the distance is a minimum grouped cluster in the same cluster otherwise it gives in
another cluster. And repeat the process until a got current cluster.
Clustering of Fruits Image Based on Color … 645
Fig. 2 Flowchart
Color feature and shape feature are extracted in complete feature vector.
646 V. Maskar et al.
Color Features: The color feature that can be used for showing different color
or mixer of color. In color feature there are different type’s colors these are also
known as RGB color. The main purpose of the this color model is for the sensing,
representation, and display of images in electronic systems as per the different color,
such as televisions and computers, though it has also been used in conventional
photography RGB is a device detect or reproduce a given RGB value differently [2].
Shape Features: The second feature that can be used as an attribute for recognition
is shaped. In machine learning, a feature is a piece of information that is relevant
for solving the computational task related to a certain application. Shape Features
may be specific structures in the image such as points, edge or objects. The shape
feature used is area, several pixels in an image is used for determining the area of a
particular object [2].
This section presents the computational analysis for investigating the performance
of the k-means algorithms for image clustering. Given algorithm is coded in python
language. The performance evaluation in the project is done for the purpose of
minimize fitness function.
Initialization
Select k cluster centroids randomly Here we apply k = 3 as shown in Fig. 4 and
stated as per Table 1.
Clustering of Fruits Image Based on Color … 647
Fig. 4 Initialization
Assignment
Assign each point to its nearest centroid cluster in this stage we assign the each and
every point nearest to centroid and calculate the distance of the point. After assigning
values to respective points we have Figs. 5, 6, 7 and 8.
Fig. 5 Assignment
using K-Means clustering for classified fruits images from their colors and shapes.In
feature we use different technique of machine for fruits clustering. And also generate
different module for clustering which is less time consuming.
Clustering of Fruits Image Based on Color … 649
References
Abstract Augmented reality (AR) is an virtual experience of the real world, some-
times across multiple sensory modalities, including visual, auditory, and haptic
(Umeda et al. in ICIIBMS 2017, track 3: bioinformatics, medical imaging and neuro-
science, Okinawa, Japan, pp 146–149, 2017 [1]) Augmented reality is expounded
largely to synonymous terms: mixed reality and computer-mediated reality. So we are
thinking to vary over the instruction into present day or computerized training with
the help of Augmented Reality. Nowadays everybody has mobile phone addition-
ally children utilize the cell phones, with the help of mobile phone we are going
to propose our framework. Considering a book there are numerous pictures yet
kids can’t envision these items in genuine so with our application we’ll founded
a framework which tells this specific picture and also the comparing data to find out
understudies effectively and adequately with no issues.
1 Introduction
The first value of augmented reality is that the manner during which components
of the digital world blend into an individual’s perception of the important world,
not as a straightforward display of information, but through the combination of
immersive sensations, which are perceived as natural parts of an environment. The
earliest functional Augmented Reality systems that provided immersive mixed reality
experiences for users were invented within the early 1990. Presently coming toward
our exploration, right off the bat we learned about the present framework, there are
numerous ways we are able to utilize Augmented Reality in education however we are
V. V. Bandgar · S. N. Patil
SKNCOE Vadgaon, Pune, India
A. A. Bahirat (B)
SVERI’s COE (Poly), Pandharpur, India
G. A. Fattepurkar
VIT, Pune, India
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 651
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_65
652 V. V. Bandgar et al.
going for the change of 2D pictures into 3D for better comprehension of understudies
with the assistance of virtual teaching. The essential estimation of enlarged the reality
is that the way where segments of the advanced world mix into a human impression
of this present reality, not as a simple showcase of knowledge, however through the
incorporation of vivid sensations, which are seen as normal pieces of a situation.
The most punctual useful AR frameworks that gave vivid blended reality encounters
to clients were designed within the mid-1990s, beginning with the Virtual Fixtures
framework created at the U.S. Commercial increased reality encounters were first
presented in stimulation and gaming organizations. During this manner, increased
reality applications have adjoin business ventures, for instance, instruction, inter-
changes, prescription, and stimulation. In training, substance can be gotten to by
filtering or review an image with a telephone or by utilizing marker less AR tech-
niques. A model significant to the event business is an AR protective cap for devel-
opment laborers which presentations data about building destinations. The mixture
of AR technology with the academic content creates new style of automated applica-
tions and acts to boost the effectiveness and attractiveness of teaching and learning for
students. Key technologies and methods are discussed within the context of education
[1]. Some of the areas where Augmented Reality plays very vital role:
Medicinal imaging has clad to be imperative for determination and treatment in our
general public. We’ve got been adequately pushing past the cutoff points of innova-
tion we had with the progression of PC innovation. As of late, medicinal preparing
frameworks utilizing computer generated realities, VRs, and blended reality, MR [2]
are created. Medicinal school understudies are required to require various restora-
tive trainings in process and achieve therapeutic aptitudes. It’s normal that utilizing
the PC vision innovation encourages improvement in therapeutic training. We built
up a restorative preparing framework utilizing enlarged reality, AR. By and huge,
there are two sorts of AR usage, marker based AR and markerless based AR [2]. In
our framework, the marker based AR was utilized in light of the actual fact that the
previous technique is easier than the latter, and therefore the last strategy is altogether
constrained by computational power. During this paper, we present our medicinal
preparing framework utilizing AR, including 3D anatomical articles and manipulable
natural interface.
Modern Education Using Augmented Reality 653
Current academic research within the MR sector has targeting technology and user-
interface research but there’s a hunt gap in studying user experiences and decision-
making, technology advancement and application development side-by-side so as to
grasp their value-in-use. The user-value drivers are numerous and will drive appli-
cation development. So far, the key value drivers are identified to be cost-saving
through out-of-home and out-of-office access, total control and high level of person-
alization, going beyond reality, personal efficacy experiences.From now of view, the
main challenge for both VR and AR technologies is to convince users that the added
value is high enough to compete with the present systems and offerings in desktops,
notebooks, tablets, smartphones and related video and game-like applications [3].
3 Design Issues
Presently allow us to see the present framework. The accompanying image demon-
strates the use of Augmented Reality in education however there’s just transformation
of 2D picture into 3D picture. So there are some shots that some understudies might
not comprehend what’s happing. So now here in Fig. 1 as you’ll see that the image of
brain in 2D is just converted in 3D. Some students could also be capable of learning
from this but some student are {going to be are} bit confused in some part so to
address this issue, we are going to propose an new system. In the wake of discussing
current structure presently it is a great opportunity to debate our proposed framework.
Our primary objective is to accomplish a full computerized help to class and easy
learning for understudies. Presently consider a bit school from rustic zone there aren’t
654 V. V. Bandgar et al.
any such an offices accessible like Augmented Reality. So our point is to manufacture
an application for those understudies from rustic zone for viable learning.
Now consider an example of a state like Maharashtra there are around 65,000
government schools are digitized but it includes digital equipment like LCD screens,
projectors, and computer machines so we are thinking grade up to create schools
digitized. Right off the bat, we want to assemble an application for our framework
we are assuming to make android application with assistance of android studio on
the grounds that the majority normal cell phones are android following making an
application now its opportunity to check it. After some review we understand that the
instruction utilizing Augmented Reality the reality is simple to grasp and effectively
able to get a handle on. For more splendid understudy there’s enormous chance to
analyze the unexplored things splendid understudies yet as poor getting a handle on
limit understudies will have the choice to induce a handle on the ideas. Presently we
are going to see some models, now consider a book of images containing pictures
of old creatures with the help of Augmented Reality understudies can see the real
3D picture of that creatures. On the off chance that an understudy can’t comprehend
the nearby planetary group with the help of increase reality we will demonstrate that
understudy the 3D liveliness of close planetary system so these are some instances of
Augment Reality in instruction. Enlarged and computer generated reality are seeing
increasingly noticeable usage in study hall and instructive settings. The configuration
and even the world of the training procedure may be changed through AR within the
years to come. Remembering that 71% of people matured 16–24 all have a mobile
phone, AR may be the subsequent huge thing in instruction. In Fig. 2 the Apple 3D
model is shown supported a target image and display it on our application.
4 Implimentation
There are main two stages in application first stage is Scanning and second one is
Displaying 3D module with some data. Figure 3 describes the system architecture.
Modern Education Using Augmented Reality 655
4.1 Scanning
In the implemented system Vuforia plugin is used to map a perfect image targets.
Like most of the work is done by vuforia like the image scanning matching with
the database. To use vuforia need to register on it’s official site and get the key for
development. Also vuforia provides its own database so after storing data we able to
get the Database file also.
For display purpose or 3D rendering the IDE used is Unity 3D, after getting the
development key and database it’s easy to handle in unity. Firstly need to create some
Targets or ImageTargets on that we can assign the images from Vuforia Database.
And then we have to take specific 3D model and place on specific ImageTarget, so
when we scan that target we will get desired 3D model. Also we playing specific
audio on specific targets.
656 V. V. Bandgar et al.
From this project the training curiosity of scholars are increased and that they learn
efficiently at school similarly as reception. It creates a learning by doing environment.
The following image describes how system is working when we scan ImageTarget
the specific 3D model is displayed (Fig. 4).
The following are some other screenshots of implemented system. If the Letter
K Image Target is Scanned the the Kangaru 3D model displayed and so on (Figs. 5
and 6).
Modern Education Using Augmented Reality 657
6 Conclusion
Fig. 5 K ImageTarget
scanned
Modern Education Using Augmented Reality 659
Fig. 6 S ImageTarget
scanned
References
1. Botella C, Baños RM, Perpiñá C, Villa H, Alcaniz M, Rey A (1998) Virtual reality treatment of
claustrophobia: a case report. Behav Res Ther 36(2):239–246
2. Umeda R, Seif MA, Higa H, Kuniyoshi Y (2017) A medical training system using augmented
reality. In: ICIIBMS 2017, track 3: bioinformatics, medical imaging and neuroscience, Okinawa,
Japan, pp 146–149
3. https://www.emergingedtech.com/2018/08/multiple-uses-of-augmented-reality-in-education/
4. Peng H, Briggs J, Wang C-Y, Guo K, Kider J, Mueller S, Baudish P, Guimbretière F, In RoMA:
ınteractive fabrication with augmented reality and a robotic 3D printer. Cornell University Ithaca,
NY, USA {hp356, jeb482, cw776, kg344, fvg3}@cornell.edu
5. Augmented Reality in Education, https://thinkmobiles.com/blog/augmented-reality-education/
OSS Features Scope and Challenges
Abstract Today oss becomes most popular in software community due to its wide
use. But some issues are also identified by users while using such softwares. Due
to free availability major advantage of such software is even if vendors stop devel-
oping such softwares other programmers can contribute to develop such softwares.
Nowdays there is big challenge to outline the policy for usage of open source soft-
ware. This paper includes challenges faced in the use of oss, features of oss, and
emerging applications of oss.
1 Introduction
Closed source and open source software are two genres of software. But open source
software are become most popular nowadays. Any user can detect open source soft-
ware code, can refine and can enrich the code to meet the growing needs of stake-
holders. Normally users do not have wisdom of code of application they use. Appli-
cation working can be directed by developers as they can operate the code due to its
free availability. Due to code availability programmers can point out flaws in it, can
add features to it in order to increase usability of that software. Most popular open
source software are linux, libreoffice etc. Oss allows programmers collaboratively
work on software to share their knowledge of code, to make changes into the source
code in order to add required features into their project. But open source software
also have issues in its use.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 661
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_66
662 M. K. Jadhav and V. V. Khandagale
2 Literature Survey
In this section we see how computer programmers need to develop open source soft-
ware for their real time projects. In early days of information technology, computers
are with already installed os and software applications. But as the software is going
to become standalone product, software vendors restricts to access their code freely
by other in order to avoid inspection of their important software details. So computer
users are become dependent on software vendors for maintenance, technical support
and inclusion of additional features of software. For this they are pay high cost to
software vendor. As a result of this there was great need of accessing the source
code freely in order to meet the increasing needs of application by modifying and
enhancing that code. So, in early 70s open source software movement was start.
Trainor [1, 2] Richard M Stallman play important role in this movement. He
introduced concept “copy left” which consists General Public License for software
which allows anyone to freely use, distribute and modify the source code at no charge.
Also introduction of Linux by Linus Torvalds forwards the open source software
movement in 1991 [2]. He implements Linux os on Intel based computers. 25% of
servers and 2.8% of desktop computers were use Linux in 2002 [3]. Still companies
were not interested for investing in open source software. In 1998 Netscape releases
source code of its web browser which motivates free software developers to improve
quality of software by working collaboratively together in a way that no company
can achieve this individually. Major open source software project Apache web server
software occupies 60% of the market [4, 5].
3 Applications of OSS
3.3 Shotcut
It is open source program for advanced video editing. It supports no destructing audio
and video editing without any quality loss [6].
3.4 Linux
It is open source operating system. It is released in 1991. Most of servers are devel-
oped in Linux. It becomes popular due to its security and less vulnerability to virus
[6].
4 Features of OSS
Most of people prefer loss because they have more control on it. As source code is
available to all freely people can examine the code to ensure that code do not doing
anything that they do not want it to do. They can make changes in source code as per
requirement of application [7].
4.2 Security
As source code of loss is freely available to all, many people are working on same
piece of code. So if any error is there in loss they all are working on that error to
remove that error. As many eyes on same piece of software, it is less vulnerable to
threats [8].
4.3 Stability
Usually many people prefer loss for long term projects. As source code of oss is
distributed publicly, so even if original vendor [8] of project stop working on source
code then also other programmers can rely on that software tools. So oss gives more
stability as compared to closed source software.
664 M. K. Jadhav and V. V. Khandagale
Non-programmers also get advantage from oss as they can use software for any
purpose as they wish which is not happen in case of closed source software users
have to use software in the way as its vendor thinks [8].
5 Challenges of OSS
It is found that one issue with oss is lack of customer support. One way to overcome
this problem is by giving answers to the questions asked by end user in forums. If
someone found something about software he must share it in forum [9].
It is challenge for business owner to outline a policy for their open source usage.
Without policy developers may use any component for any purpose which may cause
issues down the line. So establishing clear written policy is the best way to avoid
such issues.
As source code of oss is freely available to community, it may be risk for business
owner that hackers may make unwanted changes in that code. So security of that oss
may be challenged [10].
As oss is freely accessible to all, people are more consuming the software for their
business but people are less contributing towards further development of software
[9].
OSS Features Scope and Challenges 665
References
1. Trainor C (2009) Open source, crowd source: harnessing the power of the people behind our
libraries. Program-Electron Lib 43:288–298
2. https://medium.com/fossmec/the-history-of-free-and-open-source-software-for-the-third-gen
eration
3. https://portal.stupica.com/Linuxpedia/Linux_adoption.html
4. https://yourbusiness.azcentral.com/list-different-servers-3142.html
5. Appelbe B (2003) The future of Open Source Software. J Res Pract Inf Tech 35:227–236
6. https://www.techradar.com/in/best/best-open-source-software
7. Kemp R (2009) Current developments in Open Source Software. Compute Law Secure Rev
25:569–582
8. https://opensource.com/resources/what-open-source
9. https://dzone.com/articles/executive-insights-on-the-current-and-future-state-2
10. Abdullah R, Lakulu M, Ibrahim H, Selamat MH (2009) The challenges of Open Source
Software development with collaborative environment 2009. In: International conference on
computer technology and development
Text Summarization and Dimensionality
Reduction Using Ranking and Learning
Approach
Abstract Because of the exponential increment of records on the web, clients need
all the related information in a difficult situation. Finding the significant data from
such huge information is examining undertakings, hence the data recovery turns out
to be progressively crucial for looking through the applicable information viably
These oversee via Automatic content summarization. It is a procedure that perceives
the significant focuses from all the important reports to display a compact summary.
The proper text summarization and dimensionality reduction (TSDR) of summarized
text can lead to a notable reduction in accessing time for the input elements. The
proposed method produces the summarization task by dimensionality reduction with
rank (TSDRR) using training methodology. The consequence of a string of words
in a data text is appraised by the assistant of the PageRank algorithm. The subject
is first pre-processed to tokenize the determinations and perform stemming opera-
tions. Then descent-based text summarization involves selecting determinations of
high connection according to the level of the report depends on word also determi-
nation characteristics and established them collectively to make a report. The test
results determine that this approach has more reliable production than other current
classifications.
1 Introduction
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 667
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_67
668 D. Bartakke et al.
an alternate view of what is imperative. This motivated the requirement for having
an Automatic Summarizer that can play out the activity in less time and with the
slightest exertion. This drove for the examination of Automatic Summarization to
begin over 50 years back [1]. Text Summarization when all is said in done is the
way toward abridging a single article or an arrangement of related ones by summing
up the most imperative occasions, ensuring the occasions succession is coherent by
requesting them chronically. Then again automatic Text Summarization is the way
toward delivering an abbreviated variant of a text by the utilization of computers [2].
The synopsis ought to pass on the key commitments of the text. In Automatic Summa-
rization, there are two primary methodologies that are extensively utilized Extractive
and Abstractive. The primary strategy, the Extractive Summarization, removes up to
a specific utmost the key sentences or passages from the text and requests them in a
way that will create a coherent outline. The removed units contrast starting with one
summarizer then onto the next. Most summarizers utilize sentences as opposed to
bigger units, for example, sections [3, 4]. Extractive Summarization techniques are
the attention strategy on Automatic Text Summarization. The other strategy, Abstrac-
tive Summarization, includes more language-dependent tools and Natural Language
Processing (NLP) innovation. Such summarizers can incorporate words not present
in the first document [5], illustrated in Fig. 1.
Automatic summarization is the means of decreasing a collection of data compu-
tationally, to produce a subset (a summary) that describes the common significant or
appropriate message inside the opening this show in Fig. 1. The aim of this research
work is to proposed automatic text summarization using dimensionality reduction.
The objectives are:
1. To study and analyze different summary readability and learning-based text
summarization and reduction methods.
2. To assess the importance and applications of text summarization and reduction.
3. To design a novel approach for accurate, quick, and meaningful abstractions
extraction from the large text.
4. To propose a novel Text summarization framework based on effective abstrac-
tion extraction.
5. To propose the text reduction technique and combined with text summarization
to enhance accuracy.
2 Literature Review
exposition was obtaining the best sequences of a string of words securing systems
during three brands of records: news, blogs, furthermore articles.
In [7] the founder completed the recommended entrance to make the summariza-
tion assignment by unsupervised training approach. The meaning of a decision in
a data extract was decided by the representatives of the Simplified Lesk algorithm.
As an online semantic dictionary WordNet was practiced. Original, this procedure
assesses the importance of everything individual decisions from a text indepen-
dently utilizing the Simplified Lesk algorithm and provides them in contracting
methods according to their importance. Following, according to the provided section
of summarization, a remarkable character of a string of words was picked from that
required list.
Wikipedia articles obtained provided as input to the operation and extractive text
summarization was performed [8] by recognizing text features and obtaining the
decisions accordingly. The text was first pre-processed to tokenize the decisions
and conduct stemming progress. Then score the sentences using the modified text
features. Two strategies presented were preparing the commands already in the text
and identifying synonyms. These comments along with the conventional techniques
were practiced to obtain the sentences. The numbers were practiced to classify the
sentence to be in the summary outline or not with the maintenance of a neural network.
In [9] author clarified “Enhanced connected and discrete multi-objective shred
swarm optimization for text summarization” Examining down this colossal mate-
rial would be useful in the original drive for numerous assistants. So text summa-
rization structures end up being significantly in separating this huge substance.
The summaries are created in perspective on basic highlights using multi-target
approaches where sufficient composing isn’t open. Noteworthy imprisonments of
substance layout systems are versatility and execution.
In paper [10] a query-oriented text Summarization procedure used for deci-
sion extraction was introduced. In the extricate summarization method, the various
instructive decisions in the text were recognized and decided to revisit in the report.
To recognize decisions containing relevant information, a set of suitable pieces were
extricated of that text. Whatever the selected pieces of the decisions were also suit-
able, the numerous informative decisions were more precisely recognized and the
property of the produced report increases. They have determined that the application
of more becoming features commands to enhanced summaries created. To estimate
the automatically created summaries, the ROUGE pattern has been practiced.
In [11] author recommended the application of the Firefly algorithm (FF) through
that descent of reports of original Arabic records. The recommended procedure was
connected with two evolutionary methods that use genetic algorithms and harmony
exploration. A compound of informational and grammatical numbers was practiced.
The suggested FF approach was applied to decide the optimal sub-path from appli-
cant ways in the design. Several routes in the plan describe a hopeful summary, so
the FF algorithm engaged to select the optimal track. The EASC Corpus moreover
significant ROUGE toolkit obtained practiced for the examiner of the recommended
entrance.
Text Summarization and Dimensionality Reduction … 671
3 Methodology
Text Summarization implies an existing range of analysis identified together with the
Information Retrieval including Natural Language Processing districts. Text Summa-
rization occurs frequently remaining worked in the marketing department so as the
data mining, Communications communication management about text databases, for
web-based data retrieval, in word Processing accessories. Various methods contrast
in the presence of their weight formulations. Automatic text summarization holds
an essential measure for information management responsibilities. It explains the
predicament of choosing the usual relevant divisions of some text. In this paper
proposed method, TSDR without ranking sentences system performed high-quality
summarization requires sophisticated NLP techniques. The following steps to text
summary generation are:
1. Examine the Data
2. Divide Text into Sentences
3. Divide Text into Sentences
4. Text Preprocessing Vector Representation of Sentences
5. Similarity Matrix Preparation
6. Applying the PageRank Algorithm
7. Summary Extraction.
Title word: An unusual rate remains provided before the decision if this includes
information transpiring in that string of words as the principal text of the report is
dispatched via the right information. That characteristic is measured as regards:
N o.o f titlewor d
Scor e(S1) =
N o.o f totalwor d
Sentence Length: Reduce the string of wordsthat are likewise small such as time-
line or designer handles. During each string of words, the normalized period of a
string of words denotes determined because:
Sentence Position: The sentences happening beginning in the section have the
most powerful score. If the paragraph becomes p sentences, the average of every
sentence is determined:
p
S3(Sentences1) =
p
4 3 2
S3(Sentences2, Sentences2, Sentences3) = , ,
5 5 5
Text Summarization and Dimensionality Reduction … 673
Numerical data: The string of words should statistical results can emulate relevant
numerical of the report and suddenly chosen for a report. Its amount is computed as:
Inflected words: These are field-specific information including the greatest poten-
tial relationship. The degree of the product of thematic reports that happens in a
string of words over the highest quantity of thematic information in a string of words
provides the number of individual characters as:
Frequent Nouns: The relevant string of words implies that string of words that
include the greatest amount of individual names. Its composition is provided through,
The architectural flow diagram (explained in Fig. 1) of the process of text summary
generation and dimensionality reduction represent in Fig. 1. In the first phase are
preprocessing that acquired the input document from datasets and these read data
transform separate each sentence into words using segmentation and tokenization
strategy. After that remove the unnecessary word and highlighted important words
store in the matrix array list form. The analysis of input data and match similar
674 D. Bartakke et al.
keywords including each word assign a decreasing string of words to score for
word and sentences put together to generate a summary by select proper words
and sentences.
Pre-processing: In the pre-processing state is much unnecessary and excessive infor-
mation instant or noisy and misleading data, and then knowledge acquired during
the preparation period is more complicated. Data preparation and filtering measures
can take a substantial amount of processing time. Data pre-processing comprises
cleaning, Instance collection, normalization, conversion, feature extraction, and
reading, etc. The output of data pre-processing is the final training set.
There are four levels in preprocessing. The division is a strategy for dissemi-
nating a given report into sentences. Prevent words are released from the content.
Stop words are generally happening words, for example, ‘an’ a’, the’ that produces
less significance and controls clamor. The Stop words are predefined and gathered
in a cluster. Tokenization will circulate the information content into discrete tokens.
Accentuation checks, spaces, and word eliminators are the words uncovering noto-
rieties. Word Stemming is acknowledged to all novices’ words into its root structure
by pushing its prefix and postfix for a relationship with different words.
Feature Extraction: In feature descent state obtained from the original data, but
the selected content is not changed in any way. Representatives of selected content
introduce key-phrases that can be related to “tag” or index a text report or key
sentences (including headings) that collectively compose an abstract, and illustrative
images or video segments, as declared above. For text, ancestry is comparable to
the method of skimming, where the report (if available), headlines and subheadings,
personalities, the first including last paragraphs of a section, and optionally the front
and last decisions in a paragraph are read ere one wants to read the complete report
in detail.
The writing report is forwarded by collection, D = {S1, S2, …, Sk} wherever
Si imports a string of words checked in report D. The report is controlled to high-
light descent. The significant information and string of words characteristics to be
applied are determined. That commitment applies highlights such as Title infor-
mation, Sentence season, Sentence place, numerical data, Term weight, sentence
agreement, the occurrence of Thematic reports furthermore Proper Terms.
Semantic analysis: Semantic analysis is the responsibility of guaranteeing that the
reports and statements of a schedule are semantically accurate, i.e., that their purpose
is clear and compatible with how to control buildings and data types are assumed to
be practiced.
Latent Semantic Analysis (LSA): LSA does a mathematical representation of
term mode which connects the grammatical relationship among sections of textual
knowledge. LSA is produced to promote the benefits of message retrieval arrange-
ments by utilizing the “semantic” text of messages in a query as engaged in giving
through name equivalent. LSA circumvents the difficulties of synonymy, in which
various words can be applied to represent the likewise grammatical thought.
Text Summarization and Dimensionality Reduction … 675
Sentences Selection: Based upon the string of words includes all the string of words
are ranked in settling position. Sentences with the most leading number are selected
as a release summary. Certainly, the string of words is completed in the sequence
they arrive in the primary record.
Based on sentence records all the decisions are ranked in decreasing string of
words. Sentences with the most distinguished score are selected as a document
summary. Ultimately, the sentences, in summary, are organized in the string of words
they transpire in the original document.
Step 2: Division of each sentence and assign each sentence to a symbol. The
selected sentences marked as x and non-selected sentences marked as y moreover
the text of each sentence into datasets.
Step 3: Determine the amount of topic name data meaningful in both marked x
string of words and y string of words. Assemble these utilities through the database
report.
For scoring sentences, rundown portrayal approaches figure scores dependent on
how well the sentence communicates a portion of the significant subjects in the report
or joins the points. Marker portrayals score sentences dependent on joining proof
from various pointers utilizing machine learning strategies. For instance, PageRank
doles out loads to sentences by applying stochastic techniques to the chart portrayal
of the archive. Approaches for choosing sentences for the synopsis incorporate best n,
maximal minor importance, and worldwide determination. In the principal approach,
the top n positioned sentences that have the ideal rundown length are chosen. In
the second, sentences are chosen to utilize an iterative insatiable system, which
recomposes sentence scores after each progression. The third methodology chooses
sentences utilizing improvement techniques.
The suggested framework TSDRR centers around the extractive based worth posi-
tion utilizing the summarization strategy. We will talk about two new methodologies
for outlining the content. The primary strategy is displayed in Fig. 2 the recurrence
of the words dependent on the root type of the word, and furthermore the recur-
rence of its equivalent words present in the content. The subsequent strategy is
to recognize sentences containing references or references and give them a higher
weight. As appeared in Fig. 3 the summarization framework accepts contribution
as DUC2002 datasets, forms it and gives the rundown sentences. The information
record comprises of crude information to be prepared by the framework. The initial
step of the summarization is known as tokenization, which separates the sections into
sentences and every one of these sentences is additionally broken into a lot of words.
The information structure got after tokenization of sections is a rundown, containing
every component as a sentence and information structure got after tokenization of
sentences is a rundown of the rundown, containing sets of words. Tokenization is
performed through example coordinating utilizing ‘customary articulations’. Infor-
mation acquired as a lot of words is additionally broke down and stop words or most
normally happening words, similar to articles, are expelled from the arrangement of
words.
Text Summarization and Dimensionality Reduction … 677
Fig. 2 Process of generate summary and dimensionality reduction using the TSDR system
Fig. 3 Architecture of generate summary and dimensionality reduction using the TSDRR system
2002 produced 60 reference sets. The particular set sustained documents, single-
document summaries, furthermore multi-document outlines/selections, including
attitudes described by many types of models such as issue positions, biograph-
ical collections, etc. The reports were prepared including and without “sentences”
designated as determined by a transcription of the single sentence severance soft-
ware utilized for DUC 2001. These datasets general summary of the report with a
length of around 100 words or more limited was conceived. (Whitespace-delimited
tokens). The coverage metric selected ranges into account including compensated
Text Summarization and Dimensionality Reduction … 679
5 Conclusion
In this paper proposed the text summarization and dimensionality reduction TSDR
and advanced TSDRR with a ranking of sentences. In the extractive summarization
Text Summarization and Dimensionality Reduction … 681
strategy, the most educational sentences in the content are recognized and chose to
go to the outline. To distinguish sentences containing significant data, a lot of fitting
highlights are extricated from the content. The result shows the proposed system
TSTR and TSDRR are more efficient and reliable as compared to the ROUGE, Fuzzy
logic and ANFIS system. In assessment, we have utilized estimates like accuracy,
review, and F-score which are utilized essentially with regards to data recovery. Later
on, we mean to utilize more summarization-explicit systems to quantify the efficiency
of our scheme.
References
1. Wei T, Lu Y (2015) A semantic approach for text clustering using WordNet and lexical chain.
Expert Syst Appl 42
2. Bui DDA, Del Fiol G (2016) Extractive text summarization system to aid data extraction from
full text in systematic review development. J Biomed Inform 64
3. Kiyoumarsi F (2015) Evaluation of automatic text summarizations based on human. Proc Soc
Behav Sci 192:83–89
4. Johna A (2015) Vertex cover algorithm based multi-document summarization using information
content of sentences. Proc Comput Sci 46:285–291
5. Babara SA, Patil PD (2015) Improving performance of text summarization. Proc Comput Sci
46:354–363
6. Ferreira R, Freitas F, Cabral L de S, Lins RD, Lima R, Franca G, …, Favaro L (2014) A context-
based text summarization system. In: 2014 11th IAPR international workshop on document
analysis systems
7. Pal AR, Saha D (2014) An approach to automatic text summarization using WordNet. In: 2014
IEEE international advance computing conference (IACC)
8. Hingu D, Shah D, Udmale SS (2015) Automatic text summarization of Wikipedia articles.
In: 2015 international conference on communication, information & computing technology
(ICCICT)
9. Priya V, Umamaheswari K (2018) Enhanced continuous and discrete multi-objective particle
swarm optimization for text summarization. Received: 12 February 2018/Revised: 20 March
2018/Accepted: 22 March 2018 Springer Science+Business Media, LLC, part of Springer
Nature 2018
10. Afsharizadeh M, Ebrahimpour-Komleh H, Bagheri A (2018) Query-oriented text summariza-
tion using sentence extraction technique. In: 2018 4th international conference on web research
(ICWR)
11. Al-Abdallah RZ, Al-Taani AT (2019) Arabic text summarization using firefly algorithm. In:
2019 amity international conference on artificial intelligence (AICAI)
12. Thu HNT, Huu QN, Ngoc TNT (2013) A supervised learning method combines with dimen-
sionality reduction in Vietnamese text summarization. In: 2013 computing, communications
and it applications conference (ComComAp)
13. Chatterjee N, Mohan S (2007) Extraction-based single-document summarization using random
indexing. In: 19th IEEE international conference on tools with artificial intelligence (ICTAI
2007)
14. Krishnaveni P, Balasundaram SR (2017) Automatic text summarization by local scoring
and ranking for improving coherence. In: 2017 international conference on computing
methodologies and communication (ICCMC)
15. Yang Z, Yao F, Fan K, Huang J (2017) Text dimensionality reduction with mutual information
preserving mapping. Chin J Electron 26(5):919–925
682 D. Bartakke et al.
16. Liu S-H, Chen K-Y, Chen B, Wang H-M, Hsu W-L (2017) Leveraging manifold learning for
extractive broadcast news summarization. In: 2017 IEEE international conference on acoustics,
speech and signal processing (ICASSP)
17. Meena YK, Gopalanib D (2015) Domain-independent framework for automatic text summa-
rization. Proc Comput Sci 48:722–727
Properties of Extended Binary Hamming
[8, 4, 4] Code Using MATLAB
Abstract The main aim of this paper is to study various properties of extended binary
Hamming [8, 4, 4] code, when we know its generator matrix. Using MATLAB, we
can study syndrome decoding, weight of a codeword, error correction and error
detection of binary Hamming [8, 4, 4] code.
1 Introduction
In the late 1940s, Claude Shannon has started development of information theory
and coding theory as a mathematical model for communication. At the same time, R.
Hamming found a necessity for error correction in his work on computers. Already
Parity checking was used to detect errors in the calculations of the relay-based
computers of the day, and Hamming realized that a more sophisticated pattern of
parity checking allowed the correction of single errors along with the detection of
double errors. The codes that Hamming put forth, were important for theoretical
and practical reasons. In [1, 2], algorithmic approach for error correction has been
studied for few codes.
Each binary Hamming code [3] has minimum weight and distance 3 and that of
extended binary Hamming code has weight and distance as 4. In this paper, we are
going to study some properties of extended binary Hamming [8, 4, 4] code.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 683
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_68
684 N. S. Darkunde et al.
2 Preliminaries
Let Fq denote the finite field with q elements, where q is some power of a prime. A
linear [n, k]q —code is a k-dimensional subspace of Fqn . The parameters n and k are
referred to as the length and dimension of the corresponding code.
Example The subset C = {000, 001, 010, 011} of vector space F23 is [3, 2]2 linear
code. Similarly, C = {0000, 1100, 2200, 0001, 0002, 1101, 1102, 2201, 2202} is
[4, 2]3 linear code.
Let x, y ∈ Fqn , The Hamming distance from x to y, denoted by d(x, y), is defined
to be the number of places at which x and y differ.
For a [n, k]q code C containing at least two words, the nonnegative integer given
by min{d(x, y) : x, y ∈ C, x = y} is called minimum distance of C. It is denoted by
d(C).
Example For a code C = {0000, 1000, 0100, 1100} in F24 , we see that d(C) = 1.
Definition 3 (Dual Code) [3] Given a [n, k]q code C in Fqn , the subspace
The binary linear code generated by matrix G is called extended binary Hamming
[8, 4, 4] code.
686 N. S. Darkunde et al.
Communication fails due to the error in the communication channel. In this process,
receiver receives the original message with error in it. If the system knows how much
error has come with original message then that error can be removed. It is difficult
task to correct the errors in the communication; hence we have used MATLAB to
accomplish this task.
4 Conclusion
Once we know the generator matrix of any linear code, using MATLAB we can have
many of the things that can be discussed. In this paper, we have studied extended
binary Hamming [8, 4, 4] code with the help of MATLAB. The parity check matrix in
standard form, for extended binary Hamming [8, 4, 4] code is calculated and using it,
syndrome of a particular codeword is calculated. We have encoded the message and
decoded it correctly by using MATLAB. Using MATLAB, it has been verified that,
the distance of an extended binary Hamming code is 4 and it is 1-error-correcting
code.
References
1. Patil AR, Darkunde NS (2019) Algorithmic approach for error-correcting capability and
decoding of linear codes arising from algebraic geometry. In: Information and communication
technology for competitive strategies. Springer, Berlin, pp 509–517
2. Patil AR, Darkunde NS (2017) On some error-correcting linear codes. Asian J Math Comput
Res 56–62
3. Ling S, Xing C (2004) coding theory-a first course, 1st edn. Cambridge University Press
4. Huffman WC, Pless V (2010) Fundamentals of error-correcting codes. Cambridge University
Press
5. MacWilliams FJ, Sloane NJA (1977) The theory of error-correcting codes. Elsevier
6. Pratap R (2017) Getting started with MATLAB: a quick introduction for scientists and engineers,
7e. Oxford University Press.
Identification of Fake News on Social
Media: A New Challenge
1 Introduction
Fake news poses various problems in several media from derisive stories to fake news
and policy information preparation. Fake news causes huge issues for our culture. A
fabricated story by default is “fake news” but the discourse of social media latterly
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 689
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_69
690 D. V. Patil et al.
shifts its meaning. A number of them are currently using the phrase to refute the
evidence against their most common points of view.
Today, detecting fake news on social media is a major challenge in the world.
First, fake news is deliberately published to misinform readers into believing false
facts, making it impossible to identify news content based on it. But, to help distin-
guish it from the real news, we need to provide contextual details, such as user
social interactions on social media. Second, leveraging this additional information
is non-trivial in and of itself as the social connections of consumers of false news
generate data that is immense, incomplete, unstructured and noisy. The Internet and
social media made it much easier and more convenient to access news information.
Internet users may also follow the events of their interest in online mode, and mobile
device spreading makes this process much more convenient. Many scientists think
machine learning will solve the fake news problem. There’s an explanation for this:
artificial intelligence algorithms have recently started to work much better on a lot
of classification issues (image recognition, voice detection, and so on) because the
equipment is cheaper and there are bigger data sets.
Figure 1 shows general flow of followed algorithm.
2 Literature Review
A. Existing System
Today, for days, various academic bodies as well as the government working on the
topic of machine learning techniques for fraud detection have concentrated mainly on
classifying user comments and publicly accessible social media messages. In 2016,
the issue of deciding on fake news has been the subject of specific attention within
the literature during the American presidential election.
Conroy et al. [2] outline several approaches that seem optimistic towards the
aim of correctly identifying the misguiding news. They note that straightforward
content-related n-grams and shallow piece-of-speech (POS) tagging have well-tried
lean for the classification task, usually failing to account for crucial discourse data.
Preferably, these methods have been shown useful only in tandem with more complex
ways of analysis. Feng et al. [3] can attain 85–91% accuracy in deception related
classification tasks using online review corpora.
B. Proposed System
Acquainting semantic examination with Credulous Bayes and SVM classifier can
improve the calculation [7]. The greatest disadvantage of Guileless Bayes classifier
is that it thinks about all highlights of an archive, or whichever literary arrangement
being utilized, to be free despite the fact that more often than not excessively isn’t
the circumstance. This is an issue because of brought down precision and the way
that connections are not being educated if everything is thought to be isolated. As
we referenced before, one of the primary preferences of semantic investigation is
that this strategy can discover connections among words. In this manner, including
semantic investigation helps fix one of the principle shortcomings of the Innocent
Bayes classifier [8].
Adding semantic examination to SVM can improve the working of the classifier.
In “Support Vector Machines for data divided Based on Latent rules Indexing, “the
creator shows that combine the two strategies improves the productivity because of
“concentrating consideration on Help Vector Machines onto information subspaces
of the element spaces,” [7]. In the examination, semantic investigation had the option
to catch the “hidden substance of the report from a semantic perspective,” (Huang
2001), [9]. This will improve the effectiveness of SVM since the strategy would lose
less of its time characterizing good for nothing information and invest more energy
arranging pertinent information with the assistance of semantic investigation. As
laid out before, a more advantage of semantic investigation is its capacity to sepa-
rate basic information through connections between words; subsequently, semantic
examination can utilize its essential favorable position to improve SVM further [1].
3 Methodology
Online life for news utilization is a twofold edged sharp edge. From one perspective,
its minimal effort, simple access, and quick scattering of data lead individuals to
search out and devour news from online life. Then again, it empowers the across
the board of “counterfeit news”, i.e., Low-quality news with intentional bogus data.
The abundant spread of phony news has the potential for amazingly negative effects
on people and society. Along these lines, counterfeit news identification of online
networking has as of late become a developing examination that is drawing in gigantic
consideration.
Counterfeit news discovery of online life presents one of a kind qualities and
difficulties that make existing location calculations from customary news media
failed or not material. Initially, counterfeit news is deliberately composed to mislead
perusers to accept bogus data, which makes it troublesome and nontrivial to identify
dependent on news content; along these lines, we have to incorporate transitory data,
for example, client social commitment via web-based networking media like Twitter,
Facebook, and so on, to help make an assurance. Second, using this assistant data
Identification of Fake News on Social Media: A New Challenge 693
is trying all by itself as clients’ social commitment with counterfeit news produces
information that is huge, inadequate, unstructured, and loud.
3.2 Definition
(Fake News Detection) the social-news, engagement E among n users for the news
article, the task of fake news detection is to predict whether the news article a is a
fake news part or not, i.e., F: E {0, 1}.
Such that,
• There are two pieces of information expansion process, “counterfeit news” and
“genuine news”. Social occasion the phony news was basic as Kaggle gave a phony
news dataset comprising 13,000 articles uncovered all through the 2016 political
race cycle. It includes huge work around a few Locales because of it had been
the sole gratitude to do web scratching a large number of articles from changed
sites. By scratching an aggregate of 5279 articles, the genuine news dataset was
made, generally from media associations (New York Times, WSJ, Bloomberg,
NPR, and the Watchman) which are distributed around 2015–2016.
• Online news can be assembled from various sources, for example, news asso-
ciation landing pages, web indexes, and internet based life. Be that as it may,
physically choosing the truth of stories might be a troublesome errand, ordinarily
difficult to-satisfy annotators with space experience who plays out a cautious
investigation of Cases and additional confirmation, setting, and reports from Legit-
imate sources. Generally, news, information with explanations can be gathered in
the accompanying manners: Master columnists, Actuality checking sites, Industry,
indicators, and Publicly supported Laborers. In any case, there are no settled upon
benchmark Datasets for the phony news identification issue.
694 D. V. Patil et al.
Data preprocessing is a data mining technique that involves transforming raw data
into an understandable format. Real-world data is often incomplete, inconsistent,
and/or lacking in certain behaviors or trends, and is likely to contain many errors.
Data preprocessing is a proven method of resolving such issues. Data preprocessing
prepares raw data for further processing.
It is an integral step in Machine Learning as the quality of data and the useful
information that can be derived from it directly affects the ability of our model to
learn; therefore, it is extremely important that we preprocess our data before feeding
it into our model as shown in Fig. 2.
A confusion matrix is a summary of prediction results on a classification problem.
The number of correct and incorrect predictions are summarized with count values
and broken down by each class. This is the key to the confusion matrix.
Identification of Fake News on Social Media: A New Challenge 695
The confusion matrix shows the ways in which your classification model is
confused when it makes predictions. It gives us insight not only into the errors
being made by a classifier but more importantly the types of errors that are being
made as shown in Figs. 3 and 4.
News content highlights portray the Meta data identified with a bit of news. A
rundown of Illustrative news content characteristics are recorded beneath:
696 D. V. Patil et al.
Since counterfeit news endeavors to spread bogus cases in news content, the most
direct method for recognizing it is to check the honesty of incredible cases in a
news story to choose the news veracity. Information based methodologies mean to
utilize outside sources to actuality check proposed to demand news content. The
objective of certainty checking is to designate a fact incentive to a case in a particular
setting. Reality checking has pulled in raising consideration, and numerous endeavors
have been made to build up a practical computerized certainty checking framework.
Existing reality checking can arrange as master situated, publicly supporting focused,
and computational-arranged.
Identification of Fake News on Social Media: A New Challenge 697
Suppose Consider a document containing 100 words wherein the word apologist
appears 3 times.
The term frequency (i.e., tf) for apologist is then
(3/100) = 0.03.
IDF: Inverse Documents Frequency, which measures how important a term is.
While calculating TF, all terms are considered equally important.
However, it is known that specific terms, such as “is”, “of”, and “that”, may appear
many times but have little importance.
IDF(t) = log_e(Total number of file (documents)/Number of documents with term
T in it).
Assume we’ve ten million documents and also the word seems in one thousand
of those. Then, the inverse document frequency (i.e., idf) is computed as:
log(10,000,000/1,000) = 4.
Thus, the Tf-idf is the product of these two quantities: 0.03 * 4 = 0.12.
Count Vectorizer
The Count Vectorizer provides a simple way to both tokenize a collection of text
documents and form vocabulary of known words, but also to encode new documents
using that vocabulary. You can use it as follows build an instance of the Count
Vectorizer class.
If you implemented CountVectorizer on one piece of documents and then you want
to use the set of features of those documents for a new set, use the vocabulary_attribute
of our original Count Vectorizer and pass it to the new one.
Test Train Split
Train-Test Splits. The data we use is usually divided into training data and test
data. The training set contains a known output, and the model learns on this data in
sequence to be generalised to other data later on. We have the test dataset (or subset)
in series to test our model’s prediction on this subset.
Test Output
Better fitting of the training dataset, as opposed to the test dataset, usually point in
fitting. Test pieces are therefore a set of examples used only to assess the performance
(i.e. Generalization) of a fully specified classifier. Holdout sample is a sample of data
not used to fit a model, used to assess the performance. Of that model; this book uses
the term validation pieces or if one is used in the problem, test set. Instead of a
holdout sample.
Identification of Fake News on Social Media: A New Challenge 699
4 Result
As you can see by running the cells below, both vectorizers extracted the same tokens,
but obviously have different. Likely, changing the max_df and min_df of the TF-IDF
vectorizer could alter the result and lead to different features in each. Compare TF-
IDF versus bag-of-words. My intuition was that bag-of-words (CountVectorizer)
would perform better with this model.
700 D. V. Patil et al.
With the increasing quality of social media, a lot of people use news from social
media rather than traditional news media. However, social media have also been
used to spread fake news, which has strong adverse effects on individual users and
broader society. In this project, we explored the fake news problem by reviewing
existing literature in two phases: characterization and detection.
In the characterization phase, we introduced the basic concepts and principles of
fake news in both traditional media and societal media. In the detection phase, we
reviewed existing fake news detection approaches from a data mining perspective;
it includes feature extraction and model construction. We also further discussed the
data sets, evaluation metrics, and promising future directions in fake news detection
research and spread the field to other applications.
References
1. Rubin V, Conroy N, Chen Y, Cornwell S (2016) Fake news or truth? Using satirical cues to
detect potentially misleading news. In: Proceedings of the second workshop on computational
approaches to deception detection. https://doi.org/10.18653/v1/w16
2. Conroy NJ, Rubin VL, Chen Y (2015) Automatic deception detection: methods for finding fake
news. Proc Assoc Inf Sci Technol 52(1):14
3. Chen Y, Conroy NJ, Rubin VL (2015) News in an online world: the need for an automatic crap
detector. Proc Assoc Inf Sci Technol 52(1):14
4. Fake news websites (n.d.) en.wikipedia.org/wiki/Fakenewswebsite. Accessed on 6 Feb 2017
5. Conroy N, Rubin V, Chen Y (2015) Automatic deception detection: methods for finding fake
news. Proc Assoc Inf Sci Technol 52(1):1–4
6. Naive Bayes spam filtering (n.d.) https://en.wikipedia.org//wiki/NaiveBayesspamfiltering.
Accessed on 6 Feb 2017
7. Wang AH (2010) Dont follow me: spam detection in Twitter. In: Proceedings of security and
cryptography international conference (SECRYPT)
8. Memon AA, Vrij A, Bull R (2003) Psychology and law: truthfulness, accuracy and credibility.
Wiley
9. Joulin A, Grave E, Bojanowski P, Mikolov T (2016) Bag of tricks for efficient text classification
A Smart and Secure Helmet for Safe
Riding
Abstract Life is becoming more fast and hazardous while driving. Moreover life is
valuable so, we need to have some automation techniques to secure life. In this paper
we attempted to plan our thought in which the system that can recognize the person
worn or not the helmet, spot accident place and immediate response, finding whether
the person consumed liquor while riding, identify petrol level of tank and predicting
the crash or collision between the vehicles in order to avoid road side accidents. The
interaction between bike and helmet part takes place wirelessly by making utilization
of RF transmitter and receiver. RF transmitter is appended at helmet and receiver at
bike.
1 Introduction
Road accidents are incredibly increasing day by day, and in countries like India
where bikes are more in use due to low price when compared to other vehicles many
people die due to carelessness by not wearing helmet. Even though there have been
continuous awareness program from the government authorities regarding helmets
and seat belts a majority of the riders not follow the rules. In order to put an end to
accidents of bike we have planned to develop the smart helmet for bike, in which the
helmet not allow rider to ride the bike without wearing helmet. This is the best way
to make helmet is compulsory for bike riders. Other than this there are four more
features alcohol detection, identifying accidents, petrol level indication, and obstacle
detection to avoid the collision between the vehicles. Nearly a third of all fatal bike
accidents happen after riders have been drinking. Head on collisions: More than 75%
of all motorcycle crashes due to head on collision. Exceeding speed of limit: Some
32% of motorcycle accidents that resulted in death involved speeding. Left turns:
The road can turn into a danger zone when a car makes a left-hand turn. In 42% of
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 703
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_70
704 R. Kagalkar and B. Hunshal
all fatal car accidents, the other vehicle was turning left while the biker was either
heading straight, passing or overtaking. Running into objects: In some cases accident
happen without knowledge of rider due to objects or vehicles around bike.
The rest of the paper is outlined into 5 sections, where Sect. 2 presents the outline
of literature survey. The overview of proposed system and it’s irs flow diagram
given in Sect. 3. In Sect. 4 the methodologies used to developed is discussed. The
implementation and results are discussed with case studied where presented in Sect. 5.
Finaly the conclusion is made for development of work.
2 Literature Survey
The paper [1] aim is to provide a in time treatment to person by making use of
ambulance. Many a times when the rider needs an ambulance that time ambulance
may not be available and no one available at the time of accident it often causes deaths.
So present system is prepared in such a way that it gives the notifications at the earliest
so the any responsible person may take any required action for this Aurduino used
ad micro-controller. İn paper [2] aim is to save the people when the rescue workers
get harm in the disaster the Smart helmet respond to the current accident. Using
TCP and UDP socket connection communication is made between Smart helmet
and the device. It provides the disaster safety and secure helmet for the workers. In
real time problems the Smart helmet is useful through which the other person can get
the rescue worker location at the time of disaster event occur. In paper [3] building
a shield which ensures safety of the rider by implementing feature such as alcohol
detection, accident identification, location tracking and fall detection. Force sensing
resistor is used to sense the actual human touch, to ensure if the helmet is worn by the
rider. Accelerometer ADXL345 is used to measure the static acceleration of gravity.
This unit senses if the bike is falling. And bike will take decision that accident has
occurred or not. The paper [4] gives solution to help a rider in need, after the rider has
met with an accident. Raspberry Pi zero it supports multiple os such as linux based,
Raspbian OS, Windows 10 IoT cosre, Google’s Android Things, etc. Pressure sensor
it measures the pressure with which object falls. MCP 3008 IC it is a 10-bit analog
to digital converter. Google’s Firebase Cloud messaging is a crossplatform solution
for messages and notifications for Android, iOS and web applications, which is free
of cost. E Android is a mobile operating system designed for smart phones. In [5]
build protective helmet for miners. The system is used by the miners and this safety
helmet is used for detecting the health of the person and also it provides the same
information to the central office outside the mines. Helmet consists of the wireless
sensor devices which consists of the force sensor and that sensor detects the dangerous
load experienced by the miners on his head if that was heavy load then the wireless
sensors inform about the danger to the center with the help of the room manager. After
knowing the information center can help miners by sending team. The model can
be consists of the miners, Room manager and center. The center is connected to the
GUI which helps to send/Receive message from room manager. The sensor node will
A Smart and Secure Helmet for Safe Riding 705
Figure 1 shows complete functioning of proposed system that is how the decision
is made. Input to system is helmet and processing starts from this part. When the
helmet is worn then helmet is detected, if yes bike starts else bike won’t starts. Same
way if alcohol is detected if yes then bike stops else bike will be in running state. In
accident detection, the system continuously monitors and observes for an accident.
At the time accident detected if yes message is quickly sent to pre stored number
else bike is in running state. Petrol level is monitored if low petrol bike stops. And
if any obstacle is detected message is displayed on LCD. This process of execution
is represented in the form of flow chart.
706 R. Kagalkar and B. Hunshal
4 Methodologies
There are five modules are considered in the system. The heart of the system is Micro-
controller, that is Atmega 328 which has 28 pins. All sensors used in implementation
are connected to input pins of microcontroller. Figure 2 shows sytem setup of the
smart and secure helmet system.
By taking input from different types of sensors the microcontroller will produce
the output. To set some conditions and requirements data, we have developed code in
embedded C language which will be dumped on to the ATmega328 microcontroller.
Arduino software is used in order to execute embedded C code.
A Smart and Secure Helmet for Safe Riding 707
As shown in Fig. 2, push limit switch is used in order to detect the helmet is worn or
not. The push switch is attached in such a way that when person wear the helmet the
push switch is pressed at that time RF transmitter send signal that analog signal is
converted to digital by ADC after that binary 1 is generated that is transmit to bike
part where in RF receiver receives after that bike is going to start. And if RF receiver
of bike part receives binary 0 then bike will not start.
• Alcohol Detection Module
Alcohol detection is done by the MQ3 gas sensor which senses the alcohol fumes
from the breath of the person. So, if alcohol is detected then immediately bike stops
there is no chance of riding bike by consuming alcohol. MQ3 gas sensor is as shown
in Fig. 3. MQ3 is sensor which senses the gases produced by alcohol.
• Accident Detection System
Magnetic sensor will helps to detect the accident and GSM modem will immediately
send the SMS to the predefined numbers. By this we can provide proper facility to
the person who got in to an accident. When a person got into an accident at that time
helmet send the signal to bike part and then GSM receives signal to send the message
to pre-defined number to give a in time treatment to injured person.
708 R. Kagalkar and B. Hunshal
R
e
c
o
g
ni
Ɵ
o
In this module it helps the rider to know about petrol level of bike by indication. And
the levels are low level, middle level, full level and over full level. If petrol tank has
less petrol level then indication is made by buzzer and display message on LCD.
• Obstacle Detection Module
Collision of vehicle also leads to accidents to avoid that IR sensor is used at backside
of bike in order to detect obstacles at back side. So, here sensor will detect obstacle
and alert is given by the buzzer sound and message on LCD display.
Execution is done to check whether the system is working properly or not. The
performed tested and evaluated by testing the system for various focibilities of
outputs. In this testing each unit is tested to check the performance and efficiency
of the unit. To show the efficiency of whole system, with respect to all objectives
processing time and response time are calculated in the following sections.
• Helmet Detection and start of the Bike:
The rider can start the bike only if he wears the Helmet properly. Processing time of
Helmet may very according to the model as well as its size. Therefore, processing
time can be defined as, the time required to recognize the Helmet’sworking and
glowing of the Blue LED over Helmet from RED. Recognition rate is considered in
percentage, and it is with respect to successful start of the bike when a rider wears
the Helmet. In our system, the Helmet as well as bike both are fixed with Red and
A Smart and Secure Helmet for Safe Riding 709
Blue LED’s for different combinations of outcomes. The different cases considered
here are,
– If the applied Key of bike is proper then Blue LED will glow, else Red LED will
glow at Bike.
– For proper wear of Helmet, Blue LED over Helmet glows, else Red LED will
glow.
Cases to consider the Recongition rate:
– If a LED over Helmet and Bike both are Blue, it is 100% recongition rate.
– If LED of Helmet is RED or rider not having helmet and Bike is Blue, then it has
been considered as 50% recognition rate.
It means the rider either not having helmet or not properly weared it. The bike
LED is Blue because, the Key applied to start the bike is proper one.
– In case, if the rider had weared the helmet properly, but applied wrong key then
it is 50% recognition rate.
– In case, if the rider nighter weared the helmet properly, nor applied the proper key
it will be considered as 0% recognition rate.
In Fig. 3 shows an helmet recognition time of the system and in Fig. 4 shows an
Helmet processing time of the system.
• Alcohol Detection and Result Analysis:
Alcohol sensor MQ3 is used in the system to guard the rider from an accident. When
a person is trying to ride the bike at that time if alcohol fumes are detected from
breath of the person then LCD should display “ALCOHOL DETECTED” and bike
automatically have to stop. The below graphs shows the test cases with respect to
percentage of alcohol the rider has consumed. The alcohol consumption exceeds
the configured threshold value of alcohol then, bike wouldn’t start. Threshold value
considered here is 25%. The bike is configured with two LED’s. If alcohol detected
is above the threshold value, then Red LED will glow, else Blue LED will glow and
rider can run the bike.
Alarm and Indicating Unit
An alarm is used as buzzer thatbuzzes when a rider is detected by alcohol. A PS
series buzzer is used. A DC motor is used for demonstration of engine locking. For
the same DC motor is joined to pin number 9 of microcontroller. When alcohol is
detected during exheling time by the ride, the DC motor stops the working and motor
will keep on running if the sensor not detects alcohol.
Engine Locking Unit
Engine locking was developed by the using a DC motor to exhibit like a bike engine.
The DC motor operates with specific conditions; when the alcohol level crosses 40%
of threshold value the engine automatically stops. The engine runs continuously if
the alcohol level comes less than 39%. The LED is used to show when the sensor
detect alcohol. in the simulation, when the logic state is 1 the led goes on to indicate
that alcohol is present and off to show the absence of alcohol.
Figure 5 graph x-axis represents the rider samples, tested for 8 instances and the
y-axis represents the processing time in seconds.
Figure 6 graph x-axis represents the rider samples, tested for 8 instances and the
y-axis represents the recognition rate in seconds.
Due to increase in trafic as well as rider’s fast move, accidents may happen. But,
some precotionary techniques are needed to avoid the lose of life. Therefore, in
our system, vibration sensor have been used by configuring the threshold value i.e.,
15 mV/g. When the hammering over helmet crosses the set threshold value, an
alert messege is sent automatically to defined phone numbers. Recognition rate is
calculated depending on the threshold value to notify the seviority of cause happened
to the rider. The system identifies the obstacle present in front and bike of a bike. For
this, IR sensor is used to detect the threats during back and forward turns or riding
over the road. To identify an obstacle the distance considered here is four feets for
our experimentation. It can be advanced by adopting the good one.
Obstacle Detection
An IR sensor is used for detecting the obstacles arriving at back side. The sensor
is mounted over Helmet. After detecting obstacle at the back side LCD should
display “OBSTACLE IS DETECTED AT THE BACK”. The system will prompt
by displaying as in Fig. 7.
Petrol Level Detection
Continuously system should monitor petroleum level in fuel tank LCD displays
FULL, MIDDLE and LOW level. At the time of low level petrol bike should stop
automatically. Figure 8 shows the LCD display of the output.
Depending on the level of petrol in the tank we can calculate how much KM more
a rider can travel on the bike.
6 Conclusion
The research over the idea fetches us the knowledge, that the technique plays yet
the vital role in society. The systemhas an effective smart and secure helmet system
which is based on RF transmitter and receiver. There are five modules and those
are considered as the objectives of thesystem. ATmega328 microcontroller is used
for executing the objectives of whole system. The experimental test results have
been obtained and observed that the system efficiently protects and awakes the rider
by indicating the LED signals. Smart and secure helmet is usefull and efficient
application for socity to reduce the accidents.
References
1. Ahuja P, Bhavsar K (2018) Microcontroller based smart helmet using GSM and GPRS. In: 2nd
IEEE ınternational conference on trends in electronics and ınformatics. IEEE Press
2. Jeong M, Lee H, Bae M, Shin D-B, Lim S-H, Lee KB (2018) Development and application of
the smart helmet for disaster and safety. IEEE Press
3. Tapadar S, Ray S, Saha AK, Karlose R, Saha HN (2018) Accident and alcohol detection in
bluetooth enabled smart helmets for motorbikes. IEEE Press
4. Biswas JR, Kachroo S, Chopra P, Sharma S (2018) Development of an app enabled smart
helmet for real time detection and reporting of accidents. IEEE Press
5. Revindran R, Vijayaraghavan H, Huang M-Y (2018) Smart helmets for safety in mining
industry. IEEE Press, Year
6. Budiman AR, Sudiharto DW, Brotoharsono T (2018) The prototype of smart helmet with safety
riding notification for motorcycle rider. In: 3rd IEEE ınternational conference on ınformation
technology, ınformation systems and electrical engineering. IEEE Press, Indonesia
7. Long X, Cui W, Zheng Z (2019) Safety helmet wearing detection based on deep learning. In:
3rd IEEE ınformation technology,networking,electronic and automation control conference.
IEEE Press
8. Rajan PTB, Mounika M, Nivetha S, Olive Sherine J (2019) Smart helmet based on ıot for
accıdent detectıon and notıfıcatıon. Int J Emer Technol Innov Eng 5(9)
9. Shravya K, Mandapati Y, Keerthi D, Senapati RK (2019) Smart helmet for safe driving. E 3S
web of conferences
10. Gupta S, Sharma K, Salvekar N, Gajra A (2019) Implementation of alcohol and collision sensors
in a smart helmet. In: IEEE ınternational conference on nascent technologies in engineering.
IEEE Press
On Some Properties of Extended Binary
Golay [24, 12, 8] Code Using MATLAB
Abstract The main aim of this paper is to study various properties of extended
binary Golay [24, 12, 8] code, when we know its generator matrix. Using MATLAB,
we can study syndrome decoding, weight of a codeword, error correction and error
detection of extended binary Golay [24, 12, 8] code.
Keywords Linear code · Generator matrix · Parity check matrix · Extended binary
golay code · Syndrome decoding
1 Introduction
In the late 1940s Claude Shannon was developing information theory and coding as
a mathematical model for communication. In mathematical sciences and electronics
engineering, a binary Golay code is a type of linear error-correcting code used in
digital communications. The binary Golay code, along with the ternary Golay code,
are named in honour of Marcel J. E. Golay whose 1949 paper [1] introducing them has
been called, by E. R. Berlekamp, the “best single published page” in coding theory.
There are two closely related binary Golay codes. The extended binary Golay code
encodes 12 bits of data in a 24-bit word in such a way that any 3-bit errors can be
corrected or any 7-bit errors can be detected.
The other, the perfect binary Golay code, has codewords of length 23 and is
obtained from the extended binary Golay code by deleting one coordinate position.
In standard coding notation the codes have parameters [24, 12, 8] and [23, 12, 7],
corresponding to the length of the codewords, the dimension of the code, and the
minimum Hamming distance between two codewords, respectively [2, 3]. In [4, 5],
algorithmic approach for error correction have been studied for few codes. In this
N. S. Darkunde (B)
School of Mathematical Sciences, SRTM University, Nanded, Maharashtra, India
S. P. Basude
SRTM University, Nanded, Maharashtra, India
M. S. Wavare
Department of Mathematics, Rajarshi Shahu Mahavidyalaya, Latur, Maharashtra, India
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 713
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_71
714 N. S. Darkunde et al.
paper, we are going to study some properties of extended binary Golay [24, 12, 8]
code.
2 Preliminaries
Let Fq denote the finite field with q elements, where q is some power of a prime. A
linear [n, k]q —code is a k-dimensional subspace of Fqn . The parameters n and k are
referred to as the length and dimension of the corresponding code.
Example The subset C = {000, 001, 010, 011} of vector space F23 is [3, 2]2 linear
code. Similarly, C = {0000, 1100, 2200, 0001, 0002, 1101, 1102, 2201, 2202} is
[4, 2]3 linear code.
Let x, y ∈ Fqn , The Hamming distance from x to y, denoted by d(x, y), is defined
to be the number of places at which x and y differ.
For a [n, k]q code C containing at least two words, the nonnegative integer given
by min{d(x, y) : x, y ∈ C, x = y} is called minimum distance of C. It is denoted by
d(C).
Example For a code C = {0000, 1000, 0100, 1100} in F24 , we see thatd(C) = 1.
Definition 3 (Dual Code). [8] Given a [n, k]q code C in Fqn , the subspace
C ⊥ ={x = (x1, x2, x3,..., xn ) ∈ Fqn : x.c = 0 for all c = (c1 , c2 , . . . , cn ) ∈ C},
We have following properties of C and C ⊥
The binary linear code generated by matrix G is called extended binary Golay
[24, 12, 8] code.
A = [1 0 0 0 1 1 1 1 1 0 0 1;
1 1 1 0 1 0 1 0 1 0 1 0;
1 0 1 1 1 1 1 0 0 1 0 0;
1 0 1 1 1 0 0 1 0 0 1 1;
0 1 1 1 1 1 0 0 1 0 0 1;
0 1 1 1 0 0 1 0 0 1 1 1;
1 1 0 1 0 1 0 1 0 1 0 1;
0 1 0 0 1 0 0 1 1 1 1 1;
0 1 1 0 0 1 1 1 1 1 0 0;
1 0 0 1 0 0 1 1 1 1 1 0;
0 1 0 1 1 1 1 1 0 0 1 0;
1 0 1 0 0 1 0 0 1 1 1 1];
G = [I A] % Generator matrix of extended Golay code
H = mod([−A’ eye(12)],2) % Denote H as a parity Check Matrix for extended Golay
Code
d = gfweight(G) % Distance of Code
slt = syndtable(H); % Produce Syndrome look up table.
w = [1 1 0 1 0 1 0 1 0 1 0 1 1 1 0 0 0 0 0 1 1 0 1 0] % Message received
syndrome = rem(w * H’, 2);
syndrome_de = bi2de(syndrome,’left-msb’); % Convert to decimal.
disp([’Syndrome = ’,num2str(syndrome_de),’ (decimal),’,num2str(syndrome),’
(binary)’])
corrvect = slt(1 + syndrome_de,:) % Correction vector
% Now compute the corrected codeword.
correctedcode = rem(corrvect + w,2)
G=
Columns 1 through 13
1000000000001
0100000000001
0010000000001
0001000000001
718 N. S. Darkunde et al.
0000100000000
0000010000000
0000001000001
0000000100000
0000000010000
0000000001001
0000000000100
0000000000011
Columns 14 through 24
00011111001
11010101010
01111100100
01110010011
11111001001
11100100111
10101010101
10010011111
11001111100
00100111110
10111110010
01001001111
H=
Columns 1 through 13
1111001001011
0100111110100
0111110010010
0011111001100
1111100100100
1010101010110
1110010011100
On Some Properties of Extended Binary Golay … 719
1001001111100
1100100111010
0010011111010
0101010101110
1001111100010
Columns 14 through 24
00000000000
10000000000
01000000000
00100000000
00010000000
00001000000
00000100000
00000010000
00000001000
00000000100
00000000010
00000000001
d=8
Single-error patterns loaded in decoding table. 4071 rows remaining.
2-error patterns loaded. 3795 rows remaining.
3-error patterns loaded. 1771 rows remaining.
4-error patterns loaded. 0 rows remaining.
w=
Columns 1 through 13
1101010101011
Columns 14 through 24
10000011010
Syndrome = 275 (decimal), 0 0 0 1 0 0 0 1 0 0 1 1 (binary)
corrvect =
720 N. S. Darkunde et al.
Columns 1 through 13
1000001100000
Columns 14 through 24
00000100000
Corrected code =
Columns 1 through 13
0101011001011
Columns 14 through 24
10000111010
10111110010
01001001111
H=
Columns 1 through 13
1111001001011
0100111110100
0111110010010
0011111001100
1111100100100
1010101010110
1110010011100
1001001111100
1100100111010
0010011111010
0101010101110
1001111100010
Columns 14 through 24
00000000000
10000000000
01000000000
00100000000
00010000000
00001000000
00000100000
00000010000
00000001000
00000000100
00000000010
00000000001
Enter the message bit of length 12 for which you want to encode using
On Some Properties of Extended Binary Golay … 723
Golay Code = [1 0 1 1 0 0 1 0 1 0 1 0]
u=101100101010
v=
Columns 1 through 13
1011001010100
Columns 14 through 24
11001010101
Synv = 0 0 0 0 0 0 0 0 0 0 0 0
3 Conclusion
Once we know the generator matrix of any linear code, using MATLAB we can have
many of the things that can be discussed. In this paper, we have studied extended
binary Golay [24, 12, 8] code with the help of MATLAB. The parity check matrix
in standard form for extended binary Golay [24, 12, 8] code is calculated and using
it, syndrome of a particular codeword is calculated. We have encoded the message
and decoded it correctly by using MATLAB. Using MATLAB, it has been verified
that, the distance of an extended binary Golay [24, 12, 8] code is 8 and it is 3-error-
correcting code.
References
Abstract The main aim of this paper is to study the encoding process of binary
Schubert code [43,7] using MATLAB, when one knows its generator matrix. Using
MATLAB, we can study syndrome decoding, weight of a codeword, error correction
and error detection of the code.
Keywords Linear code · Generator matrix · Parity check matrix · Schubert code ·
Encoding
1 Introduction
In the late 1940’s due to Shannon, Hamming and Golay, the approach to error correc-
tion coding was taken by modern digital communications. Error is the most afflicted
part in any type of communication. In this receiver receives the original message with
error in it. If the system knows how much error has come with original message then
that error can be removed. So much research work has been done in the field of error
correction and detection of incoming messages. In this paper we examine the family
of Schubert unions, in particular, the binary Schubert code which were defined in [1]
and how they are used in practice with the help of MATLAB. We have designed some
representatives of generator matrix and the encoding as well as decoding process have
been discussed. Using MATLAB, the error correcting capability of defined codes in
paper [2] have been verified. Linear error correcting codes associated to Schubert
varieties which is also known as sub-varieties of Grassmannian were introduced by
Ghorpade and Lachaud [3]. These Grassmannian codes were studied by Ryan [4,
5] and Nogin [2, 6]. The upper bound for minimum distance of Schubert code was
studied in [3].
M. S. Wavare (B)
Department of Mathematics, Rajarshi Shahu Mahavidyalaya, Latur, MS, India
N. S. Darkunde · S. P. Basude
School of Mathematical Sciences, SRTM University, Nanded, MS 431606, India
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 725
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_72
726 M. S. Wavare et al.
3 Preliminaries
Let Fq denotes the finite field having q elements, where q = p h , p a prime and
h a natural number. We denote Fqn as the n-dimensional vector space over Fq . For
any x ∈ Fqn , the support of (x), is the nonzero entries in x = (x1 , x2 , . . . , xn ). The
support weight (or Hamming norm) of x is defined by, x = |sup p(x)|.
More generally, if W is a subspace of Fqn , the support of W, Supp (W) is the set of
positions where not all the vectors in W are zero and the support weight (or Hamming
norm) of W is defined by,
Encoding Using the Binary Schubert Code … 727
W = |supp (W )|
A linear [n, k]q -code is a k-dimensional subspace of Fqn . The parameters n and k
are referred to as the length and dimension of the corresponding code. The minimum
distance d = d(C) of C is defined by,
d = d(C) = min{x : x ∈ C, x = 0}
More generally, given any positive integer r, the rth higher weight dr = dr(C) is
defined by
Note that d1 (C) = d(C). It also follows that di ≤ d j when i ≤ j and that
dk = |sup p(C)|, where k is dimension of code C. Thus, we have 1 ≤ d1 = d <
d2 < . . . < dk−1 < dk = n. The first weight d1 is equal to the minimum distance
and the last weight is equal to the length of the code.
An [n, k]q -code is said to be nondegenerate if it is not contained in a coordinate
hyperplane of Fqn . Two [n, k]q -codes are said to be equivalent if one can be obtained
from another by permuting coordinates and multiplying them by nonzero elements of
Fq. It is clear that this gives a natural equivalence relation on the set of [n, k]q -codes.
The matrix whose rows forms basis for linear code known as generator matrix
where as the generator matrix of dual of any linear code C is known as parity check
matrix for linear code C.
Then the linear code generated by this matrix G is known as binary Schubert code
of length 18. This code is a binary [18,5,7] linear code.
Now in the next section let us consider the binary Schubert code of length 43,
which will be constructed in the similar way of binary Schubert code of length 18
[1], So for the next section we have used a binary Schubert code of length 43 and
dimension 7 to discuss the encoding of the message.
clc
I=eye (7); % Identity matrix of order 7
X= [ 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1;
0 0 0 0 0 0 0 0 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0;
0 0 0 0 0 1 1 1 0 0 1 1 0 1 1 1 1 1 1 0 0 0 0 1 0 0 0 0 1 0 1 0 0 0 0 1;
0 0 0 0 1 0 1 1 0 1 0 1 1 0 1 1 0 0 1 1 0 1 0 1 0 1 0 0 0 0 1 0 0 0 1 0;
0 1 1 1 0 0 0 0 1 0 0 0 1 1 1 1 0 1 0 1 1 0 0 0 1 0 0 1 0 1 0 0 0 1 0 0;
1 0 1 1 0 1 0 1 0 0 0 0 0 1 1 1 0 1 0 0 0 0 0 1 1 0 1 0 1 1 0 0 1 0 0 0;
1 1 0 1 1 0 0 1 0 0 0 0 1 0 1 1 0 0 0 1 0 0 0 1 1 1 1 1 0 0 0 1 0 0 0 0];% Matrix X
G=[I X]; % The generator matrix for binary Schubert Code of length 43
H=mod ([-X’ eye(36)],2); % Parity Check Matrix for the defined Code
u=input (‘Enter the message bit of length 7 for which you want to encode using the
defined v=mod(u*G,2) % Encoding of message u
Synv=mod(v*H’,2) % Syndrome of encoded message
Encoding Using the Binary Schubert Code … 729
u=input (‘Enter the message bit of length 7 for which you want to encode using the
defined Enter the message bit of length 7 for which you want to encode using the
defined Code
u = [1 1 0 0 1 0 0 1]
v=
Columns 1 through 13
1001001110100
Columns 14 through 26
1001010001110
Columns 27 through 39
1101101001100
Columns 40 through 43
1101
Synv =
Columns 1 through 13
0000000000000
Columns 14 through 26
0000000000000
Columns 27 through 36
0000000000
% Use a Binary length 43
n = 18; k = 5; %Length and Dimension
clc % Clearscreen
I=eye(7); % Identity matrix of order 7
X= [ 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1;
0 0 0 0 0 0 0 0 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 0 0;
0 0 0 0 0 1 1 1 0 0 1 1 0 1 1 1 1 1 1 0 0 0 0 1 0 0 0 0 1 0 1 0 0 0 0 1;
0 0 0 0 1 0 1 1 0 1 0 1 1 0 1 1 0 0 1 1 0 1 0 1 0 1 0 0 0 0 1 0 0 0 1 0;
0 1 1 1 0 0 0 0 1 0 0 0 1 1 1 1 0 1 0 1 1 0 0 0 1 0 0 1 0 1 0 0 0 1 0 0;
1 0 1 1 0 1 0 1 0 0 0 0 0 1 1 1 0 1 0 0 0 0 0 1 1 0 1 0 1 1 0 0 1 0 0 0;
730 M. S. Wavare et al.
1 1 0 1 1 0 0 1 0 0 0 0 1 0 1 1 0 0 0 1 0 0 0 1 1 1 1 1 0 0 0 1 0 0 0 0]; % Matrix X
G=[I X]; % The generator matrix for extended binary Schubert Code of length 43
H=mod([-X’ eye(36)],2); % Denote H as a parity Check Matrix for defined Code
slt= syndtable(H); % Produce Syndrome look up table.
w = [1 0 0 1 1 1 1 1 1 0 0 1 1 1 1 0 0 1 0 1 1 1 0 0 0 0 0 0 1 1 1 1 1 0 1 0 1 0 1
syndrome = rem(w * H’,2);
syndrome_de = bi2de(syndrome,’left-msb’); % Convert to decimal.
disp([‘Syndrome = ‘,num2str(syndrome_de), ‘ (decimal),’,num2str(syndrome),’
(binary)’])
corrvect = slt(1+syndrome_de,:) % Correction vector
% Now compute the corrected codeword.
correctedcode = rem(corrvect+w,2).
5 Conclusions
Thus, using the MATLAB, we have encoded the message information of length 43
with the help of generator and parity check matrix of given Schubert code [1] of
length 43, But due to limitation of size of matrix one cannot construct syndrome
table and can’t decode it, which we can observe from the decoding program at the
end in above section.
References
1. Wavare Mahesh S (2019) Codes associated to Schubert varieties G(2,5) over F 2 . New Trends
Math Sci 7(1):71–78
2. Yu D (2017) Nogin (1997), Spectrum of codes associated with the grassmannian G(3; 6). Probl
Inf Transm 33(2):26–36
3. Ghorpade SR, Lachaud G (2000) Higher weights of grassmann codes, coding theory, cryp-
tography and related areas. In: Buchmann J, Hoeholdt T, Stictenoth H, Tapia-Resillas H (eds)
Springer, Heidelberg, Germany, pp 120–131
4. Ryan CT (1987) An application of grassmannian varieties to coding theory. Congr Numer
57:257–271
5. Ryan CT (1987) Projective codes based on grassmann varieties. Congr Numer 57:273–279
6. Yu D (1996) Nogin, codes associated to grassmannians, arithmetic, geometry and coding theory.
In: Pellikan R, Perret M (eds) S. G. Vl_adut_, Walter de Gruyter, Berlin, New York, pp 145–154
7. Hu_man WC, Pless V (2010) Fundamentals of error-correcting codes. Cambridge University
Press
8. MacWilliams FJ, Sloane NJA (1977) The theory of error-correcting codes. Elsevier
9. Ling S, Xing C (2004) Coding theory-a first course, 1st edn. Cambridge University Press
Encoding Using the Binary Schubert Code … 731
10. Patil AR, Darkunde NS (2019) Algorithmic approach for error-correcting capability and
decoding of linear codes arising from algebraic geometry. Springer, Information and Commu-
nication Technology for Competitive Strategies, pp 509–517
11. Pratap R, Getting started with MATLAB: a quick introduction for scientists and engineers, 7e.
Oxford University Press
Data Mining Techniques for Privacy
Preservation in Social Network Sites
Using SVM
Abstract The rising numbers of users over the network have also raised the privacy
risk and incidents of various types of theft and attacks. Hence, the social networks
have been the major victims. The users over these networking sites share the infor-
mation under various attributes like gender, location, contact information etc. This
personal information can get compromised due to malicious act that severely violates
the integrity of the data and privacy protection policy. As a result, it has become
mandatory for a service provider to offer privacy protection before publishing any
kind of data over the network. In this research, we have proposed data mining tech-
niques for user’s privacy preservation in social network sites using Support Vector
Machines (SVMs). Social media datasets ARNET and SDFB are used for the anal-
ysis of privacy preservation models by calculating Average Path Length parameter.
Finally, the proposed model shows 1.72% and 1.46% less information loss and 1.42%
to 5.09% reduction in APL with these datasets as compared to previous works.
Keywords Privacy · Data mining · Privacy violations · SVM · Social network sites
1 Introduction
In today’s digital scenario, every day to day activity leads to large amount of data. Data
mining is used to analyse this data and with the help of various types of methods,
knowledge is discovered which is further used for many applications in different
fields. Big amount of data about different individuals can lead to privacy breach if
accessed by unauthorised person. Sensitive data may include health data, financial
data, academic data, location based servicing data, social media data etc. [1]. To
protect this data, data mining methods have to be combined with privacy preservation
techniques so that knowledge can be taken out without exploitation of personal
information related to user. Privacy Preserving Data Mining (PPDM) is extraction
of useful patterns from large set of data while preventing the release of sensitive
data of an individual [2]. Privacy of data can be viewed from 4 different roles [2,
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 733
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_73
734 V. Kalunge and S. Deepika
3]. According to provider of data, data should be provided to the collector taking
into consideration both requirement and sensitivity of data. Data is collected by data
collectors from different data providers and it is stored in various warehouses and
databases. Modification of data is done in such a way that protection is ensured as
well as relevance of data is not lost. Data Miner uses this stored data and applies
mining algorithms to find various patterns, rules and relationships. These mined
results should also remain hidden from unauthorised users. Decision makers use
these results to take the decisions for various purposes after analysis (Fig. 1).
Data from various data sources is extracted, transformed and then loaded into the
integrated source called data warehouse. Processing is done on this warehouse using
data mining algorithms. To preserve the privacy, privacy preservation techniques
have to be applied. With this, sensitive data of user is protected as well as required
results can also be found. Privacy preserving techniques provide protection during
process of mining as well as secure the results produced by data mining [4]. Scheme
of PPDM is shown in the figure below: This study introduces concept of Privacy
Preserving Data Mining, various roles in PPDM and basic foundation of IST.
2 Literature Survey
In this section, various existing techniques related to Privacy Preserving Data Mining
are discussed on the basis of work done by different authors in this area. Literature
study had shown that have been a constant research going on for the improvement
of k-anonymity based privacy protection. Some of the important studies have been
summarized in this section. Lin and Chen [5] focussed the issues related to violation
of privacy preservation measures and postulated a PPSVC (privacy-preserving SVM
Classifier) that could transform traditional SVM into a privacy protecting classifier.
The effectiveness of the proposed design was demonstrated by the offered defence
against adversarial attacks without compromising the classification accuracy of the
classifier [5]. Okada et al. [6] proposed a k-anonymity based technique that appends
the noise to edges of social network being studied via k-neighbourhood subgraphs.
The design was implemented such that it suppressed any change in the magnitude
of edge length between nodes. Experimental evaluation had demonstrated that the
proposed design could successfully maintain the edge length of the k-anonymity
based graphs [6]. Tsai et al. [7] studied the privacy preservation issues of anonymizing
the shortest path. The privacy preservation is addressed with the concept utilizing
Data Mining Techniques for Privacy Preservation in Social … 735
3 Proposed Methodology
The proposed design involves optimization of the sensitive data that is to be dissem-
inated over the web. Enhanced k-anonymity based privacy protection is achieved
by passing information to Cuckoo Search (CS) for optimization followed by clas-
sification carried on with Support Vector Machine (SVM). The privacy protection
offered by the proposed design is evaluated with experimental calculation of APL
and Information loss against larger data sets. The steps of proposed methodology are
explained with block diagram in Fig. 2.
In the proposed design, two databases are used representing larger data sets of
social networking sites. ARNET database offers comprehensive data mining covering
736 V. Kalunge and S. Deepika
social academic networks. The initial citation network was demonstrated by 629,814
papers covering 632,752 citations [14]. SDFB is Six Degrees of Francis Bacon that
corresponds to digital information of relationships of early modern social network.
The dataset summarizes the information covering 171,419 relationships established
between 15,824 EMSN. Further, data set is divided into 109 different labelled groups
whose relationships are summarized under 64 categories [15].
Cuckoo Search is a nature inspired algorithm that is inspired with the reproduction
and egg laying behavior of cuckoo birds. To reach an optimal solution CS observes
following assumptions:
• Cuckoo randomly searchers for the nest to lay eggs in the nests.
• Only the best qualities of eggs are preserved and the nest containing them is further
carried to next generation.
• Moreover there exists the probability that the host cuckoo could identify the
foreign egg.
SVM is used as a classifier to the output of the Cuckoo Search Algorithm. As
SVM is a binary classifier, hence in order to provide training to the classifier, the
data is being divided into two classes namely C1 and C2. C1 has values which were
placed in noisy nodes 1 and C2 has values which were placed in noisy nodes 2. The
average APL of every node is passed as the training data.
Data Mining Techniques for Privacy Preservation in Social … 737
4 Results
where (pi, pj) represents the shortest path length existing between two points that is
divided by the total number of possible paths. The observed average path length of the
proposed design using ARNET dataset against two existing studies is summarized
in Table 1. In 2019 Namdarzadegan and Khafaei had implemented Cuckoo based
optimization [13] while Macwan et al. in [16] demonstrated the effectiveness of
k-anonymity model [16] for privacy preservation of social networks.
Comparative analysis of APL with the existing studies for ARNET dataset is
shown in Fig. 3 where node count is plotted on X-axis ranging from 10 to 100
nodes against the APL values on Y-axis. It is observed that the line graph corre-
sponding to the proposed work exhibited lowest APL as compared to the existing
work of Namdarzadegan and Khafaei [13] and Macwan et al. [16]. The graph shows
that average APL of 0.397 is observed for Namdarzadegan and Khafaei, 0.434 for
Macwan et al. with lowest APL of 0.3865 for the proposed design. This means that
738 V. Kalunge and S. Deepika
the proposed design is 1.05% better than Namdarzadegan and Khafaei work and
4.72% better than Macwan et al.’s work.
Next, APL comparison is also carried on with SDFB dataset and the corresponding
observed APL values are listed in Table 2. In column 2 lists APL values of the
proposed work against node count varying from 10 to 100. Also, the APL values
corresponding to Namdarzadegan and Khafaei and Macwan et al. work are listed in
column 3 and column 4 respectively. The table shows that magnitude of the APL
values is lowest for the proposed work in comparison to the existing works.
Figure 4 shows the APL comparison for SDFB dataset. It is observed that for SDFB
dataset our proposed design achieved an average APL value of 0.382 in comparison
to the average APL of 0.397 demonstrated by Namdarzadegan and Khafaei [13] and
Data Mining Techniques for Privacy Preservation in Social … 739
0.434 by Macwan et al. [16]. Overall, our design exhibited 1.42% and 5.09% lower
APL as compared to Namdarzadegan and Khafaei and Macwan et al., respectively.
Information Loss
It is another parameter that evaluates the effectiveness of the proposed design in
achieving privacy preservation. It represents the percentage of information that get
compromised and lost as a result of data transfer. A better privacy protection model
always results in lower information loss. In the current study, it is calculated using
both ARNET and SDFB datasets.
Table 3 compares the information loss for ARNET dataset of the proposed design
740 V. Kalunge and S. Deepika
with information loss observed by Song et al. who had implemented random k-
anonymous method to achieve privacy protection [11]. Table shows that compara-
tively lower information loss is observed for the proposed work over all the node
counts.
Information loss for ARNET over 10 to 100 nodes is compared in Fig. 5. The
line graph plots node count on X-axis against the information loss values plotted on
Y-axis. It is observed that information loss of the proposed design varied from 30.8
to 32.4% with an average information loss of 31.56%. Information loss observed for
privacy preservation method of Song et al. exhibited an average value of 33.28%.
In other words, our proposed design demonstrated 1.72% lower information loss for
ARNET dataset.
Further, information loss is also calculated for SDFB dataset and the resultant
parametric values of information loss are summarized in Table 4. Information loss
of the proposed design is given in column 2 and Song et al. is given in column 3.
It is also observed that information loss of the proposed design for SDFB dataset
lies between 31.2 and 32.6% which is little higher than information loss that was
observed for ARNET dataset.
Figure 6 compares the information loss of SDFB dataset for our proposed work
and Song et al.’s work. The graph shows that the proposed work demonstrated an
average information loss of 31.82% while Song et al.’s work demonstrated an average
information loss 33.28%. It means that the proposed work exhibited 1.46% lower
information loss corresponding to the effectiveness of our proposed design.
5 Conclusion
Today’s time is the time of Big Data. Useful information extracted from the large
amount of data helps to take very important decisions in various fields and orga-
nizations but they can lead to privacy breach of an individual. The need of hour is
to protect sensitive data of an individual and hence, data mining algorithms have to
be combined with privacy preservation techniques. The present work had proposed
an enhanced k-anonymity based privacy protection where cuckoo search optimiza-
tion is used. The fitness function based improved privacy protection model is further
cross validated using SVM. QoS parameters namely, information loss and APL were
evaluated for ARNET and SDFB datasets over the node count ranging from 10 to
100 to show the effectiveness of the implemented techniques. It is observed that APL
of the proposed work for ARNET dataset is 1.05 and 4.72% while for SDFB dataset
is 1.42 and 5.09% lower when compared with Namdarzadegan and Khafaei’s and
Macwan et al.’s work, respectively. Similarly, information loss of the proposed work
also showed decrease of 1.72 and 1.46% for ARNET and SDFB dataset as compared
to information loss observed with Song et al.’s work. Hence, the proposed design
proved to be highly effective enhanced k-anonymity based privacy preservation.
References
1. Mendes R, Vilela JP (2017) Privacy-preserving data mining: methods, metrics, and applications.
IEEE Access 5:10562–10582
2. Bhandari N, Pahwa P (2019)Comparative analysis of privacy-preserving data mining tech-
niques. In: International conference on innovative computing and communications. Springer,
Singapore, pp 535–541
3. Xu L, Jiang C, Wang J, Yuan J, Ren Y (2014) Information security in big data: privacy and
data mining. IEEE Access 2:1149–1176
4. Siraj MM, Rahmat NA, Din MM (Feb 2019) A survey on privacy preserving data mining
approaches and techniques. In: Proceedings of the 2019 8th international conference on
software and computer applications. ACM, pp 65–69
5. Lin KP, Chen MS (2010) On the design and analysis of the privacy-preserving SVM classifier.
IEEE Trans Knowl Data Eng 23(11):1704–1717
6. Okada R, Watanabe C, Kitagawa H (2014) “A k-anonymization algorithm on social network
data that reduces distances between node”. In: 2014 IEEE 33rd international symposium on
reliable distributed systems workshops. IEEE, pp 76–81
7. Tsai YC, Wang SL, Kao HY, Hong TP (2015) Edge types versus privacy in K-anonymization
of shortest paths. Appl Soft Comput 31:348–359
8. Maheshwarkar B, Patidar P, Rawat MK, Maheshwarkar N (2016) “K-AMOA: K-anonymity
model for multiple overlapped attributes”. In: Proceedings of the second international confer-
ence on information and communication technology for competitive strategies. ACM, p
83
9. Tsai YC, Wang SL, Song CY, Ting IH (2016) Privacy and utility effects of k-anonymity on
association rule hiding
10. Casino F, Domingo-Ferrer J, Patsakis C, Puig D, Solanas A (Sep 2013) Privacy preserving
collaborative filtering with k-anonymity through microaggregation. In: 2013 IEEE 10th
International conference on e-business engineering. IEEE, pp 490–497
Data Mining Techniques for Privacy Preservation in Social … 743
11. Song F, Ma T, Tian Y, Al-Rodhaan M (2019) “A new method of privacy protection: random
k-anonymous”. IEEE Access
12. Shailaja GK, Rao CG (2019) “Opposition intensity-based cuckoo search algorithm for data
privacy preservation”. J Intell Syst
13. Namdarzadegan M, Khafaei T (2019) “Privacy preserving in social networks using combining
cuckoo optimization algorithm and graph clustering for anonymization”. Asian J Res Comp
Sci 1–12
14. Tang J, Zhang J, Yao L, Li J, Zhang L, Su Z (2008) “Arnetminer: extraction and mining of
academic social networks”. In: Proceedings of the 14th ACM SIGKDD international conference
on knowledge discovery and data mining. ACM, pp 990–998
15. SDFB Team, six degrees of Francis Bacon: reassembling the early modern social network.
www.sixdegreesoffrancisbacon.com. Accessed URL https://www.kaggle.com/rtatman/six-deg
rees-of-francis-bacon
16. Macwan KR, Patel SJ (2017) k-Degree anonymity model for social network data publishing.
Adv Electr Comp Eng 17(4):117–125
Near Field Communication (NFC)
Technology and Its Application
R. D. Kulkarni
1 Introduction
NFC stands for Near Field Communication, which is a short distance, low data rate
and low cost protocol. It is complementary to WIFI and Bluetooth and is a subset
of 13.56 MHz RFID. NFC is deployed to enable contactless transactions for data
exchange and simplified setup of more complex communications such as Bluetooth
and WIFI. For instance, version 4.0 and 4.1 of Bluetooth low energy has no secure
way of exchanging the temporary key for encryption without the help of NFC [1].
Hence, NFC is poised to play an important role in the connectivity for the “Internet
of Things.” In addition, NFC is widespread in smart phone markets, as in 2015,
50% of smartphones were compatible with NFC. NFC is gaining presence in more
and more applications, from POS terminals for tap & pay applications to consumer
environments with tap & pair application to home appliances and healthcare for tap
& exchange applications.
R. D. Kulkarni (B)
SVERI’s COE, Pandharpur, India
e-mail: [email protected]
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 745
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_74
746 R. D. Kulkarni
The difference between NFC and RFID can sometimes raise confusion, as the
terms are sometimes used synonymously. However, NFC is actually an extension
of RFID. NFC is designed to build on RFID by enabling more complex exchanges
between devices [1]. The root cause of the confusion is the fact that it’s still possible
to read passive RFID tags with an NFC reader and to write to a limited amount of
memory. RFID was designed for identification. RFID tags can hold a small amount
of data, and you can read and write to them from RFID readers, but the amount of
data is limited to a few thousand bytes [2]. An RFID is passive when it receives
power from the reader, but it is active when it has its own power source.
An RFID exchange involves two factors: a target and an initiator. The initiator,
a tag reader or reader/writer, starts the exchange by generating a radio field and
listening for responses from any target in the field. The target, a tag, responds when
it picks up a transmission from an initiator. It will respond with a unique identifier
number (UID) [2].
RFID has various standards defined by ISO (International Standards Organiza-
tion) [2]. A standard defines radio frequency, data transfer rates, data formats, and
(optionally) layers of the protocol. It can be rather specific—for instance the ISO-
11784 standard was developed for animal tracking! The most common one is the
ISO-14443 standard that was developed for use with payment systems and smart
cards. Many NFC applications were developed based on this standard.
As said previously, NFC can be thought of as an extension of RFID and hence
shares many properties (active or passive communication modes, how to initiate
communication and hardware layer). But, instead of just delivering static data from
memory, an NFC target can generate unique content for each exchange and deliver
it back to the initiator. For example, if you’re using NFC to exchange address data
between two phones, the NFC target device can be programmed to only provide
limited information if it has never seen this particular initiator before.
NFC is becoming ubiquitous, but due to many terminologies and standards, taking
the first steps can be overwhelming. This article aims to clarify and simplify those
steps for newcomers.
2 Communication Ways
NFC Modes
The NFC forum defines three communication modes, as in Fig. 1 and Table 1:
3 Security Concerns
As mentioned, most NFC scenarios are required to deal with sensitive data like
credit card numbers, bank account details, account balances, personalized tickets or
other personal data. For data storage and wireless data transfer security is therefore an
essential issue. NFC thereby provides several mechanisms for security and immunity.
First of all, there is obviously a certain physical security due to the touch-to-connect
principle. As a matter of fact, the NFC technology only provides data transfer between
two devices or between a device and a tagged object when the distance between the
two items falls below a certain range. The usual transmit power of radio frequency
and the receive sensitivity of NFC readers that fulfil the previously described ISO
specifications are only strong enough to operate up to a range of a few centimetres [3].
That means that e.g. at a supermarket checkout the customer needs to place his phone
directly over the reading device. Data skimming, that is capturing and intercepting
transferred data by a distant attacker, is hence not possible that easily. Misuse is
however possible in the form of relay attacks. Such attack is basically accomplished
via an attacker serving as a man-in-the-middle who is forwarding transmitted data
between a reader, e.g. a reading device for NFC payments, and a target transponder,
e.g. a NFC device serving as a credit card, that is actually out of the reading range. The
following example illustrates the procedure: An attacker places his modified NFC
phone on top of the NFC reader at the POS, e.g. at a checkout in the supermarket.
Near Field Communication (NFC) Technology and Its Application 749
The phone however forwards the data, e.g. via Bluetooth, to his distant accomplice
holding another modified NFC phone. He in turn holds his phone—without attracting
attention—next to a contactless NFC smart card, e.g. NFC credit card, of the attacked
target person. The accomplice’s phone operates as NFC reader simulating the actual
reading device at the POS. The reader at the POS assumes the target person’s card
is close, and unknowingly debits his account. The additional data exchange and
forwarding via the attacking person-in-the-middle however naturally takes some
extra time. A possible countermeasure for such relay attack could for example be
built upon a quite short timeout threshold that avoids transactions if the response
time is too slow. The concept of Google Wallet is however also secured against such
attacks as thereby a PIN has to be entered in order to activate the phone’s NFC
broadcast hardware and to activate the Secure Element that is storing all the sensitive
data. The Wallet PIN also prevents unauthorized usage of the payment card in case
the NFC phone is stolen.
For the wireless channel communication itself the NFC specifications do not
ensure any secure encrypting mechanisms. On higher layers however, NFC appli-
cations can of course use industry-standard cryptographic protocols like SSL/TLS
based methods to encrypt the data that is to be transferred over the air and that is
stored in the Secure Element. Other possible NFC vulnerabilities involve the manip-
ulation of NFC tags [3]. Existing passive NFC tags, e.g. on a smart poster, could
be replaced by spoofed tags such that a modified NDEF message is read by a NFC
reading device. Possible attacks could for example replace the URI record with a
malicious URL, e.g. in order to make the user load a phishing website that steals
the users credentials. Experiments show that one can modify the NDEF message of
a smart poster such that the visualized content on the phone, e.g. the website URL,
is hardly distinguishable from the original one. Furthermore, it is even possible to
create a malformed NDEF message that causes the some applications to crash. This
form of manipulation could be used by attackers to debase the relationship between
a user and the pretended service provider. The Signature Record Type that was previ-
ously addressed in this paper and that signs a whole NDEF message can however
serve as a countermeasure against URL spoofing and similar attacks. By manipu-
lating the signed tag payload a signed NDEF message will lose its integrity and will
be recognized as not-trusted.
4 Application of NFC
5 Conclusion
NFC is a wireless technology for short range data transmission. It operates in three
modes via two types of communication. RFID and magnetic induction are the key
factors in implementing NFC. Few forms of implementing this technology is through
NFC tags and NFC payment [7]. Through NFC Payments transactions are secure and
carried out with just a tap. The SE is responsible for security, authenticity and data
confidentiality with respect to payments. NFC tags will play inevitable role in future
smart devices for more integrated functions, smart transportation, aviation industry,
shipping, manufacturing industry for automation of particular tasks. Integrating NFC
technology with our modern data communication and transaction process ensures
convenience, time saving, energy efficiency and most importantly improved security.
References
1. NFC-enabled cellphone shipments to soar fourfold in next five years | IHS Online Newsroom,
press.ihs.com, (2014) [Online]. Available: https://press.ihs.com/press-release/design-supply-
chain/nfcenabled-cellphone-shipments-soar-fourfold-next-five-years
2. Klaus F (2003) In RFID handbook aplications, technology
3. Haselsteiner E, Breitfuß K (2006) Security in near field communication (NFC). In: Workshop
on RFID security
4. https://www.cnet.com/how-to/how-nfc-works-and-mobile-payments
5. Langer J, Roland M (2010) Anwendungen und technik von near field communication (NFC),
Anwendungen und Tech. von Near F. Commun. NFC, pp 205–241
6. https://nfc-forum.org/resources/what-aretheoperating-modes-of-nfc-devices/
7. EMVCo (Nov 2011) https://www.emvco.com/
8. El Madhoun N, Bertin E (2017) Magic always comes with a price: utility versus security for
bank cards. In: IEEE cyber security in networking conference (CSNet’17)
9. Sherringham K, Unhelkar B (2009) In handbook of research in mobile business. 2nd edn
10. Agbinya JI, Masihpour M (2012) Power equations and capacity performance of magnetic
induction communication systems. J Wirel Pers Commun 64(4):831–845
Performance Analysis of Convolution
Neural Network with Different
Architecture for Lung Segmentation
Abstract As lung cancer is one of the significant causes of death, there is a need for
the development of algorithms for early detection of these cancers. Early detection
of lung cancer helps to provide appropriate treatment and reduce morbidity. Accu-
rate segmentation of the lung is an essential step in every computer-aided diagnosis
(CAD) system to provide an accurate lung CT image analysis. This study is focused
on the design of the appropriate architecture of the convolution neural network (CNN)
using suitable combinations of CNN blocks to improve lung segmentation efficiency.
Based on the scientific intuition, three CNN architectures are proposed for effective
segmentation of lung parts from CT images. These CNN architectures are varied by
the depth of down sampling of images as 32 × 32, 16 × 16 and 8 × 8. The perfor-
mances of these CNN are obtained as under segmentation or over-segmentation by
comparing the segmented lung part with ground truth lung images. This performance
analysis shows the segmentation efficiency greatly affected by appropriate selection
of downsampling of these images.
1 Introduction
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 753
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_75
754 S. P. Pawar and S. N. Talbar
early detection of lung cancer helps to minimize the severity of cancer by reducing
its morbidity [1]. As Computer Tomography is the best method to detect the presence
of lung cancer, the proposed algorithm is demonstrated using CT scan images. Lung
segmentation is one of the important steps in the process of computer-aided diagnosis
(CAD) multi-stage process. The proposed study is focused on the development of
the suitable architecture of CNN for lung segmentation.
ILD (Interstitial lung diseases) is a group of diseases that contain 150 histo-
logical diagnoses associated with disorders of the lung parenchyma [2] that even-
tually affect breathing and cause progressive scarring of lung tissues. ILD lung
tissue patterns intensity is similar to the chest region which becomes a hurdle in
lung segmentation. Thus, most existing methods fail [3] to include such ILD lung
tissue patterns in segmentation which leads over-segmentation or under segmenta-
tion problem. To overcome this limitation, we proposed a robust method for accurate
lung segmentation.
The lung segmentation task is crucial to find out the accurate boundary of lung
parenchyma so that we cannot mislead lung nodules. There are various approaches
that have been proposed for lung segmentation. Initially, [1, 4, 5] grey-level thresh-
olding based approach for segmentation of lung from the non-lung region. Brown
et al.[1] presented an automated knowledge-based method to segment lung regions
from CT image based on volume, shape, relative position of the organ. Sun et al.[5]
developed a 3D region-growing algorithm to segment lung volume based on a fuzzy
logic approach and 3D morphological closing operation. Threshold-based techniques
ignore lung regions that have more attenuation, such as in the presence of interstitial
lung disease. However, these conventional methods failed in the presence of dense
abnormalities such as lung nodule, fibrosis etc.
In literature, methods proposed in [2, 6] developed to include dense abnormalities
to get accurate lung segmentation. Pu et al. [6] presented an adaptive border marching
algorithm that includes juxtapleural nodules, but this algorithm failed to include
abnormal lung tissue patterns. Further, Prasad et al. [7] proposed fixed threshold-
based techniques that included lung boundary curvature features which can help to
include abnormal lung tissue patterns. However, the performance of this method
depended on accurate rib segmentation.
Another shape-based approach developed for lung segmentation by Ye et al. [3]
for the inclusion of abnormal lung tissue patterns. They proposed a two-step segmen-
tation method to extract lung region in the first step they used to extract initial lung
mask using fuzzy thresholding method in the second step lung contour chain code
was used to obtain a complete mask. However, this method led over-segmentation
in the inclusion of large abnormal lung tissue patterns.
Further, Shen et al. [8] have proposed A bidirectional chain coding method
combined with a support vector machine (SVM) classifier is used for inclusion of
juxtapleural nodules. However cascade nature of the method, system performance
depends on scan quality of CT image, classifier and threshold.
In the existing literature, intensity-based approaches have been proposed [9, 10]
intensity-based method methods that rely on robust feature extraction. Hosseini-Asl
et al. [9] have proposed incremental constrained nonnegative matrix factorization to
Performance Analysis of Convolution Neural Network … 755
extract robust features. Further, in [10] the author proposed a 3D lung segmentation
method based on the new Nonnegative Matrix. Factorization approach to impose
incremental constraints on it. Method [9, 10] includes juxtapleural nodules but they
failed in the case of dense abnormalities.
Above lung segmentation methods failed in case of inclusion of lung tissue
patterns. Therefore, there is a need to develop a strong learning mechanism for
lung segmentation.
This paper is focused on the design of c-GAN based approach Convolution neural
network with different architecture for lung segmentation. In this study, three different
CNN networks are proposed with architectures varied by the depth of downsampling
of images as 32 × 32, 16 × 16 and 8 × 8. The lung segmentation performances of
these networks are studied to form the segmentation map of the CT slice without any
post-processing of these images. The performance of these networks is studied for
segmentation of lung in the presence of lung nodule abnormalities.
encoded the input lung CT slice into feature maps and followed by downsampling
the encoded feature maps.
A multi-scale dense feature extraction (MSDFE) module [11] which consists of
four inception blocks connected serially through the dense connection between is
utilized for robust feature extraction.
The MSDFE plays an important role in dense feature extraction based on which
the depth of downsampling varies. To understand the influence of these downsam-
pling on the segmentation performance, three CNN architectures with the depth of
downsampling of images as 32 × 32, 16 × 16 and 8 × 8 are studied. Figure 1 shows
the CNN network which comprises four MSDFE which downsamples the images to
32 × 32 whereas Fig. 2 shows the network with six MSDFE and Fig. 3 shows eight
MSDFE which downsamples the images to 16 × 16 and 8 × 8, respectively. In these
three networks half of the MSDFE modules obtain multi-scale features processed
through the simple convolution layer (filter size of 3 × 3) having stride factor =
2 whereas the remaining half MSDFE modules, to maintain the symmetry in the
network, to obtain multi-scale feature maps processed through the simple decon-
volution layer (filter size of 3 × 3) having up-sampling factor = 2. Therefore, the
network proposed in Fig. 1 downsamples the images of 128 X 128coming out of
Fig. 4 Lung segmentation results for under segmentation cases a Input lung CT slice b CNN (8 ×
8) c CNN (32 × 32) d CNN (16 × 16) e Ground truth
qualitative performance analysis, these cases from the database are presented in the
form of under-segmentation, over-segmentation and complex cases. Figure 4 shows
the under segmentation cases for three different networks proposed and compared
with the ground truth. The red colour indicates the under-segmented part whereas
blue colour represents the over segmented part. For the cases shown in the image,
the network with downsampling 16 × 16 shows good performance of segmentation
showing close to the ground truth images whereas other two networks fail to outline
the complex areas of the lung and do not include some of the parts of the lungs. Even
in some cases, the network with 8 × 8 downsampling performs poorly as compared
to 32 × 32 downsampling network which may be due to densely mapping of the
features resulting in the exclusion of some of the complex parts of lungs.
Figure 5 shows the over-segmentation cases from the testing of segmentation
efficiency of three networks proposed. For these cases also the network with down-
sampling to 16 × 16 gives fairly good results as compared to the other two networks.
However, these two networks give better segmentation than under-segmentation
cases. The performance variation can be greatly seen when qualitative analysis of
critical cases is carried out, as shown in Fig. 6. In this section, it can be clearly
seen that the network with downsampling to 16 × 16 outperformed the other two
networks. In these cases, the other two networks overestimated the complete region,
which is not part of the lung.
Performance Analysis of Convolution Neural Network … 759
Fig. 5 Lung segmentation results for Over Segmentation cases a Input lung CT slice b CNN (8 ×
8) c CNN (32 × 32) d CNN (16 × 16) e Ground truth
Fig. 6 Lung segmentation results for complex segmentation cases a Input lung CT slice b CNN
(8 × 8) c CNN (32 × 32) d CNN (16 × 16) e Ground truth
4 Conclusion
In this study the effect of downsampling of images in the CNN network on the
lung segmentation is studied using three CNN architectures with varied depth of
downsampling of images as 32 × 32, 16 × 16 and 8 × 8. The qualitative performance
of these three networks is studied for the segmentation cases of under-segmentation,
760 S. P. Pawar and S. N. Talbar
over-segmentation and complex cases. The qualitative analysis shows that the CNN
network with six inception layers which downsamples images to 16 × 16 gives far
better performance as compared to the other two networks.
References
1. Brown MS, Mcnitt-Gray MF, Mankovich NJ, Goldin JG, Hiller J, Wilson LS, Aberie D (1997)
Method for segmenting chest CT image data using an anatomical model: preliminary results.
IEEE Trans Med Imaging 16:828–839
2. van Rikxoort EM, de Hoop B, Viergever MA, Prokop M, van Ginneken B (2009) Automatic
lung segmentation from thoracic computed tomography scans using a hybrid approach with
error detection. Med Phys 36:2934–2947
3. Ye X, Lin X, Dehmeshki J, Slabaugh G, Beddoe G (2009) Shape-based computer-aided
detection of lung nodules in thoracic ct images. IEEE Trans Biomed Eng 56:1810–1820
4. Brown MS, Goldin JG, McNitt-Gray MF, Greaser LE, Sapra A, Li K-T, Sayre JW, Martin K,
Aberle DR (2000) Knowledge-based segmentation of thoracic computed tomography images
for assessment of split lung function. Med Phys 27:592–598
5. Sun X, Zhang H, Duan H (2006) 3d computerized segmentation of lung volume with computed
tomography. Acad Radiol 13:670–677
6. Pu J, Roos J, Chin AY, Napel S, Rubin GD, Paik DS (2008) Adaptive border marching algorithm:
automatic lung segmentation on chest ct images. Comput Med Imaging Graph 32:452–462
7. Prasad MN, Brown MS, Ahmad S, Abtin F, Allen J, da Costa I, Kim HJ, McNitt-Gray MF,
Goldin JG (2008) Automatic segmentation of lung parenchyma in the presence of diseases
based on curvature of ribs. Acad Radiol 15:1173–1180
8. Shen S, Bui AA, Cong J, Hsu W (2015) An automated lung segmentation approach using
bidirectional chain codes to improve nodule detection accuracy. Comput Biol Med 57:139–149
9. Hosseini-Asl E, Zurada JM, Gimel’farb G, El-Baz A (2015) 3-d lung segmentation by
incremental constrained nonnegative matrix factorization. IEEE Trans Biomed Eng 63:952–963
10. Hosseini-Asl E, Zurada JM, El-Baz A (2014) Lung segmentation based on nonnegative matrix
factorization. In: 2014 IEEE international conference on image processing 415 (ICIP). IEEE,
pp 877–881
11. Isola P, Zhu J-Y, Zhou T, Efros AA (2017) Image-to-image translation with conditional adver-
sarial networks, In 2017 IEEE conference on computer vision and pattern recognition (CVPR).
IEEE, pp 5967–5976
12. Ulyanov D, Vedaldi A, Lempitsky V (2016) Instance normalization: the missing ingredient for
fast stylization, arXiv preprint arXiv:1607.08022
A Secure Data Sharing Platform Using
Blockchain and Fine-Grained Access
Shamsundar Bhimade, Prashant Bhandare, Amey Bhatlavande,
and Bharati Deokar
1 Introduction
However, it is very crucial to know the three W’s for sharing purpose, such as what,
when, and where. These questions need to be very much clear before initiating the
data sharing process. There is still some scope to work on, how the data set owner
should be given incentives or reward. This research provides secure sharing and
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 761
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_76
762 S. Bhimade et al.
2 Literature Survey
The major underlying problem in research data sharing is the fear of researchers
regarding misuse and misinterpretation of data. This is because data sharing
approaches are still immature in the context of trust, which is slowly going to be
established among research community. To tackle this issue, various solutions are
proposed, for example, protection of identities of every individual and controlled
access to the data rather than making all the data open access. Still these solu-
tions cannot provide trust, immutability to digital data, and traceability regarding
data usage. Cloud servers store the excessive amount of data, which is a centralized
authority. There are various type of risks associated with a central authority, such as
single point failure. To avoid such failure, third parties are involved to provide data
backups. To eliminate third party for developing a trust based model, a blockchain
is introduced to provide trust and transparency. Decentralized storage is a solution,
which allows storage of data independently on multiple nodes of the network in the
form of a distributed ledger. The problem is the storage and processing limitation of
network nodes. For this purpose, interplanetary file system (IPFS) is adapted, which
is a peer to peer architecture [4]. There is no risk of single point failure. It is similar
to web3, but with different features.
On the security of public key protocol Proposes public key encryption protocol,
describes the various techniques to encrypt the public key [1]. First complete group
key management scheme which can supports all these functions yet preserves effi-
ciency. The proposed scheme is based on the new concept of access control poly-
nomial (ACP) that efficiently and effectively support full dynamics, flexible access
control with fine-tuned granularity, and concealment.New scheme is protected from
various attacks from both external and internal malicious parties [2]. Achieving
secure role based control on encrypted data in cloud achieved through RBAC. RBE
scheme allows RBAC policies to be apply for the encrypted data stored in public
clouds. RBE-based hybrid cloud storage architecture provides facility of an orga-
nization to store data securely in a public cloud, while maintaining the sensitive
information related to the organization’s structure in a private cloud [3].
One approach to encrypt documents satisfying different policies with different
keys using a public key cryptosystem such as attribute-based encryption, and proxy
re-encryption is called broadcast group key management (BGKM), and then give a
secure construction of a BGKM scheme called ACVBGKM. Major advantage of the
BGKM scheme is that adding users/revoking users can be performed efficiently by
updating only some public information. BGKM used for an efficient approach for
fine-grained encryption-based access control for documents stored in an untrusted
cloud file storage [4]. MONA proposed a new secure multi-owner data sharing
scheme, for multiple groups in the cloud. They applied the group signature. and
dynamic broadcast encryption techniques, any cloud user can secretly share data
with others. The storage overhead and encryption computation cost of our scheme
are independent with the number of revoked users. Also they analyze the security of
scheme with difficult proofs, and demonstrate the efficiency of scheme in experiments
764 S. Bhimade et al.
3 Proposed System
The data owner define an access policy to enforce dissemination conditions i.e.
read–write conditions. Then he encrypts data for a set of receivers, and outsources
the ciphertext to CSP for sharing and and spreading the conditions. The data co-
owners tagged by data owner can add some access policies to the encrypted data
with CSP and generate the renewed ciphertext. The data distributor can access the
data and also generate the re-encryption key to disseminate data owner’s data to
others if he satisfies enough access policies in the ciphertext. The data accessor can
decrypt the initial, renewed and re-encrypted ciphertext with her or his private key.
Figure 1 indicates different part of system &connection between them.
A Secure Data Sharing Platform Using Blockchain … 765
In the proposed scheme, members are people with the same interests and want to
share data in the cloud using encryption technique. The most worrying problem when
users store data in the cloud server is the confidentiality of the outsourced data. In the
system, users of the same group conduct a key agreement. Subsequently, a common
conference key can be used to encrypt the data that will be uploaded to the cloud
to ensure the confidentiality of the outsourced data. Attackers or the semi-trusted
cloud server cannot learn any content of the outsourced data without the common
conference key. This system uses a technique called group signatures, which allows
users in the same group to anonymously share data in the cloud.
Cloud server provider i.e. CSP provides users with seemingly unlimited storage
services. In addition to providing efficient and convenient storage services for users,
the cloud can also provide data sharing services. However, the cloud has the charac-
teristic of honest but curious. In other words, the cloud will not deliberately delete or
modify the uploaded data of users, but it will be curious to understand the contents
of the stored data and the user’s identity. Figure 2 indicates how data flows between
different parts of the system.
4 Algorithms
with128-bit keys. Rijendeal was founder. In this drop we are using it to encrypt the
data owner file. Figure 3 Shows analysis of existing system & proposed system in
terms of performance.
Input:
Table 1 gives performance analysis between existing system & proposed system.
Table 1 Performance
Existing system Proposed system
analysis
(ms) (ms)
Key generation 500 450
time
Encryption time 1500 1200
Decryption time 10 9
768 S. Bhimade et al.
5 Conclusion
In this paper, a blockchain-based secure data sharing and delivery of digital assets
framework is presented. The main aim of this proposed scenario is to provide data
authenticity and quality of data to customer as well as a stable business platform for
owner This system is design for secure data sharing scheme, for dynamic groups
in an untruth cloud. A new type authentication system, which is highly secure, has
been proposed in this system. User is able to share data with others in the group
without disclose identity privacy to the cloud. It also supports efficient user revocation
and new user joining. User revocation can be done through a public revocation list
without updating the private keys of the remaining users, and new users can directly
decrypt files stored in the cloud before their participation. System also provides the
new double encryption technique for data security. New re-encryption provides tight
authentication.
References
1. Dole D, Yao AC (1983) On the security of public key protocols. IEEE Trans Inf Theory
29(2):198–208
2. Zoo X, Dai Y-S, Bettino E (2008) A practical and flexible key management mechanism for
trusted collaborative computing. In: Proceedings IEEE conference computer common. pp
1211–1219
3. Zhou L, Varadharajan V, Hitchens M (2013) Achieving secure role-based access control on
encrypted data in cloud storage. IEEE Trans Inf Forensics Secur 8(12):1947–1960
4. Nabeel M, Shang N, Bertino E (2013) Privacy preserving policy based content sharing in public
clouds. IEEE Trans Know Data Eng 25(11):2602–2614
5. Liu X, Zhang Y, Wang B, Yang J (2013) Mona: Secure multi owner data sharing for dynamic
groups in the cloud. IEEE Trans Parallel Distrib Syst 24(6):1182–1191
6. Zhu Z, Jiang Z, Jiang R (2013) The attack on Mona: secure multi owner data sharing for
dynamic groups in the cloud. İn: Processing ınternational conference ınformation science cloud
computer. pp 185–189
7. Khan BUI, Olanrewaju RF (2015) SSM: secure-split-merge data distribution in cloud ınfras-
tructure. İn: 2015 IEEE conference on open systems (ICOS). August 24–26, Melaka,
Malaysia
8. Xu J, Wen Q, Li W, Ji Z (2013) Circuit ciphertext-policy attribute-based hybrid encryption with
verifiable delegation in cloud computing. IEEE Trans Parallel Distrib Syst 24(6):1182–1191
9. Liu JK, Au MH, Huang X, Lu R, Li J (2013) Fine-grained two-factor access control for
web-based cloud computing services. IEEE Trans Parallel Distrib Syst 24(6):1182–1191
10. Sultan NH, Barbhuiya FA (2016) A secure re-encryption scheme for data sharing in unreliable
cloud environment 978-1-5090-2616-6/16 IEEE https://doi.org/10.1109/SERVICES.2016.16
11. Pasupuleti S, Ramalingam S, Buyya R (2016) An efficient and secure privacy-preserving
approach for outsourced data of resource constrained mobile devices in cloud computing.
J Netw Comput Appl 64:12–22
12. Sun W, Wang B, Cao N, Li H, Lou W, Hou Y, Li H (2014) Privacy preserving multi-keyword text
search in the cloud supporting similarity based ranking. IEEE T ParallDistr 25(11):3025–3035
13. Peters GW, Panayi E (2016) Understanding modern banking ledgers through blockchain tech-
nologies: future of transaction processing and smart contracts on the internet of money. İn:
Banking beyond banks and money. Springer, New York, NY, USA pp 239–278
Efficient and Interactive Fuzzy Type
Ahead Search in XML Data
Abstract This paper we are written for XML document to store XML data in XML
formats for Security purpose. Here we compute the problem of efficiently creating
ranked result for keyword search query in XML document. In old methods there are
Xlink, Xpath and Xquery are query methods available to search data in XML file
of XML DB. Here the method new users are not able to understand syntax of query
when accessing the query. In this steps first write query, put forward to the system
and retrieve relevant results. In case of keyword search there is the fuzzy type ahead
search over XML data that user write a keyword search on fly way and access a new
information pattern, This method are optional to old methods. The users didn’t need
to know knowledge of XML query languages and its syntax. We are also adding a user
study confirming that keyword-based search in SQL for a range of DB retrieval task.
The query time, the text index carry keyword-based searches with giving interactive
answer. Successful keyword search is valuable for top-k in XML document, these
are user simply manage, semantic and steer into documents.
1 Introduction
In old keyword based search system in XML data, A user write a keyword query
and submit it to system, Retrieves information. In fact particular person know about
language that what is Xpath and Xquery languages, What are the syntax, notation
etc., of them Because without syntax, no one able to retrieve data. We are study of
effective search in XML data. The system is searching XML data on the users type
in doubt keyword. It is allows to user discover data as they have type, If even in
available minor errors in their keyword. We are proposing effective index structure
and top-k algorithms to achieve a more interactive rate. We observe effective ranking
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 769
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_77
770 L. Dethe et al.
functions and early massacre techniques to progressively know the top-k relevant
results [1, 2].
The today’s day most of the transactions on the internet XML are used to storing
and retrieving reason. Lots of leading products developed companies are use XML
metadata structure. This paper progress with a goal to manage XML information. It
is helps in storing and retrieving relevant answers. In this case consumer has limited
knowledge of the data, frequently the users be aware of left in dark when issuing
the query, and has to use a try and see draw near for publishing, finding, managing,
retrieving data from DB in XML formats and updating storing data in Document of
XML. There is special modules of this paper.
The one modules is a SQL manager, which help to retrieve and manage data from
XML data in XML DB and we implement keyword search in XML data in XML
database. The user security and management are another modules. DB servers is
Client–Server based DB. It is the more easy to retrieve, user-friendly and easy to
access the DB for both programmer and the client of understand. Actually It is used
to create database, report, query and the table [3].
2 Literature Survey
The method used frequently is Auto complete which is predicts expression that the
user may have type in standard on the unfinished string the user has been type. Here
the trouble with Auto complete. The system delicacy a query as a single string, If it
consist several keywords. There is one answer to this problem set by Bast and Weber
is that Complete Search in textual documents. It can find approximate answers by
allowing keywords search in query, that come out at any seats in the answer. Fuzzy
search can gives user instant replay as users type in keyword. It doesn’t need users to
type in complete keywords. The Fuzzy search can assist users browse the information
that users save typing effort and efficiently search the results.
We also considered fuzzy search in relational DB. XML data in a Fuzzy search
mode and it is not negligible to swell existing method to support fuzzy search in
XML data, Because XML has parent–child bonds. We need to recognize absolute
XML sub trees that confine such structural bonds from XML data to result the queries
with keyword. TASX find the XML data on the fly as user’s type in query keywords
still in the occurrence of the errors of their input keywords. Every query with many
keywords needs to be answered strongly. The leading challenge is search-efficiency.
This short running-time requirement is mainly difficult when the backend repos-
itory has a large amount of data. We suggest successful index structures and algo-
rithms to answer keyword queries in XML data. Efficient ranking method and timely
extinction techniques to gradually find out top-k answers.
XML stands for Extensible Markup language. The word “Extensible” imply that
a developer can expand his ability to describe a document, and describe meaningful
tags for his purpose XML is used to generate vibrant content. Databases are study of
SQL-SERVER, ORACLE, My SQL, XML are done in the portion of manipulating the
Efficient and Interactive Fuzzy Type Ahead Search in XML Data 771
stored facts by their respective query language. XML database assists professionals
and the corporate to trace and maintain the data into the database. For using the
over specified database corporate has to pay valued amount as per the company rules
and regulations for receiving the registration from the authorized database compa-
nies. Installation cost, maintenance cost and the accomplishment cost can affect the
company’s production price. The XML database is a platform self-governing server
database and can be used at no cost provided by the Sun Microsystems [4].
In XML nearby are two types Xpath and Xquery. Xpath is declarative language
for XML that give a simple syntax for addressing a fraction of on Xml file. Xpath set
of element can be retrieved by specifying a index like path with zero or more state
place on the lane. Xpath delight an a XML document as a rational tree with nodes
for each part, attribute text, processing instruction, comment, namespace and root
[1, 5]. The essential of the addressing method is the context node (begin node) and
location path which show a path from one point in an XML document to a further.
Xpointer can be used state on complete location or relative location. Location of path
is composed of a chain of steps joined with “/” each travel down the previous step.
Xquery is fit in feature from query language for relational scheme (SQL) and Object
oriented scheme (OQL) [4].
Xquery holds up operations on document sort and can negative, take out and
reform documents. W3c query running group has planned a query language for
XML called Xquery. Values always state a series node can be a document, element,
attribute, text, namespace. peak level path state are ordered according to their location
in the unique hierarchy, top-down, left–right array [6]. The significant parts are Data-
Centric document and Document-Centric file. Data-centric file Xpath are compound
for known. It can invent both in the database and outer surface the database. These
files are used for communicate data among companies. These are mainly processing
by engine; they have quite regular arrangement, fine-gained data and no combine
content.
Document-Centric are document typically designed for human utilization, they
are usually composed openly in XML or some other plan (RTF, PDF, SGML) which is
then transformed to XML. Document-Centric require not have regular arrangement,
bigger gained data and plenty of mixed substance [3, 7]. In this paper to study
of previous technology that they are working LCA (lowest common ancestor) [8],
ELCA (Exclusive Lowest common Ancestor) [8], MCT (minimum cost tree) [6, 9]
and begin new technology Top-k algorithm [1, 5, 10] recognize fairly accurate best
ranking solution in system in XML document more effectively and professionally.
on the foundation of tables and with esteem to their queries. For accessing, creating
and maintaining the DB. Users should have the agreement from the server. Server
allowance the permission and later that client (user) can perform what he desires to
perform. Client can analysis only the encrypted from of data, Because of the entire
data are preserved in the XML DB in decrypted from what a client cannot able
perceive that. For the safety point of view. It has exacting user with their passwords,
who are the approved persons who can access the DB.
This is the query analyzer DB to which several users can access the DB at the
same time with no limits. It is a platform independent DB and more cost-effective
than any other DB. We recommend the index to improve search routine. We can use
“arbitrary access” depends on the index to do an early killing in the algorithm. That
is, set an XML element and a key keyword. We can got the relative score of the
keyword and the element uses the index, Except retrieving inverted lists. Fagin et al
algorithm have proven that the threshold-based algorithm using arbitrary access is
best overall algorithm that properly find the top-k results.
It is very costly to build the union lists of each input keywords as there may be
several predicted words and many reversed lists as an alternative, we can produce
a partial virtual list on the fly of user. We only utilize the element in the partial
effective list to calculate the top-k results. The partial virtual record can stay away
from accessing all the element of inverted lists of forecasted words. It is only needs
to retrieve those with large scores, and if we have calculated the top-k results using
the partial retrieved elements, we can do an early killing and do not require to visit
other element on the inverted element lists.
This system huge number of safety options give to data and users. Administrator
has the majority of job to form user and allows agreement to maintain DB. When
the users are entering username and private key into the system to login for use, He
will do work on the particular data to store data into document format, access data
from document file present in DB, all together he departs to temp directory where
document is stored only on his data observe into temp folder not extra person data,
Because when users log out them file delete from folder preserve DB safety of every
users. One more is data traveling form one user account to document DB or other
person, Data is encrypted and preserve consistency into particular network.
The XML server consists of different modules. The GUI Clients build the connection
with server with help own username and password. The SQL manager establishing
the connection with respective DB, parse the input query in proper syntax to display
answers Grid or in File format. The entire syntax’s for DDL and DML queries are
amassed in “syntax.xml” file.
The Example syntax: Data storage space with the popularity of XML of the server
require to effort with and lay up XML data. The Example XML file format is:
Efficient and Interactive Fuzzy Type Ahead Search in XML Data 773
<content>
<author>Laxman</author>
<title>The fuzzy search</title>
</content>
Association builder of xml server establish relation among two XML documents,
applying the concept of primary key. The user executive helps for managing users
by editing users, Creating users, Deleting users, Editing/Assigning their password.
The Server will give its own verification features to validate users using facility
of XML Encryption/Decryption, unauthorized persons or intruders cannot retrieve
important document. In this segment three significant XML query and keyword look
for methodologies are explained. The major problem connected to Xpath and Xquery
are their difficulty concerned in the syntax for query. The compared to Xpath and
Xquery, LCA-based interactive search [11] and MCT [6] are superior and excellent.
below part give detailed information on the above discussed methods as shown in
Fig. 1.
keyword query Q = {k1, k2, k3, k4,…, kl}, a MCT of query Q and node N is the sub
tree available at N, and for every keyword ki e Q, If node N is a qussi-content node
of the ki , that sub tree contain the pivotal path for ki and node particular node N.
We have to first understand the predicated words for every input keyword. After,
we will design the MCT for each node in the XML document tree depends on the
predicated words and revisit the good ones with the larger score. The important
advantage of this is even if a node doesn’t have descendent nodes that contain all the
keyword in the query input, This node could immobile be considered as a potential
result [12, 13].
To Reduce the limitation of LCA based method exclusive LCA (ELCA) [12, 13] is
designed. It tells that an LCA is ELCA if it is immobile an LCA after without its
LCA descendants. Finding the separate LCA of every contain node is called ELCA.
XU and papakonstantion [13] designed a binary-search based method to competently
find ELCAs.
This paper we are first checking it that how top-k search based algorithm are benefi-
cial. However ranking the results of keyword it needed to LCA and MCT with them
picky score [11, 14]. The parameterized top-k algorithm separated in two different
stages. The first one is a structure of algorithm that on a problem occurrence construct
a structure of possible size and the another stage is an enumerating the algorithm
that gives the k best answers to the instance depends on the structure. We are devel-
oped new methods that bear efficient enumerating algorithms. We design the relation
among fixed-parameter tractability and parameterized top-k algorithm [1, 4].
Efficient and Interactive Fuzzy Type Ahead Search in XML Data 775
Now we converse how to design the MCT for a node N as result to the query.
Naturally, we first estimate the significance among node N and every input keywords
and then join this relevance score as total score of MCT. We need to target on different
methods to measure the relevance of node N to a query keywords and join relevance
scores [3, 7, 15].
A. Ranking the Sub Tree
There are have only two different ranking function to calculate the rank and score
between node N and keyword ki .
Case 1: N has keyword ki .
The score of that node N and keyword ki is computed by
where,
1. tf (ki , N)—Number of incidence of ki in sub tree rooted N
2. idf (ki )—Ratio among no. of nodes in XML to no. of nodes that have keyword
ki
3. itl(n)—Inverse term length of |N/Nmax| = node with max terms s-set to 0.2
Case 2: Node N doesn’t have keyword ki, but its descendent got ki . The ranking
depends on ancestor and descendent relation. The another ranking function to find
the score between N and kj is:
where
4 Conclusion
We evaluate the effectiveness of the calculating the prefixes on the trees. This are like
to a query keywords. The proposed efficient incremental algorithm to react to single-
keyword queries that are declared as prefix states. considered special algorithms for
computing the results to a query with several keywords. Well-planned algorithms are
developed for incrementally calculating results to queries by using cached answers
of previous queries in sort to get a high interactive rate on large data sets. The
LCA-based methods to interactively find out the predicted results and the designed
a minimum cost trees based search methods to capably know the nearly all related
results. We devise a fuzzy search to additional improve search routine.
References
1. Feng J, Li G (May 2012) Efficient fuzzy type-ahead search in XML data. Proc IEEE Trans
Knowl Data Eng 24(5)
2. Sivapuja S, Mohiddin SK, Srikanth Babu S, Srikar Babu SV (March–April 2013) Efficient
searching on data using forward search. J Emerg Trends Technol Comput Sci 2(2)
3. Bast H, Weber I (2006) Type less, find more: fast autocompletion search with a succinct index.
In: Processing Asnn. int’l ACM SIGIR conference research and development in information
retrieval (SIGIR) pp 364–371
4. Andrew Eisenberg IBM (2002) Advancement in SQL/XML jim meton oracle corp
5. Willimson H (2009) The complete reference of XML. The McGrew-Hill Companies Inc., New
York
6. Ding B, Yu JX, Wang S, Qin L, Zhang X, Lin X Finding top-k min-cost—connected tree in
database”.The Chinese university of Hong Kong China
7. Li G, Feng JH, Zhou L, Interactive search in XML data department of computer science and
technology, Tshinghua national laboratory for information science and technology, Tsinghua
university. Beijing, China, 100084
8. Teradat YX (2003) Yannis papakonstantion university of California. Efficient LCA based
keyword search in XML data. ACM copyright
9. Bourret R (2005) XML and database, independent consultant, felton, A 18 woodwardia Ave.
Felton CA, USA Spring, 95018
10. Li D, Li C, Feng J, Zhou L SAIL: structure-aware indexing for effective and progressive
top-k keyword search over XML document. Department of computer science, University of
California, Irvine, CA USA pp 92697–3435
11. Liu Z, Chen Y ( 2007) Identifying meaningful return information for Xml keyword search.
Proc ACM SIGMOD In Conf Manage Data 329–340
Efficient and Interactive Fuzzy Type Ahead Search in XML Data 777
12. Bast H, Weber I (2007) The complete search engine: interactive, efficient, and towards Ir&db
integration. In: Biennial conference innovative data systems research (CIDR) pp 88–95
13. Ji S, Li G, Li C, Feng J (2009) Efficient interactive fuzzy keyword search. In: Processing Int’l
conference world wide web
14. Chen L, Kanj LA, Meng J, Xia G, Zhange F (2012) Parameterized top-k algorithm,
communicated by D-Z DU
15. Harel D, Tarjan RE (1984) Fast algorithms for finding nearest common ancestors. J SIAM
Comput 13(2):338–355
IOT Based Interactive Motion Detection
Security System Using Raspberry Pi
Abstract The Internet of Things (IoT) is the network of physical devices with elec-
tronics, software, sensors and other objects which consists of an embedded system
that enables to collect and exchange data. It is mainly intended for transferring
user data in real time security monitoring system, e.g. to monitor and control traffic
and road condition. Unquestionably, everyday day-life and behavior of potential
subscribers have adhered to IOT. From this perspective domestic and work areas will
have effects of IoT. It is anticipated that there will about 50 billion internet-enabled
devices by 2020. The aim of this paper is to introduce security alarm system for
detecting motion and get an image or notification using low processing power when
motion is detected. The user is alerted by sending snapshots through mail or notifi-
cation via text message. In case of unavailability of the network service, Raspberry
Pi will store the data locally and send that data when the internet is available. Rasp-
berry Pi is a low-cost device as compared to other available present systems. It is a
small sized computer used to process a captured image or data as and when motion
is detected. Passive Infrared (PIR) sensors are used to detect the motion; the image
is captured through the camera and provisionally stored in the raspberry pi module.
This system is suitable for small personal area i.e. personal office cabin, bank locker
room, parking entrance.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 779
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_78
780 G. Khare et al.
1 Introduction
A British technology pioneer, Kevin Ashton, through his work of Auto-ID Center
at the Massachusetts Institute of Technology (MIT), introduced a term Internet of
Things in 2002 [1]. Nowadays the Internet of Things is used in numerous research
areas like in the fields of home automation, transportation, agriculture, automobiles,
and healthcare etc. Also, the IoT will change everyday life and help in situations like
managing airports’ passenger flows, smart homes, heating buildings, caring for the
elderly, military, defense. Home automation brings up the application of computer
and information technology for controlling home appliances and domestic features
[2]. This results from convenience, energy efficiency and safety benefits along with
improved quality of life. The vision of IoT is “from anytime, anywhere, connectivity
for anything” [3]. In the existing system, CCTV cameras are used for surveillance but
it incurs the high cost due to more hardware and storage for continuous recording of
activities. Also, human intervention is required to detect any unauthorized activity.
The IoT enables physical objects to see, hear, think and perform jobs to share
information and to coordinate decisions. The type of communication that we
experience today is either human–human or human–machine but the IoT assures
machine-machine communication. The overall idea of the IoT is illustrated in
Fig. 1, in which every internet enabled device communicates with other indepen-
dent services, wherein each domain sensors interact directly with each other.The
basic idea behind IoT is to permit an autonomous exchange of useful information
using internet-enabled devices.
The IoT uses low-cost computing devices shown in Fig. 1.where there are less
energy consumption and limited impact on the environment. Through IoT, billions of
devices will be connected to each other including cars, phones, jet planes, appliances,
wearable gear etc.
There are many challenges to the use of IoT. Since every device needs IPv4 to
address for communication. Currently, the IPv4 has only 4.3 billion unique addresses,
which will be exhausted soon and we need to adapt to IPv6 [4]. Therefore one of
the challenges to be met is limited availability of IPv4 address. Another challenge in
IoT would be data storage as billions of devices are collecting and transferring the
data and that needs to be stored for which massive storage space is required. While
dealing with IoT, more and more authenticate information is being collected from
the devices. Hence, we have to make sure that the security policies are in place to
protect the data from the hands of hackers [5]. Now a day the need of people is to
monitor and control their business activities or house while moving across the globe.
To achieve an economical and safe way of monitoring and the controlling Internet
of things can be utilized effectively. The focus of IoT is to allow things to become
smarter, more reliable and more autonomous. Sensors have to play a key role in the
field of security systems. The use of sensors depends upon the type of application
like ultrasonic sensor, photoelectric sensor, temperature sensor, and passive infrared
(PIR) sensor.
2 Proposed Algorithm
The existing system has many disadvantages. It is difficult to implement and very
complex. It requires more human effort and hardware. So, the entire system is expen-
sive. These can be overcome by the use of interactive motion detection security system
using Raspberry Pi in IOT.
This system consists of three layers—the first layer is for motion detection. In
the second layer actions are processed according to Python script and in third layer
recorded snapshot or video is sent to the user.
Raspberry Pi (shown in Fig. 2) is a credit card sized computer and also known as
Model B + have the following features:
• 700 MHz ARM CPU
• 512 MB SDRAM
• Ethernet RJ45
• 2 × USB 2.0
• HDMI and Composite RCA.
Raspberry Pi works same as that of a standard PC. It needs a keyboard for entering
the command, a monitor for the display unit, an operating system, and a power supply.
Internet connectivity is provided via Ethernet or Wi-Fi. The Raspberry Pi board
consists of an inbuilt processor and a graphics chip, main memory (RAM), different
interfaces and connectors. It runs on a LINUX based operating system Raspbian
OS [6]. Raspbian has a desktop environment similar to windows and Mac called
Lightweight X11 Desktop Environment (LXDE), so it provides an easy transition
782 G. Khare et al.
for those not familiar with LINUX command line. External storage device can’t be
used to boot the Raspberry Pi. Some of the pin connections are essential, others are
optional but all Raspberry Pi models have the same CPU named BCM2835 which
is cheap, powerful, and it does not consume a lot of power [7]. Raspberry Pi has a
wide range of usage.
In this security system, PIR sensor (Fig. 3) is connected to the MCP 3208 ADC
via Serial Peripheral Interface (SPI) protocol and detecting a range of PIR sensor is
6 m. It covers the 180° angle.
Python is chosen as the main programming language as it is suitable for real-world
applications. Also, it is fully fledged and easy to learn [8]. Python script compares the
last frame and the current frame of the live video. If there is any mismatch then the
motion flag is set and further events are processed. Upon execution of python script,
snapshots are captured from the camera until the motion is detected. For storing the
statement temporarily an external SD card is used. As there could be a chance that the
person would attempt to damage the device, captured images uploaded immediately
to the external server. Figure 4 shows the GSM Module.
GSM Module Needs.
1. Raspberry Pi 3
2. GSM/GPRS TTL UART Modem-SIM800L
3. A 5 V battery or power supply
4. Connecting wires
5. A mobile phone
6. An extra SIM card.
Initially, the user has to insert a SIM card into GSM modem with suitable connec-
tion with Raspberry Pi. The SIM used in this module should have enough balance
to send the text message and care should be taken that it must be placed in such a
way that it having an appropriate range of a particular network. Give power supply
to GSM module and wait for few seconds for initialization of SIM.
Figure 5 shows a block diagram of the given proposed system. A Passive Infrared
(PIR) sensor is connected to the general I/O pins of the raspberry pi [9].
When the motion is detected, the camera module records the snapshot or video for
the assigned time or until motion is detected. An external SD card is used for storing
the snapshot temporarily in the raspberry pi. An internet connection is required for
sending email and text message notifications. If the server is not available then the
snapshots are stored in the raspberry pi [10].
784 G. Khare et al.
When the experimental set up is initiated and motion is detected through the PIR
sensor, the web camera has captured a series of images and these snapshots are sent to
the Gmail account of the registered user. Also, the notification through text message
has been given to the same user on mobile. Hence, it can be inferred that the security
system is successfully implemented with more reliability and efficiency.
4 Conclusion
Raspberry Pi provides a smart, efficient and economic platform for implementing the
security system and home automation. This security system mainly focuses on motion
detection and it enables the user to monitor their homes and buildings even from
remote locations. This system could be an alternative for expensive security systems
being used in the present days. This system does not need any special modifications
in the infrastructure at the place where it is installed. It can be implemented without
much difficulty.
References
4. Atzori L, Iera A, Morabito G (2010) The internet of things: a survey. Comput Netw
54(15):2787–2805
5. Nilssen A (2009) Security and privacy standardization in the internet of things, in eMatch’09—
future internet workshop, Oslo, Norway
6. Patil N, Anand K Ward: IOT and raspberry PI based environmental monitoring application
7. Pawar AA, Rangole JS (Feb 2, 2016) Review paper on raspberry Pi and wi-fi based home
automation server 3
8. Ajeeth, Sandhyaraani M (June 3, 2016) Security system for industries using raspberry Pi and
IOT. 2
9. Soundhar Ganesh S et al (2015) Raspberry Pi-based interactive home automation system
through internet of things. 3(3)
10. Anitha G et al (2015) An internet of things approach for motion detection and controlling home
appliances without cloud server. 3(8)
Commercially Successful Rural
and Agricultural Technologies
Cognitive Intelligence of Internet
of Things in Precision Agriculture
Abstract In the recent scenario, the technology of the Internet of Things (IoT)
acting a vital part in Precision agriculture, Military, Engineering applications. The
main resource of our country is the agriculture field. IoT is widely adopted in the
Precision Agriculture field to count the dissimilar environmental constraints such as
soil moisture, humidity, temperature, PH value of soil, etc. for increasing the yield
of the crop. While using the IoT in Precision Agriculture it aided to decrease the
consumption of the natural assets (freshwater, clean air, healthy soils, etc.) used
in farming. Therefore, the purpose of this work is to implement the several IoT
technologies accepted for Precision agriculture. This work has also points to the
different communication technologies and wireless sensors available for Precision
Agriculture. The proposed system will very helpful to our farmers because these
technologies applied for utilize the limited resources to increase the yield of crop.
These technologies applied cognitively at exact location and at exact time so quantity
and quality of crop will definitely increase.
1 Introduction
The important field of any nation is farming. The cardinal source of economy is
agriculture. As concerned to the Indian economy, agriculture is the main sector. The
Indian agriculture sector employs almost 50% of the countries. India is acknowledged
to the world’s major creator of pulses, rice, wheat, and spice yields that they produce,
which trusts on plant growth and the yield they acquire.
R. K. Patil (B)
Department of Technology, Shivaji University, Kolhapur, India
S. S. Patil
College of Engineering, Karmveer Bhaurao Patil, Satara, India
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 789
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_79
790 R. K. Patil and S. S. Patil
Plant growth and yield depend on three parameters one is monitoring the land,
second is continuous alert for identify the diseases and third is drip irrigation manage-
ment. The Indian farmers manually worked for these three parameters. Due to this
human errors will occur and yield will be reduced.
Cognitive intelligence of Internet of Things (IoT) in agriculture applications acting
a vital role to meet the above parameters for increasing yield through resulting of
improving the efficiency by effective use of inputs like soil, fertilizers, and pesticides,
monitoring the livestock, predict the pest and diseases, scanning storage capacities
like water tank level, monitor the crops are fed and watered well.
In the current scenario as concern to our Indian farmers whose agriculture land has
limited and financially weaker, the agriculture field wants such tools and expertise that
increases the effectiveness of quality of production and decreases the environmental
effect on the yield.
The Internet of things along with the Wireless Sensor Network (WSN) in agricul-
ture may place forward the central solution role to precision agriculture. Precision
agriculture is the technique of applying the correct amount of input (water, fertilizer,
pesticides, etc.) at the accurate location and at the right time to improve produc-
tion and increase quality while protecting the environment [1]. WSN is defined as a
gathering of nodes structured into a supportive network.
Internet of Things (IoT) stand-in a important role in Precision agriculture. IoT
is an internetworking of physical kits, automobiles, constructions, and other items
embedded with electronics, software, sensors, actuators, and network connectivity
that assists these devices to gather and exchange the information.
2 Literature Review
Agriculture techniques. We used three research database as follows for finding the
existing related work in the domain of precision agriculture using WSNs.
• IEEE (www.ieee.org)
• Springer (www.springerlink.com)
• Science Direct (www.sciencedirect.com).
Yuthika Shekhar et al. [2] proposes irrigation management system, the moisture
and temperature data are took using KNN (K Nearest Neighbour) cataloging machine
learning algorithm for getting the sensor data for irrigating the soil with water.
P Rajalakshmi and S. Devi Mahalakshmi introduced the Sensors are the essential
components for precise agricultural applications. Depending on the content of the
moisture, obtained from soil moisture sensor, the motor automatically switches on.
If the moisture content is very low in soil then automatically sprinkles on to the roots
of plants [3].
Karan Kansara et al. [4] presents the irrigations practiced during winter, was far
overdid the irrigation water want. While in the summer, the irrigation accepted have
Cognitive Intelligence of Internet of Things in Precision … 791
covered 44% of irrigation water requests. The revision reveals a high variety in the
practice of irrigation between different principles and different farms.
Avatade and Dhanure [5]. This paper defines the system construction on hardware
and software design. Jia Uddin et al. [6] introduce the uses ARMs and RF modules
and the system will be associated using these modules. The main essential aspect
of this system is the radio frequency module which is used to send and receive the
data to the controller. They are mentioned that this system uses three nodes that
communicate with each other and irrigate the field involuntarily without man power.
The goal of this project is to modernizing agriculture technology.
Bai and Liang [7] presents finest consumptin of water is the major idea of this
irrigation system to lessen water utilization. This project uses soil moisture sensors
to find the amount of water in their cultivation and also uses sensor to notice water
level in the tank.
3 Methodology
• Adruino Uno
• Soil moisture sensor
• ESP8266 wifi module
• Motor driver PCB for Arduino.
Adruino Uno: The Arduino Uno is a prototype platform (Open Source) based on
easy to use hardware and software. Arduino Uno is microcontroller board based on
the ATmega328. It covers everything required to backing the microcontroller.
Soil moisture sensor: It is used to count the moisture level in the soil. It can
calculate the ideal soil moisture content for different species of plants. We can place
the points of the sensor into soil and check the moisture level, but the position of the
prongs should be right.
ESP8266 wifi module: It is a self- contained SOC with integrated TCP/IP protocol
stack that can provide any microcontroller admittance to your WiFi network. The
ESP8266 is skilful of either hosting an application or offloading all Wi-Fi networking
purposes from alternative application processor.
Motor driver PCB for Arduino: The shield and its consistent Arduino library
create it easy to regulator two bidirectional, high power, brushed DC motors with an
Arduino or Arduino clone.
792 R. K. Patil and S. S. Patil
or data by giving them the best prediction. In my project, we can use naïve Bayes
classifier.
The suggested system can be reflected of having different levels, to gather sensed
data, and actuate the end devices based on sensed data. The end devices are interfaced
with the Arduino.
A. DHT 11: It has both temperature and humidity sensor. DHT 11 has the best
structures are its simple connection, better quality, less costly, and interference
capability and precise calibration. It can shield the transmission area up to 25 m.
B. Soil Moisture Sensor: It sense the amount of water content in the soil can be
counted.
C. Light Dependent Resistor (LDR): It sense the light inside the greenhouse area.
If the light intensity changes then LDR responds with variable resistance. The
value of resistance of LDR reduces as the light intensity growths.
D. Inlet and Exhaust Fans: Inlet and Exhaust fans are used to push the heat outside
from the greenhouse and pull fresh air from outside into the greenhouse. This
maintains the suitable airflow essential for the in greenhouse crops.
E. Water Pump: The water pump reacts when the soil moisture level below or
above the specified threshold. The water pump starts and switches on or off
respectively according to conditions.
F. ThingSpeak: It is an Internet of Thing analytics platform service that deals you
to collective, visualize, and analyze live data streams in the cloud. You can also
drive data to ThingSpeak from our Arduino, generate instant visualizations of
live data. It is possible without setting up servers or developing web software,
ThingSpeak allows engineers and scientists to make a prototype and form IoT
systems.
G. Dropbox Capture a web page screenshot on web page Save Record.
H. For identification of disease, image processing is used. Following algorithm is
used for the same.
• The picture of camera is taken.
• The median filter is performed by taking the size of majority of the vectors
inside a veil and arranging the extents.
• The photo is given as a input for the system.
• The framework will utilize the picture handling strategies.
• After enhancement, division is finished then compare with thresholding and
take appropriate decision for increase yield of crops through IoT.
The projected system works fine and carried effective outcomes. The finale devices
like light source, water pump, and fans inside the greenhouse activated according
to the threshold condition of parameters like temperature, humidity, soil moisture
values. The evidence cached from the Arduino pass to the Thingspeak server and
794 R. K. Patil and S. S. Patil
data pictured in the server. The values collected by the sensor DHT11 temperature
and humidity and the soil moisture sensor. In field, the humidity values are different.
The discrepancy is in the form of increasing and decreasing and stable sometimes. So
whenever the humidity value less than or equal to the threshold value the fan switched
on automatically. In Field, the temperature values are different. The discrepancy
is in the form of increasing and decreasing and stable sometimes. So whenever
the humidity value greater than or equal to the threshold value the fan activated
on automatically In field, the soil moisture values are varied. The discrepancy is
in the form of increasing and decreasing and stable sometimes. So whenever the
humidity value larger than or equal to the threshold value the water pump switched
on automatically. In the same way, field visualized in the Thingspeak. The light will
be on and off according to the LDR sensor module threshold value condition.
5 Conclusion
The data collected in the server is visualized. The collected data is to be used for
formulating the good crop yield and appropriate crops. The forthcoming scope of
this project is to add extra sensors for better maintenance.
References
1. https://www.hypertech.co.il/product/smart-agriculture/-SmartAgriculture-HyperTech
2. Shekhar Y, Dagur E, Mishra S (2017) Intelligent IoT based automated irrigation system. Int J
Appl Eng Res ISSN 0973-4562 12(18):7306–7320
3. Rajalakshmi P, Devi Mahalakshmi S (2016) IOT based crop-field monitoring and irrigation
automation. Intell Syst Control (ISCO), 2016 10th Int Conf IEEE https://doi.org/10.1109/ISCO.
2016.7726900
4. Kansara K, Zaveri V, Shah S, Delwadkar S, Jani K (2016) Sensor based automated irrigation
system with IOT: a technical review, (IJCSIT). Int J Comput Sci Inform Technol 6(6):5331–5333
5. Avatade SS, Dhanure SP (May 5, 2015) Irrigation system using a wireless sensor network and
GPRS. Int J Advan Res Comput Commun Eng 4(5)
6. Uddin J, Taslim Reza SM, Newaz Q, Uddin J, Islam T, MyonKim J (2012) Automated irrigation
system using solar power, ©2012 IEEE
7. Bai D, Liang W (Oct 2012) Optimal planning model of the regional water saving irrigation and
its application. In: 2012 International Symposium on Geomatics for Integrated Water Resources
Management (GIWRM). IEEE, pp 1–4
8. Case study on software data reduction techniques use for effective bug triage published on Dec
13, 2017 https://www.irjet.net/archives/V3/i11/IRJET-V3I11255.pdf
9. A part of bug report for bug 284541 in Eclipse. www.researchgate.net › figure › A-part-of-bug-
report
Flooring: A Risk Factor for Fall-Related
Injuries in Elderly People Housing
Abstract Flooring is one of the built features which is playing a crucial role in
elderly housing and also reported as one of the major cause for fall-related injuries
among elderly people. A house with in appropriate flooring leads to falls which
is a major serious accident to people in old age (Designing supportive spaces for
the elderly with the right floors, https://professionals.tarkett.com/en_EU/node/des
igning-supportive-spaces-for-the-elderly-with-the-right-floors-1092, [1]). The study
aimed to ascertain the gap between existing flooring design and necessities of the
elderly people. The elderly women aged sixty and above were the subjects of the
study. The sample was drawn from Kurnool district of Andhra Pradesh. Most of the
elderly people houses were not provided with flooring that has minimum chances
for falls. The elderly people preferred to have non-slippery flooring, provision of
sound-absorbing materials and floor with different colors at various levels to avoid
risk of falls in houses.
1 Introduction
A fall is defined as an event which results in a person coming to rest accidentally on the
ground or other lower level or floor level. Now a day’s falls are the second major cause
for injuries and death among the elderly people worldwide and reported as major
public health problem [2, 3]. Flooring in houses is also one of the environmental
risk factors which include home threats. Home hazards such as cluttered flooring
and floor furnishings predict falls among elderly people [4]. Carrying heavy objects,
walking on slippery floors and poor lighting may increase the risk of falls [5]. Kitchen
is the place where most non-fall injuries (31%) occurs and in contrast, fall-related
injuries happen most frequently (20%) in the bedroom [6].
U. F. Rupanagudi (B)
Department of Family Resource Management, APGC, Acharya N. G. Ranga Agricultural
University, Guntur, Andhra Pradesh, India
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 797
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_80
798 U. F. Rupanagudi
The risk level of fall-related injuries among the elderly may be influenced by
physical, sensory and cognitive levels related to combination of environments and
housing requirements that are not designed for an ageing population [3]. Caring for
the housing of elderly people is becoming one of the greatest challenges in society.
Elderly people feel safer when the house has provided with physical features that
are specifically designed for elderly people keeping in view their needs [7]. Shiny
textured and dark-colored flooring will influence elderly people perception of the
environment as unsafe [8]. Physical features in interiors of the house that support
safety and independence need to be carefully considered while designing the house
for the elderly [9]. The present study aimed to identify the gap between existing
flooring conditions and requirements of the elderly regarding flooring design that
assists the elderly people to live safely and comfortably without any risk in the
houses.
The study was conducted at Kurnool district of Andhra Pradesh state in India. A
total sample of 60 elderly women in the age of sixty and above who were living in
a separate residence without children with or without spouse was the criteria for the
selection of sample for the study. The study was conducted during the year 2019–
2020. Purposive sampling method was adopted to draw the sample. Standard design
guidelines that ease the elderly people to have a movement without any trouble
proposed by various researchers assisted as a base for identifying the features to
measure the present flooring in the houses of the elderly. The flooring was physi-
cally observed and evaluated against the recommended guidelines. An interview cum
observation schedule was developed. Depending on the presence and absence of the
design feature, scoring was given. Score 3,2,1 were given in case the existing feature
was ‘above the recommended guidelines’, ‘exactly as per the recommended guide-
lines’ and ‘below the recommended guidelines’ respectively. Design requirements
of flooring specified by respondents were quantified in terms of essential, preferred
and neutral with scores 3, 2 and 1. The probable score each respondent can get was
between 8 and 24. The results were interpreted such that the higher the score, higher
was the probability of the flooring as per the recommended guidelines. Frequencies
and percentages were calculated for the existing flooring and needs of the elderly
people. Chi-square statistical analysis was carried out to find out the gap between
existing flooring and design requirements of the elderly people.
Flooring: A Risk Factor for Fall-Related Injuries in Elderly … 799
Flooring provided had to be non-slippery both outside and inside the house to avoid
the risk of falls. Slip-resistant floor finishes should be used and shiny reflective floors
such as marble, glazed tiles and the like should be avoided [10–14].
Flooring should visually contrast with the walls, handrails and sanitary fittings.
Avoid using transparent glass materials and mats on floors. The home should be
provided with a large amount of clear floor space. Loose carpets should be avoided
because of the risk of tripping. Provision of bedside carpets in the bedroom should be
avoided [10, 14]. Guiding blocks provided at the walkway should be of red chequered
tile with smooth bricks or stone finish [11].
Floor patterns such as stripes that could be mistaken for steps should not be used for
floors in corridors. The slope of floors should not be greater than1:20 in case of unlev-
elled floor surfaces. If the greater slope is adopted, the floor should be designed as a
ramp. For people with low vision, lines of brightly coloured fluorescent tape should
be placed on the floor surface to assist mobility in poorly lighted areas. Carpeting
should be avoided in circulation areas. Carpets provided in the circulation spaces
should not be deeper than 0.4 inches and has to be securely fixed and firm cushion.
Backing and exposed edges of carpets should be fastened to the floor surface and
trimmed along the entire length of the exposed edge [12]. Flooring should be smooth
and levelled for materials such as ceramic tile or brick or stone. Door thresholds and
minor changes in the floor level should be avoided [13]. Floor surface provided has
to be of matt finish. Projections and variation in floor level that obstruct the easy
access for persons with disabilities should be avoided [10, 12, 14].
Eight standard design guidelines were identified to assess the design of flooring in
houses. Slightly less than three fourth of the houses had no non-slippery floors and
68.33% had loose carpets or rugs in the house. No house was provided with a wooden
floor to accommodate wheel chair users and no provision was made to assist people
with low vision. The majority (91%) of the houses had flooring that contrasts with the
wall. Three-fifths of the houses had avoided bedside carpets in the bedroom. Ninety-
five per cent of the houses had no sound-absorbing materials for walls and floors to
avoid echoes and 90% of houses had no differentiation of color of floor levels and
areas of different surfaces. The existing flooring features in elderly housing were
presented in the Table 1.
800 U. F. Rupanagudi
No extra effort was taken to make flooring non-slippery in the existing elderly
houses. Similar findings regarding bathroom flooring were reported in the study [15].
The provisions to make the floor safe and assist the elderly to lead a comfortable life
in the old age were not made. The flooring was one of the components in the housing
highly neglected to suit to the needs of the elderly.
Table 1 (continued)
Recommended Position of existing flooring against the guidelines Total
Design Above the Exactly as per the Below the
guidelines recommended recommended recommended
guidelines guidelines guidelines
N % N % N % N %
Different 2 3.33 4 6.67 54 90 60 100
colors were
used for
different floor
levels, zonings
or areas of
different
functional
surfaces
Chi-square analysis was done to find out the association between existing flooring
design and needs of the elderly. The null hypothesis formulated was.
H0 There exists no significant association between existing flooring design
and requirements of elderly people.
More than half (66.67%) of the elderly preferred to have flooring as per the recom-
mended guidelines. Provision of non-slippery flooring in the house, wooden flooring
to accommodate wheelchair users, floor color that visually contrasts with handrails,
walls to enable the elderly to have easy movement were recommended by various
802 U. F. Rupanagudi
4 Conclusion
The elderly felt flooring as one of the important feature that should be considered
while designing the house for elderly people to reduce the risk of fall-related injuries.
The design of flooring was not found as per the recommended guidelines in most of
the houses. The recommended guidelines can be adopted for housing while planning
a house for elderly people, wheelchair users, people with various abilities to avoid
risk of falls in houses.
Flooring: A Risk Factor for Fall-Related Injuries in Elderly … 803
Table 3 Association between the existing flooring design and requirements of the elderly people
n = 60
Existing The design requirements regarding flooring
flooring design Neutral Preferred Essential Total
features
N % N % N % N %
Below the 1 1.67 17 28.33 1 1.67 19 31.67
recommended
guidelines
Exactly as per 1 1.67 40 66.67 0 0.00 41 68.33
the
recommended
guidelines
Total 2 3.33 57 95.00 1 1.67 60 100.00
χ2 value 2.5579
Probability 0.2332
value
References
1. Designing supportive spaces for the elderly with the right floors. https://professionals.tarkett.
com/en_EU/node/designing-supportive-spaces-for-the-elderly-with-the-right-floors-1092
2. Sibley KM, Voth J, Munce SE, Straus SE, Jaglal SB (2014) Chronic disease and falls
in community-dwelling Canadians over 65 years old: a population-based study exploring
associations with number and pattern of chronic conditions. BMC Geriatr 14:22
3. World Health Organization. https://www.who.int/news-room/factsheets/detail/falls#:~:text=
Falls%20are%20the%20second%20leading,greatest%20number%20of%20fatal%20falls
4. Northridge ME, Nevitt MC, Kelsey JL, Link B (1995) Home hazards and falls in the elderly:
the role of health and functional status. Am J Public Health 85(4):509–515
5. Stevens M, Holman J, Bennett N (2001) Preventing falls in older people: impact of an inter-
vention to reduce environmental hazards in the home. J Am Geriatrics Soc 49(11):1442–1447
6. Carter SE, Campbell EM, Sanson-Fisher RW, Gillespie WJ (2000) Accidents in older people
living at home: a community-based study assessing prevalence, type, location and injuries.
Aust N Z J Public Health 24(6):633–636
7. Zamora T, Alcantara E, Artacho MA, Cloquell V (2008) Influence of pavement design
parameters in safety perception in the elderly. Int J Ind Ergon 38(11):992–998
8. Bamzar R (2018) Assessing the quality of the indoor environment of senior housing for better
mobility: a Swedish case study. J Housing Built Environ 34:23–60
9. Engineer A, Esther MS, Bijan N (2018) Designing interiors to mitigate physical and cognitive
deficits related to aging and to promote longevity in older adults: a review. Gerontology 64:612–
622
10. Design guidelines for the elderly and elderly with frailty, https://www.lwb.gov.hk/en/consult_p
aper/BFA_ch6.pdf
11. Guidelines and space standards for barrier-free built environment for disabled and elderly
persons. https://cpwd.gov.in/Publication/aged&disabled.PDF
12. National building code (2016) Requirements for accessibility in the built environment for elders
and persons with disabilities. Bureau of Indian standards, New Delhi, vol 1 pp 49–136. https://
standardsbis.bsbedge.com/
13. Parker WR (1987) Housing for the elderly. In: DeChaira J, Callender J (eds) Time-saver
standards for building types 2nd edn. McGraw-hill international editions. pp 87–101
804 U. F. Rupanagudi
14. Guidelines for the planning of houses for senior citizens. https://www.housinglin.org.uk/_ass
ets/Resources/Housing/Support_materials/Other_reports_and_guidance/
15. Nagananda MS, Sengupta A, Santhosh J, Anand S, Rehman SMK, Khan AM, Rautray P,
Gharai D, Das LK (2010) Circumstances and consequences of falls in community-dwelling
older women. Wseas Trans Biol Biomed 7(4):287–305
Analysis of Construction Readiness
Parameters for Highway Projects
Abstract Highways contribute greatly to the economy and growth of a country. For
highway building projects to complete on time and on budget, it is therefore necessary.
However, when the project begins prematurely, interruption often occurs, resulting
in delays, which have had numerous negative results on all project shareholders.
Previous origins are one of the causes of construction hold up. Before studying the
information, however, there is little understanding of whether a highway is premature
or ready to build. The purpose of this research is to find out the criteria consumed in
action to determine whether a road construction project is prepared or not. To this
end, interviews are conducted and analyzed with sixteen practitioners working on
road construction projects. Important attached (1) building readiness can be evaluate
even during the launch phase; and (2) failure to comply with construction prepara-
tion requirements can lead to job delay, wasteful operation, reworking, and labour,
equipment or material deficiencies. The study provides to the current knowledge
by defining criteria which point out whether a road project is ready to be built or
not. Learnings from this study will prevent premature initiation of road construction
projects by industry.
1 Introduction
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 805
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_81
806 H. R. Godbole and R. C. Charpe
negative impact on India’s economic growth due to hold up and cost annul [2]. Conse-
quently, it is important for economic and social growth to recognize approaches for
improving incentives to have a successful road construction project.
When finished exact time, within the budget and for stakeholder gratification, a
construction project has been commonly recognized as fruitful. Since at least one
stakeholder is benefitting from early construction, almost always project teams are
under pressure to start constructing, whether or not they are really ready to start.
Premature building begins when a determination is taken by at least one party to
start building at a cost which surpass an adequate indulgence of a party and may
lead to an interruption of construction. Over rusts of costs, sliding schedules, extra
time and poor output are among the most usual and major jolts on projects that do
not start ready [3]. The industry must therefore prevent early start-ups in a building
project which will inevitably have various undesirable impacts. The preparation for
construction is outlined as the series of actions and process to be finished or essentially
finished before construction in order to begin and carry the operations productively.
Many projects are started before they get ready, which leads to decreased capacity
and displeasing project execution (premature start). Construction-ready projects are
reduced on normal by 22% in schedule, 29% in productivity improvements, 20%
in cost savings, 7% in rework savings and 21% less in change in comparison with
projects that are not ready for construction [4]. Moreover, sufficient readiness for
construction can help to avoid many of the factors that lead to sequence works and
ultimately improve project performance [4]. Furthermore, before starting construc-
tion it is important for industry experts to assess the preparation of the sector. This
research examines the parameters used in practice to distinguish between a ready or
unprepared highway projects. To accomplish this goal, this article discusses research
questions concerning the following: What metrics can be used to disseminate between
whether or not a road project is ready for construction? By interviewing seventeen
industry professionals who are employed on road building projects in India, the
writers address the issue. The data derived from the interviews were then investi-
gates with themes. At last, a series of criteria for construction readiness is devel-
oped for road building’. The instruction from this research could prevent industry
escape premature beginning of road building projects by inspecting the construction
readiness of a project.
2 Background
Building readiness is one of the subjects that researchers have researched. A prepara-
tion evaluation identifies potential challenges for the implementation of new policies,
structures and processes in supervisory conditions. The objective of a readiness eval-
uation is to decide if latent obstacles to accomplishment exist and to allow personals
Analysis of Construction Readiness Parameters for Highway … 807
or teams, before starting a project, to overcome these barriers. Assessing the project’s
readiness to build helps the project team avoid premature commencement by quan-
titatively control whether a project is ready for construction with an improvement
in building cost performance, scheduling and development efficiency [4]. Previous
research has identified criteria for assessing building preparation in general to avoid
premature construction starts. However, the characteristics of different kinds of
construction projects vary. Of example, road buildings are horizontal, require more
territories and are physically larger than houses. It also includes working around a
wide and demanding range of current conditions, including roads, wetlands, under-
water or overland networks. The construction preparation requirements common to
highway projects are therefore worth evaluating.
In India researchers and industry professionals have carried out numerous studies,
because of the significance of road construction for the growth of an economic and
societal region. The main five key advance aspect for Butterworth Outer Ring Road
BORR Expressway in India were identified as an efficient communication system,
excellent project governance, public and private sector accountability and commit-
ment and competitive acquisition and delegation ability [5]. Meanwhile, inadequate
planning, temperature, bad spot management, insufficient site analysis and under-
ground infrastructure are the top five circumstances that assist to the retardation of
road building projects. Five aspects of green highway jargon were pointed out as
general understandings by stakeholders: conservation and habitat protection, energy
and emission mitigation for life cycles, recycling, recycling, re-use and recycling,
wetlands based storm water management, and overall corporate benefit [6]. Different
studies have examined the common issues in road projects in India as well as the
various aspects of road projects, because this is a major subject. There is still a defi-
ciency of investigation into how the road project should be decided whether or not
it is ready to start construction. This research will therefore occupy the space by
analyzing the criteria to determine the quality of road construction projects.
3 Methods
Data collection consists of the collection of data from interviews with professionals
going on road-building projects. The analysis of the collected data is based on qual-
itative approaches. The following sections discuss how the construction readiness
parameters can be collected and analyzed. Figure 1 summarizes the methodology of
this study for examining the practical parameters of whether a road project is ready
to begin construction or not.
808 H. R. Godbole and R. C. Charpe
data were transcribed, interpreted, re-read and documented by the reporters. The
second phase is the elementary application creation. The writers coded the data for
as many possible topics and trends. The writers then discussed, spoke over, and
decided on any coding enhancements and/or changes. The third stage is to search
for subjects situated on the primary codes. The writers also updated the codes from
the second and the original data from the first stages during the process of creating
the themes. Phase four is to discuss the subjects. The authors tested, identified and
optimized sub-themes constantly in order to ensure that data is saturated, checked
whether the themes work with coded extracts and the whole data set. Phase 5 is
the identification and naming of the subjects. To order to ensure the participants are
responsive to individually coded answers, the authors continue to reverse and reverse
the thematic codes and interpretation of the interview. Finally, the performance of
the study should be recorded (phase six).
4.1.1 Approval
In accordance with interview results, the project has received permission from various
legal bodies, such as regional authorities, utilities, the Department of Work and Safety
(DOWS) and the Construction Industry Development Board (CIDB), as part of the
criteria used to determine if development was ready or not. Project approval refers
to the acquisition of permits by the appropriate authorities to ensure that the plan
complies with the standards set in the building regulations and that local authorities
are monitoring certain phases of construction. Officers conduct inspections on the
building site during construction, in order for work to proceed in quality and compli-
ance as expected. If the inspection officers are not informed of permits, they may
stop their work until a decision has been made to minimize effectiveness and detain.
(a) Received letter of award from client. As drafted acceptance that the
entrepreneur has been fortunate and will obtain the job, the customer shall
send the letter of award to the contractor. Other problems may occur without a
letter, as the owner has given the project to the contractor, due to the absence
of a written and signed contract. There is no agreement without a contract, no
safety for the builder and no assurance that the builder is paid for the work he
does. In fact, no agreement is reached on the terms of the contract when the
project manager actually checks them; the cost for the job is often paid and
can’t persuade the customer to change the terms and conditions more appro-
priate or reasonable before starting the work. If a buyer is too expensive, their
cash flow may be affected. The staff may therefore refuse to work because they
can’t pay their wages on time, as a result of the job stopping. Buildings that
would induce loss of productivity without workers could not be achieved.
(b) Acquisition of land completed. Land accomplishment is the procurement
method for the project of the necessary or mandatory property. It is one of the
key elements in the process of soil development. The land procurement method
gives the public sector with a legal procedure to acquire land for advanced
projects to benefit the country. The construction of roads may require local
people to acquire land for the country’s infrastructure and economic devel-
opment. It is often difficult for many reasons to acquire land from the state,
mainly in conjunction with compensation or land prices, thereby lengthening
the duration of the scheme. If the project has started, the building work will
stop until the purchase of the land is completed and productivity losses result.
(c) Adequate money. Funding for a building project is important. Throughout the
building, project director must make sure that the project has sufficient money.
Failure to finance them could lead employees and caters to decline to work
Analysis of Construction Readiness Parameters for Highway … 811
and to afford equipments on the site which could have a negative impact on
building work. However, a shortage of funds will involve the cash flow of the
project and result in delays in site acquisition, resulting in project delays as a
whole [11].
(d) Acquire project insurance. Building insurance is a method of trade for the
fixed payment based premium to secure the interests of the parties involved
with building projects. In the construction sector, it is an important technique
of leading endanger. Construction is a risky and typically volatile activity that
presents a significant risk of personal injury, real property harm and economical
damage. Insurance shortage can cause workers to reject jobs because they may
not have adequate money to pay medical charges if accidents occur.
(e) Date and period of construction confirmed official commencement. The
starting date is the time when the project schedule starts to work. The construc-
tion time is defined in comparison as the time frame given to the contractor by
the owner to complete the project under normal working conditions. Project
managers need to monitor the start and length of construction and schedule
building work accordingly. Failure to plan a project may lead to the delayed
project [12]. Failure to plan means that the schedule to be followed by workers
is not established. When the workers work on projects, they won’t have a clear
picture. There are no deadlines to reach, so that the workforce’s loose environ-
ment reduces productivity. It means that the project is not finished on a late
basis.
(f) The nearest authority. Building site theft can cause different project problems.
Therefore, to prevent crime at the construction site, work with local authorities
is essential. Project managers should verify and inform local authorities of
the site. For fact, on weekends and holidays, the project managers can also
seek the authority to monitor the site if the contractor is afraid of fraud. To
addition to the direct costs of repairing the stolen goods, the cost of renting the
stolen equipment is also borne and revenue is destroyed. The project schedules
can also be delayed due to interruptions due to a shortage of equipment and
material.
(g) Meeting with consultant and client. The client is the creator of the project,
and contractors include architects, structural engineers and others who are
hired by the customer to carry out expert tasks for a project. Stakeholders will
discuss the outstanding questions on the project and provide feedback during
this meeting. Productivity may be diminished due to unresolved problems.
Furthermore, there could be no disruption to the project timeline to certain
actions on the project.
(h) Verified construction work plan. A work plan is a series of actions or tasks
that have to be carried out well for a successful project. In fact, a job plan
shows all the tasks of a project that are involved and responsible for each task
and when the tasks are completed. It can help the contractor remain organized
during the project work and help him complete the project in time and on
budget. Starting buildings without a work plan could lead to several problems,
including reduced productivity and sliding schedules.
812 H. R. Godbole and R. C. Charpe
(i) Client and consultant approved construction work plan. In order to ensure
no conflict between parties, customer and consultant should know the work plan
that has been established by the contractor Failure to understand the project
work plan could cause stakeholder conflict. If litigations are not managed
adequately, they will lead to project delays, a mentality of teams that is not
decided, project costs are higher and business relations are damaged [13].
However, there may be work stoppage due to the lack of competitiveness and
delay due to a disagreement between stakeholders. Therefore, before beginning
a job, the customer and contractor will approve the work plan.
(a) Drawings approved by consultant. Until project management can start the
project, it is important that the contractor will review and authorize the design
drawing. The approval of a consultancy can ensure compliance with designs,
specifications and standards of construction. During the construction period,
buyers often change the design, which at the same time affects the entire
project life. Until the latest drawing problem from the architect, the contractor
can do the construction work. The job cannot therefore be completed within
the contract period resulting from delays and reduced productivity. Therefore,
drawings must be accepted by the contractor prior to construction in order to
prevent work interruptions.
(b) Governments approved drawings. Authorities are accepting plans to ensure
the safety of new building for human beings and as laid down in the planning
rules. The process of approval tests if the structure is secure and structurally
sound and whether the approved building materials are safe for human life.
Starting with a drawing approved still by an authority might pose a risk to the
plan for a project. If officials at the site check and realize that the drawing is
not allowed, they can stop the works causing delays and lost productivity.
(c) Everything is provided in complete drawings. To guarantee that the project
will be done in a timeline, full sketches are essential. Errors and omissions
may result in project delays causing the drawings to be incomplete. The plan is
vague during building because significant details are not shown in the drawing;
workers cannot continue the work during construction work. The research must
therefore stop immediately until it is possible to obtain details of that particular
drawing which lead to loss of productivity and delay.
(d) Verified variations in drawings and designs. Differences between tenders and
drawings will result in project delay. The bidders (contractors) are selected at
the time of tender. For building purposes, the design drawing shall be provided
by the architect. Sometimes these sketches vary. There are variations. These
differences contribute to rework, quantity adjustments and construction delays
and defects. In addition, refurbishments result in reduced productivity [14].
Analysis of Construction Readiness Parameters for Highway … 813
4.2.1 On-Site
(a) Site office. The building site includes office facilities to provide project
managers with housing, rooms for meetings, and the site records storage. The
site office is also important to ensure that all documentation of the project
is preserved safely. When the project begins before the site office is built,
documents including project files and sketches can be seen anywhere. Lack of
project documents such as reports or sketches may cause a project schedule to
be disrupted. Of example, before the work can restart, project managers need
to find the missing sketches which lead to lost productivity and delay.
(b) Site condition as per the contract. Project managers should ensure that the
conditions of the site are identical to those stated in the contract and that no
different conditions are observed. The site condition of the project manager
is different when the site condition varies significantly from the terms of a
contract with the owners of the project or what is normally expected in the site
area. The site condition is different. Differing conditions at the site will cause
the project to stall due to work on the site until the relevant stakeholders have
made decisions [15].
(c) Verified utility cables and relocate if necessary. Transfer of the services
means the modification, substitution or relocation as required by the road-
building project: elimination and reconstruction of the facility, procurement
of the appropriate right-of - way, displacement or rearrangement of existing
infrastructure, alteration of the form of utilities and the introduction of the
necessary safety and protection measures [16]. Due to construction, damage
to underground utilities can result in weeks of delays. For example, hitting
a water pipe unintentionally will cause the project to stop momentarily. The
flood will splash and inundate the field, making the site impossible to work.
Public issues and legal problems will also be posed which will prolong the
wait.
(d) Verified traffic diversion and control system. The diversion of traffic system
is designed to prevent traffic around the building site being disrupted by the
project, and to reduce the disruption of traffic if desired. The traffic control
system, on the other hand, relates to a traffic management system in and around
building sites. In terms of forbidding an incident at a building site, these criteria
may be related. Vehicles are often the main cause of deaths and injuries to
workers and public members when reversing, loading and unloading. The root
cause of staff and construction vehicle accidents is inadequate coordination and
monitoring. Accidents involving a building vehicle may result in an eventual
overrun of the time due to an accident. A building workers accident may also
lead to work shortages, which ultimately lead to construction delays.
(e) Safety signboard. (e) In any working environment, safety signs are key. The
primary importance of a safety sign is to ensure that workers are well aware of
814 H. R. Godbole and R. C. Charpe
the potential hazards in particular situations and environments. The safety signs
must be posted. Without signs, there would be a lack of direction from many
workers in times of crisis, so companies could face major legal problems should
accidents occur. Because of insufficient satellite safety accidents involving
construction workers can reduce labour and efficiency, resulting in disruption
of work and eventually a delay in planning [17]. Furthermore, the number of
workers involved in job disruption can be impacted by injuries. Sometimes it
takes more time than expected to find a substitute for the injured workers.
(f) Provision of CCTV and checked building engine parking space. CCTV is
a network which sends TV signals to a limited set of screens and is widely
used for crime prevention in shops and public places. Conversely, the checked
construction machinery parking space is concerned with the design of construc-
tion machinery parking space. To order to prevent the occurrence of crime
on the construction site, CCTV deployment and parking inspection for the
construction machinery can be connected. Robbery and vandalism today on
commercial construction sites.
Industry can influence productivity and income runoff. This is a concern. The
deployment of security cameras is one of the approaches used to prevent crime
at the construction site [18]. In the allocated parking space at the site, in order
to prevent crime, construction equipment should also be stored. Theft will
easily target computers without parking at the site that are left unattended in
remote areas. Companies must pay to replace or temporarily hire the damage
equipment to support their projects in due course. Nevertheless, there are
other consequences, not least the resultant downtime due to the lack of equip-
ment available. It is therefore essential to ensure that pre-construction parking
area for building machinery is provided and that CCTV was installed prior to
construction commencement in order to prevent an equipment shortage.
(g) Verified temporary utilities on site. Several services like water, electricity,
internet connection, lighting, etc., need to be temporarily established before
work on a construction site can begin. Building site can interrupt work on
a building project without electricity, causing a delay. To order for project
managers to coordinate with clients about the job, Internet and telephones are
required on the construction site. One reason behind delays in the Ethiopian
construction industry is the lack of readily available on-site services [19].
(h) Adequate work force. Building workers mean those employed in the building
industry. Workers’ shortages cause project planning interruptions. For example,
inadequate operators on the site will disorder construction, as the workforce
is insufficient for building equipment to operate. In fact, given the preparation
of material, an insufficient worker will control the length of the projects. The
lack of foreign and local workers is one of the most common factors leading
to delays in the Indian building industry [20].
Analysis of Construction Readiness Parameters for Highway … 815
(a) Verified nearest material supplier and quarry. The provider of materials
means any individual providing equipment, materials, services and to be used
in connection with the building contract work. On the other side a quarry is a
place where rocks, sand and gravel are excavated from the ground, dimension
stone, slate building aggregate. The geographical location and delivery lead
time transport and managerial charges may be affected by material supplier or
quarry. The option of a source or quarry close to the site, for example, can help
ensure that the material arrives at the correct time. Providing remote suppliers
or quarries may mean longer delivery times and additional costs for carriage.
Unless a supplier delivers material on time, the processes can be interrupted and
the project conclusion postponed. One of the reasons influencing productivity
is the availability of materials on site.
(b) Equipment availability. A variety of specialized equipment is necessary for
road construction. A lack of equipment affects labour productivity as equip-
ment supports the construction process is used to support the workforce. The
digger is used, for instance, to excavate and move a large object. Building
material could not be relocated without this resulted in a delay of the project
and work halt. In the Indian construction industry, machinery supply is also
one reason for delay [21].
5 Conclusion
References
1. Chandra A, Thompson E (2000) Does public infrastructure affect economic activity?: Evidence
from the rural interstate highway system. Reg Sci Urban Econ 30(4):457–490
2. Bank Negara India (2018) Bank Negara India quarterly bulletin Q4
3. Griego R, Leite F (2016) Premature construction start interruption: how awareness could
prevent disputes and litigations. J Leg Aff Dispute Resolut Eng Constr 9(2):04516016
4. Ibrahim MW (2018) Improving project performance by mitigating out-of-sequence work and
assessing construction readiness. The university of Wisconsin-Madison
5. Ghazali FM, Rashid SA, Sadullah AM (2017) The critical success factors for public-private
partnership highway construction project in India. J Eng Technol (JET) 8(1)
6. Nusa FNM, Nasir S, Endut IR (2018) Awareness of green highway concept and terminology: a
perspective of on-site personnel in Indian highway construction industry. Advan Trans Logistics
Res 1(1):475–487
7. Turner DW III (2010) Qualitative interview design: a practical guide for novice investigators.
Qual Rep 15(3):754–760
8. Braun V, Clarke V (2006) Using thematic analysis in psychology. Qual Res Psychol 3(2):77–
101
9. Rahman RA, Ayer SK (2017) Prevalent issues in BIM-based construction projects. In:
Proceedings of joint conference on computing in construction, vol 1. pp 645–652
10. Radzi AR, Bokhari HR, Rahman RA, Ayer SK (2019) Key attributes of change agents for
successful technology adoptions in construction companies: a thematic analysis. In: Computing
in civil engineering 2019: data, sensing, and analytics. American Society of Civil Engineers,
Reston, VA, pp 430–437
11. Abdul-Rahman H, Berawi MA, Berawi AR, Mohamed O, Othman M, Yahya IA (2006) Delay
mitigation in the Indian construction industry. J Constr Eng Manage 132(2):125–133
12. Marzouk MM, El-Rasas TI (2014) Analyzing delay causes in egyptian construction projects.
J Adv Res 5(1):49–55
13. Cheung SO, Suen HC (2002) A multi-attribute utility model for dispute resolution strategy
selection. Constr Manage Econ 20(7):557–568
14. Jarkas AM, Bitar CG (2011) Factors affecting construction labor productivity in Kuwait. J
Constr Eng Manage 138(7):811–820
15. Amarasekara WD, Perera BA, Rodrigo MNN (2018) Impact of differing site conditions on
construction projects. J Legal Aff Dispute Resolut Eng Constr 10(3):04518006
16. Jung YJ (2012) Evaluation of subsurface utility engineering for highway projects: benefit–cost
analysis. Tunn Undergr Space Technol 27(1):111–122
17. Durdyev S, Omarov M, Ismail S (2017) Causes of delay in residential construction projects in
Cambodia cogent. Engineering 4(1):1291117
18. Berg R, Hinze J (2005) Theft and vandalism on construction sites. J Constr Eng Manag
131(7):826–833
19. Gebrehiwet T, Luo H (2017) Procedia engineering, vol 196. p 366–374
Analysis of Construction Readiness Parameters for Highway … 817
20. Abdul Kadir MR, Lee WP, Jaafar MS, Sapuan SM, Ali AAA (2005) Factors affecting
construction labour productivity for Indian residential projects. Struct Surv 23(1):42–54
21. Sambasivan M, Soon YW (2007) Causes and effects of delays in Indian construction industry.
Int J Proj Manage 25(5):517–526
Biogas Generation by Utilizing
Agricultural Waste
Abstract Biogas is produced using agricultural waste and cow dung with help of the
anaerobic co-digestion in a biogas digester. Generally cow dung is used as the raw
material for the Biogas generation. In this biogas production using different compo-
sition of the cow dung and agricultural waste as right now cow dung is indispensable
member for the biogas production with agricultural waste. Also other factors such
as retention time, temperature can also be observed while taking different compo-
sitions. Theoretical maximum biogas yield is calculated for different raw material
composition for biogas production. It is an extensive study on design of the biogas
digester and how agricultural waste can be used to produce biogas. Design of experi-
ment was performed with computer software to predict the biogas yield and to decide
important factor responsible for biogas generation.
1 Introduction
Energy is an important factor for the economic progress and further improving the
standard of the living of the people. Due to ever increasing demand of the energy for
industrial as well as commercial sector it has become a major issue. Due to environ-
mental impact of currently used sources of energy as the natural gas, fossil fuels, coal
and crude oil there is need for using alternate renewable sources. Renewable energy
systems are considered as a favorable key that plays an essential part in solving
the issue on sustainable resources as it provides clean and efficient. Energy sources
that are considered renewable are those that are naturally and continually refilled by
P. Chaudhari (B)
K. J. Somaiya College of Engineering, Vidyavihar, Maharashtra, India
e-mail: [email protected]
S. Thakker
Department of Mechanical Engineering, K. J. Somaiya College of Engineering, Vidyavihar,
Maharashtra, India
e-mail: [email protected]
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 819
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_82
820 P. Chaudhari and S. Thakker
nature such as solar, hydro, wind and biogas and are very environment-friendly due
to the fact that these resources produce no carbon emissions.
In this paper Minitab software was used to make a design of experiment for
finding out the optimization of the biogas yield with an agricultural waste, theoretical
maximum biogas potential was also calculated to few agricultural waste. This method
of finding design of experiment model can be utilized for various other agricultural
wastes to optimize the process of biogas generation.
2 Literature Review
The chemical composition of the raw material to be fed to biogas reactor has to be
known to apply this model. This model (refer Fig. 1) utilizes only carbon, hydrogen
oxygen and sulfur as obtained from ultimate analysis of the raw materials. This
ultimate analysis values are obtained from published values for ultimate analysis of
these agricultural wastes.
Below are the values obtained from literature review for the raw material as
bagasse, rice straw and cow dung. As it is a biological data composition mainly
depends on the condition they are harvested and are going to be different as per
conditions. The balance between simplicity and effective production of the biogas
are to be achieved with help of this model.
This model can predict biogas output if it’s considered that reaction goes into
completion. This model is simplified and is used to estimate the theoretical maximum
biogas potential with help of ultimate analysis composition of raw materials. The
composition of various biomasses is shown in Table 1.
Different existing models utilize different approaches to predict the biogas yield.
Buswell and Hatfield developed the stoichiometric formula to predict the quantity
of methane from information of the chemical composition of an agricultural waste
Table 1 Composition of
Raw material Rice straw Bagasse Cow dung
various biomasses
C 34.11 49.2 49.04
H 6.53 4.69 6.43
O 59.19 0.18 2.25
N 0.17 43 41.94
S 0.1 0.02 0.34
A + C1 B C2 C + C3 D + C4 D + C5 F
Where A and B are reactant and C, D, E, F are the products and constant of the
reactions are:
b c 3d e b c 3d e
C1 = a − − + + C2 = a + − − −
4 2 4 2 8 4 8 4
a b c 3d e
C3 = − + + + C4 = d C5 = e
2 8 4 8 4
It can be seen that reaction can be applied to any biomass which is used to produce
the biogas with anaerobic digestion. This model makes and assumption that these
elements are the sole components of the raw biomass. Ultimate masses of the feed-
stock are taken to calculate the maximum methane potential, which are used as the
variables.
a · mass · ultimate a · mass · ultimate
a= =
mmC 12.0107
b · mass · ultimate b · mass · ultimate
b= =
mm H 1.0079
Biogas Generation by Utilizing Agricultural Waste 823
3.2 Calculations
By using Boyle modified formula we can estimate the biogas yield from the elemental
composition of biomass. The theoretical biochemical methane potential (TBMP) of
the material is calculated in ml CH gVS−1 as:
22.4 × a + b8 − 4c − 3d8 − 4e
TBMP = (1)
12.017a + 1.0079b + 15.999c + 14.0067d + 32.065e
4 Design of Experiment
In this study Biogas yield from glass silage is optimized and compared with the
mechanical beating of the glass silage as well as varying temperatures in the
controlled conditions. An experimental data required for the DOE was taken from
the paper by FATMA A. ALFARJANI [7]. In this paper anaerobic digestion of glass
silage is performed and beating time and temperature were taken as main factors.
It is often required to find the conditions, which would optimize the process of
requirement. We require the process input parameters values for which we will obtain
optimized response.
Design matrix is generated by the Minitab software as per the conditions (refer
Table 3) fed to the software then experiments are performed according to the design
matrix. Glass silage was used as the biomass waste in the given paper and face
centered composite design was used for design matrix generation.
Experiment was performed for the values given by the design matrix and Table 4
shows that biogas yield in cc for collections as it was collected over span of 24 days
with taking reading every third day so it shows collections.
Biogas Generation by Utilizing Agricultural Waste 825
If the data obtained from the experiments is plotted on the graph then we get the
better picture of the effect of the temperature and beating time on the biogas yield.
Figure 2 illustrate contour plot for the effect of beating time and temperature on
826 P. Chaudhari and S. Thakker
biogas yield for the seventh collection. Which shows that biogas yield is higher for
lower beating time and as beating time is increased biogas yield decreases. As for
temperature higher the temperature higher will be the biogas yield.
5 Conclusion
Biogas Generation is studied with Anaerobic Digestion. It was found that maximum
theoretical potential reduces as the cow dung in reduced and agricultural waste
is added. In case of Cow dung maximum TBMP was calculated. For Rice husk
maximum biogas yield affects more negatively than that of bagasse. Design of experi-
ment Response Surface Method showed that Temperature is the major factor affecting
the biogas yield than beating time. Optimal condition obtained at high tempera-
ture and low beating time, as beating time increases it affects negatively on yield.
Experiments can be performed for the different agricultural waste and for varying
compositions to find optimum composition and condition for biogas production.
Biogas Generation by Utilizing Agricultural Waste 827
References
1. Kigozi R, Muzenda E, Aboyade AO (2014) Biogas technology, current trends, opportunities and
challenges. In: Proceedings of the 6th international conference GREEN technology, renewable
energy and environmental engineering. Cape Town, South Africa, pp 27–28
2. Ilaboya IR, Asekhame FF, Ezugwu MO, Erameh AA, Omofuma FE (2010) Studies on biogas
generation from agricultural waste; analysis of the effects of alkaline on gas generation. Worl
Appl Sci J 9:537–545
3. Vats N, Khan AA, Ahmad K (2019) Observation of biogas production by sugarcane bagasse
and food waste in different composition combinations. Ener 185:1100–1105
4. Almomani F (2019) Prediction of biogas production from chemically treated co-digested
agricultural waste using artificial neural network. Fuel 280:118573–118586
5. Kucukkara B, Yaldiz O, Sozer S, Ertekin C (2011) Biogas production from agricultural wastes
in laboratory scale biogas plant. J Agri Machina Sci 7:373–378
6. Achinas S, Gerrit JWE (2016) Theoretical analysis of biogas potential prediction from
agricultural waste. Resou Effic Techno 2:143–147
7. Alfarjani F (2012) Design and optimization of process parameters in bio-gas production
systems. Dublin City University, Diss
8. Buswell AM (1939) Anaerobic fermentations. Bulletin (Illinois State Water Survey). 32
9. Mittal S, Ahlgren E, Shukla PR (2019) Future biogas resource potential in India: a bottom-up
analysis. Renew Ener 141:379–389
10. Guarino G, Carotenuto C, Cristofaro F, Papa S, Morrone B, Minale M (2016) Does the C/N ratio
really affect the bio-methane yield? a three years investigation of buffalo manure digestion.
Chem Engg Trans 49:463–468
11. Chandratre SJ, Chaudhari V, Kulkarni B, Mahajan B, Bavaskar KP (2015) Biogas production
from local agricultural waste by using laboratory scale digester. Res J Rec Sci 4:157–165
12. Mogal V (2016) A review on quality by design. Pharm Biol Eval 3:313–319
Effect of Change in the Resilient Modulus
of Bituminous Mix on the Design
of Flexible Perpetual Pavement
Abstract High modulus bituminous mix is one of attractive option to increase load-
bearing capacity of pavement structure against conventional structural distresses,
such as rutting and fatigue crack. A comparative study was carried out between five
combinations of flexible pavement mentioned in IRC guidelines with conventional
perpetual pavement design criteria. All five combinations were subjected to high
modulus bituminous mix of 3000 MPa to 8000 MPa. The perpetual pavements were
designed by using mechanistic-empirical design software IITPAVE. The results were
further validated using the design software KENPAVE and WESLEA. The study
investigates the effect of high modulus bituminous mix in conventional perpetual
pavement design with respect to life cycle cost analysis and overall pavement thick-
ness. It is found that use high modulus bituminous mix results in substantial decrease
in pavement thickness.
1 Introduction
A perpetual pavement is a long lasting pavement designed to last for about 50 years.
It does not undergo major structural rehabilitation or reconstruction and needs only
periodic surface renewal in response to distresses confined to the top of the pave-
ment [1]. The mechanistic empirical approach is generally observed for analysis
and design of perpetual pavements. Various limiting values of strain for different
layers of pavement are considered while designing or analysing the performance of
perpetual pavements. As a perpetual pavement structure is subjected to only non-
structural deteriorations, only a periodic surface renewal is generally required. The
critical characteristic of perpetual pavements lies in the fact that it is never replaced
entirely and removed. The point that only the surface layer is renewed with the base
structure staying in place, there is a considerable reduction of construction materials
[2]. There were no official standard provisions for the design of perpetual pavement
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 829
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_83
830 S. Kulkarni and M. Ranadive
in India until the year 2012. However, with the latest publication of the guidelines
by the Indian Roads Congress [1] the design to satisfy 50 year design periods can be
adopted with mechanistic pavement design methodology.
The tensile strain acting in the horizontal direction at the bottom of asphalt layer
and the compressive strain acting in the vertical direction at the top of subgrade are
considered as the critical strains in any pavement structure. The general principle
of perpetual pavement design is to keep these strains in some particular limits, and
many researchers have adopted a maximum limit of 70 microstrains (µ) and 200
microstrains (µ) for horizontal tensile strain and vertical compressive strain respec-
tively [3]. Indian Roads Congress has proposed the strain values of 80µ and 200µ
respectively for the fatigue and rutting endurance limit [4]. The Endurance Limit
(EL) is a level of strain below which there is no cumulative damage over an infinite
number of cycles. A bituminous layer experiencing strain levels less than EL should
not fail due to fatigue. Similarly, if compressive strain acting in the vertical direction
at the top of subgrade is less than the EL, it should not fail due to rutting. Signifi-
cance of EL is that, such a limit would provide a thickness limit for the pavement and
increasing the thickness beyond this limiting thickness would provide no increased
structural resistance to fatigue and rutting damage.
In IRC 37:2018, five different categories of flexible pavements are given. The five
trial combinations proposed are:
Effect of Change in the Resilient Modulus of Bituminous … 831
For the present analysis, the fatigue model suggested in the Indian Road Congress
(IRC 37:2018) guidelines was adopted. The model was calibrated in the research
scheme R-56 (1999) studies using the pavement performance data collected during
the research scheme R-6 (1995) studies and R-19 (1994) studies 90% (Eqs. 1 and 3)
reliability levels.
1 3.89 1 0.854
Nf = 0.561 × C × 10−4 × ( ) ×( ) (1)
εt MRm
where
C = 10 M , and
M = 4.84 × [(V be / V a + V be ) – 0.69]
Va = percent volume of air void in the mix used in the bottom bituminous layer,
Vbe = percent volume of effective bitumen in the mix used in the bottom bituminous
layer, Nf = fatigue life of bituminous layer (cumulative equivalent number of 80 kN
standard axle loads that can be served by the pavement before the critical cracked
area of 20% or more of paved surface area occurs), 1t = maximum horizontal tensile
strain at the bottom of the bottom bituminous layer (DBM) calculated using linear
elastic layered theory by applying standard axle load at the surface of the selected
pavement system, MRm = resilient modulus (MPa) of the bituminous mix used in
the bottom bituminous layer, selected as per the recommendations made in these
guidelines.
832 S. Kulkarni and M. Ranadive
The factor ‘C’ is an adjustment factor used to account for the effect of variation
in the mix volumetric parameters (effective binder volume and air void content) on
the fatigue life of bituminous mixes and was incorporated in the fatigue models
to integrate the mix design considerations in the fatigue performance model. For
this paper, the Va , Vbe and ‘C” factor values considered are 3.5%, 11.5% and 2.35
respectively as recommended in IRC guidelines. So Eq. 1 becomes
1 3.89 1 0.854
Nf = 1.32 × 10−4 × ( ) ×( ) (2)
εt MRm
For the present analysis, the rutting model suggested in the Indian Road Congress
(IRC 37 2012) guidelines was adopted as shown in Eq. 3. The model was calibrated
in the research scheme R-56 (1999) studies using the pavement performance data
collected during the research scheme R-6 (1995) studies and R-19 (1994) studies at
90% reliability levels.
1 4.5337
NR = 1.41 × 10−8 × ( ) (3)
εv
where NR is the cumulative number of repetitions for rutting failure, 1v is the vertical
strain. According to the concept of fatigue endurance limit, if the tensile strain caused
by the traffic in the bituminous layer is less than 70 microstrains, as per tests conducted
in laboratories at 20 °C in USA, the bituminous layer never cracks. Similarly, if the
vertical subgrade strain is less than 200 micro strain, rutting in subgrade will be
negligible. For a pavement temperature of 35 °C, the endurance limit is about 80µ
and 200µ [4]. Hence in this study, the design of pavement is carried out by trial
and error with IITPAVE till endurance limit of the strain values closest to 80µ and
200µ respectively for the fatigue and rutting endurance limit are obtained. Thus, it
will also avoid overly conservative design. For inputs to the IITPAVE, KENPAVE
and WESLEA software, a single axle dual wheel assembly was considered for the
analysis. The standard axle load considered was 80 KN. The contact radius was
assumed as 15.5 cm with a tire pressure of 0.56 MPa. Subgrade California Bearing
Ratio (CBR) is assumed as 10% as per IRC guidelines. The proposed pavement
combinations were designed using guidelines given in IRC 37. In this study, as shown
in Table 2, minimum thickness permissible for sub-base, base and CRL is considered
for all the trial sections and bituminous pavement layers are kept as variable ones
for perpetual pavement design. It was adopted to provide sufficient stiffness in the
upper pavement layers as per the concept of perpetual pavements. Combinations with
CTSB and CTB were checked against fatigue cracking as per the guidelines in IRC
37.
Effect of Change in the Resilient Modulus of Bituminous … 833
High Modulus Bituminous mix which was originally originated from France brought
a significant change in the selection of bituminous materials by road construction
and maintenance agencies [5]. This mix requires stiffer bitumen with a penetration
grade of 10/20 or 15/25 along with an aggregate gradation with good interlocking
properties and accommodating the high amount of bitumen, nearly 6% [6]. These
High Performance bituminous mixes are the mixes that would give value of more
than 3000 MPa of resilient modulus 35 °C. The modulus and poisons ratio of different
pavement layers as mentioned in Table 1 are considered for analysis here (Table 2).
Availability of hard binder is a primarily concern issue as most of the refineries
in India as refineries are producing VG-40 binder which is harder grade at present.
Though it is possible to produce much harder by air blowing process, these binders
will lead to premature cracks [7]. On the other hand, hard binder can be obtained
through different techniques: asphalt derived additives and polyolefin addition to
normal binder Selection of modifiers to produce hard binders is based on ability
Table 1 Recommended material properties for structural layers as per IRC guidelines
Material type Elastic/Resilient modulus (MPa) Poisson’s ratio
Bituminous layer with VG40 3000 to 8000 0.35
Cement treated base 5000 0.25
Crack relief layer 450 0.35
Cement treated sub-base 600 0.25
Unbound granular layers 0.2(h*)0.45 MR Subgrade 0.35
Subgrade 17.6 × (CBR)0.64 = 76.82 0.35
* Thickness of sub-base layer in mm
Table 2 Sample table of thickness of pavement layers for different trial combinations and
corresponding pavement responses from IITPAVE (Resilient modulus of bituminous mix considered
as 3000 Mpa)
Trial Sub-base Base CRL/SAMI DBM BC Total 1t 1v
combination (mm) (µs) (µs)
A 200 (GSB) 150 (WMM) 245 50 645 80 167.3
B 200 (CTSB) 100 100 (CRL) 140 50 590 80 172.9
(CTB)
C 200 (CTSB) 100 SAMI 135 50 485 13.64 199
(CTB)
D 200 (GSB) 100 100 (CRL) 150 50 600 80 171.4
(CTB)
E 200 (CTSB) 150 – 210 50 610 78.53 171.9
(Granular
base course)
834 S. Kulkarni and M. Ranadive
Table 3 Pavement thickness calculations for different modulus and pavement combinations
MRm (MPa) Trial combination and corresponding thickness (mm)
A B C D E
DBM BC DBM BC DBM BC DBM BC DBM BC
3000 245 50 140 50 135 50 150 50 210 50
3500 235 50 130 40 125 50 140 50 190 50
4000 220 50 120 50 120 50 130 50 175 50
4500 205 50 110 50 110 50 120 50 165 50
5000 190 50 100 50 105 55 110 50 155 50
5500 180 50 95 50 100 50 105 50 145 50
6000 170 50 90 50 95 50 95 50 135 50
6500 160 50 80 50 95 50 90 50 130 50
7000 155 55 75 50 90 50 85 50 120 50
7500 145 50 70 50 90 50 80 50 115 50
8000 140 50 65 50 85 50 75 50 110 50
of the binder that yield a binder which is less susceptible to temperature and offers
resistance to major distresses and durability. In general the modifiers used to produce
hard binders should enhance the stiffness of the binders that provides higher perma-
nent deformation resistance at higher temperatures as well as resistance to fatigue
cracking at intermediate and low-temperature domains. From the literature it is found
that Ethylene vinyl acetate (EVA), Styrene Butadiene Styrene (SBS), crumb rubber,
hydrated lime, Gilsonite etc., can improve stiffness of bituminous mix [8].The present
study as depicted in Tables 3 and 4 try to evaluate effect of increased modulus in
case of perpetual pavement design ref Fig. 1.
Life-cycle cost is considered to include the initial cost of construction, cost of reha-
bilitation or reconstruction, and cost of future maintenance regime. In the present
study, LCCA is carried out for 50 years for comparing the overall cost associated
with construction of one-kilometer length of perpetual pavements with 4 lanes i.e. of
14 m width using Net Present Value (NPV) method. The following steps are involved
for the analysis.
1. Determine the initial construction cost.
2. Develop a maintenance plan and find out maintenance cost.
3. Determine life-cycle costs using the NPV method.
Effect of Change in the Resilient Modulus of Bituminous … 835
Table 4 MRm and corresponding pavement thickness for perpetual pavement design
MRm (MPa) Trial Combination
A B C D E
Total Pavement Thickness (mm)
3000 645 590 485 600 610
3500 635 580 475 590 590
4000 620 570 470 580 575
4500 605 560 460 570 565
5000 590 550 455 560 555
5500 580 545 450 555 545
6000 570 540 445 545 535
6500 560 530 445 540 530
7000 555 525 440 535 520
7500 545 520 440 530 515
8000 540 515 435 525 510
675
650
Pavement Thickness (mm)
625
600
575
A
550
B
525
500
C
475 D
450 E
425
400
3000 3500 4000 4500 5000 5500 6000 6500 7000 7500 8000
Modulus of Bituminous Mix
The discount rate has been adopted as 10% whereas inflation rate has been adopted
5%. Based on inputs from field engineers dealing with road construction and officials
working with estimates in Government Departments, it was brought to light that while
estimating the future cost of road construction we should assume that the inflation
in price happen only during 50% of the tenure of design period of the pavement.[9].
According to the recommended guidelines [10], a layer of 25 mm BC is to be provided
once every five years in case of flexible pavements. The cost comparison for a 1 km
stretch of 14 m for was done using the standard rates in Indian Rupees as per the state
government’s public works department schedule of rates for year 2019–20 as shown
836 S. Kulkarni and M. Ranadive
Table 6 Year-wise maintenance schedule and economics for perpetual pavements (millions INR)
Year Maintenance for proposed rigid Cost per K.M Inflation @5.0% per annum NPV
pavements
5th Overlay of 25 mm 2.54 3.24 2.01
10th Overlay of 25 mm 2.54 4.14 1.60
15th Overlay of 25 mm 2.54 5.28 1.26
20th Overlay of 25 mm 2.54 6.74 1.00
25th Overlay of 25 mm 2.54 8.60 0.79
30th Overlay of 25 mm 2.54 10.98 0.63
35th Overlay of 25 mm 2.54 14.01 0.50
40th Overlay of 25 mm 2.54 17.88 0.40
45th Overlay of 25 mm 2.54 22.82 0.31
50th Reconstruction and major repair work in 50th year as per concept of Perpetual Pavement
Total 8.50
Table 7 Sample cost calculation of pavement reconstruction (after 50 years) for perpetual pavement
with MR = 3000 MPa (millions INR)
Perpetual pavement design Initial construction cost Inflation @5.0% per annum NPV
combination
A 36.17 122.48 11.30
B 28.40 96.17 8.88
C 25.63 86.79 8.01
D 28.71 97.22 8.97
E 33.38 113.04 10.43
Effect of Change in the Resilient Modulus of Bituminous … 837
Table 8 Sample life-cycle cost analysis per KM for perpetual pavements with respect to NPV for
period of 50 years with MR =3000 MPa (millions INR)
Type of pavement Initial construction Maintenance Major maintenance Total
(Overlay) (Reconstruction)
A 36.17 8.50 11.30 55.97
B 28.40 8.50 8.88 45.78
C 25.63 8.50 8.01 42.14
D 28.71 8.50 8.97 46.18
E 33.38 8.50 10.43 52.32
pavement reconstruction and life cycle cost analysis are shown in Tables 7 and 8
respectively. Life-Cycle Cost Analysis ref Fig. 2 per KM for all the combinations is
shown in Table 9.
Total Cost in million INR
60.00
55.00
50.00 A
45.00 B
40.00 C
35.00 D
30.00 E
3000 3500 4000 4500 5000 5500 6000 6500 7000 7500 8000
Modulus of Bituminous Mix
Table 9 Life-cycle cost analysis per KM for perpetual pavements with respect to NPV for period
of 50 years
Type of Modulus of bituminous mix (MPa)
pavement 3000 3500 4000 4500 5000 5500 6000 6500 7000 7500 8000
Total lifetime cost in INR millions
A 55.97 55.03 53.61 52.20 50.64 49.84 48.90 47.95 47.48 46.39 45.92
B 45.78 44.84 43.89 42.95 42.00 41.53 41.06 40.12 39.64 39.17 38.70
C 42.14 40.88 40.41 39.32 38.69 38.07 37.90 37.90 37.28 37.28 36.66
D 46.18 45.23 44.29 43.35 42.40 41.93 40.99 40.52 40.04 39.50 39.10
E 52.32 50.28 48.87 47.92 46.54 45.59 45.24 44.62 43.67 43.20 42.29
838 S. Kulkarni and M. Ranadive
5 Conclusion
References
1. Indian Road Congress (IRC 37) (2018) Guidelines for design of flexible pavement
2. Ranadive MS, Kulkarni SS (2016) Perpetual pavement for rural roads: a concept. In: National
conference on fifteen years of Pmgsy. Transportation engineering group, civil engineering
department, Indian institute of technology Roorkee, India, p 23
3. Walubita LF, Scullion T (2010) Texas perpetual pavements new design guidelines. Texas
department of transportation and the federal highway administration
4. Indian Road Congress (IRC SP 53) Guidelines on use of modified bitumen in road construction
(Second revision)
5. Serfass JP, Bense P, Pellevoisin P (1997) Properties and new developments of high modulus
asphalt concrete. Eighth international conference on asphalt pavements
6. Geng H, Clopotel CS, Bahia HU (2013) Effects of high modulus asphalt binders on performance
of typical asphalt pavement structures. Constr Build Mater 44:207–213
7. Corte JF (2001) Development and uses of hard-grade asphalt and of high-modulus asphalt
mixes in France. Transp Res Circular 503:12–31
8. Goli A (2018) Rheological investigations on high modulus asphalt binders. In: International
conference on pavement and computational approaches
9. Gupta G et al (2018) Life-cycle cost analysis of brick kiln dust stabilized perpetual pave-
ments for lowering greenhouse gas emissions in India. Urbanization challenges in emerging
economies. ASCE
10. Ministry of Road Transport and Highways Government of India (2013) Specification for road
and bridge works
11. Government of Maharashtra (2019–20) PWD schedule of rates
Analysis of Three-Stage Open Pan Heat
Exchanger Working on Dual Fuel
for Jaggery Making
Abstract From ancient times the jaggery production has been done by the same
traditional process. Jaggery industry provides good employment to the village people
in the rural area. But efficiency for the traditional plant is minimum due to the wall heat
losses in the furnace, stack losses through a chimney. The efficiency of the jaggery
plant is very much less looked upon as compared to other worldwide industries.
Different researchers work on these problems to enhance the thermal performance
of the plant and to minimize the extra usage of bagasse. The introduction of a thermic
fluid heater along with an open pan heat exchanger to the replacement of conventional
furnace in the industry, improves production and efficiency. Besides the thermic fluid
heat exchanger, multiple pans are also introduced for a better heat transfer rate. From
the new method developed the total cycle time reduces up to 40% of the total time
in the traditional method.
1 Introduction
India is a large producer of jaggery. Jaggery is used in many real-life applications like
Ayurveda and medical fields. Jaggery contains 60–85% of sucrose, 3–10% moisture
and 10–15% of invert sugar and also jaggery consist of vitamins, fibers, protein,
glucose, minerals, content, etc. [1]. Not only jaggery but Sugarcane is also serving
as a green biofuel in a developing country [2]. India is the main merchant’s exporters
of Jaggery throughout the world. India sent out 3,41,120.09 MT of Jaggery and
dessert shop items to the world for the value of Rs. 1633.05 Crore in the year 2019–
20. So it is important to protect this industry by solving the problem associated with
it [3].
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 839
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_84
840 A. N. Kore and S. S. Lakade
is used for the transfer of heat to the juice through indirect contact of flue gasses where
wall heat loss and stack losses are present.
So most of the researchers are working on this problem to minimize all these losses
by designing controlled fire furnace, varying the height of the chimney [11], changing
the material of bricks [12], providing dryers to bagasse but still the efficiency of the
plant is not improved by more than 48%.
Jaggery plant with modifying furnace ordinary masonry bricks by fire bricks at
bottom, by changing chimney cross-section from square to circular that optimized
height of the chimney.Which causes less bagasse consumption & more jaggery
production with a better quality of jaggery. It saves around 208 kg bagasse for 8
quintal production per day [13].
2 Project Identification
3 Objective
4 Methodology
One end of TFH is connected to the pan using pipes through which hot thermic fluid
gets flowing and returns to the TFH through a pump and reservoir.
Thermic Fluid: Thermic fluid is heat transfer oils. These are high-performance
products intended for use in the closed indirect heating system. These oils have very
good heat transfer efficiency. The recent liquid goes through the containers of heat
money handler within which heat vitality is going to be changed. HPCL markets
thermal fluid under the name of HYTHERM.
HYTHERM 600—It is the finest quality oil base stocks and is formulated with
elite extra substances to boost execution at higher temperatures. Ability to provide
maximum heat transfer rate in indirect closed fluid flow system up to 320 °C operating
temperature.
Pan–Pan is one of the important elements in the modified plant. It is used to
collect and heat sugarcane juice indirect way through the thermic fluid. There are
three types of pans conferred i.e., clarification, evaporation and crystallization pans.
This is made by using stainless steel. The warmth thermic fluid handler tubes are
organized deep down of pan. The initial pan is rectangular in size and therefore the
arrangement is limpeting. The rest of the two pans are in a circular cross-section with
the spiral arrangement of tubes are available at the bottom of the pan.
Reservoir–Reservoir is used to accommodate excess volume of hot thermic fluid
when it gets heated from thermic fluid heater which is situated before the pump.
Pump: Pump is used to circulate thermic fluid in closed loop cycle. It is use to
force the thermal fluid into the thermic fluid heater. Some amount of force is needed
to pass the thermal fluid in to heater that is completed by pump as shown in Fig. 1.
In the system the pump pushes the thermic fluid to TFH from the reservoir. In the
thermal fluid heater, the bagasse is burnt in combustion chamber and to heat the
thermic fluid up to 220 °C. The blower is used to enhance the combustion process by
providing a controlled airflow rate to the thermic fluid heater. Complete combustion
provides the utmost energy to the thermal fluid and it saves the additional bagasse
consumption. After heating the thermal fluid within the thermic fluid heater it passes
through the main supply header. This header is further connected to inlet sub supply
header through which fluid is travelled to the bottom of the pan1, pan 2 and pan 3
i.e. clarification pan, evaporation pan and crystallization pan respectively through
the pipe. In the first pan, juice temperature is increased up to 100 °C during this
process, scum present in the juice gets removed and clean juice passed to the pan
2 for further process. In pan 2 temperature of juice is increased up to 108 °C. This
juice is further transferred to pan 3 where the temperature of juice is increased up to
striking temperature i.e. 118–120 °C as per the sugarcane juice quality.
This different temperature of juice is maintained by controlling the mass flow rate
of the thermic fluid which is flowing from the bottom of the pan through the tubes.
Analysis of Three-Stage Open Pan Heat Exchanger Working … 843
Inlet and outlet temperature of thermic fluid in the supply header and return header is
maintained constant i.e., 220 °C and 208 °C respectively to avoid thermal imbalance.
This low temperature fluid then return to TFH through main return header and
reservoir as shown in Fig. 1. In second cycle temperature rise for thermic fluid in TFH
required is relatively very less as compared to first cycle so less energy generation
in TFH. Result of this bagasse consumption is relatively very less for this plant.
6 Experimental Procedure
Trials were conducted on a 50 TCD three pan jaggery unit. Daily trials on six batches
per day are taken to measure process time and corresponding temperature and brix of
juice. 1400 lit juice was taken in to pan for every trial. As thermic fluid requires more
time for heating during start of the day, So for every first batch of the day, 120 min to
844 A. N. Kore and S. S. Lakade
130 min were required for completion of process. For later batch it took up to 90 min
to 95 min. Consecutive eleven days trials were conducted on a plant to measure the
process time of jaggery completion and to measure corresponding temperature and
brix of juice. Inlet and outlet temperature of the thermic fluid is maintained at 220–
208 °C during the trial. Digital Refractometer was used to measure the brix of juice
and digital infrared thermometer gun was used to measure the temperature of juice.
Trial was conducted on 1400 lit of sugarcane juice in a rectangular stainless steel
pan1 of 0.5 mm thickness under constant heat input. Heat is provided to sugarcane
juice through hot thermic fluid flowing through hollow pipe attached at the bottom of
the pan1, where temperature of juice is increased up to 100 °C. During this process
scum present in the juice gets separate out and clean juice get transferred in to pan2.
In pan2 evaporation of juice is carried out and maximum amount of water present
in the juice was evaporated and temperature of juice increased up to 108 °C. At the
end relatively concentrated juice gets transferred to pan3 where temperature of juice
increased up to striking temperature of jaggery. For every process in the pan, time
required for its completion and brix of sugarcane juice was measured and compared
with conventional plant.
Figures 2 and 3 depicts that time required for the process completion in conven-
tional plant was around 160 min whereas modified plant need around 95 min. and
it is measured with respect to temperature of juice and brix value at the start and
end of the process of every pan. It shows that to reach striking temperature of juice
conventional plant took more time as compared to modified plant.
For 50 TCD plant average dry bagasse developed after extraction of juice is around
15,000 kg and requirement of bagasse for conventional jaggery making process is
around 18,400 kg, which is more than that of available bagasse as shown in Fig. 4.
So it is necessary to buy additional bagasse to run the plant. Whereas in case of
modified plant it consumes only 7640 kg of bagasse per day. Saved bagasse can be
sold to earn additional revenue for jaggery manufacturer.
140
120
Temperature C
100
O
80 Clarification
60 Evaporator
40 Crystalizer
20 Conventional
0
0 30 60 90 120 150 180
Time in min
Brix Vs Time
90
80
70
Brix in % 60
50
40
30
20
10
0
0 50 100 150 200
Time in min
Conventional Plant Modified Plant
12000
10000 7640
8000
6000
4000
2000
0
Modified Conventional
Types of Plant
7 Conclusion
Acknowledgements Special thanks to Mr. Dhananjay Tekawade, Jaggery Consultant. The exper-
imental work is supported by Cybas Agro Neutraceuticals, Hatvalan Kangaon road, Tal:Daund,
Pune-412215, Maharashtra, India.
846 A. N. Kore and S. S. Lakade
References
1. Sardeshpande VR, Shendage DJ, Pillai IR (2010) Thermal performance evaluation of a four
pan jaggery processing furnace for improvement in energy utilization. Ener 35:4740–4747
2. Adewuyi A (2020) Challenges and prospects of renewable energy in Nigeria: a case of
bioethanol and biodiesel production. Ener Rep 6:77–88
3. APEDA: India export jaggery per year. [Online]. Available: https://agriexchange.apeda.gov.in/
IndExp/PortNew.aspx
4. Rao KSS, Sampathrajan A, Ramjani SA (2003) Efficiency of traditional jaggery making
furnace. Mad Agri 90:184–185
5. Kulkarni SN, Ronge BP (2018) Case study on heat and mass balance of single pan jaggery
plant. Int J Adv Mech Engg 8:143–151
6. Kore A (2018) Unpublished field work at Sangola report
7. Khattak S, Greenough R, Sardeshpande V, Brown N (2018) Exergy analysis of a four pan
jaggery making process. Ener Rep 4:470–477
8. Singh RD, Baboo B, Singh AK, Anwar SI (2009) Performance evaluation of two pan furnace
for jaggery making. J Inst Eng Agric Eng Div 90:27–30
9. Rao P, Das M, Das S (2007) Jaggery—a traditional Indian sweetener. Ind J Tradit Knowl
6:95–102
10. Manjare A, Hole J (2016) Exhaust heat recovery of jaggery making furnace. Int J Sci Res
5:1349–1351
11. Shiralkar KY, Kancharla SK, Shah NG, Mahajani SM (2014) Energy improvements in jaggery
making process. Ener Sust Dev 18:36–48
12. Arya P, Jaiswal UK, Kumar S (2013) Design based improvement in a three pan jaggery making
plant for rural India. Int J Eng Res 2:264–268
13. Madan KH, Jaiswal UK, Kumar JS, Khanna SK (2004) Improvement in gur (jaggery) making
plant for rural areas. J Rur Technol 1:194–196
The Performance and Emission Analysis
of Diesel Engine with Sunflower Biodiesel
1 Introduction
Diesel engines are commonly used in many sectors like transportation, marine
engines, power generation & in agricultural due to their better thermal efficiency,
lower fuel consumption. Hence demand for the fossil diesel fuel is increasing rapidly.
The main problem associated with the CI engine is particulate matter (PM) emission
and oxides of nitrogen (NOx). The reduction of PM and NOX is a difficult task [1]
in conventional CI engines. Numbers of researchers have suggested that bio-diesel
as the replacement completely or can be used partially blended with mineral diesel,
because use of bio-diesel reduces the exhaust emissions. Bio-diesel consists more
oxygen, less carbons and sulphur by weight than the mineral diesel [2]. Different
countries such as USA, Germany, etc., were utilizing bio-diesel mixed with diesel.
The neat palm oil not used as an edible oil in Malaysia it is used as the Palm bio-diesel
as alternative to diesel [3, 4].
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 847
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_85
848 A. S. Joshi and S. Ramesh
Biodiesel have certain advantages like intimacy to diesel properties, local avail-
ability & renewable in nature, good lubricity, safe to store and lack of difficulty in
transportation, low exhaust emissions [5]. However it is having disadvantages like
cold flow & low heat content. This deficiency can be eliminated by choosing of
proper feedstock material for bio-diesel production which is done by the process of
transesterification and use of proper catalyst in the transesterification process. The
use bio-diesel noted that increase in the emission of NOx [6]. Different conclusions
were made for increase of NOx, like advancement in injection [7], high temperature
of flame [8], change in fuel properties like density, volatility and iodine number [9],
and rapid burning due to availability of oxygen [10]. It is also identified that bio-
diesel is capable to reduce soot emission [11, 12] because of sulphurs & aromatic
compounds are not present [13], and existence of fuel-bonded oxygen [14]. Thus it
is important to study performance, combustion and exhaust emissions of bio-diesel
and embrace as alternative fuel for CI engines.
Bio-diesel effectiveness as a substitute fuel is considered on the basis of three
different characteristics. Combustion is a vital process helpful for testing the viability
of the fuel as well as performance efficiency. The main performance characteristics
are thermal efficiency, specific fuel consumption, torque etc. as per as emission
characteristics are concern, Environmental clearance is a key aspect for selection of
fuel to fulfill environmental regulatory norms which are measured for exhaust gases.
The characteristics are mainly divided into three subsections;
1. Performance
2. Emission and
3. Combustion.
1.1 Performance
Efe Şű et al. [17] prepared bio-diesels from five vegetable oils (soybean, canola,
corn, sunflower and hazelnut) by transesterification methanol is utilized as alcohol
and KOH as a catalyst; three stage water washing was did for filtering out contami-
nants from bio-diesel. Uyumaz [18] used transesterification method to prepare bio-
diesel from mustard oil; methanol is utilized as alcohol; reaction temperature was
60 °C methanol mustard oil ratio was 20%. Shrivastava et al. [19] prepared bio-diesels
from two different feedstocks Roselle and Karanja by transesterification method;
methanol is utilized as alcohol, KOH as a catalyst; bio-diesel is washed with pure
water for the removal of methanol content. At last bio-diesel was heated up to 100 °C
for one hour to remove moisture. Mishra et al. [20] produced oil from Simarouba
Glauca by two step transesterification; in first step was acid esterification methanol
is utilized as a alcohol and H2 SO4 as an catalyst; in second step of alkaline ester-
ification methanol and KOH was used as catalyst. Nair et al. [21] produced Neem
bio-diesel by two step transesterification in first step NaOH was used as catalyst; in
second step H2 SO4 is utilized as catalyst in both steps methanol is utilized as alcohol.
Raman et al. [22] derived bio-diesel from rapeseed oil by trans-esterification using
hexane solvent and alkaline catalyst method; NaOH is used as catalyst and methanol
as alcohol.
Goga et al. [23] produced oil from rice bran oil by single stage alkaline transes-
terification process; KOH was used as catalyst and methanol as alcohol. Nathagopal
et al. [24], have prepared Calophyllum inophyllum methyl ester (CIME) from Calo-
phyllum inophyllum oil by three stage transesterification process; first stage was
acid catalyzed Esterification with methanol and concentrated H2 SO4 as catalyst;
second stage esterification KOH was used as an catalyst; third stage was purification,
produced CIME washed with pure (distilled) water followed by heating to remove
moisture. Higher alcohol blends were prepared by addition of higher alcohols (n-
pentanol and n-octanol) on volume basis. Abed et al. [25], have prepared bio-diesel
by transesterification from waste cooking oil (sunflower oil), using methanol and
NaOH as catalyst. Asokan et al. [26], prepared bio-diesel from juliflora oil by two step
transesterification; in acid transesterification, juliflora oil was mixed with methanol
850 A. S. Joshi and S. Ramesh
and sulfuric acid (catalyst), followed by alkali transesterification with methanol and
NaOH as catalyst. From number of studies it is understood that bio-diesel may be
produced from different feedstocks under different conditions and by using different
catalysts.
The literature also reveals that, the different investigators have conducted exper-
iments on different engines using different bio-diesel blend ratio from 5 to 100% by
weight/mass. This makes the difficult to compare the performance. For this reason,
the literature review on engine performance is presented in tabular format for every
one of the engine separately for better understanding. Table 1 illustrate the exper-
imental investigations on the performance and emission of different engines. The
tables describes, author name, engine parameters, fuel blends and interpretation along
with the comments.
2 Methodology
The characterization and testing of sunflower made bio-diesel is done for the compli-
ance with Indian standards. Bio-diesel mixtures of B10 (10 bio-diesel, 90% diesel),
B20 (20% bio-diesel, 80% diesel), B30, B40, B00 etc. are set-up in volume premise.
The physio-chemical characteristics of SFME are enlisted in Table 2.
Table 1 The performance and emission analysis of different engines
S. No. Author Title Fuel blend Interpretation Comments
1 ManKee Lam et al. “Cultivation of microalgae Microalgae Microalgae requires less More economica compare to
for bio-diesel production: A land as compare to other other feedstock
review on upstream and feedstock
downstream processing”
2 M. Vijay Kumar et al. “The impacts on Antioxidant and oxygenated Combustion, performance All characteristic are
combustion, performance additives and emissions competitive
and emissions of bio-diesel characteristics are found out
by using additives in direct
injection diesel engine”
3 Babban Yadav et al. “Performance evaluation Microalgae, fossil diesel and Slightly lower power & Compatible wit petro diesel
and emission characteristics Soybean Methyl Ester torque, More SFC, lowered
of microalgae fuel in No2 & PM
combustion engine”
4 Xiaolei Zhanga et al. “The potential of microalgae Microalgae, traditional Review paper More economica compare to
in bio-diesel production” bio-diesel (vegetable oils); other feedstock
5 Rajendra Pawar et al. “A comprehensive review Vegetable oil and animal oil Increase BSFC & BTE and Within th acceptance range
The Performance and Emission Analysis of Diesel Engine …
Fig. 2 Bio-diesel
production scheme
Table 2 The physical and chemical properties of blends of SFME and diesel EN14214
Property Method Specifications Reference B10 B20 B30 B40 B100
diesel
Flash point °C EN ISO 3679 min. 120 68 75 86 97 110 183
Kinematic EN ISO 3104 3.5–5 3.18 3.21 3.26 3.31 3.35 4.7
viscosity
40 °C, mm2/s
Density EN ISO 3675 860–900 832 835 839 843 848 868
15 °C, kg/m
Calorific – – 44.8 43.6 42.2 41.7 40.3 33.5”
value MJ/kg
Table 3 Specifications of
Number of cylinders 4
Deutz F4L912 diesel engine
Bore/stroke (mm) 100/120
Displacement (l) 3.770
Compression ratio 19
Maximum rated speed (rpm) 2500
Mean piston speed (m/s) 10
Power (kW) @ 2350 rpm 51
Mean effective pressure (bar) 6.9
Maximum torque (Nm) @ 1450 rpm 238
Minimum idle speed (rpm) 650”
Table 4 Precision of
Parameter Precision value
parameters by measuring
system Speed (rpm) 0…6000 ± 2%
Torque (Nm) 0…1000 ± 2%
Fuel consumption (kg/h) 0…50 ± 2%
Intake air temperature (°C) 0…50 ± 5%
Inlet pressure (kPa) −50…300 ± 5%
Exhaust temperature (°C) 0…800 ± 5”
3 Results
Exhaust Gas Temperature indicates the heat discharge rate of a fuel in combustion
time. The combustion with bio-diesel has been increased because of more oxygen
levels in it. As shown in Fig. 4, that EGT increases with the load for the all mixtures
of SFME. It can be correlated with the research work of Godiganur et al. [28] of
6-cylinder turbo-charged diesel engine with mahua bio-diesel mixtures showed that
EGT increased with the engine load. Datta et al. [29] experimented on 2-cylinder,
The Performance and Emission Analysis of Diesel Engine … 855
4-stroke, diesel engine with jatropha bio-diesel mixtures showed that EGT increased
with the increased flame temperature of bio-diesel.
Carbon monoxide emission is due to the incomplete fuel combustion. We have
observed that carbon monoxide emission decreases at low-load and it increases
abruptly for all fuels (Fig. 5). The similar results were carried out by Nabi et al. [30]
on a single cylinder with cotton bio-diesel that the Carbon monoxide emissions were
decreased by 24% in comparison with diesel; Çelikten et al. [31] 4-cylinder diesel
engine with soybean Bio-diesel observed a reduction by 28% of carbon monoxide
emissions.
The total combustions in a combustion chamber increases carbon di-oxide emis-
sion. We have obtained that carbon di-oxide emission increases with the loads for the
tested fuels (Fig. 6). Huzayin et al. [32] observed that the carbon di-oxide emission
increases for jojoba bio-diesel mixture of each engine load. Fontaras et al. [33] for
mixtures of soybean the carbon di-oxide emission increases by 14% in the case of
B100 and 9% in the case of B50. Aydin and İlkiliç [34] observed that it is reduced
by 16% of carbon di-oxide emission of rape-seed bio-diesel.
In general the emission of hydrocarbons (HC) depends mainly by compositions
and combustion characteristics of the fuels tested. If combustion is improved the HC
emissions decrease and vice versa. Because of the high content of oxygen in the bio-
diesel it is expected that HC emission will decrease for blends of SFME and diesel. It
is observed that the decrease of HC emissions depends of the percentage of bio-diesel
in the blend (Fig. 7). Sahoo et al. [35] found a reduction in HC emissions by 20.64,
20.73 and 6.75% using bio-diesel of karanja, jathropa and polanga. Tsolakis et al.
856 A. S. Joshi and S. Ramesh
[36] observed a reduction of nearly 50% for rape-seed bio-diesel compared with low
sulfur diesel.
Nitric oxide (NO) and nitrogen dioxide (NO2 ) are formed in the process of oxida-
tion of the nitrogen during combustion and depend of the combustion temperature
and oxygen content. The tests show an increase in NOx emissions with the increase
in engine load for all blends (Fig. 8). It is found highest for SFME because of
high oxygen content which results in complete combustion causing high combustion
temperature. Gumus [37] found an increase of NOx emissions for apricot seed kernel
oil methyl ester by 10% compared to diesel fuel. Aydin and İlkiliç [34] reported an
increase of NOx emissions 16.7% with B20 blend and 11.8% with B100 for bio-diesel
of rape-seed.
4 Conclusion
The 4-cylinder diesel engine worked successfully while testing with bio-diesels of
Sunflower Methyl Ester and its mixes. These mixtures are examined to obtain the
physio-chemical characteristics. It was observed that the Exhaust Gas Temperature
was more for unadulterated bio-diesel. The more oxygen level increased the temper-
ature of combustion which was resulted in the increase of temperature for the various
mixtures tested. Carbon monoxide and Hydrocarbon emissions were very high for the
858 A. S. Joshi and S. Ramesh
diesel fuel and low for the mixtures of Sunflower Methyl Ester. It was observed that
the NOX was very high for pure Sunflower Methyl Ester and it’s mixtures because
of high volatility, low heat content and high viscosity as compared to the diesel fuel.
References
1. Yamaki Y et al (1994) “Application of common rail fuel injection system to a heavy duty diesel
engine”. SAE paper 942294
2. Xue J et al (2011) Effect of biodiesel on engine performances and emissions. Renew Sustain
Energy Rev 15(2):1098–116
3. Sharma YC, Singh B (2009) Development of biodiesel: current scenario. Renew Sustain Energy
Rev 13:1646–51
4. Palit S et al (2011) Environmental impact of using biodiesel as fuel in transportation: a review.
Int J Global Warming 3(3):232–256
5. Misra RD, Murthy MS (2011) Blending of additives with biodiesels to improve the cold flow
properties combustion and emission performance in a compression ignition engine—a review.
Renew Sustain Energy Rev 15:2413–2422
6. Sun J et al (2010) Oxides of nitrogen emissions from biodiesel—fuelled diesel engines. Prog
Energy Combust Sci 36:677–95
7. Thangaraja J et al (2014) Experimental investigations on increase in biodiesel—NO emission
and its mitigation. Proc Inst Mech Eng, Part D: J Automobile Eng 228:1274–1284
8. Weiss B et al (2007) A numerical investigation in to the anomalous slight NOx increase when
burning biodiesel; a new (old) theory. Fuel Process Technol 88:659–667
9. Canakci M (2009) NOx emissions of biodiesel as an alternative diesel fuel. Int J Veh Des
50:213–228
The Performance and Emission Analysis of Diesel Engine … 859
10. Xue J et al (2011) Effect of biodiesel on engine performances and emissions. Renew Sustain
Energy Rev 15:1098–116
11. Ogunkoya D et al (2016) Investigation of the effects of renewable diesel fuels on engine
performance, combustion, and emissions. Fuel 140:541–554
12. Suh HK, Lee CS (2016) A review on atomization and exhaust emissions of a biodiesel fueled
compression ignition engine. Renew Sustain Energy Rev 58:1601–20
13. Yunuskhan TM et al (2014) Recent scenario and technologies to utilize non-edible oils for
biodiesel production. Renew Sustain Energy Rev 37:840–51
14. Boehman A et al (2003) Fuel formulation effects on diesel fuel injection, combustion, emissions
and emission control. In: Proceedings of the DOE diesel engine emissions reduction conference,
newport, RI, August 24–28, 2003. US: DEER, pp 1–9
15. Ayoola AA, Anawe PAL, Ojewumi ME, Amaraibi RJ (2016) Comparison of the properties
of palm oil and palm kerneloil biodiesel in relation to the degree of unsaturation of their oil
feedstocks. Int J Appl Nat Sci 5(3):1–8
16. Ghadge SV, Raheman H (2006) Process optimization for biodiesel production from mahua
(Madhuca indica) oil using response surface methodology. Biores Technol 97:379–384
17. Saleh HE (2009) Experimental study on diesel engine nitrogen oxide reduction running with
jojoba methyl ester by exhaust gas recirculation. Fuel 88:1357–1364
18. Singh S, Prabhahar M Experimental investigation of performance and emission characteristics
using chlorella algae bio diesel as alternative fuel
19. Efe Ş et al (2018) Comparative engine characteristics of biodiesels from hazelnut, corn,
soybean, canola and sunflower oils on di diesel engine. Renewable energy, (Accepted
manuscript)
20. Uyumaz A (2018) Combustion, performance and emission characteristics of a DI diesel engine
fuelled with mustard oil biodiesel fuel blends at different engine loads. Fuel 212:256–267
21. Rathore Y et al (2019) Experimental investigation of performance characteristics of compres-
sion—ignition engine with biodiesel blends of Jatropha oil & coconut oil at fixed compression
ratio. Heliyon 5:e02717
22. Shrivastava P et al (2019) An experimental evaluation of engine performance and emission
characteristics of CI enine operated with roselle and Karanja biodiesel. Fuel 254:115652
23. Mishra SR et al (2018) Impact of simarouba glauca biodiesel blends as a fuel on the performance
and emission analysis in an unmodified DICI engine. Renew Energy Focus 26
24. Raman LA et al (2019) Experimental investigation on performance, combustion and emission
analysis of a direct injection diesel engine fuelled with rapeseed oil biodiesel. Fuel 246:69–74
25. Goga G et al (2019) Performance and emission characteristics of diesel engine fueled with rice
bran biodiesel and n-butanol. Energy Rep 5:78–83
26. Soni T, Gaikwad A Waste pyrolysis tire oil as alternative fuel for diesel engines
27. Tutunea D, Bica M, Dumitru I, Dima A (2014) Experimental researches on biodiesel prop-
erties. SMAT 2014 3rd international congress science and management of automotive and
transportation engineering 23rd–25rd of October, Craiova, Tome I, ISBN 978-606-14-0864-1,
Universitaria, pp 295–300
28. Godiganur S, Murthy CHS, Reddy RP (2009) 6BTA 5.9 G2–1 Cummins engine performance
and emission tests using methyl ester mahua (Madhuca indica ) oil/diesel blends. Renew Energy
34:2172–2177
29. Datta A, Palit S, Mandal BK (2014) An experimental study on the performance and emission
characteristics of a CI engine fuelled with Jatropha biodiesel and its blends with diesel. J Mech
Sci Technol 28(5):1961–1966
30. Nabi MN, Rahman MM, Akhter MS (2009) Biodiesel from cotton seed oil and its effect on
engine performance and exhaust emissions. Appl Therm Eng 29:2265–2270
31. Çelikten I, Koca A, Arslan MA (2010) Comparison of performance and emissions of diesel
fuel, rapeseed and soybean oil methyl esters injected at different pressures. Renew Energy
35:814–820
32. Huzayyin AS, Bawady AH, Rady MA, Dawood A (2004) Experimental evaluation of diesel
engine performance and emission using blends of jojoba oil and diesel fuel. Energy Convers
Manag 45:2093–2112
860 A. S. Joshi and S. Ramesh
1 Introduction
In any mechanical systems, friction and wear loss is one of many factors in energy
consumption. To decrease friction and wear, oil lubricant has been studied to be used
as lubricant additives that have promising effects on friction and wear reduction in
automotive. Oil lubricant of many compositions and sizes have demonstrated certain
degrees of friction modifying and anti-wear effects. We recently concluded that in
the boundary lubrication region, the addition of oil lubricant could reduce the friction
coefficient up to 65%, and wear use as high as 74%. Such lubricants combining with
a base oil and dispersed Jatropha oil lubricant emerged as a new class of lubricants,
the bottleneck for further development, however, is the aggregation of Jatropha oil
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 863
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_86
864 M. R. Sanjay and L. B. Abhang
lubricant in a base oil. A stable suspension of Jatropha oil lubricant is essential for
a usable lubricant.
Selected Jatropha oil lubricant is found better among all added into the oil. Among
those that were added into oils, Jatropha oil lubricant has received much attention
and exhibited excellent applications for their excellent friction and wear properties
[1]. The reduction of wear depends on interfacial conditions such as normal load,
geometry, relative surface motion, sliding speed, surface roughness, lubrication, and
vibration. Chemical additives in lubrication fluid control anti-wear properties, load-
carrying capacities, and friction under the specified boundary lubrication conditions.
3 Problem Statement
1. Bio-lubricant easy available and more cooling property, good frication charac-
teristic.
2. Less surface area and good lubrication property.
3. It may cause friction due to high temp or increase the cycle time of replacement
of part.
4. Chances of slippage are more while higher applied force chance to an accident
occurs.
5. Increase life of Piston Ring.
4 Necessity of Work
The principal motivation for formulating new additives using Jatropha oil lubricant
as a promising solution for improving the Tribological behaviour is that Jatropha oil
lubricant has the potential to offer significant tribological benefits of both solid and
Experimental Analysis of Effect of Bio-lubricant Between... 865
liquid lubrication and extend the life of the mechanical components. The decline
of the friction coefficient between the rings/liner assemblies plays a critical role
in improving engine performance and fuel efficiency. Bio lubricant additives play
a significant role in the formation of a tribofilm layer on the worn surfaces via a
physical or chemical absorbed mechanism to enhance protection of the worn surfaces
and create a rolling effect between sliding surfaces.
5 Objective of Work
6 Design Methodology
A piston ring is one of the essential parts of the Diesel/Petrol engines. It is an open-
ended ring that fits into a groove on the outer diameter of a piston in a reciprocating
engine such as an internal combustion engine or steam engine [3]. The principal
function of the piston rings is to form a seal between the combustion chamber and
the crankcase of the engine. The goal is to prevent combustion gases from passing
into the crankcase and oil from passing into the combustion chamber the three main
functions of piston rings in reciprocating engines are:
1. Sealing the combustion/expansion chamber.
2. Supporting heat transfer from the piston to the cylinder wall.
3. Regulating engine oil consumption [4].
The materials used to make Piston Rings are one of the most critical factors in its
Performance. Thus we have selected Cast Iron bronze and nylon material to make
piston Ring.
1. Cast Iron: Cast iron is an iron-carbon alloy with a carbon content great than
2%. As the most common ring material, cast iron has a low—friction, wear
866 M. R. Sanjay and L. B. Abhang
6.4 Experimental
The Cygnus friction and wear test machine are used to evaluate the friction and wear
characteristics of lubrication in this experiment. The Cygnus friction and wear test
machine are designed to study friction and wear in dry or lubricated sliding over a
wide range of speed, load and temperature. It is a tri-pin-on-disc machine which is
conducted by using three pins on a disc as testing specimens. Specifications of the
Cygnus friction and test machine is mentioned in Table 1 [3].
The Cygnus friction and wear test machine are connected with a computer having
embedded a block diagram based application construction program, Visual Designer.
Visual Designer allows developing custom data acquisition, analysis, display in the
form of a graphical chart or numerical meter. The programs can be controlled simply
by drawing a program’s data structure in the form of a block diagram. The block
diagram can also contain textual comments, allowing the process being monitored or
controlled to be documented. The block diagram can also contain textual comments,
allowing the process being monitored or controlled to be documented [3].
With the help of Cygnus Friction and wear testing Machine I have taken the following
reading.
Figure 1 Show the curves of pins wear as a function of sliding time for various
Jatropha oil blended with lubricant SAE 40. The values of linear pin wear under
2000 rpm and 30 N loads for each pin vary from 0.02 to 0.05 mm. It was observed
that the higher or maximum wear occurred at the beginning of the experiment for
some of the test specimens. It is clear from the graph that maximum wear occurred
for JBL40 and for JBL10 wear is minimum. We also can observe from the graphs
that except JBL40, for each JBL, pin wear decreases gradually and continuously. At
the beginning of the test, we can see the wear rate was fast in the period that is called
the running-in period.
Chart Title
Wear (mm)
0.1
0.05
0
1 2 3 4 5 6 7 8 9 10 11 12
Sliding Time (m/MIn)
0% of jatropha 10% of jatropha 20% of jatropha
30% of jatropha 40% of jatropha 50% of jatropha
Fig. 1 Linear pin wear as a function of sliding time for various bio-lubricants
868 M. R. Sanjay and L. B. Abhang
Chart Title
0.35
0.3
0.25
0.2
COF
0.15
0.1
0.05
0
1 2 3 4 5 6 7 8 9 10 11 12
Sliding Time (m/MIn)
0% of jatropha 10% of jatropha
20% of jatropha 30% of jatropha
40% of jatropha 50% of jatropha
Figure 2 shows the curves of friction coefficient plotted against the sliding time
for various Jatropha oil-based bio-lubricants. The results in Fig. 2 depict that the
lubricant regimes that occurred during the experiment were the boundary lubrication
where the value of friction coefficient (µ) for boundary lubricant is in the range of
0.001–0.2 except for JBL50. For 0% of Jatropha oil-based bio-lubricants, it can be
seen that the coefficient of friction is highest at the beginning, and then it falls rapidly
up to minimum value compared to all samples. This phenomenon can be explained
by attributing oxide layer on the aluminium surface. Lubricant additives react with an
oxide layer, which causes to form a thick tribo-film on the aluminium surface. At the
initial stage, the shear stress of the tribo-film is high. Thus, the coefficient of friction
is high early. With continuing the sliding, the aluminium pin is eroded, and a new
metal surface is exposed. Since the fresh metal surface has a lower tendency to react
with lubricant additives, the formation rate of tribo-film decreases than the beginning.
Therefore, shear stress becomes lower than beginning and constant throughout the
process.
Figure 3 Shows the relation of the averages oil temperature of varies percentage
of Jatropha oil-based bio-lubricants with the sliding time, respectively. The range
of temperature is about 20–100 °C. From the graphs, we can see that lubricant
temperature increases with increasing sliding time for each percentage of Jatropha
oil-based bio-lubricants. The maximum temperature rise occurred for JBL40, and
minimum temperature rise occurred for JBL 10. The progressive increase in temper-
ature period is known as the running-in period during the asperities of the sliding
surface are progressively cut off. From the graphs, it can be noticed that the 30 and
Experimental Analysis of Effect of Bio-lubricant Between... 869
Chart Title
60
40
Temp 20
0
1 2 3 4 5 6 7 8 9 10 11 12
Sliding Time (m/MIn)
0% of jatropha 10% of jatropha 20% of jatropha
30% of jatropha 40% of jatropha 50% of jatropha
40% of Jatropha oil Based bio-lubricant are produced more heat than other while
10% of Jatropha oil-based bio-lubricants is generated lower heat.
7.1.4 Viscosity
Figure 4 show that the viscosity of Jatropha oil-based bio-lubricants decreases expo-
nentially for both 40 °C and 100 °C operation temperatures. The Table 3 shows the
viscosity grade requirement for the lubricants set by the International standard orga-
nization (ISO) while Fig. 4 shows the behaviour of the viscosity of Jatropha oil-based
bio-lubricants at 40 °C and 100 °C operation temperatures. Comparing Table 2 and
Fig. 4, it can be stated that at 40 °C operation temperature 10, 20 and 30% of Jatropha
oil-based bio-lubricants meet ISO VG100 requirement.
But 40 and 50% of Jatropha oil-based bio-lubricants do not meet the criteria.
However, it can also be noted that all bio-lubricants have a higher viscosity than the
expected value.
Chart Title
200
150
Viscosity
100
50
0
120% 100% 80% 60% 40% 20%
Axis Title
40 C 100 C
Fig. 4 The viscosity of various percentages of Mineral Oil for temperature 40 °C and 100 °C
operations
870 M. R. Sanjay and L. B. Abhang
The accuracy of the model was developed by substituting the experimental data in
the model equation at various conditions of test load, sliding velocity and quantity of
oil and the corresponding predicted values of weight loss of piston ring both under
SAE 15W40 and Jatropha oil were obtained in the following Tables 2 and 3.
The Table 2 shows that the predicted values for the piston ring weight loss and
the actual values of weight loss of piston ring were quietly closed with each other
under SAE15W40 oil. This validated the credibility of the model developed for
establishing a correlation between the process variables and eight loss of piston ring
under SAE15W40 oil.
The Table 3 shows that the predicted values for the piston ring weight loss and
the actual values of weight loss of piston ring were close with each other under
Jatropha oil, thus, validating the credibility of the model developed for establishing a
correlation between the process variables and eight loss of piston ring under Jatropha
oil.
From the experimental layout for both wears under SAE15W40 oil and Jatropha
oil, it was concluded that the combinations of process parameters such as test load
of 60 N, sliding speed of 1200 cm/min and 30 ml of oil yields a low value of weight
loss of piston ring. But Jatropha oil yields a low value of weight loss of piston ring
as 0.0034 mg when compared to the SAE15W40 oil which yields the weight loss of
the piston ring as 0.004 mg at the same process parameter combinations.
8 Conclusion
References
1. Adhvaryu A, Liu Z, Erhan SZ (2005) Synthesis of novel alkoxylated triacylglycerols and their
lubricant base oil properties. Indus Crops Prod 21(1):113–119
2. Shahabuddin M et al (2012) Tribological characteristics of amine phosphate and acety-
lated/butylated diphenylamine additives infused bio lubricant. Energy Educ Sci Technol Part
A 30:89–102
3. Shahabuddin M et al (2013) Comparative tribological investigation of bio-lubricant formulated
from a non-edible oil source (Jatropha oil). Indus Crops Prod 47:323–330
872 M. R. Sanjay and L. B. Abhang
4. Siniawski MT, SN, Adhikari B, Doezema LA (2007) Influence of fatty acid composition on the
tribological performance of two vegetable-based lubricants. J Synthetic Lubricant 24:101–110
5. Wolff A (2014) Simulation-based study of the system piston-ring-cylinder of a marine two-stroke
engine. Tribol Trans 57:653–667
6. Arumugam S, Sriram G (2012) Effect of bio-lubricant and biodiesel-contaminated lubricant on
tribological behaviour of cylinder liner-piston ring combination. Tribol Trans 55:438–445
7. Baker CE, Theodossiades S, Rahnejat H (2012) Influence of in-plane dynamics of thin
compression rings on friction in internal combustion engines. J Eng Gas Turb Power
Garbage Monitoring and Collection
System Using RFID Technology
Abstract Now each day we face the matter of garbage which is scattered everywhere
in buildings also as in villages the Gathering of that garbage isn’t done on time. So,
we’ve proposed the system of garbage pickup & Monitoring. This system, it’s alleged
to be collect the bins from every registered home. The worker should collect Bin from
the house and scan the RFID tag with the RFID Reader. The RFID Reader is going to
be interfaced with the Wi-Fi module. After scanning of RFID tag the message "Bin
collected" are going to be delivered to the respective customer. If any home is going
to be missed for collection of Bin, then the actual customer can call to the customer
care with the given toll-free number. After interaction with care, the message “Bin
collection is remaining is sent to the nearby bin collector worker.
Keywords Ultrasonic sensors · RFID Reader and tag · GSM · Garbage bin
1 Introduction
As we observe the present days, the waste is spread everywhere and overflows it
from the bins kept at the important areas. These areas are often railway stations,
our homes or schools Itcreatesbadhealthconditionforthepeoplethatlivenearthedust-
binareas, to avoid such a situation, we are getting to design bins collecting system
means it contains frequency identification (RFID) CARD reader. Garbage pickup
System minimizes and improves the efficiency by including costs per km, per hour
to transfer the waste after collection from various locations of the town or villages.
Recent days, many of us live in cities just for their simple earning money all together
with the ways, and lots of people are coming from the agricultural areas for the oppor-
tunities which are largely found in urban areas. Thanks to that massive crowd of indi-
viduals, there’s a considerable amount of garbage is produced a day. By observing
all the issues were getting to propose a system which can collect the waste from
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 873
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_87
874 A. A. Kadam et al.
every home which is registered already. The system works faithfully and doesn’t
skip collecting garbage from any home.
Advantage:
• It saves money as the overall budget of the project in a smaller amount thanks to
the database.
• RFID avoids the disadvantage of barcode scanning, which needs line-of-sight
access to every barcode, and also it scans just one item at a time.
• RFID eliminates the likelihood of errors which are created by humans.
• The RFID reader can scan the number of tags.
• RFID is often integrated with scanning, and readers are fixed.
2 Literature Survey
for wireless transmission [9]. P. Ramchandar Rao, S. Sanjay Kumar, Ch. Rajendra
Prasad They used Garbage monitoring system it’s an improvement of normal dustbin
by changing it to be smart using sensors. Garbage monitoring system may be an idea
of which makes a traditional dustbin smart using ultrasonic sensors for garbage
level detection. It display sends the message to the priority department person to
update the status of the Bin using GSM model [10]. The workers are carrying RFID
Tags with them. The RFID Reader is attached at every home.The RFID Reader is
interfaced with the Wi-Fi module. At the initial stage of the Gathering, the worker
will collect the waste from a particular home. He will scan the RFID Tag with the
RFID reader.The reader is going to be connected to the Wi-Fi module, which gives
knowledge about the collection of Bin. After that, the message “Bin collected at
your home” is shipped to the customer. If the Gathering of Bin isn’t wiped out the
correct way or the Gathering of bins refused, there’ll be an alternate solution for
the particular problem. The customer is provided with a toll-free number to register
their complaints. If things occurred like Bin is skipped or not collected properly,
they could register their complaints through the toll-free number. The GPS system
is employed to trace the method. If any complaint is registered by the customer, the
monitoring system will send a message to the nearby worker, and he will follow an
equivalent flow for collection and Monitoring of the waste. The method is going to
be continued until the last home’s Bin is going to be collected. The Fig. 1 show the
flow chart of the proposed system to manage collection and Monitoring of registered
customers bin from their home.
3 Methodology
In the whole process of “Garbage Collection and Monitoring System,” we are using
the pairs of RFID Reader and RFID tag. Therein we’ve homes which are registered.
The workers are carrying RFID Tags with them. The RFID Reader is attached at
every home. The RFID Reader is interfaced with a Wi-Fi module. At the initial stage
of the Gathering, the worker will collect the waste from a particular home. He will
scan the RFID Tag with the RFID Reader. There are going to be connected to the
Wi-Fi module, which gives knowledge about the collection of Bin. After that, the
message “Bin collected at your home” is shipped to the customer. If the Gathering of
Bin hasn’t wiped out a correct way or the Gathering of Bin is refused, there’ll be an
alternate solution for the particular problem. The customer is provided with a toll-
free number to register their complaints. If things occurred like Bin is skipped or not
collected properly, they could register their complaints through the toll-free number.
The GPS system is employed to trace the method. If any complaint is registered by
the customer, the monitoring system will send a message to the nearby worker and
he will follow an equivalent flow for collection and Monitoring of the waste. The
method is going to be continued until the last home’s Bin is going to be collected.
The fig no 3 show the flowchart of the proposed system to manage collection and
Monitoring of registered customers bin from their home.
4 Conclusion
This paper concludes that by using this smart garbage monitoring system using RFID
over IOT’s we can easily dispose the waste present within the garbage bins as early
as possible in comparison to the previous methods without it affecting to the people
and keep the environment clean.
References
1. Smith TF, Waterman MS (1981) Identification of common molecular subsequences. J Mol Biol
147:195–197. Samudhana Satyamanikantha (July-2017). Smart garbage monitoring system
using sensors with RFID over the internet of things. Journal of Advanced Research in Dynamical
and Control Systems (JARDCS), SaveethaUniversity,Thandalam, Chennai,TamilNadu, India.
2. Nanavati S, Moinuddin K, Lokesh KS, Narayana A (2017) Waste managementin IoT-enabled
smart cities: a survey. Int J Eng Comput Sci ISSN: 2319–7242(IJECS)
3. Baradi V, Rawool D, Deshputre R, Ghadigaonkar H (2018). Intelligent garbage collection using
IoT. Int J Innov Adv Comput Sci IJIACS ISSN2347–8616
4. Dr. Sathish kumar N, Vijayalakshmi B IOT based smart garbage alert system using
arduinoUNO.2016IEEE region10 conference (TENCON) proceedings of the international
conference
5. Abdurahman F, Aweke S, Assefa C (2016) Automated garbage monitoring system using arduino
IOSR J Comput Eng (IOSR-JCE)
Garbage Monitoring and Collection System Using RFID Technology 877
6. Yusof NM, Jidin AZ, Rahim MI (2017) Smart garbage monitoring system for waste
management. MATEC web of conferences 9701098 (2017)
7. Kesthara V, Khan N, Praveen SP, Mahesha C, Murali N (2015) Sensor based smart dustbin for
waste segregation and status alert. Int J Latest Technol Eng Manage Appl Sci (IJLTEMAS)
8. Sirichai Watanas ophonand Sarinee Ouitrakul (2014) Garbage collection roboton the beach
using wireless communications, 2014 3rd international conference on informatics, environ-
ment, energy and applications
9. SoukaynaMouatadid ZFA (2011) RFID Based Solid Waste Collection Process, International
Conference in Institute of Electrical and ElectronicsEngineers(IEEE)
10. Ramchandar Rao P, Sanjay Kumar S, Rajendra Prasad Ch (2017) Garbage monitoring system
using arduino. Int J Trend Sci Res Dev (IJTSRD)
A Survey on Mental Health Monitoring
System Via Social Media Data Using
Deep Learning Framework
Abstract In today’s society Depression, Stress and Gloominess are some of the
most broadly perceived and increasing mental issue influencing us. The presence
of a system that is automatically capable of identifying a users mental state is of
great benefit. Due to users spending a lot of time on social media using that to
check his well-being will be helpful in many ways. There are various algorithms
such as Random Forest, SVM, ANN, CNN, RNN present using which this can be
achieved. Sentiment Analysis and deep learning techniques could provide us robust
algorithms and structure for a target also a chance for observing mental issues which
are, specifically of depression and stress. In this paper different ways of dealing
with depression shown on social media platform are studied. This will enable in
achievement of better understanding of the various mechanisms used in depression
detection.
1 Introduction
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 879
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_88
880 S. Banerjee and N. F. Shaikh
At present, various methods for depression and stress detection do exist some of
the diagnosis rely on self-reporting linked with a informed assessment of health care
practitioners [4]. The presence of an automatic system will not only help the various
health care professionals but it will also provide a cost efficient and immediate if not
exact diagnosis. Sentiment and deep learning or machine learning technologies on
social media platforms could help to tackle such objectives. Even though the aim of
these systems is not outright replacing any professional they surely do support their
work and give immediate aid. There are various other methods of identification of
a depressed user but nowadays users spend a lot of time accessing social media so
if there is a mechanism that will utilize this data it will cause in immediate benefit
[5, 6]. The organization of the paper is as follows: Sect. 2 gives the related work
and limitations and last section concludes the paper with future work followed by
references.
2 Related Work
User’s feelings about various topics may be understood using social media data.
Along these lines, systems, such as, checking and suggestion frameworks (RS) can
gather and dissect this information [1]. There are systems that can identify a user’s
stress with the help of his social media posts. Different deep learning algorithms
such as CNN, RNN, LSTM-RNN prove to be very useful in such cases [1]. The
main advantage considered here is informing some of the authorized persons about
the depression detected this will result in taking steps which might be necessary
to help the stressed and depressed user and might save them from ending up to
take any drastic steps. Also motivational messages are sent to depressed users thus
cheering them up and relaxing them the system also has a very easy to use interface
that proves to help the users in navigating and performing the needed function [1].
The recommendation system consisted of 360 different messages of relaxing, happy,
motivational etc. categories. The sentiment analysis here comprises of building a
sentiment metric the sentiment score consists of a range of −5 to +5. The users
data from their Facebook is collected and this data is used for identification of the
depressed user. Different ontological networks are considered from user data such
as their activity, sleep patterns etc. The use of CNN and BILSTM is done for the
depression identification here along with sentiment analysis inclusion thus providing
with a more accurate system [1].
A Survey on Mental Health Monitoring System … 881
In social media networks there may be emojis too which have also been considered in
some of the works. Work has been done mainly on social media sites like Facebook
and Twitter. There is work done on other platforms like reddit too which have longer
texts and may thus give more insights about the user [7]. The dataset used by them
comprised of 1293 depressed and 548 standard tweets. The subreddits available for
depression help and other subreddits involving friends and family are included in it for
the dataset building. Total of 68 different features could be taken here. Extensive work
for obtaining the feature set was done and various methods have been considered. For
n-gram modeling unigram and bigram models have been considered, for linguistic
dimensions LIWC and for Topic modeling LDA is considered. Most related words
for different categories like job, depression, tired, friend etc. are also identified. Next,
for the purpose of classification many algorithms have been used here classifiers like
Logistic Regression, SVM, Adaboost, MLP are analyzed. The main finding here
were words related to a depressed users messages are, preoccupation with oneself,
sadness, hostile, anxiety, anger or suicidal thoughts, with a most emphasis on the
present and future [7]. For single feature case SVM using bigram model provided the
best result. It has been observed that MLP using combined features models presented
the best solution when compared to other combinations such as unigram models and
Adaboost algorithms [7]. The work provided a way to get relation between user’s
language and him facing depression. Also the combination of n-gram, LDA and
LIWC was studied and also their impact individually was considered too [7].
With the help of social media we may identify negative words which ultimately result
in the formation of depressive words. Twitter data is considered here, the dataset built
is a large one. Data collection is over a span of 3 months and the target was kept as
Australian elections. Next the collected data cleaning and frequently occurring word
removal process is carried out. Various unigram models may be prepared to carry
out this process [2]. Here the words are given scores and using this identification
of the depressed words using the scores and hence identification of the depressed
user is done [2]. The tweets are all rated, the rating is done based on its strength.
There are two scales, the positive range consisting of +1 to +5 and the second being
the negative range varying from −1 to −5 [2]. Such methods helps to provide a
simple yet efficient mechanism for the monitoring of a users mental health. Out of
all users 13 depressed users with lowest scores of sentiment were identified. For these
users further data analysis was carried out, LDA method could separate tweets into
different types. Then the negative sentiments identification and standard deviation
representation of the same was studied. Human behavior is not constant happiness
882 S. Banerjee and N. F. Shaikh
and sadness varies always for normal users, but for a stressed user this may not be the
behavior. A stressed user’s standard deviation could indicate a continuous sadness
state. Hence words which when used in Tweets give a more probability of user being
depressed could be identified. Here, however the detailed analysis on the positive
sentiments has not been carried out [2].
Using Depression levels can influence the probability of taking drastic steps such as
committing suicide out of sheer stress. Such unfortunate incidents can be prevented
if early detection of such issues can be done. This can be achieved by presenting
students as well as parents with questionnaires which may include questions like:
if they are regular to school, if they have mood swings also their age groups, their
appetite, sleeping habits etc. [3]. The set of students were between age group of
15–29 and there were many factors that caused in the presence of depression in their
life. The various causes for it included failure, bullying, pressure from parents etc.
The option to fill the answers are from value of ‘0’ to ‘3’ where ‘0’ denotes not at
all and ‘3’ represents always. The data collection process is performed using API
for reddit and standard data set for Twitter. There are multiple subreddits in reddit
that have a lot of user interaction and engagement. The data responses considered
here are around 619 out of this almost 18 different features could be identified. The
feature extraction is based on Tf-idf and using the questionnaires data points as
reference. Now the data from the questionnaires can be treated to obtain features
like age, regularity etc. which is given to classifier, along with this their Twitter data
is also monitored to detect the presence of stress or depression in Tweets [3]. The
data is used for both training and testing purpose. Machine learning algorithms such
as Random Forest, XGBoost, SVM are considered for the analysis. The maximum
accuracy was obtained with the help of XGBoost here. As per the data obtained
around 48% of people faced hopelessness and around 30% students felt bad about
themselves [3].
Social media websites are the place where users are nowadays expressing their
emotions the most and in the most honest manner [5] Facebook and Twitter are
the main sites where users express themselves and also connect. There has been
work done on data based on these sites, using machine learning algorithms the users
may be classified into: normal, mild, severe or extremely depressed [5]. The gap
A Survey on Mental Health Monitoring System … 883
between expression on social media site and actual thoughts may also be fulfilled by
building multiple choice questions that will help to understand and detect depressed
users [5]. Both Facebook and Twitter data are collected, then various algorithms are
used for the purpose of depression detection. The step after getting the data using tool
is using Natural Language Toolkit in order to get a clean version of it. The further
analysis deals with the presence of the formation of a datasplit in order to get the
train data. The system is considered on 100 twitter user, their data generation on one
week is considered for the analysis process. The system considered single user based
on the user name and their individual result of being depressed or not is carried out.
The system gives access to the users friends and family to know about the persons
mentals health. The model also may be implemented by the social media sites itself
to provide help to their own users. SVM, Naive Bayes and NLP techniques are used
for the identification of depression [5].
If the depressed user is motivated and made to feel better his complete potential
will be utilized and that person will not take any drastic life threatening steps too.
As per requirement there may be different approaches of data collection, it can be
questionnaires asked to the person, the posts they put on social media, or even text
used in verbal communication as well as expressions on face [8]. The facial detection
may be done for a user it will but depend on many factors such as the lighting or
the angle of getting image. Social Media data however depends on many lesser
parameters. Hence this will give a simpler way to collect the data and will have many
lesser dependencies to be considered also there is no requirement of any additional
resource here thus the cost is also minimized. Also data collection from Twitter may
be done using tool like Twint, set of positive and negative Tweets may be collected
by this. A sample of data collected with the help of this is presented and given as
reference. Various machine learning algorithms such as SVM, Decision Trees, Naive
Bayes Classifier, KNN Classifier are considered for the detection of a depressed user
[8]. The usage of deep learning algorithms has been suggested it will be useful for
accuracy improvement [8].
The diagnosis of depression is in-fact a major issue, across the world people do
not always have access to facilities which are readily available for their access [4].
With the help of twitter data here a system has been provided, the twitter data is
used to construct graphs with relation between depression and symptoms. Statistics
and NLP is used to identify presence of depression in the tweets [4]. The social
media platform redddit is considered a total data of 135 users is used. Different
884 S. Banerjee and N. F. Shaikh
word embedding were checked such as, fastText, GloVe, word2vec also multiple
dimensions were considered. The closest neighbours for the word ‘depression’ is
identified and we can also obtain ‘insomnia’, ‘suicidal’ etc. are the terms that have the
closest association with ‘depression’. The sentence formation itself is a process that
is of a lot of depth. There are different parts of a sentence which have their associated
meanings and the way in which they are present are also having a significant impact
on the meaning. Some of the adjectives presence may be having imapact on the
system for example ‘very’ will add more weight to the sentence and will make it
more positive or more negative depending on its context of usage. A deep learning
mechanism with the use of CNN proves to be very successful in extracting deep
features and obtaining depressed users from the given set of users [4]. Different
ERDE models are considered here and CNN models provide a good result with
metadata association but the collection of metadata is a difficult task on its own [4].
Doctors usually work with patients via questioning to detect depression however
this data is not available for common public. So depression symptoms are not very
clearly disclosed [6]. With the help of twitter data here a system has been provided,
the twitter data is used to construct graphs with relation between depression and
symptoms. Statistics and NLP is used to identify presence of depression in the tweets.
Initially in the data collection phase a total of 120 depressed tweets were collected
in a span of around 15 days. The study of depression based network generation here
considered total of 64 thousand tweets after all cleanup activities. Work in this area
to obtain a relationship between depression and various symptoms associated with it
is done in detail [6]. The words associated with depression are thus identified, words
like anxiety, insomnia, hate have the maximum similarity in context of depression
detection [6]. The words are identified for their frequencies and also their degree
which implies to their number of occurrences. The word score may be utilized in
this process. Word2Vec mechanism is implemented and also a semantic graph with
the center as “depression” is prepared and other words weights are calculated with
reference to it [6]. There are algorithms that are there to identify the similar words with
respect to depression and they are merged together, next their statistical distribution
is found out. The merge of medical system in a social media system is the motive
here, the medical systems will be benefited if this system is utilized in them [6].
A Survey on Mental Health Monitoring System … 885
Social media tools devices are wide spread however there may be other mechanisms
which can be used for depression detection. Humans Electroencephalography signals
can also provide a mechanism to provide an understanding of depressed vs non
depressed users [9]. It is important to monitor stress and depression as it also has
an impact on a persons general physical health. A stressed user finds it difficult to
carry on with his daily activities at ease. The study conducted comprised of people
in the age group of range 18 years to 30 years. All regular users were considered
and smokers were excluded as they consider it as a coping mechanism of stress
already. The volunteers are asked to be in different states and their EEG signals are
considered. Simple actions like read watch video are used to capture the reactions.
Also stress and depression may be a factor to cause other issues to, issues like blood
pressure increase and other lifestyle disorders may be caused due to the existence on
stress and depression in a person [9]. The EEG signals of the users may be utilized
with Wavelet transforms in order to obtain a system that is capable to identify if a
user is depressed or not [9]. The features are extracted with the help of DWT, a total
of 5 level decomposition too are used. ANNs are further trained to classify efficiently
the depressed user. The process of back propagation is carried out based on which
the values of next weight is kept on being updated every time. The main issue is the
cost and availability of EEG capture procedure in this research. Also simple human
behaviors like eye blinking or hand movements too cause in error in value of EEG
recording [9].
Stress and depression can also be identified with the help of analyzing a users inter-
actions on the social media sites [10]. Various user level and Tweet level attributes are
considered and they are used for analysis. The data is provided to a CNN which helps
in identification of the depressed users [10]. Here not just own tweet was considered
but also comments that a user puts on others posts are considered for analysis [10].
The features considered are based on both user level as well as tweet level aspects
here real world huge dataset is considered for analysis purpose. The use of the users
Tweet content as well as the social interaction factors are considered for the analysis.
The different features considered may be positive negative words or emoticons also
use of adjectives can be studied. The different social attention parameters which
were considered included their comments and likes. The users Social Tie analysis
is performed too and what they like or retweet are also considered along with his
own posts [10]. A factor graph model is provided along with CNN in order to obtain
various relationships amongst the user behaviours. CNN model gives a good result
when applied and provides a good accuracy. The social structure of the friends of
886 S. Banerjee and N. F. Shaikh
the stressed users appeared to be less connected when compared to the non stressed
users social structure. Also stressed users had more association to anger and sadness
where as very less association to family, friend and leisure when compared to a non
stressed user [10].
An alternative way to identify stress may be by analyzing a user’s daily phone usage
activities, few more parameters may be used to enhance the process. Parameters
such as weather data, personality traits may be considered [11]. User spends his
most time with his phone so this analysis should provide us a good platform to
understand the current users state of mind [11]. Statistical Models are prepared from
the various features that are collected and the models are studied with reference to
the impact it has based on other parameters such as current weather in the given area,
the personality traits of the user. The user may have permanent personality issues
which also influences the results [11]. A multiple factor based system that is capable
for the identification of stress finding could be achieved. A total of 111 subjects were
considered and their seven months data was taken into account for the work, they
belonged to different backgrounds and had different lifestyles. The data collection is
done based on many ways including self reporting, call and sms logs and proximity
data using bluetooth data. Obtaining daily stress from a users smart phone usage
activity provides a good way for stress monitoring, it is essential to identify stress
and help the users take necessary help to overcome their issue. The non availability
of proximity data is an issue in this case, it might not be easy to obtain it and use it
[11].
3 Open Issues
Mental health is a very important aspect of well-being of humans hence its a field
of extensive research area. In the above section, few of the approaches which have
been implemented in many situations in order to achieve the same purpose have been
mentioned. Many different algorithms have been studied which provide depression
detection in social media platform. In the existing works mostly data mining and
machine learning algorithms have been used and major work is done some aspects
of user behaviour, sentiment analysis may also be included in a more complete
manner. User’s personality type may also be considered in analysis, more features
related to occupation etc. may be included too. Also more number of social media
platforms may be included for getting a better understanding of a users behavior.
A Survey on Mental Health Monitoring System … 887
4 Conclusion
In this survey, identification of online users with depression and stress that is threat-
ening to people’s mental well-being is studied. Social media account privacy is a
concern in many platforms. It has been observed that Facebook and Twitter are the
social media platforms that have most works done on. This also identifies that in
many situations deep learning’s performance is better than other algorithms. It has
also been seen that in many cases CNN, RNN have better accuracy compared to
algorithms like random forest.
References
1. Rosa RL, Schwartz GM, Ruggiero WV, Rodríguez DZ (2019) A knowledge-based recom-
mendation system that includes sentiment analysis and deep learning. IEEE Trans Indus Inf
15(4):2124–2135
2. Tao X, Dharmalingam R, Zhang J, Zhou X, Li L, Gururajan R (2019) Twitter analysis for
depression on social networks based on sentiment and stress. IEEE Int Conf Behav Econ
Socio-Cult Comput 6:1–4
3. Jain S, Narayan SP, Dewang RK, Bhartiya U, Meena N, Kumar V (2019) A machine learning
based depression analysis and suicidal ideation detection system using questionnaires and
twitter. In IEEE students conference on engineering and systems, pp 1–6
4. Trotzek M, Koitka S, Friedrich CM (2020) Utilizing neural networks and linguistic metadata
for early detection of depression indications in text sequences. IEEE Trans Knowl Data Eng
32(3):588–601
5. Al Asad N, Pranto MAM, Afreen S, Islam MM (2019) Depression detection by analyzing
social media posts of user. In IEEE international conference on signal processing, information,
communication & systems, pp 13–17
6. Long Ma, Wang Y (2019) Constructing a semantic graph with depression symptoms extraction
from twitter. IEEE Conf Comput Intell Bioinf Comput Biol, pp 1–9
7. Tadesse MM, Lin H, Bo X, Yang L (2019) Detection of depression-related posts in reddit social
media forum. IEEE Access 7:44883–44893
8. Narayanrao PV, Kumari PLS (2020) Analysis of machine learning algorithms for predicting
depression. Int Conf Comput Sci Eng Appl, pp 1–4
9. Berbano AEU, Pengson HNV, Razon CGV, Tungcul KCG, Prado Seigfred V (2017) Classi-
fication of stress into emotional, mental, physical and no stress using electroencephalogram
signal analysis. IEEE Int Conf Sig Image Proc Appl, pp 11–14
10. Huijie Lin J, Jia JQ, Zhang Y, Shen G, Xie L, Tang J, Feng L, Chua TS (2017) Detecting stress
based on social interactions in social networks. IEEE Trans Knowl Data Eng 29(9):1820–1833
11. Bogomolov A, Lepri B, Ferron M, Pianesi F, Pentland AS (2014) Daily stress recognition from
mobile phone data, weather conditions and individual traits. Proc ACM Int Conf Multimedia
22(14):477–486
An Approach to Heart Disease Prediction
with Machine Learning Using Chatbot
Abstract Cardiovascular diseases are one of the leading causes of mortality and
morbidity worldwide which is a major concern to be dealt with, but it is difficult to
identify cardiovascular diseases in the earlier stages because of several contributory
factors which posses risk such as high blood pressure, high cholesterol, diabetes, and
many other factors. Prediction of cardiovascular diseases is considered as one of the
important aspects of healthcare analysis given a large amount of raw data available,
waiting to be converted into valuable information. On the other hand, the majority
of internet users are adopting one or more messenger platforms which enable us to
deploy AI-based conversational bots that can be developed to cater to individual users
to provide user satisfaction, high level of accessibility and flexibility. The aim of this
paper is to demonstrate the use of Machine Learning models using Google Cloud
Platform to assist users in the prediction of cardiovascular diseases using a chatbot
with the help of Flutter framework on Android and other messenger platforms such
as Telegram and Facebook.
1 Introduction
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 889
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_89
890 C. Nanaware et al.
CVD development process in later years is not inescapable [2]. Detection of CVD
mostly requires a clinical diagnosis but chances of incorrect diagnosis and treatment
are probable. Additionally, not all tests contribute towards effective diagnosis, the
majority rely on doctors expertise and experience. With the availability of abundant
raw data in the healthcare sector and machine learning advancing at an astounding
speed, we can use the data to support doctors in leveraging the diagnosis accuracy
with the help of machine learning techniques. Supervised Machine Learning model
takes input data and trains the model over the data with the help of the target namely
the label we want to predict thus resulting in a model that attempts to predict the
desired label. With the progress of NLP, and the internet users adopting one or more
messenger platforms, it increased the usage of chatbots, especially the ones which are
integrated within the messenger platform. A chatbot is a computer program with AI
capabilities that interacts with the user through text or audio. Chatbots are an evolu-
tion of question answering systems which can be effectively used in any domain that
is feasible to interact for a quick response. Chatbots offer responses faster than other
means of communication and are of higher quality, it is a manifestation of customer
ease. Chatbots are trained to generate responses to users queries and this is the most
important aspect to train bot further to adapt to end-users and their preference when it
comes to designing personalized chatbot. Chatbots mainly classify user requests into
intents and entities to give an appropriate answer to specific questions asked. Intents
are used to classify exactly what the user is asking and to process the request as per
model trained on intents. Entities are used used to process the request after under-
standing the intent of user such as what or where related to which things or person
the user is asking about [3]. Multiple platforms allow us building chatbots using their
NLP base to train custom bots such as Dialogflow, a service offered by Google which
runs on Google Cloud Platform. It is a natural language processing (NLP) platform
which can be used to build conversational applications and provide experiences for a
company’s customers in various languages and on diverse platforms [4]. Dialogflow
is used to develop different digital assistants which act as a conversational agent to
identify dialog states and respond with the corresponding state which is stored in the
dialog tree as to how the conversation flows between user and agent. Additionally,
as the complexity of big data increases, smaller enterprises struggle to accommodate
the storage and computing needs required for training and deployment of machine
learning models. This is where public cloud providers come into play. Cloud makes
it easy for such companies to experiment with various machine learning capabili-
ties and scale up as projects are pushed to production and demand increases. Cloud
offers elasticity and multiple services to set an infrastructure without having to invest
heavily and create cloud-native applications. Cloud providers such as Google Cloud,
AWS, Microsoft Azure provide many options for implementing intelligent features
in enterprise applications that don’t need deep knowledge of machine learning or
AI theory. BigQuery, a service by Google Cloud Platform is a fully managed and
serverless data warehouse that enables cost-effective, scalable and fast analysis over
petabytes of data. It is a Software as a Service(SaaS) that offers to query using
ANSI SQL. BigQuery has built-in machine learning capabilities that help to train
ML models with good accuracy using ANSI SQL. BigQuery can also be used to
An Approach to Heart Disease Prediction with Machine … 891
train and test models as it is a flexible ML platform that is easy to rely on due to SQL
related queries.
This research paper dedicates section II for system design for the proposed method,
section III deeply elaborates the proposed techniques and whereas section IV eval-
uates the performance of the system and finally section V concludes the paper with
traces of future enhancement.
2 System Design
Figure 1 shows system design and explains how different software and technologies
are implemented in this project.
3 Proposed Methodology
The goal of the prediction methodology is to design a machine learning model that
classifies whether patients have chances of heart disease or not, and deploy the model
on the Android/iOS platform with the help of chatbot using Flutter Framework and
Telegram. The dataset that is used to train the model is the Cleveland Heart Disease
Dataset made available to download through UCI Repository which contains a total
of 76 attributes but only a subset of 14 attributes is used for the prediction. The
search was in the main direction to the causes or the main factors which have a
strong influence on the heart. Some factors are constant such as sex, age, and family
background. Among many parameters, there are some varying parameters like blood
pressure, heart rate, cholesterol, chest pain that are considered mainly in the dataset.
Figure 2 shows features which are important and contribute in prediction.
The next step is data discretization and preprocessing techniques in the form of
Data Transformation, Data cleaning, Data Reduction. After Preprocessing the data
was split into training and testing dataset and was applied to algorithms such as
Decision Tree, Logistic Regression and KNN. The model accuracy was compared to
determine the best performing model which is later used to establish communication
with Chatbot. To generate the model we used BigQuery, a Google Cloud service that
enables users to create and execute various machine learning models in BigQuery
using standard SQL queries. Dialogflow, another service offered by Google was used
to create a design of the conversational agent. Dialogflow We integrated our chatbot
using Dialogflow fulfilment with Cloud Function so that it can interact with our model
in BigQuery.Using Dialogflow we deployed the chatbot on Telegram with the help of
Telegram API Token generated by Botfather which is used to set up Telegram bots.
Further, the chatbot was integrated with our Flutter application using Dart language
and Dialogflow API with the help of flutter_dialogflow package in Flutter. Flutter
authenticates the users with the help of Firebase Authentication, enabling users to
log in using Google, thus providing ease of access.
4 Results
The proposed methodology has been implemented using Google Cloud Platform,
Dialogflow and Flutter. To deploy the most accurate classifier we compared the
accuracy of Logistic Regression, Decision Tree and KNN algorithms.It was found
that out of Logistic Regression, KNN and Decision Tree algorithms, Logistic Regres-
sion performed the best with an accuracy of 85.14% while Decision Tree and KNN
performed with an accuracy of 74.72% and 84.61% respectively. Table 1 shows the
comparison between Logistic Regression, Decision Tree and KNN algorithms.
We can describe the performance of the classifier using a confusion matrix that
allows the visualization of the performance of an algorithm. In Table 2 the number
of incorrect and correct predictions are summarized together with their respective
count values and broken down by each class. It gives us insight not only into errors
being made by a classifier but more importantly the types of errors that are being
made by the classifier.
Classification rate/Accuracy:
Classification rate or Accuracy of a classifier is given by the following relation:
TP + TN
Accuracy =
TP + TN + FP + FN
5 Conclusion
Heart Disease is one of the major concerns for society today due to its high mortality
rate. The complexity of manually detecting the heart disease can be leveraged using
Machine Learning algorithms. In this paper, we discussed the accuracy of Decision
An Approach to Heart Disease Prediction with Machine … 895
Tree, KNN and Logistic Regression. The accuracy of the Decision Tree is 74.72%,
KNN is 84.61% and Logistic Regression is 85.14%. The Logistic Regression clas-
sifier performed with the highest accuracy than suggested in previous papers hence
we suggest Logistic Regression be used for further prediction of diseases. Further,
this system also can be used in the form of ensembles i.e. combinations of multiple
techniques. This would increase further accuracy and high performance of the system.
896 C. Nanaware et al.
References
1. https://www.who.int/en/news-room/fact-sheets/detail/cardiovascular-diseases-(cvds)
2. Karvonen MJ (1988) Prevention of cardiovascular disease among the elderly. Bull World Health
Organ 66(1):7–14
3. Liu B, Xu Z, Sun C, Wang B, Wang X, Wong DF, Zhang M (2018) content-oriented User
Modeling for Personalized Response Ranking in Chatbots. IEEE/ACM Trans Audio Speech
Lang Process 26(1):3–9
4. Gupta AK, Kumar P, Joshi S (2020) Conversational agent dialog flow interface. United States
patent application publication, US2020/0004874 A1,2-6, Jan 2
Automated Early Detection of Diabetic
Retinopathy
1 Introduction
Diabetic Retinopathy (DR) is a standout among the most successive reasons for
visual debilitation in created nations and is the main source of new instances of visual
deficiency in the working age populace. By and large, almost 75 individuals go dazzle
each day as an outcome of DR A viable treatment for DR require early finding and
consistent checking of diabetic patients, however this is a testing undertaking as the
malady indicates couple of manifestations until it is past the point where it is possible
to give treatment [3].
Diabetic Retinopathy is an eye issue that can cause visual deficiency. Little veins
in the back of the eye are called retinal veins. Indications of Diabetic Retinopathy are
gliding spot in vision, obscured vision and blocked vision. At the point when sugar
level in blood builds, blood vessels in the back of the eye ends up frail and in view
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 897
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_90
898 S. Shegdar et al.
of this vessel releases the blood and lipo-proteins liquid [5]. After that liquid ends
up skimming spot in vision with the goal that Diabetic patient cannot see anything
totally through the vision. In the event that we don’t do the treatment of this ailment
on the time then it might be conceivable of complete vision misfortune or visual
deficiency. On the off chance that we distinguished early the indication of Diabetic
Retinopathy, it is conceivable to keep extra loss of vision.
• Four stages of Diabetic Retinopathy are as follows:
As a cost effective way to handle the healthcare resources, systematic screening for
DR has been identified. An important screening tool for early DR detection is the
emergence of automatic retinal image analysis [13]. This can save both cost and time,
as it reduces the manual workload of grading as well as diagnostic cost and time.
Machine learning is a family of computational methods that allows an algorithm to
program itself by learning from a large set of examples that demonstrate the desired
behavior, removing the needs to specify rules explicitly. The accuracy, sensitivity
and specificity of the algorithm for detecting Diabetic Retinopathy (DR) can help
ophthalmologists and physicians raise the red flag and thus provide early treatment
to patients and bring in a more preventive care which can bring down the burden on
healthcare resources. Machine learning can thus help the old adage-prevention better
than cure, by predicting who is more liable to be at risk of DR or not.
2 Literature Survey
Cotton Wool Spots. RGB images is converted into YCbCr. Y plane is used for
detection of blood vessels, optic disc and exudates. After candy edge detection,
image will converted into binary to perform Skeletonization operation. DCT is
used for feature extraction. Author has proposed DCT (Discrete Cosine Trans-
form) for feature extraction. Extracted feature goes to SVM classifier. After that
SVM Classifier categories into Normal, DR and Glaucoma.
5. R. Radha and BijeeLakshman [7], Retinal image analysis using morpholog-
ical process and clustering technique. This paper proposes a method for the
Retinal image analysis through efficient detection of exudates and recognizes
the retina to be normal or abnormal. Morphology operators are applied to the
enhanced image in order to find the retinal image ridges. A simple thresholding
method along with opening and closing operation indicates the remained ridges
belonging to vessels. The clustering method is used for effective detection of
exudates of eye.
6. Yun et al. (Neural Network): Automatic classification of different stages
of diabetic retinopathy-mild nonproliferative retinopathy, moderate non-
proliferative retinopathy, severe non-proliferative retinopathy and proliferative
retinopathy using neural network from six features extracted from the retinal
images and accuracy obtained was 72.00 [11, 12].
3 Proposed System
The above Fig. 2. Shows the overall structure of the proposed system. It includes all
the modules which we are going to implement in this project. The architecture gives
idea of input to the system, processing on that input and what will be the output of
the project.
Feature Extraction
Image 1.Removal of optic Classification
Processing disc and blood vessels KNN
Techniques 2.Exudates detection Classifier
3. Micro-aneurysms
detection
Fig. 2 Proposed systems for detection and classification of different stages of diabetic retinopathy
Automated Early Detection of Diabetic Retinopathy 901
• Image Database:
The Messidor database [10] includes 1200 eye fundus shading numerical photos
of the back shaft obtained by 3 ophthalmologic workplaces using a shading video
30CD camera on a Topcon TRC NW6 non-mydriatic retinograph with a 45-degree
field of view. The pictures were caught implementing 8 bits for each shading plane
at 1440 X 960, 2240 X 1488 or 2304 X1536 pixels. 800 images were acquired with
pupil dilation (one drop of Tropic amide at 0.5%) and 400 without dilation. The
1200 pictures are bundled in 3 sets, one for every ophthalmologic division, utilizing
the TIFF position. What’s more, an Exceed expectations document with therapeutic
judgments for each picture is given.In this work, we utilize the pictures of only one
ophthalmologie division 1,69 with moderate NPDR (grade 2), and 149 with serious
NPDR (grade 3).
4 Implementation
4.1 Module 1
Image Processing:
The aim of this stage is to enhance those features that will allow, in the next stage,
classify the data set of fundus images. We have investigated the performance of two
image processing techniques: the canny edge detector and the histogram equalization
method.
(1) The Canny edge detector:
Edge detectors are local image processing techniques to detect edge pixels (pixels at
which the intensity of an image function changes abruptly). Particularly we have used
Canny algorithm because its performance is superior in general to the traditional edge
detectors [8]. This method consists of the following basic steps: Smooth the input
in age with a Gaussian filter, compute the gradient magnitude and angle images:
apply non-maxima suppression to the gradient magnitude image and use double
thresholding and connectivity analysis to detect and link edges.
Canny edge detection is used to find out the areas where intensity changes abruptly.
In the above Fig. 3., first image shows the original image and in second one it shows
the areas where intensity change abruptly i.e. the areas where defects are present.
(2) Histogram Equalization:
The histogram equalization is a widely used nonlinear method designed for the
enhancement of images. It is a non-parametric method to match the cumulative
distribution function of some given image to a reference distribution. This method
902 S. Shegdar et al.
This process determines the density of blood vessel in a retinal image. For that, the
RGB picture is changed to its CMY portrayal and the magenta segment is isolated.
On the magenta component, morphological operations (i.e. erosion, opening, and
dilation) hide blood vessels. The difference between the magenta component and
the resulting image of the morphological processing is binarized after a histogram
matching that increases its contrast, as shown in Fig. 4. The noise existing in the
binarized image is reduced through dilation and erosion operations. At last, the
density of white pixels (i.e. blood vessels) is computed.
(2) Microaneurysms
Microaneurysms are small lumps in the blood vessels, looking as little and round
shape spots near to tiny blood vessels. In order to determine the number of microa-
neurysms, the green component is extracted and the blood vessels are concealed
utilizing the commotion diminished picture of the previous process as shown in
Fig. 5. Basically, the pixels corresponding to blood vessels are painted with the
average retina color. And then, a disc-based dilation operation is applied to detect
the microaneurysms. It is important to find the difference between the resulting image
and the image of edges to finally remove those edges. Toward the end, the conceivable
microa-neurysms are shifted by utilizing morphological tasks, a number of pixels to
get the actual micro-aneurysms.
5 Result Analysis
The evaluation of our proposal is implemented in python and it has two sections DR
detection and DRNP grade classification. The main idea of this is to detect DRNP
grades. We trained the KNN classifier with all features of the images and then tasted
it. According to the features the system will detect the type as DR or NPDR and
mild, moderate and severe. With reference to Table 1.
The metrics used to evaluate the performance of the machine learning methods
were accuracy, precision, recall, and f-measure, defined as follows:
References
1 Introduction
A study was made on the prevailing ELV from different countries for comparing it
with the Indian scenario. From the findings, it had been realized that the ELV solution
could not be the sole solution towards reducing, recycling waste, and eco-friendly
process. However, this can be one of the various steps that must be taken and are also
crucial. The advantages are often reaped only in concurrence with the development
of the infrastructure of our country. While marketers plan to capture high revenues
at all the stages, they alter their sales techniques noticeably throughout the product
life cycle. Some economists believe that somehow sales of a brand-new product can
be predicted—right from the instant of its entry to its withdrawal from the market.
We will assess if the PLC (Product Lifecycle Concept) is sensible within the Indian
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 905
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_91
906 V. K. Bupesh Raja et al.
automobile industry and can showcase the Lifecycle Analysis of the models sold by
various OEMs in India [1].
2 Economic Analysis
For motor recycling, the car is sent to scrap. The vehicle is demolished, and its parts
are sent for sale in the market. In the process, the scrapyard owner earns more profit
than the vehicle owner [2].
The cost of a remanufacturing engine is less than half the price of the new engine.
Remanufacturers tend to earn an outstanding amount of money. By recycling the
engine and resale parts to the scrapyard, owners can earn a fair amount of money.
Also, consumers can save money on engine replacement. Apart from this, setting up
this new industry can open doors for new job opportunities [2].
Many individuals keen on a new car. Although the feeling of driving a novel car is
exciting and exhilarating—there is nothing like getting behind the wheel of a car that
you love and cherish and have driven for years [2, 3].
Like a human being, cars have a lifecycle, which starts well before the primary time
you drove it and ends well after the last. Here is the fascinating truth about cars: The
lifecycle of a car begins and ends in a very factory. During the manufacturing phase,
countless materials are accustomed to making many different parts. Metal is required
for the vehicle’s body, plastic is employed within the interior and parts of the body,
glass is required for the windows, and cloth and leather are used to upholster the
interior [2, 3].
Product Lifecycle of Automobiles 907
In the different phases of the product life cycle, firms will focus on different aspects
of promoting and sales.
3.2 Growth
Firms need to maximize growth to increase product sales from small retailers to big
supermarkets.Firms can change marketing from niche areas to a more mass-market
[3, 4].
3.3 Maturity
With peak penetration, the firm may seek to extend prices to increase profitability.
However, if the market is extremely competitive, the firm may feel the requirement
to keep prices low to defend market share.The firm may seek to enhance the product
to achieve market differentiation and extend the period of maturity[3, 4].
908 V. K. Bupesh Raja et al.
3.4 Decline
In the decline phase, the firm may feel it is best to let the product go—e.g. diesel cars
cannot solve problems with pollution and damage to its brand reputation. However,
with an iPhone, Apple lets old models go to get replaced by the successive model.
Decline and discontinuing the product is often the simplest way to force customers
to shop for an upgrade—next time their contract expires [3, 4].
This can be set apart into different sections like pre-assembly, assembly, use, post-use.
4.1 Pre-Assembly
4.2 Assembly
4.3 Use
4.4 Post-Use
1. Transportation to scrapyards.
2. Toxic wastes from scrapping.
3. Energy used for dismantling/scrapping.
4. Transportation of recycled parts to the place of reuse [5, 6].
5 Conclusion
There has been an unexpected gush in Indian Automobile Industry in the past two
decades. It has attracted global attention to investing in Indian Market. The product
life cycle in an automobile is significant because, on this cycle, many ancillary
industries depend. It’s expected that the automotive industry will play a vital role in
helping the economy to continue India’s growth.
References
1. Comparative analysis of scrap car recycling management policies: The 7th International confer-
ence on waste management and technology conference (2012) Kun Yue, Procedia Environmental
Sciences 16:44–50
2. Broch F, Warsen J, Krinke S (2015) Implementing life cycle engineering in automotive devel-
opment as a helpful management tool to support design for environment. In Sonnemann G,
Margni M (eds) Life cycle management. LCA compendium—the complete world of life cycle
assessment. Springer, Dordrecht, pp 319–329
3. Ribeiro C, Ferreira JV, Partidário P (2007) Life cycle assessment of a multi-material car
component. Int J LCA 12(5):336–345
4. https://www.ausmed.com/cpd/articles/product-life-cycle-relates-healthcare
5. Sabadka D, Molnár V, Fedorko G (2019) Shortening of Life Cycle and Complexity Impact on
the Automotive Industry. TEM J 8(4):1295–1301
6. Agnihotri D, Chaturvedi P (2013) Indian automobile industry: a life-cycle. VSRD Int J Bus
Manage Res 3(8):323–329
Analysis of Heavy Metal Pollutants
in the Sediments from Coastal Sites
of Al-Hodiedah Governorates, Yemen
1 Introduction
Heavy metals term is used to describe more than metal that are metals or metalloids
(elements that have both metal and nonmetal characteristics). The heavy metals
include Chromium, Arsenic, Cadmium, lead, Mercury, and Manganese. Generally
the heavy metals have densities above 5 g/cm3 . They cannot be degraded or destroyed
and are persistent in all parts of the Environment. The Human activities affect the
natural geological and biological redistribution of heavy metals through pollution of
the air, water, and soil. All the primary anthropogenic sources of heavy metals are
point sources such as mines, foundries, smelters, and coal-burning power stations
as well as diffuse sources such as combustion by products and car exhaust. Humans
also affect the natural geological and biological redistribution of heavy metals by
M. H. Nomaan
Amran University, Amran, Yemen
D. B. Panaskar
SRTM University, Nanded, Maharashtra, India
R. S. Pawar (B)
SVERIs College of Engineering, Pandharpur, India
e-mail: [email protected]
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 911
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_92
912 M. H. Nomaan et al.
altering the chemical form of heavy metals released into the environment. Such
change often affect a heavy metal’s toxicity by allowing it to bio-accumulate in
plants and animals, bio-concentrate in the food chain, or attack specific organs of
bodies. They are named after those metals between atomic number 21 (scandium)
and atomic number 84 (polonium), except for aluminum, which has atomic number
13, but it is also considered a heavy metal [1]. Heavy metals can also bio-accumulate
in the aquatic flora and fauna, such as plankton, fish, etc. Therefore, an early detection
of high heavy metal concentrations in water bodies such as rivers or wetlands are
vital for nature conservation [2]. The main transport of heavy metals in water bodies
is mainly through sediment [3], in the form of suspended solids.
Recently, several studies have been carried out describing the concentration and
distribution of heavy metals along the Yemeni Red Sea region. They have focused
on the existence of heavy metals in fish, mussels and shoreline sediments. In general,
data revealed that the levels of cadmium (Cd), cobalt (Co), crome (Cr), cupper (Cu),
ferrous (Fe), manganese (Mn), nickel (Ni), lead (Pb), and zinc (Zn) in the region were
generally within the range of levels reported for other regions in the world [4, 5].
However, the concentrations of Cd, Cr, Mn, Ni, Pb and Zn in some fish collected from
As Salif and Al Hodiedah sites were higher than those in other sites in the region. It is
largely believed that human activities attribute to such high concentrations, which was
mainly due to human activities in both sites. Concentrations of some metals, including
Zn, Cu, Mn and Fe in both shrimps and sea snails collected were higher more than in
other fish collected from the same sites. Another study, conducted by DouAbul and
Hebba [6], revealed that the concentration of heavy elements in fish, molluscs and
shoreline sediments in the region were rather low concentrations. Additionally, the
shrimps (P. semiculatus) collected showed a lower concentration (5.93 ppm) of Cu,
compared with that (29–171 ppm) showed by Karbe [7]. However, in most recent
studies [8], the mean concentrations of some heavy elements in dissolved particulates
and sediments, collected from several location in the region were slightly higher than
those found in other parts of the world. Sources of these metals were most likely
natural origin.
Salinity was also measured at each sample location. For each location, surface sedi-
ment samples were collected at 5 cm depths after removing any plant debris and
large stones. The sampling locations include areas of different environmental back-
grounds like fish aquaculture, recreational areas, ports, industries and urban areas.
All samples preserved in the Central Laboratory of the University of Aden at the time
of processing for digestion and analysis. A total of 15 samples from sediments were
collected from the sites during low tide in December 2011. The scientific method
used to collect sediment samples, where the study samples were collected from the
province of Al-Hodeidah five regions are as follows: Hodiedah port, Al-kateef shore,
Cornish location, Almehwate site and Al-Manjer location (Fig. 1) [9].
YEMEN
SANAA
Al Hudaydah
Scale
0 100 200 km
Aden
100 200 mi
0 0
42 44
0
16
1 Sanaa
2
3
4
Red Sea
5
Scale
0 50 100 km
1
Sampling Station
0
14
The trace elements from sediments were determined according to Oregioni and
Aston, [10] procedure. The concentration of heavy metals like Cu, Pb, Zn, Cd, Co,
Ni, Cr, Fe and As for all samples of Al-Hodiedah were determined. Heavy metal
analysis was performed on the < 63 µm fraction of the sediment which had been
separated by sieving after drying and grinding. The determination of Heavy metals
in particulate and sediment samples was done according to the procedure described
by Sturgeon [11]. The Perkin-Elmer 2380 atomic absorption spectrophotometer was
used for heavy metal analysis. The general method involves atomization of samples
by thermal sources and the absorption of a specific wavelength by the atomic source
as it is excited. The radiation used is a hollow cathode lamp containing, as its cathode,
the same element under analysis.
Heavy metals that will be referred to in this study as contaminants of concern that
will form the focus of this study are As, Cd, Cr, Cu, Co, Fe, Ni, Pb, and Zn. These
metals are known to be present in landfills and their presence poses serious problems
to human health and environmental concern. Such focus is because of their known
toxicity with regard to human health and ecology or their ability to bio accumulates
and moves through the food chain. Bioaccumulation processes may adversely affect
food resources by making them unfit for human consumption [12]. The Heavy Metal
analysis of sediments from Al-Hodiedah Governorate during year 2011 has been
presented in Table 1.
Arsenic of sediment samples varies from 0.27 to 1.22 µg/g in the year 2011 Table
1. The minimum As has been recorded from Cornish location, while maximum As
has been recorded from Hodiedah Harbour. The average As and Standard Deviation
of the sediment is 0.6957 µg/g and 0.3490 respectively Fig. 2.
Cadmium of sediment samples varies from 7.8 to 9.5 µg/g in the year 2011 Table 1.
The minimum Cd has been recorded from Al-Manjer location, while maximum Cd
Table 1 Heavy metal concentration (µg/g dry wt.) of sediment samples from Al-hodiedah governorate
Zones Heavy metal concentration during 2011
As Cd Co Cu Cr Fe Ni Pb Zn
µg/g dry wt
Hodiedah harbour 1.22 9.5 31.25 23.45 117.25 2591.75 15.83 62 52
Al-Kathib shore 0.63 9.25 26.7 18 46.2 2526 13.33 62 39.54
Cornish location 0.27 7.82 25.25 14.87 58.65 2541.18 17.15 63 36.33
Almehwate Site 0.53 7.94 25.55 21.45 87.14 2379.25 18.33 65 48.74
Analysis of Heavy Metal Pollutants in the Sediments …
1.2
1
As (μg/g)
0.8
0.6
0.4
0.2
Hodiedah Al-Kathib Shore Cornish Location Almehwate Site Al-Manjer
Harbour Location
Al Hodiedah Governorate Locations
9
Cd (μg/g)
7
Hodiedah Al-Kathib Shore Cornish Location Almehwate Site Al-Manjer
Harbour Location
Al Hodiedah Governorate Locations
has been recorded from Hodiedah Harbour. The average Cd and Standard Deviation
of the sediment is 8.462 µg/g and 0.8398 respectively Fig. 3.
Cobalt of sediment samples varies from 23.1 to 31.25 µg/g in the year 2011 Table 1.
The minimum Co has been recorded from Al-Manjer location, while maximum Co
has been recorded from Hodiedah Harbour. The average Co and Standard Deviation
of the sediment is 26.37 µg/g and 3.0225 respectively Fig. 4.
Analysis of Heavy Metal Pollutants in the Sediments … 917
29
Co (μg/g)
25
21
Hodiedah Al-Kathib Shore Cornish Location Almehwate Site Al-Manjer
Harbour Location
Al Hodiedah Governorate Locations
21
Cu (μg/g)
17
13
Hodiedah Al-Kathib Shore Cornish Location Almehwate Site Al-Manjer
Harbour Location
Al Hodiedah Governorate Locations
Copper of sediment samples varies from 14.87 to 23.45 µg/g in the year 2011 Table
1. The minimum Cu has been recorded from Cornish Location, while maximum Cu
has been recorded from Hodiedah Harbour. The average Cu and Standard Deviation
of the sediment is 19.22 µg/g and 3.3192 respectively Fig. 5.
Chromium of sediment samples varies from 46.2 to 117.25 µg/g in the year 2011
Table 1. The minimum Cr has been recorded from Al-Kathib Shore location, while
918 M. H. Nomaan et al.
100
80
Cr (μg/g)
60
40
20
Hodiedah Al-Kathib Shore Cornish Location Almehwate Site Al-Manjer
Harbour Location
Al Hodiedah Governorate Locations
maximum Cr has been recorded from Hodiedah Harbour. The average Cr and
Standard Deviation of the sediment is 73.102 µg/g and 28.9914 respectively Fig. 6.
Iron of sediment samples varies from 2379.25 to 2677.71 µg/g in the year 2011 Table
1. The minimum Fe has been recorded from Almehwate location, while maximum Fe
has been recorded from Al-Manjer location. The average Fe and Standard Deviation
of the sediment is 2543.18 µg/g and 109.12 respectively Fig. 7.
2600
Fe (μg/g)
2500
2400
2300
Hodiedah Al-Kathib Shore Cornish Almehwate Site Al-Manjer
Harbour Location Location
Al Hodiedah Governorate Locations
17
Ni (μg/g)
14
11
Hodiedah Al-Kathib Shore Cornish Location Almehwate Site Al-Manjer
Harbour Location
Al Hodiedah Governorate Locations
Nickel of sediment samples varies from 13.33 to 19 µg/g in the year 2011 Table 1.
The minimum Cd has been recorded from Al Kathib Shore, while maximum Cd has
been recorded from Al-Manjer Location. The average Cd and Standard Deviation of
the sediment is 16.728 µg/g and 2.2505 respectively Fig. 8.
Lead of sediment samples varies from 62 to 65 µg/g in the year 2011 Table 1.
The minimum Pb has been recorded from Al-Manjer location, Al Kathib Shore and
Hodiedah Harbour, while maximum Pb has been recorded from Almehwate Site.
The average Pb and Standard Deviation of the sediment is 62.8 µg/g and 1.3039
respectively Fig. 9.
Zink of sediment samples varies from 36.33 to 52 µg/g in the year 2011 Table 1.
The minimum Zn has been recorded from Cornish location, while maximum Zn has
been recorded from Hodiedah Harbour. The average Zn and Standard Deviation of
the sediment is 42.788 µg/g and 7.1122 respectively Fig. 10.
920 M. H. Nomaan et al.
64
Pb (μg/g)
62
60
Hodiedah Al-Kathib Shore Cornish Location Almehwate Site Al-Manjer
Harbour Location
Al Hodiedah Governorate Locations
50
Zn (μg/g)
40
30
Hodiedah Al-Kathib Shore Cornish Location Almehwate Site Al-Manjer
Harbour Location
Al Hodiedah Governorate Locations
4 Conclusion
As can be seen from the tables from previous chapter mean Heavy metals level in all
samples from Al-Hodiedah Governorate in the following order: Fe > Cr > Pb > Zn
> Co > Cu > Ni > Cd > AS. Arsenic, copper, Nickel and Zinc has been found with
lowest concentration in marine sediments of Costal Al-Hodiedah Governorate. All
concentrations are low and can be attributed to natural background levels. Therefore,
it can be deduced that they poses little or no treat to this most of this Environ-
ment. Cadmium, cobalt, Chromium, Iron and lead has been found with the highest
concentration in the marine sediments of Costal Al- Hodiedah Governorate. All the
concentrations can be attributed to natural background levels. Therefore, they pose
threat to the surrounding environment.
Analysis of Heavy Metal Pollutants in the Sediments … 921
References
Abstract Rapid and accurate detection of Infectious disease at rural setup is one of
the most challenging tasks. Advancement in molecular biology-based techniques had
revolutionized the field of diagnosis. Among several advantages, these techniques
are restricted to tertiary healthcare centers due to high cost and specific operating
procedures. We had developed a fast, battery-operated, portable device which can be
clubbed with conventional PCR for screening of Infectious diseases at rural setup.
The proof of concept was tested on standard reference samples of Mycobacterium
tuberculosis (H37RV strain) DNA as a positive control and ATCC Strain of E.coli
as a negative control in duplicates by clubbing the system with Conventional PCR
machine (For Amplification) and compared the result with Real-Time PCR and Fluo-
rescence microscopy techniques. The results of the detection system were found
consistent with conventional diagnostic techniques and can be used for screening of
infectious diseases at rural setup.
1 Introduction
Infectious diseases are one of the oldest health problems associated with mankind and
consuming millions of people every year. Pathogenic microorganisms are responsible
for infectious diseases and can be transmitted between individuals and population,
thereby threatening the general public health and, potentially, the economy [1]. In
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 923
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_93
924 P. Y. Navalsing et al.
order to solve the problem of over increasing rates of these infectious diseases, rapid
and specific identification of the causative agent plays a crucial role in successful
treatment and curing of the infected individuals. Conventional methods for iden-
tification of microorganisms associated with infectious diseases involves culturing
and sub-culturing of organisms, which is time-consuming and labor-intensive [2]
[3]. After discovery of the Nucleic Acid amplification and detection technologies
(PCR assays), various methods have been developed for the detection and character-
ization of microorganisms. These assays are highly sensitive & specific and can be
utilized for screening of all classes of microorganisms which causes various infec-
tious diseases [4, 5]. Nucleic acid based assays are continuously expanding in various
fields and are being adopted by clinical laboratories for identification of the disease
specific pathogens accurately in a time bound manner [6–9]. These methods had
identified numerous microbial pathogens responsible for acute and chronic diseases
of the mankind [10–17]. Nucleic acid based testing (NAT) relies on simple princi-
ples of direct amplification and detection of pathogen specific nucleic acid sequences,
and have better sensitivity than immunological and cell culture based detection tech-
niques [18]. These techniques offer very simple and effective way for screening of
infectious diseases such as TB and HIV, which are very difficult to diagnose by
conventional microbiological methods. Nucleic acid based testing’s are being devel-
oped for screening of drug resistance associated with particular pathogens and had
great advantages as compared to conventional antibiotic susceptibility testing. In spite
of the various advantages of Nucleic acid testing (NAT) based assays, these tests are
restricted in centralized laboratories due to use of high-end technologies and complex
operating procedures [19]. A Currently available instrument for nucleic acid detection
are bulky and expensive, which are available only at sophisticated tertiary healthcare
centers, and are not adopted currently by primary healthcare centers [20]. Due to
lack of skilled manpower and limited financial resources in rural areas of developing
countries well established commercially available NAT systems, cannot be deployed
at rural setup through the network of centralized clinical laboratories for serving
the need of poor strata of the society [21, 22]. The major challenge faced by devel-
oping world in diagnostic field is the development of robust, portable and low-cost
assays which will assist the detection of disease specific genetic biomarkers reliably
in places which are far away from the tertiary healthcare centers [23]. ASSURED
(affordable, sensitive, specific, user-friendly, fast, equipment free, and delivered) has
been invented for more than 10 years to define the ideal test to meet the needs of the
developing world [24].
In order to address these problems, various efforts are being made to develop
simple and low cost NAT-based assays and have focused mainly on the amplifica-
tion and detection steps for the detection of nucleic acids which can suite resource
limited settings [25]. Lot of simpler and cost effective nucleic acid amplification
technologies are emerging in the field of medical diagnostics, but the problem of
detection is still. Here we introduce a very simple and robust, but sensitive fluores-
cence method for end point detection of amplified nucleic acid using a miniaturized
low cost optical detection system which can be clubbed with low cost conventional
DNA amplification systems and can be operated at remote areas as well, for detection
Development of Low Cost PCR Product Detection System … 925
of Nucleic acids. Real time detection involves real-time monitoring of the amplifi-
cation reactions which need highly sensitive and accurate optical instrumentation
which leads to increase overall cost of the equipment on the other hand endpoint
detection technique needs simple instrumentation and provide less complex outputs
processing for interpretation which reduces the overall cost of detection. Hence we
had worked out on the development of simple and robust endpoint amplified nucleic
acid detection system for rural parts of the developing nations. We had conducted the
pilot study on standard reference positive and negative samples and found that the
detection system results matches with existing real-time PCR assay and fluorescence
detection systems.
Development of the end point amplified nucleic acid detection system is based on
the principle of fluorescence signal excitation and emission from the targeted nucleic
acids and fabricated by using off the shelf components in order to reduce develop-
ment cost of nucleic acid detection system [26]. We had utilized 6—fluorescein
amidite (FAM) dye specific filters and light source for excitation and emission of
these molecules as these molecules are attached to the probes in most of the commer-
cially available Nucleic acid based amplification kits. The basic design of the DNA
detection system was developed as per the basic principle of excitation and emission
of light by fluorescence molecules attached to Target DNA molecules which is as
shown in following Fig. 1.
Here we are amplifying the targeted Nucleic acid with the help of conventional
PCR protocol by using sets of primers and FAM Labelled probes and then visualizing
the samples post amplification on PCR Product Detection system, if tube shaped
fluorescence is observed inside the detector then the sample is positive and if tube
shaped fluorescence is absent then sample is considered as negative for the targeted
nucleic acid.
DNA was extracted from the standard reference samples by using commercially
available HiPurA™ Mycobacterium tuberculosis DNA Purification Kit (Cat No.
MB545) from HiMedia.
Development of Low Cost PCR Product Detection System … 927
The protocol has been standardized to amplify the control samples of Mycobac-
terium tuberculosis (Positive control) and E.coli (Negative Control) by using Hime-
dias Mycobacterium Tuberculosis real time Probe based PCR kit (Cat No. MBPCR
108) on Conventional low cost PCR machine for detection of the Tuberculosis. The
amplification protocol was standardized as per the kit insert. The mastermix was
prepared by mixing 10 µL of 2 × PCR Taq Mixture (MBT061), 2 µL of Primer Mix
for M. tuberculosis (DS0132A), 0.5 µL MTB Probe (FAM Labelled) (DS0253) and
5 µL of Extracted Template DNA, the volume of reaction mixture is then balanced
by adding 2.5 µL of Molecular Biology Grade Water for PCR (ML065), and the
prepared reaction mixture was amplified on Himedias Wee-16™ Thermal Cycler by
subjecting the samples to PCR cycles as Initial denaturation for 95 °C for 15 min
and 40 cycles of Denaturation at 95 °C for 15 s and annealing at 60 °C for 1 min.
After amplification of samples they are quickly transferred to detection system for
end point analysis, After placing the samples in the sample resting slot the samples
were excited by using monochromatic source of light and then there after the signal is
visualized by naked eyes in order to conclude the results as if tube shaped fluorescence
is observed inside the filter of detection system then sample is considered as positive
and if tube shaped fluoresce is absent inside the filter then samples is labelled as
negative for Tuberculosis. The detection system is useful for screening purpose.
Post analysis of samples on the detection system the samples were analyzed on
FL Auto Fluorescence Microscope (EVOS® FL Auto Imaging System; Cat No:
AMAFD1000) under 2X magnification for detection of visible signal generated from
amplified samples. The significance behind performing Fluorescence microscope
analysis is to test proof of concept as well as to compare the results of the detection
system. If tube shaped fluorescence signal is observed under the microscope, then
samples are considered positive for targeted nucleic acids and if fluorescence signal
is absent, then they are labeled as negative for the targeted nucleic acid.
a b c d
Fig. 3 Results of samples on developed PCR Product detection system; Image A & B denotes-
PCR Tube shaped fluorescence signal for positive control sample 1 & 2 Indicated by Arrow; Image
C & D denotes- No fluorescence Signal from Negative control sample 3 & 4
After completion of the conventional PCR assay the samples are quickly transferred
inside the detection system and the results are captured for both positive and negative
samples as shown in the Fig. 3.
Clear visible tube shaped fluorescence Signal is observed in both the positive
control 1 and 2 in triplicate indicated by arrow in image A and B and no fluores-
cence was seen in negative control 3 and 4 indicated by image C and D of Fig. 3,
which indicated that the results are consistence with the Real Time PCR assay and
fluorescence microscopy results. Thus the developed detection system in the present
study can be directly utilized for detection of FAM Labelled amplified nucleic acid
for specific screening of infectious diseases at rural setup. The developed detection
system in present study is mainly consist of low cost laser as excitation source and
fluorescence filter for filtration of excited light emitted from sample by eliminating
light directly coming from laser (Excitation source), the whole assembly is created by
using of the shelf components in order to minimize the cost of the detection system.
The basic operating procedure to use this system will be, amplification of Pathogen
specific sequences by using FAM labelled probes by conventional PCR machines
and then subsequent detection of the amplified samples by using our PCR product
detection system. This system can be used for identification of both bacterial and
viral pathogens by using commercially available FAM probes based real time PCR
kits.
The samples were successfully amplified by real time PCR (HiMediaInsta Q48TM
Real time PCR machine) by targeting mpt64 gene specific region of Mycobacterium
tuberculosis bacteria by using Hi-media Mycobacterium tuberculosis real time FAM
Probe based PCR kit, the amplification curve for the samples is as shown in the
Fig. 4. The curve A and B of Fig. 4 represents the successful amplification of mpt46
gene targeted by the FAM labelled probes for detection of Tuberculosis. The results
930 P. Y. Navalsing et al.
Fig. 4 Real Time PCR Amplification plot for the Positive and Negative control samples on HiMedi-
aInsta Q48 System; Curve A& B denotes- Amplification plot of H37RV strain as a Positive control
samples; Line C & D denotes- Amplification plot of E.coliATCC Strain as a Negative control
samples
shows that, threshold cycles for the positive control (H37RV Strain) was found to be
19.3 and 19.5 respectively and for negative control samples flat line was observed in
comparison to positive samples which indicates that no amplification was observed
in the negative samples as shown by C and D overlapped lines of Fig. 4, which
indicates the samples are negative for targeted mpt64 sequences as it is specific for
Mycobacterium Tuberculosis, hence no amplification is observed inside the E.coli
samples (Negative control).
a b c d
Fig. 5 Images for PCR Products analyzed on EVOS® FL Auto Fluorescence microscope, under
2x magnification; Image a & b denotes - PCR Tube shaped fluorescence signal for positive control
sample 1 & 2; Image C & D denotes- No fluorescence Signal from Negative control sample 3 & 4
4 Conclusion
In-order to screen the infectious diseases in rural setup in developing nations, the
systems should be developed which will be easy to operate, robust, low cost, light
weight (portable) and particularly battery operated, so that paramedical persons can
operate it. Several researchers had made an effort to develop such systems for the
endpoint PCR product detection but they are complex. Other researchers are trying to
develop low cost lab on chip detection systems but currently they are under develop-
ment phases our system offers a great reliability over other methods as the results can
be directly visualized by naked eyes and there is no need of photo detectors as well as
complex optoelectronic components which leads to increase in the overall cost of the
detection system. The only limitation associated with our study, we neither performed
culture and PCR to screen statistically sufficient number of samples patients to be
considered as gold standard for comparison with our detector which in the near future
we will be tested. The inclusion of either of it would have enabled us to calculate
sensitivity and specificity of the DNA end point detector for assessing the routine
applicability of and the system will be optimized for screening of various infectious
diseases other than Tuberculosis.
References
1. Chen H, Liu K, Li Z, Wang P (2019) Point of care testing for infectious diseases. Clinica Chim
Acta 493:138–147
2. Fantahun M, Kebede A, Yenew B, Gemechu T, Mamuye Y, Tadesse M, Yaregal Z (2019)
Diagnostic accuracy of Xpert MTB/RIF assay and non-molecular methods for the diagnosis
of tuberculosis lymphadenitis. PLoS ONE 14(9):
932 P. Y. Navalsing et al.
Abstract When a vibrating system is damped with more than one type of damping,
it is necessary to determine which of these types of damping are more effective to
control the resonant response. In such a case, identification of damping parameters
from the responses of a vibrating system becomes an important factor. Therefore
when the system is damped due to coulomb friction, viscous friction, it is necessary
to develop theoretical and experimental methods for identification of these damping
parameters from the responses of the vibrating system. As such, for identification
of coulomb and viscous friction (and also with particle damping) parameters, it is
proposed to develop methods irresponsible for the control of resonant response of
vibrating systems. The paper contains experimental setup and results about these
different types of mechanical vibrations.
1 Introduction
S. T. Bodare (B)
SKN Sinhgad College of Engineering, Korti, Pandharpur, Maharashtra, India
S. D. Katekar · C. Chaudhari
Mech Engineering Department, Sandip University, Nashik, Maharashtra, India
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 935
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_94
936 S. T. Bodare et al.
after particular time intervals and steady-state input, a transient input, or a random
input. The periodic input can be varied as a harmonic or a non-harmonic disturbance.
Example is loading machines shaking due to an imbalance. In case of transportation,
vibration caused by an engine or uneven road is an example too. The vibration of a
building due to an earthquake also causes vibrations. In the case of linear systems,
the frequency of the steady-state vibration response resulting from the application of
a periodic, harmonic input is equal to the frequency of the applied force or motion.
The vibrations are said to be damped, when the energy of a vibrating system is
gradually dissipated by friction, and other resistances. in such case of vibration the
amplitude of vibrations gradually decrease or change in frequency or intensity or
cease and the system rests in its equilibrium imposition. One of the examples of this
type of vibration is the vehicular suspension dampened by the shock absorbers. Den
Hartog first solved the forced response of a SDOF with both viscous and dry friction
damping. Viscous and coulomb friction is more important in different applications of
mechanical vibrations. These applications are like dynamics, and control problems.
Most friction dampers are used to reduce the resonant forces by providing sliding
contact between points experiencing relative motion due to vibration, thereby dissi-
pating resonant vibration energy. The above study is conducted to determine various
types of vibrations and effect of forced vibrations over mechanical elements such as
spring. Also study of viscous, coulomb and particle damping, and has been done.
This study will be useful in investigation of damping performance over a wide range
of excitation frequency and amplitude. Also, this study contributes in investigation
of frequency response analysis for various types of damping combinations.
2 Literature Review
Liang and Feeny [1] this gives a method for determination of Coulomb and Viscous
friction coefficients from responses of a harmonically excited dual-damped oscillator
with linear stiffness. The method of identification has been depended on existing
analytical solutions of non-sticking responses excited near resonance.
M.R. Duncan [2] have investigated the performance of a single particle vertical
impact damper over a range of forcing oscillation amplitudes and frequencies, mass
ratios, structural damping ratios, impact damper lid heights, and damper/structure
coefficients of restitution. Previous studies have not examined such a large parameter
space and instead have focused primarily on conditions at the structure’s natural
frequency and periodic particle trajectories.
Levitan [3] analyzed the motion of a system (refer Fig. 1) with harmonic displace-
ment of the base as shown in figure. The friction forces in the model proposed by
him are in between the base and the mass. Also in this paper, analytical solution
for the response of the support-excited system has been presented. The solution
to the equation of motion has been developed through the application of a Fourier
series to represent the frictional force opposing the relative motion between mass
and supporting structure.
Determination of Viscous, Coulomb and Particle Damping … 937
Den Hartog [4] has presented an exact solution for the steady-state vibration of a
harmonically excited oscillator damped by combined dry and viscous friction (refer
Fig. 2). The system, as shown in figure, it consists of a forced excited mass m with
friction forces acting in between ground and them. The quite a few experimental
tests to verify solutions have been performed to find out the forced response of a
single-degree-of-freedom system with both viscous and dry friction damping.
Hundal [5] has done experimental study a base-excitation frictional oscillator
in which close form analytical solutions of the equation of motion were obtained.
The consequences of these experiments have been unfilled in non-dimensional form
as magnification factors in opposition to frequency ratios as functions of Viscous
and Coulomb friction parameters. It has been shown that the mass motion may
be continuous or it has a single stop during each cycle, depending upon system
parameters. The response of a single degree of freedom spring-mass system with
Viscous and Coulomb friction, with harmonic base excitation, has been determined.
Mao et al. [6] have examined the characteristics of Particle damping with respect
to a simple single-mass impact damper and a dry-friction damper. The analysis of
3 Methodology
When the vibrating system has both viscous and coulomb friction damping, it is
observed that the coulomb friction has noticeable effect on the resonant amplitude.
However, it is not clear from the steady state response analysis how much is the
contribution of the viscous friction in the system towards the reduction of resonant
amplitude [8, 9]. In more than few circumstances, such as in control systems, it is
necessary to know the content of the viscous and coulomb friction damping. in such
case, the identification of these two types of friction damping in a vibrating system
will play a significant role when the method used to control the resonant amplitude
of a system is the use of damping.
or
where,
m, c, f (x) and k are the mass, viscous damping coefficient, Coulomb friction
force, and spring constant respectively, y(t) = Y0 cos(ωt + ∅) is the base excitation
motion.
One can write the Eq. (2) as,
c
ẍ + ẋ + ωn2 x − x f = ωn2 Y0 cos(ωt + ∅) (3)
m
where,
x f = Fk0 denotes the equivalent “friction displacement” and mk = ωn2 denotes
un-damped natural frequency. The peak values of non-sticking response occur at the
turning points and the peak amplitude x0 of the response is given as,
2
x0 xf 1 xf
= −G + −H 2 (4)
Y0 Y0 q2 Y0
and,
2
ω ω 2
q= 1− + 2ξ (7)
ωn ωn
Y0
x0 = −Gx f + (8)
q
The above equation is nothing but equation of line of Y0 vs. X0 in which the slope
and intercept are q and Gx f respectively.
If the system is excited at two excitation levels with the same frequency (close
to the damped natural frequency) with input output pairs denoted as, (Y1 , X1 ), and
(Y2 , X2 ) one can write using Eq. (8) as,
Y1
x1 = −Gx f + (9)
q
Y2
x2 = −Gx f + (10)
q
Subtracting Eq. (9) from Eq. (10) and after rearranging yields,
Y2 − Y1
q= (11)
x 2 − x1
x1 Y2 − x2 Y1
xf = (12)
G(Y1 − Y2 )
An expression for ξ can be obtained by putting Eq. (7) into Eq. (11) with the
condition ω ≈ ωn ≈ ωd
Y2 − Y1
ξ= (13)
2(x2 − x1 )
sinh(ξ π)
G≈ (Approximately) (14)
cosh(ξ π) − 1
When more than two input-output amplitudes pairs are measured, the intercept
land slope of Eq. (8) is estimated by a least-squares fit of the (Yi, xi) data. It should
be noted that implementing the identification process, the system is required to be
excited lat the resonance.
Determination of Viscous, Coulomb and Particle Damping … 941
4 Experimental Set up
i. An experimental set up has been designed and developed to obtain input output
amplitude relationship for SDOF Harmonically excited system with Viscous,
Viscous and Coulomb friction, and Viscous, Coulomb friction and Particle
damping
ii. For this purpose, design and development of viscous fluid damper, dry friction
damper and particle damping system has been carried out.
942 S. T. Bodare et al.
iii. A schematic of the experimental test set-up is shown in the Fig. 3 of the
“C” frame. The frequency response curve of the system is obtained using
the accelerometer pick up with its necessary attendant equipment and a FFT
analyzer, over a small range of excitation frequency.
In this case, the SDOF spring mass system with different damping types, system
has been excited from 5 Hz to 12 Hz using electro dynamic exciter and the steady
state response X has been measured using an accelerometer and FFT analyzer. Only
in case of PVD the readings are taken with increase in the number of balls from no
ball to full of balls with regular intervals.
5 Theoretical Analysis
(a) The non-sticking response of a Single Degree Of Freedom (SDOF for the short)
system subjected to harmonic excitation with both Viscous and Dry friction
has been determined in the approach of Cheng and Zu.
(b) Also the necessary expressions for viscous friction and dry friction parameters
have been obtained by establishing the input output amplitude relationship
near resonance of the SDOF system with viscous and Coulomb damping and
subjected to harmonic mass excitation and using two input output amplitude
pairs the values of dry and viscous friction has been determined in the approach
of Liang and Feeny.
(c) Using more than two input output response pairs, the values of Viscous and
Coulomb friction damping parameters have been estimated using a least square
fit of input output response parameters.
Determination of Viscous, Coulomb and Particle Damping … 943
In this case, the SDOF spring mass system has been excited from 5 Hz to 12 Hz using
electro dynamic exciter and the steady state response X has been measured using an
accelerometer and FFT analyzer.
The value of system damping ratio ξs has been determined as,
f2 − f1
ξs =
2 × fn
5.2 Spring
The springs selected are helical compression with both ends squared and ground.
The dimensions of springs are obtained using standard formulae of design of helical
compression spring (refer Table 1). The material of springs is selected as high grade
spring steel.
6 Results
Using the experimental set up, the frequency response analysis has been carried out
for a harmonically based excited SDOF system. Table 2 shows Damping Rations of
Various Combinations
1. System with mass m and stiffness k and system damping (S).
2. System with m. k and system damping plus viscous damping (S + V).
3. System with m, k and system damping with coulomb damping (S + C).
4. System with m, k, and system damping S with Viscous Damping and Coulomb
damping (S + V+C).
5. System with m, k, coulombs damping, viscous damping, and particle damping
(S + V+C + P).
944 S. T. Bodare et al.
7 Conclusion
In viscous, coulomb friction & particle damping, the coulomb friction damping is
non-linear damping so it should be minimized, for that purpose we found coulomb
friction parameters & the result obtained are nearly the same in theoretical and exper-
imental method. The unique aspect of PID is that high damping has been achieved
by absorbing the kinetic energy of the structure as opposed to the more traditional
methods of damping that uses elastic, where strain energy stored in the structure
is converted to heat. From the above literature study, it is necessary to understand
contents of viscous damping, coulomb damping, and particle damping in a system.
The dominant damping identification in a system is useful to understand the behavior
of system.
References
1. Liang JW, Feeny BF (2004) Identifying coulomb and viscous friction in forced dual-damped
oscillators. ASME J Vib Acoustics 126:118–125
2. Steven E (2003) Olson: an analytical particle damping model. J Sound Vib 264:1155–1166
3. Duncan MR, Wassgren CR, Krousgrill CM (2005) The damping performance of a single
particle impact damper. J Sound Vib 286:123–144
4. Mao (2004) Simulation and characterization of particle damping in transient vibrations. ASME
J Vib Acoustics 126:202–211
5. Den Hartog JP (1931) Forced vibrations with combined coulomb and viscous friction. Trans
ASME 53:107–115
6. Hundal MS (1979) Response of a base excited system with coulomb and viscous friction. J
Sound Vib 64:371–378
Determination of Viscous, Coulomb and Particle Damping … 945
7. Levitan ES (1960) Forced oscillation of a spring-mass system having combined coulomb and
viscous damping. J Acoust Soc Am 32:1265–1269
8. Perls Thomas A, Sherrard Emile S (1956) Frequency response of second order systems with
combined coulomb and viscous damping. J Res Natl Bureau Standard 57:45–64
9. Liang JW, Feeny F (1999) The estimation of viscous and coulomb damping in harmonically
excited oscillators. Proceedings of DETC 99, Las Vegas
10. Elliott SJ, Tehrani MG, Langley RS (2005) Nonlinear damping and quasi-linear modelling
Phil. Trans R Soc A.37320140402
11. Duifhuis H (2012) 2012 Cochlear mechanics: introduction to a time domain analysis of the
nonlinear cochlea. Springer, New York
Assessment of Godavari River Water
Quality of Nanded City, Maharashtra,
India
Abstract Water resources are sources of water that useful or potentially useful to
humans. Uses of water are including domestic, agricultural, industrial, recreational as
well as environmental activities. Virtually, all these human uses require fresh water.
The water samples from the Godavari River of the five sampling site are taken and
analyzed for the physicochemical parameters such as Colour, Odour, Temperature,
pH, Electrical Conductivity (EC), Total Solids (TS), Total Dissolved Solids (TDS),
Total Suspended Solids (TSS), Total Hardness (TH), Calcium (Ca), Magnesium
(Mg), Chloride (Cl), Alkalinity (TA), Dissolved Oxygen (DO), Chemical Oxygen
Demand (COD), Biological Oxygen Demand (BOD) and Turbidity, etc. parameters
in seven weeks. Godavari River water is one of the important sources of water in
Nanded city. As the Nanded is drought-prone area alteration within major or minor
in the characteristics of River water results in great attention of day to day life of
citizens of Nanded. There are five samples taken weeks of March and April months
of Godavari River to access the quality of water and the results are compared with
WHO standards.
P. R. Shaikh (B)
School of Earth Sciences, S.R.T.M. University, Nanded, Maharashtra, India
G. Deore
Matoshri College of Engineering, Nanded, Maharashtra, India
A. D. Pathare
Pravara Rural Engineering College, Loni, Maharashtra, India
D. V. Pathare
Padmashri DVVP Institute of Technology and Engineering (Polytechnic) College, Pravaranagar,
Loni, Maharashtra, India
R. S. Pawar
SVERIs College of Engineering Pandharpur, Solapur, Maharashtra, India
e-mail: [email protected]
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 947
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_95
948 P. R. Shaikh et al.
1 Introduction
Water is well known as a natural solvent. Before, it reaches the consumer’s tap,
it comes into contact with different pollutants, including organic and inorganic
matter, chemicals, and other contaminants which pollute them. Many public water
systems, treated water with chlorine to kill disease-producing contaminants that may
be present in the water. Although disinfection is an important step in the treatment of
potable water, the taste and odor of chlorine is objectionable. The disinfectants that
are used to prevent disease can create byproducts which may pose significant health
risks. The water cycle explains interactions between the atmosphere, hydrosphere,
and lithosphere. The water or hydrologic cycle is a major driving force on our planet.
A river is a natural watercourse, usually freshwater, flowing towards an ocean, a lake,
a sea, or another river. In a few cases, a river simply flows into the ground or dries
up completely at the end of its course, and does not reach another body of water [1].
2 Study Area
0
40
Nanded Taluka
India Nanded District
19011'0"
300
Maharashtra 19009'0"
200
0 0 200 Km
0 1000 Km 10 190 07'0"
Scale
Nanded City
Scale
0 2 Km
77017'0" 0
77 19'0" 77021'0"
Shree Nagar
Vazirabad
Old Mondha
4
3
1
Index
5 Sample Number
19 o08' 23"
For the study of River water quality we have select the Godavari River at Nanded city
because the growth and development of the city affect the water quality of this River.
As Nanded city is situated in Marathwada region of Maharashtra (draught zone).
Population of this city is mainly dependent on the River water to sustain. Therefore
any major alteration in the physico-chemical characteristics of River water of this
region affects the day to day activities of peoples in this region adversely.
950 P. R. Shaikh et al.
There are 5 locations selected as per the pollution sources and total 7 times (Weeks of
March and April Months) water samples were collected from the study of River water
quality. Samples are collected in the polythene bottles. The collected samples are
transfer to the laboratory for further analysis. The physico-chemical parameters such
as Colour, Odour, Temperature, pH, Electrical Conductivity (EC), Total Solids (TS),
Total Dissolved Solids (TDS), Total Suspended Solids (TSS), Total Hardness (TH),
Calcium (Ca), Magnesium (Mg), Chloride (Cl), Alkalinity (TA), Dissolved Oxygen
(DO), Chemical Oxygen Demand (COD), Biological Oxygen Demand (BOD) and
Turbidity, etc. are analysed. The physico-chemical parameters are checked by the
titrometric method [3, 5].
The water samples from the Godavari River of the five sampling site are taken. The
samples are filtered with the help of filter paper and analyzed for the physico-chemical
parameters such as Colour, Odour, Temperature, pH, Electrical Conductivity (EC),
Total Solids (TS), Total Dissolved Solids (TDS), Total Suspended Solids (TSS), Total
Hardness (TH), Calcium (Ca), Magnesium (Mg), Chloride (Cl), Alkalinity (TA),
Dissolved Oxygen (DO), Chemical Oxygen Demand (COD), Biological Oxygen
Demand (BOD) and Turbidity, etc. parameters in seven weeks as shown in the Table 1.
The all samples show the greenish colour, which is the indicator of algal blooms.
The odour is objectionable, because of the mixing of sewage in the River water. The
temperature is varies from 35 to 38 °C. It is increases because of the summer season.
4.1 pH
The pH of samples in seven weeks of March and April months is ranges from 7.4
to 7.8. The pH value of all the weeks of two months falls in Neutral category. The
average and standard deviation of pH is 7.52 and 0.1 respectively. The pH of sample
no. 1 is higher as compare to the other samples of first three weeks ref Fig. 2.
The EC of samples in seven weeks of March and April months is ranges from 1952
to 2178 µS/cm. The average and standard deviation of EC is 2080.9 and 77.12
respectively. The EC of sample no. 1 is higher as compare to the other samples and
Assessment of Godavari River Water Quality of Nanded City … 951
Table 1 Physico-chemical parameters of Godavari river water samples in the weeks of March and
April months
Parameters Unit Min Max Avg StDev
pH – 7.4 7.8 7.52 0.1
EC µS/cm 1952 2178 2080.9 77.12
TS mg/l 2253 3394 2743.2 520.6
TSS 1000 2000 1411.4 489.76
TDS 1249 1394 1331.4 49.35
TH 154 302 191.2 33.8
Ca 36.07 84.17 47.18 11.31
Mg 16.52 25.33 20.23 2.4
Cl 99.4 205.9 148.97 21.89
TA 130 150 137.57 5.86
DO 3.2 5 3.81 0.43
COD 78 128 104.71 12.25
BOD 31 51 41.66 4.84
Turbidity NTU 220 663 370.09 81
7.8
7.7
7.6
pH
7.5
7.4
7.3
7.2
12345123451234512345123451234512345
2100
µS/cm
2000
1900
12345123451234512345123451234512345
lower in sample no. 5 of all the sampling weeks. From the Fig. 3 it is seen that the
EC is gradually increased during the sampling periods.
The TS of samples in seven weeks of March and April months is ranges from 2253 to
3394 mg/l. The average and standard deviation of TS is 2743.2 and 520.6 respectively.
The TSS of samples in seven weeks of March and April months is ranges from 1000
to 2000 mg/l. The average and standard deviation of TSS is 1411.4 and 489.76
respectively. The TDS of samples in seven weeks of March and April months is
ranges from 1249 to 1394 mg/l. The average and standard deviation of TDS is 1331.8
and 49.35 respectively. All the samples are falls in above the highest desirable limit,
but below the maximum permissible limit given by WHO [6, 7]. From the Fig. 4, 5
and 6 it is seen that the TDS is gradually increased during the sampling periods.
The TH of samples in seven weeks of March and April months is ranges from 154 to
302 mg/l. The average and standard deviation of TH is 191.2 and 33.8, respectively.
The all samples falls hard water category. According to WHO [6, 7], all samples falls
Assessment of Godavari River Water Quality of Nanded City … 953
3100
mg/l
2600
2100
1600
12345123451234512345123451234512345
2000
1500
mg/l
1000
500
0
12345123451234512345123451234512345
above the highest desirable limit, but below the maximum permissible limit except
one sample. It can be seen that the river water samples are more total hardness
content because of mixing of sewage. From the Fig. 7 it is seen that the TH is
954 P. R. Shaikh et al.
1500
Graphical Representation of TDS
1400
mg/l
1300
1200
12345123451234512345123451234512345
320
Graphical Representation of TH
mg/l
220
120
12345123451234512345123451234512345
gradually increased during the sampling periods. The sample no. 4 TH content is
higher in the 8 March, 15 March and 22 March sampling period and in the sample
no 2, TH content is higher in next all sampling periods.
Assessment of Godavari River Water Quality of Nanded City … 955
100
Graphical Representation of Ca
mg/l
50
0
12345123451234512 345123451234512345
The Ca of samples in seven weeks of March and April months is ranges from 36.07 to
84.17 mg/l. The average and standard deviation of Ca is 47.18 and 11.31 respectively.
All samples falls below the highest desirable limit, only two samples falls above the
highest desirable limit but below the maximum permissible limit given by WHO [6,
7]. From the Fig. 8 it is seen that the Ca is gradually increased during the sampling
periods. The sample no. 4 Ca content is higher in the 8 March, 15 March and 22
March sampling period and in the sample no 2, Ca content is higher in next all
sampling periods.
The Mg of samples in seven weeks of March and April months is ranges from 16.52 to
25.33 mg/l. The average and standard deviation of Mg is 20.23 and 2.4 respectively.
All samples falls below the highest desirable limit given by WHO [6, 7]. From the
Fig. 9 it is seen that the Mg is gradually increased during the sampling periods. The
sample no. 4 Mg content is higher in the 8 March, 15 March and 22 March sampling
period and in the sample no 2, Mg content is higher in next all sampling periods.
956 P. R. Shaikh et al.
30
Graphical Representation of Mg
mg/l
20
10
12 3451 2345 1234 512 3451 2345 1234 5123 45
4.7 Chlorides
The Cl of samples in seven weeks of March and April months is ranges from 99.4 to
205.9 mg/l. The average and standard deviation of Cl is 148.97 and 21.89 respectively.
According to WHO, all samples falls below the highest desirable limit except one
sample. It can be seen that the river water samples are more chloride content because
of mixing of sewage. From the Fig. 10 it is seen that the Cl is gradually increased
during the sampling periods. The sample no. 5 Cl content is higher except first two
sampling periods.
The TA of samples in seven weeks of March and April months is ranges from 130 to
150 mg/l. The average and standard deviation of TA is 137.57 and 5.86 respectively.
From the Fig. 11 it is seen that the TA is gradually decreased during the sampling
periods.
Assessment of Godavari River Water Quality of Nanded City … 957
180
mg/l
130
80
123451234 51234512345123451 234512345
140
120
123451 23451234512 345123451234 512345
The DO of samples in seven weeks of March and April months is ranges from 3.2
to 5 mg/l. The average and standard deviation of DO is 3.81 and 0.43 respectively.
958 P. R. Shaikh et al.
6 Graphical Representation of DO
mg/l
2
1 23 45 12 34 51 23 45 12 3 45 12 34 51 23 45 12 34 5
From the Fig. 12 it is seen that the DO is gradually decreased during the sampling
periods. The sample no. 5 DO content is higher in the first 4 sampling period and in
sample no 3, DO content is higher in next all sampling periods.
The COD of samples in seven weeks of March and April months is ranges from 78
to 128 mg/l. The average and standard deviation of COD is 104.71 and 12.25 respec-
tively. According to WHO [6, 7], all samples falls below the maximum permissible
limit. From the Fig. 13 it is seen that the COD is gradually increased during the
sampling periods.
The BOD of samples in seven weeks of March and April months is ranges from 31 to
51 mg/l. The average and standard deviation of BOD is 41.66 and 4.84 respectively.
According to WHO [6, 7], all samples falls below the maximum permissible limit.
From the Fig. 14 it is seen that the BOD is gradually increased during the sampling
periods.
Assessment of Godavari River Water Quality of Nanded City … 959
150
Graphical Representation of COD
mg/l
100
50
123 4512345 1234512 3451234 5123451 2345
45
mg/l
35
25
1 234 512 34 512 345 12 345 123 451 23 451 234 5
600
500
NTU
400
300
200
12 34512 34512 34512 34512 34512 34512 345
4.12 Turbidity
The Turbidity of samples in seven weeks of March and April months is ranges from
220 to 663 mg/l. The average and standard deviation of Turbidity is 370.09 and 81
respectively ref Fig. 15.
5 Conclusion
Godavari River water is one of the important sources of water in Nanded city. As the
Nanded is drought prone area alteration within major or minor in the characteristics
of River water results in great attention of day to day life of citizens of Nanded. There
are five samples taken weeks of March and April months of Godavari River to access
the quality of water and the results are compared with WHO standards. All the values
are found within limit (except EC, TDS, and TH). The Chloride value is found little
higher. This indicates that River water is contaminated with sewage pollution, fecal
matter. Probable solutions are suggested to avoid the further contamination.
1. Analysis of physico- chemical parameters of water shows that all the parameters
are within (except EC, TDS, and TH) Standard limit provided by WHO.
2. River water of this region is contaminated with sewage waste
3. Disposal site for sewage waste should be out of the premises of the city to avoid
further contamination of River water of the area.
Assessment of Godavari River Water Quality of Nanded City … 961
References
1. Verma S (2009) Seasonal variation of water quality in Betwa river at Bundelkhand region, India.
Glob J Environ Res 3(3):164–168
2. Srivastava A, Kr R, Gupta V, Agarwal G, Srivastava S, Singh I (2011) Water quality assess-
ment of Ramganga river at Moradabad by physico-chemical parameters analysis. VSRD-TNTJ
2(3):119–127
3. Panaskar DB, Wagh VM, Pawar RS (2014) Assessment of groundwater quality for suitability
of domestic and irrigation from Nanded Tehsil, Maharashtra, India. SRTMUs J Sci 3(2):71–83
4. Pawar RS, Panaskar DB (2014) Characterisation of groundwater in relation to domestic and
agricultural purposs, Solapur Industrial Belt, Maharashtra, India. J Environ Res Dev (JERAD)
9(01):102–112
5. APHA, AWWA, WEF (1985) Standard methods for the examination of water and wastewater,
16th ed. American Public Health Association. Washington DC, pp 6–187
6. World Health Organization (WHO) (1998) International standards for drinking water quality-
geneva, WHO
7. World Health Organization (WHO) (2002) Guideline for Drinking Water Quality, 2nd ed. Health
criteria and other supporting information. World Health Organization, Geneva, pp 940–949
Groundwater Quality Assessment
in an Around Thermal Power Plants
in Central India
Abstract More than hundred million tons of coal fly ash is produced annually in
India from combustion of coal in power plants. It is expected that about 150 million
tons of coal ash will be produced due to burning of coal in power plants by the year
of 2015. This will require about 30,000 hectare of land for the disposal of ash. One
of the biggest problems due to disposal of large quantities of coal ash is the possible
leaching of different hazardous pollutants, including TDS, fluoride and sulphate. The
present study investigated the leaching of soluble compounds in ash ponds and its
impact on the groundwater quality in the sub-watershed around the ash ponds in the
vicinity of Koradi and Khaperkheda near Nagpur in Maharashtra, India. A network of
twenty three observation wells set up for monitoring of water level and groundwater
quality for major cations and anions during pre-monsoon season, 2010. The results
indicate that the SO4 concentration is very high (>1000 mg/L) in the samples which
is much closer to the ash ponds. The TDS and Fluoride concentration is also elevated
with respect to BIS standards in few samples.
Keywords Coal fly ash · Power plants · Ash ponds · Groundwater quality · Koradi
1 Introduction
In India, groundwater is the principal source for drinking, irrigation and domestic as
well as industrial requirements especially in the rural areas, even in peri urban areas,
where centralized water supply is not available; groundwater is used for drinking and
domestic purposes. Hence, Protection of the groundwater resource assumes impor-
tance as it constitutes 96% of the available freshwater globally [1]. The huge amount
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 963
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_96
964 V. U. Deshmukh et al.
of coal consumption for electricity generation used in thermal power plants in most
part of India. Due to this, lot of waste in the form of ash is generated and is often
mixed with water and disposed in ponds in the form of slurry. The ash is rich in the
major constituents namely, CaO, MgO, Na2 O, K2 O, SiO2 , Fe2 O3 , MnO, TiO2 and
P2 O5 . Besides, the ash is also rich in trace elements like Cu, Pb, Zn, Ni, Co etc.
[2]. The ponds do not have a secured liner to prevent the seepage of contaminants
to the aquifer. Hence, it is essential to study the likely threat to groundwater regime
from such sources of groundwater pollution. Among the various sources of ground-
water pollution, the ash ponds of the Thermal power plants have been one of the
major concerns recently both on a global as well as national level. The soil and water
contamination from ash disposal ponds has been a subject of much research the world
over [3–5]. The contamination is a matter of much concern as huge amounts of ash
are generated on a daily basis and they are continuously disposed on the ground. The
leachate can ultimately get into the aquifer and contaminate it. The present work is an
attempt to study the impact of the ash ponds on the groundwater resources. It seeks to
adopt a holistic approach in examining the groundwater contamination vis-à-vis the
ash ponds. The study has been undertaken in a sub-watershed surrounding the ash
ponds of Koradi and Khaparkheda power plants near Nagpur in Maharashtra. The
groundwater quality is assessed vis-à-vis the BIS [6] (Bureau of Indian Standards)
guidelines for standards of drinking water. The focus is on the major cations /anions
and fluoride.
2 Study Area
The study area (57.8 km2 ), covers the ash ponds of the Koradi and Khaparkheda
Power plants ref Fig. 1. It is located approximately 15 km from the Nagpur city
and it is bounded by longitudes 79 ° 05´ E to 79 ° 15´ E and latitudes 21 ° 10´
N to 21 ° 17´ N (SoI Toposheet No. 55 O/3 and 55 O/4 on 1:50,000 scale). The
region experiences an arid climate with temperature varying from 10º to 46 °C. The
average annual rainfall is approximately 1,100 mm [7]. The area receives rainfall
from the southwest monsoon during the period June to September. The ash from
Koradi Thermal Power Plant is disposed in Pond 1 and Pond 2, whereas Pond 3
receives the ash from Khaparkheda power plant. The ash from the power plant is
generated in two different modes. It is disposed as fly ash in the atmosphere and
as ash slurry in ash ponds. The ash slurry is made by mixing the bottom ash with
water and transported to the pond through pipelines. Pond 1 and Pond 2 are spread
over 254 ha whereas Pond 3 is spread over 282 ha. The Koradi Power Plant and the
Khaparkheda Power Plant are in operation since 1974 and 1989 respectively. The
ponds do not have liners at the bottom as well as the flanks.
Geologically, the area comprises formations of Sausar group (quartzite’s, marbles,
schist’s and gneisses), sedimentary formations (sandstone, shale and clay) equivalent
of Kamthi stage of Gondwana super group [8, 9]. The southern part of the study area
is covered by the Archeans whereas the northern part is covered by the Gondwanas.
Groundwater Quality Assessment in an Around Thermal Power … 965
A network of twenty three observation wells both dug wells as well as hand pumps
were set up in the study area during pre monsoon 2010 season. The pre-monsoon
sampling was done during May months and analyzed for major cations (Na, K, Ca,
Mg), anions (CO3, HCO3, Cl, SO4), fluoride following standard protocol [10] and pH
and temperature were measured in the field. The samples were filtered by Whatman
filter paper (no. 42) prior to the analysis in the laboratory. Water level was measured
with the help of an electric contact-gauze (KL 010 A) with a buzzer. The reduced
level of the observation wells has been measured by surveyors. The reduced level
of the observation wells has been used to obtain the groundwater level (above mean
sea level). Subsequently, the groundwater level (amsl) contours were prepared for
identifying the groundwater flow direction in the study area. The QA/QC was ensured
by analyzing the samples in duplicate manner for effective accuracy of results and
maintaining various quality criteria like analysis of externally supplied standards,
analysis of reagent blanks and calibration with standards during the laboratory. As
a minimum requirement of the accuracy in the analytical instruments used in the
analysis, three different dilutions of the standards were measured when an analysis
is initiated. Subsequently, the calibration curve was plotted by analyzing one or
more standards within the linear range. The concentration of the specific parameters
was obtained from the calibration curve against the corresponding wavelength. The
majority of analyzed groundwater samples had ionic charge balance <2–3%.
966 V. U. Deshmukh et al.
The groundwater table varied between 3.4 m to 15.7 m (bgl) in pre-monsoon season,
2010. A groundwater level contour (amsl) indicates ref Fig. 2 that the groundwater
flows towards the south-east direction.
A close look towards groundwater quality ref Table 1 showed that the sulphate is
one of the problematic parameter contaminating the groundwater in the study area.
The sulphate concentration varied in between 6 – 943 mg/L (pre 2010). It is indicated
that the higher sulphate concentration (>400 mg/L) continuously observed in KD-8,
KD-9 and KD-10 which immediate downstream of the ash ponds ref Fig. 3. In 2010,
TDS concentration varied in range of 127 – 2322 mg/L for pre-monsoon season. The
fluoride concentration continuously increasing in KD-23 which is much closer to
the fly ash carrying stream. The literature said that, due to the ash pond effect there
is chance of fluoride contamination in the groundwater [11]. The piper diagram ref
Fig. 4, reveal that the distribution pattern of anions and cations for pre-monsoon and
post monsoon seasons of the study period. The chloride and sulphate anions play
major contribution followed by sodium and potassium in pre monsoon season 2010.
Chemical characteristics ref Table 2 and trace elements ref Table 3. of pond
ash analyzed by XRF, the results showed that the major constituent in the pond
ash samples is Si, Ca, Al, K, Na, Mg, S, Ti and P as prominent elements in the
KD-1
KD-2 KD-3
KD-4
KD-5
21.24 KD-21
KD-11
KD-10
Latitude
KD-9
KD-23 KD-20
KD-19
21.22 KD-7 KD-22 KD-15 KD-18
KD-8 KD-6 KD-17
KD-12
21.2 KD-13
Pre-10
Fig. 2 Water table (amsl-m) contour plot of pre monsoon season, 2010
Groundwater Quality Assessment in an Around Thermal Power … 967
form of oxides, silicates and alumino-silicates. A result indicates that the pond ash
comes into the Class F category depending on the percentage of SiO2 (53.38%)
followed by Al2 O3 (25.70%) and Fe2 O3 (5.32%) respectively. The concentration of
Mg (2800 mg/kg) is increased in the respective pond ash followed by P (850 mg/kg),
Mn (310 mg/kg) and Cr (220 mg/kg).
968 V. U. Deshmukh et al.
Fig. 4 Piper diagram of pre-monsoon season, 2010. (Green and blue colour represent dug wells
and hand pumps respectively)
5 Conclusion
Combustion of coals in thermal power plants is one of the major sources of envi-
ronmental pollution due to generation of huge amount of ashes which are disposed
of in large ponds in the vicinity of thermal power plants. During the study period
it was observed that parameters namely TDS, Sulfate and fluoride concentration
were significant in few samples. The major concern is about SO4 concentration
which increased in the few samples in the downstream of ash ponds. The severity of
fluoride concentration with respect to groundwater quality is also alarming around
the ash ponds. It is desired that regular monitoring in the critical season i.e. pre-
monsoon season need to be carried out to generate long term data for trend analysis
and compliance measures.
References
1. Shiklomanov IA, Rodda JC (2003) World water resources at the beginning of the twenty-first
century, UNESCO/IHP int. Cambridge University Press, Hydrology Series, Cambridge, UK,
p 2003
2. Deshmukh AN, Shah KC, Shrivastava BN (1994) Impact of rainy season (monsoon) on fly
ash dispersal—a case study of Koradi thermal power plant. Maharashtra Gondwana Geol Mag
8:1–17
3. Theis TL, Westrick JD, Hsu CL, Marley JJ (1978) Field investigations of trace metals in
groundwater from fly ash disposal. J Water Pollut Control Fed 50:2457–2469
4. Theis TL, Richter RO (1979) Chemical speciation of heavy metals in power plant ash pond
leachate. Environ Sci Technol 13:219–224
5. Theis TL, Gardner KH (1990) Environmental assessment of ash disposal. Crit Rev Environ
Control 20:21–42
6. BIS (1991) Drinking water specifications IS: 10500. Bureau of Indian standards, New Delhi
7. Central Ground Water Board (CGWB) (2009) Report on district groundwater management
studies in parts of Nagpur district, Maharashtra. Central Region, Nagpur
8. Suryanarayana K (1965) A new lower gondwana basin between Nagpur and Kampee. Curr Sci
34:610–621
9. Suryanarayana K (1968) The nara-pawangaon extension of the bhokara basin between Nagpur
and Kamptee. Ind Geo Sci Asso, Hyderabad 8:67–78
10. APHA (2005) Standard methods for analysis of water and waste water, 21st (edn) Washington,
DC, 2005
11. Deshmukh AN, Shah KC, Appulingam S (1995) Coal Ash: a source of fluoride contamination,
a case study of koradi thermal power station, district Nagpur, Maharashtra. Gondwana Geol
Mag. 9:21–29
Micro, Nano Manufacturing, Fabrication
and Related Applications
XML Based Feature Extraction
and Process Sequence Selection
in a CAPP System for Micro Parts
Abstract In this work, a process sequence module in a CAPP system using feature-
based modeling (FBM) for prismatic and axis-symmetric parts is developed. It
extracts feature information in Extensible Markup Language (XML) format and
thereon generates the process sequence for efficient manufacturing of the micro
parts. The system consists of two components: (1) development of FBM through
design by feature (DBF) approach and automatically extract the feature information
and verify micro machine tool capability feature details and (2) Process sequence
determination to produce the micro features in prismatic and axis-symmetric micro
parts using a knowledge-based system (KBS) approach. A CAPP system for pris-
matic and axis-symmetric parts can be realized only with the incorporation of many
micro features and the development of other activities in process planning.
1 Introduction
To manufacture the micro parts, a set of instructions that involve design specification
are given and this task is termed process planning. CAPP is the tie between CAD
(Computer-aided design) and CAM (computer-aided manufacturing) as it furnishes
the plan of the process to be used in producing a designed part. Normally, CAPP
incorporates various activities depending on the sophistication of the tool, such as
feature extraction from the part model, tool, fixtures and machine selection, process
and sequence determination, set up planning, and machining parameters. This work
aims to employ a feature-based modeling approach (FBM) to extract features in
XML format and sequence the selected process parameters using object-oriented
programming and a knowledge-based system (KBS) approach for prismatic and
axis-symmetry parts [1, 2].
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 973
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_97
974 G. Gogulraj and S. P. Leokumar
2 Review of Literature
Process planning translates the design information of a product into process instruc-
tions for effective manufacturing. With the development of CAD and CAM systems,
process planning evolved into CAPP systems that can act as a bridge. While the latter
creates a process plan of a new part based on the analysis of the geometry, material,
and other miscellaneous factors that affect manufacturing [3, 4]. Among the several
activities of CAPP, feature recognition is considered to be the dreariest job as it fails to
comprehend the geometrical specification of a part drawing given by the CAD system
in terms of its engineering meaning. It started in the late 1990s, an object-oriented
approach to interface CAD and CAPP systems developed. Further on, FBM and
based models have been developed and much research has been contributed towards
their development. FBM is a model that allows the designer to specify features in
engineering terms rather than geometric terms and besides can store non-graphic
information [5]. Sivankar et al. [6] developed an intelligent CAPP system based on
FBM and neural networks (NN) which recommends appropriate manufacturing oper-
ations and their sequence. The most commonly used FBM approaches are designed
by feature (DBF) and standard for the exchange of product model data (STEP). STEP
based feature modeling has been presented for prismatic parts [7]. A high-level 3D
solid model was used as a base for the part of the design and the designed part was
extracted in XML format. It is regarded that XML-based feature extraction ensures
trouble-free integration with the decision-making system [8–10].
3 Methodology
To successfully integrate CAD with CAPP, a component model that contains infor-
mation more than just topology and geometry is essential. High-level information
such as a selection of materials, tolerance, surface finish, etc. is limited in the
existing CAD system. Both high and low-level process parameters are extracted
by the pre-processor inbuilt in modeling software and are converted to XML data.
By converting process parameters into XML complex feature recognition processes
can be neglected, which is one of the advantages of this work. The feature extraction
is automated by the object-oriented approach developed in our previous work [11].
XML files are converted to an appropriate input format to be fed to the process
planning task by the system-designed post-processor. Feature and model information
is automatically extracted from the XML file using an algorithm developed with an
object-oriented approach and the extracted information is sent to the post-processor.
The function of the post-processor is to convert the extracted features and model
parameters to a format suitable for verification. This is done by randomly searching
for the features and model parameters in the XML file using a developed algorithm.
The signal to the post-processor to extract parameters from the XML file is given
by selecting the extract option shown in the user interface. Complex features are
eliminated by an intelligent and interactive system developed as shown in Fig. 4. A
part CAD model is an elementary input for a CAM system. For the layout, details
have been introduced, the definition of the XML schema and an algorithm has been
developed to extract function and model parameters as shown in Figs. 1, 2 and 3.
In addition to extracting features, the proposed system also performs verification
of micro machine tool capability feature detail in an extract XML file in Fig. 2.
It offers expert advice when feedback is beyond the capabilities of the machine
tool. It contains the proposed methodology and the pseudo-code for the extraction
and verification of the part feature. When confirmed, Process plan generation and
process parameter generation feature and model information further used for decision
making.
Fig. 1 Methodology for feature extraction and verification for micro machine tool capacity from
XML files
6 Implementation
Table 1 Different micro-machining operations and their process sequence for prismatic parts
1. Prismatic parts
Sl.no Sequence Sequence Justification
1 1. End milling-(Roughing) The open-source operation of the end
2. End milling-(Finishing) milling operation is initially performed
2 1. End milling-(Roughing) Open source end milling is initially
2. End milling-(Finishing) performed, followed by closed source
3. Micro-drilling-(Roughing) drilling. Open source slot milling is done
4. Micro-drilling-(Finishing) first to avoid burr deposition in the drilled
hole
3 1.Ball End milling-(Roughing) Open source milling is initially done to
2.Ball End milling-(Finishing) form a chamfer at the edge. To avoid burr
3.End milling-(Roughing) deposition in the machined part, slot
4.End milling-(Finishing) milling is performed first, followed by
5.Micro-drilling-(Roughing) sequential drilling by hole diameter (from
(For different hole sizes in Cad model larger to smaller diameter)
operation sequence = D1 > D2 > D3 > D4)
6.Micro-drilling-(Finishing)
and (4) user interface [Figs. 1 and 2]. Based on the decision made by the decision-
making environment, the process sequences are generated step by step manner. The
decision-making environment receives the input from the user about the micro-
machining process through the user interface. Here, the visual studio environment
plays a vital role in providing the decision-making environment to select the process
sequencing required for the part.
The developed system has a menu bar where feature extraction and plan gener-
ation are available. Once selected, a pop-up window will generate a red color label
and feature name and add-on dimension limits exceed for feature extraction and
on selecting the extraction option the feature extraction takes place automatically.
The design of the system will not allow the user to input values that are beyond
the specification of the micro-machine tool and if the value is beyond the specifi-
cation the system will give expert advice for alternate options as shown in Fig. 3.
The features that are highlighted in red color and feature name to add on dimension
limits exceeded are beyond the micro-machine tool specification, so from this, the
user can identify that the values have to be modified accordingly.
In the process sequence, the user will be allowed to select the order in which
the sequences are to be executed. The process sequence module will be carried out
following the material of the part and surface finish. After the sequence is set, the tool
and fixture module will select the respective tools and fixtures required concerning
the process sequence module and process parameter module by the material and
geometry of the part. After all the parameters are fixed the developed system will
generate the process plan output in a suitable format for low-level manufacturing as
shown in Fig. 4.
XML Based Feature Extraction and Process Sequence Selection … 979
7 Conclusion
Table 2 Different micro-machining operations and their process sequence for Axis-Symmetry
parts
2. Axis-Symmetry parts
Sl.no Sequence Sequence Justification
1 1. Facing A facing and plain turning
2. Plain turning-(Roughing) operation are initially performed
3. Plain turning-(Finishing) to reduce the extra material size,
4. Step turning-(Roughing) followed by a gradual step-by-step
5. Step turning-(Finishing) turning operation with different
cutting depths
2 1. Facing Facing and plain turning
2. Plain turning-(Roughing) operations are carried out
3. Plain turning-(Finishing) sequentially followed by step
4. Step turning-(Roughing) turning with varying depth of
5. Step turning-(Finishing) cutting. Micro-drilling is
6. Micro-drilling-(Roughing) performed as a final operation to
7. Micro-drilling-(Finishing) facilitate a change of plane (as the
workpiece must be removed from
the spindle and placed on the vice)
3 1. Facing Facing and plain turning
2. Plain turning-(Roughing) operations are carried out
3. Plain turning-(Finishing) sequentially followed by step
4. Step turning-(Roughing) turning with varying depth of
5. Step turning-(Finishing) cutting and taper turning with the
6. Taper turning-(Roughing) required dimension. Micro-is
7. Taper turning-(Finishing) performed as a final operation to
8. Micro-drilling-(Roughing) facilitate a change of plane (as the
9. Micro-drilling-(Finishing) workpiece must be removed from
the spindle and placed on the vice)
4 1.Facing Facing and plain turning
2.Plain turning-(Roughing) operations are carried out
3.Plain turning-(finishing) sequentially followed by step
4.Step turning-(Roughing) turning with varying depth of
5.Step turning-(Finishing) cutting. Grooving is performed to
6.Grooving-(Multi-pass) reduce the diameter of a particular
7.Micro-drilling-(Roughing) region. Micro-is performed as a
8.Micro-drilling-(Finishing) final operation to facilitate a
change of plane (as the workpiece
must be removed from the spindle
and placed on the vice)
(continued)
980 G. Gogulraj and S. P. Leokumar
Table 2 (continued)
2. Axis-Symmetry parts
Sl.no Sequence Sequence Justification
5 1. Facing Facing and plain turning
2. Plain turning-(Roughing) operations are carried out
3. Plain turning-(Finishing) sequentially followed by step
4. Step turning-(Roughing) turning with varying depth of
5. Step turning-(Finishing) cutting and taper turning with the
6. Taper turning-(Roughing) required dimension. Grooving is
7. Taper turning-(Finishing) performed to reduce the diameter
8. Grooving-(Multi-pass) of a particular region. Micro-is
9. Micro-drilling-(Roughing) performed as a final operation to
10. Micro-drilling-(Finishing) facilitate a change of plane (as the
workpiece must be removed from
the spindle and placed on the vice)
6 1. Facing Facing and plain turning
2. Plain turning-(Roughing) operations are carried out
3. Plain turning-(Finishing) sequentially followed by step
4. Step turning-(Roughing) turning with varying depth of
5. Step turning-(Finishing) cutting. Grooving is done next to
6. Grooving-(Multi-pass) reduce the diameter of a particular
7. Micro-slot-(Roughing) region, followed by a
8. Micro-slot-(Finishing) micro-milling operation to avoid
9. Micro-drilling-(Roughing) burr deposition in micro holes.
10. Micro-drilling-(Finishing) Micro-drilling is performed as a
final operation to facilitate a
change of plane (as the workpiece
is to be removed from the spindle
and placed on the vice)
7 1. Facing Facing and plain turning
2. Plain turning-(Roughing) operations are carried out
3. Plain turning-(Finishing) sequentially followed by step
4. Step turning-(Roughing) turning with varying depth of
5. Step turning-(Finishing) cutting and taper turning with the
6. Taper turning-(Roughing) required dimension. Grooving is
7. Taper turning-(Finishing) done next to reduce the diameter
8. Grooving-(Multi-pass) of a particular region, followed by
9. Micro-slot-(Roughing) a micro-milling operation to avoid
10. Micro-slot-(Finishing) burr deposition in micro holes.
11. Micro-drilling-(Roughing) Micro-drilling is performed as a
12. Micro-drilling-(Finishing) final operation to facilitate a
change of plane (as the workpiece
is to be removed from the spindle
and placed on the vice)
XML Based Feature Extraction and Process Sequence Selection … 981
systems. Through heuristic methods, the database for the sequence of the process to
be followed for each operation was developed for both the part models under study.
The visual Studio environment was used to integrate the two components namely
feature extraction and process sequencing through the user interface.
The system developed is interactive, intelligent, user friendly, and can be extended.
Nevertheless, the present system is proposed for prismatic parts and axis-symmetric
parts that possess features like micro mills, turns, and grooves. With the addition
of various micro features and incorporation of other CAPP activities like parameter
selection, tool and fixture selection, setup selection, and automatic generation of NC
part program for the available model information, a complete CAPP system can be
realized in the future for prismatic and axis-symmetric parts.
References
1. Rahman M, Senthil Kumar A, Prakash JRS (2001) Micro milling of pure copper. J Mater
Process Technol 116:39–43
2. Shunmugam MS, Mahesh P, Bhaskara Reddy SV (2002) A method of preliminary planning for
rotational components with c-axis features using genetic algorithm. Comput Indus 48(3):199–
217
3. Amaitik SM, Kiliç SE (2005) STEP-based feature modeller for computer-aided process
planning. Int J Prod Res 43(15):3087–3101
982 G. Gogulraj and S. P. Leokumar
4. Asad ABMA, Masaki T, Rahman M, Lim HS, Wong YS (2007) Tool-based micro-machining.
J Mater Process Technol 192–193:204–211
5. Lai X, Li H, Li C, Lin Z, Ni J (2008) Modelling and analysis of micro-scale milling considering
size effect, micro cutter edge radius, and minimum chip thickness. Int J Mach Tools Manuf
48:1–14
6. Sivasankar R, Asokan P, Prabhakaran G, Phani AV (2008) A CAPP framework with optimized
process parameters for rotational components. Int J Prod Res 46(20):5561–5587
7. Amaitik SM (2013) Fuzzy logic models for selection of machining parameters in capp systems.
Int J Comput Inf Technol 2(2):279–285
8. Leo Kumar SP, Jerald J, Kumanan S (2013) “An expert system for cutting tool selection
for micromachining processes”. International conference on precision, meso, micro and nano
engineering, Calicut, pp 133–139
9. Leo Kumar SP, Jerald J, Kumanan S (2014) An intelligent process planning system for micro
turn-mill parts. Int J Prod Res 52:6052–6075
10. Leo Kumar SP, Jerald J, Kumanan S (2015) Feature-based modelling and process parameters
selection in a capp system for prismatic micro parts. Int J Comput Integr Manuf 28:1046–1062
11. Leo Kumar SP, Jerald J, Kumanan S, Prabakaran R (2017) A review on current research aspects
in tool-based micromachining processes. Mater Manuf Process 29:1291–1337
To Study and Optimize the Effects
of Process Parameters on Taper Angle
of Stainless Steel by Using Abrasive
Water Jet Machining
Nomenclature
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 983
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_98
984 M. K. Gawade and V. S. Jatti
1 Introduction
The material used is iron based alloy, which is manufactured commercially having
grade type 304 stainless steel sheet with dimension 330 mm × 100 mm × 6 mm. Its
yield strength is 205 MPa, tensile strength is 515 MPa, hardness 92 HRB. Chemical
composition of type 304 stainless steel specification is shown in Table 1.
Experiment was performed on Water Jet German—3015 Machining facilitated
with CNC equipment. Table 2 shows process machining parameters. The experiments
were carried under water pressure 3100 bar and the abrasive material was garnet 80
mesh size. Figure 1 shows the abrasive water jet machine.
The DOE is a systematized technique to establish the correlation between control
factors and final response. Table 3 shows the 3 control factors levels.
RSM (response surface methodology) is a statistical method used to find the
relationship of control factors on response process. RSM is used to design with
Table 2 Parameters of
Sr No. Parameters Values
AWJM machine
1 Diameter of nozzle 1.1 mm
2 Orifice diameter 0.35 mm
3 Water pressure 2500–3100 bar
4 Abrasive flow rate 100–300 g/min
5 Traverse speed 50–200 mm/min
minimum number of experiments performed and also used to obtain optimum condi-
tions to produce desirable responses. Box-Behnken design (BBD) model were used
for 3 level and 3 factors based on the experiments were performed. The relationship
between the independent variables and response variables [4] is given in Eq. 1.
k
k
k
Y = β0 + βi X i + βii X ii2 + βi j X i X j (1)
i=1 i=1 i, j=1,i= j
Wt − Wb
Taper angle(θ ) = tan−1 ( )degree (2)
2∗t
where, Wt is width cut at top in mm, Wb is width cut at bottom in mm, t- thickness
in mm.
From Eq. (3), it has been observed that parameter SOD and TS have positive
effects and AFR has the negative effect on taper angle.
Further, to check the most significant parameter affecting on taper angle is deter-
mined by ANOVA (Analysis of Variance) about 95% confidence interval (CI). It is
a computational method helps us to evaluate the significance of every control factor
on response factor.
Table 5 shows the ANOVA results. The F value is the fisher’s statistical test and
P value is the probability of significance acts in accordance with coefficient of R-sq
and R-sq (adj) is determined. This coefficient implies the acceptability and suitability
of the model. It is observed that, process variables having lager F value and P value
less than 0.05 implies that the variable is significant. For the value of AFR, P value
less than 0.05 implies that the variable is significant statistically. The value of R-sq
is 89.47% and R-sq (adj) is 70.53% for taper angle implies accuracy and fitness of
model. Higher the value of R-sq satisfies the accuracy and fitness of model.
For better analysis, we have plotted the graphs of probability and residuals for
taper angle of stainless steel material. The graph of normal probability shows that
residual are distributed normally have a close fit to line. Versus fit graph implies
there is randomly distribution of residuals. Versus order graph implies observation
a constant variance. This graphs shows clarity that the observation are reliable and
has 95% confidence interval (CI). Figure 3 shows the residual graph.
Figure 4 shows the effects of various process parameters on taper angle. It is
observed that there is rise in taper angle as traverse rate increases this is because as
traverse speed increases cutting action gets reduced at bottom width of cut this leads
(a)
(b)
(c)
(d)
To Study and Optimize the Effects of Process Parameters … 989
to increases in the value of taper angle. It is observed that as flow rate of abrasives
increases there is decrement in taper angle. Also, as stand-off distance increases there
is increment in taper angle is due to increase in focus area of jet results increase in
width of cut.
Figure 5 shows the surface plots, to find the effect of different control factors,
surface plots are also generated. Figure 5a shows interaction between AFR and SOD
keeping value constant of TS at 120 mm/min. It is noticed that for lower value of SOD
and higher value of AFR there is decrement in angle of taper and similarly for higher
SOD and lower AFR there is rise in angle of taper, this is due to abrasive particles
which cannot pierce sufficiently inside the specimen material [6]. Figure 5b shows
the interaction between SOD and TS keeping value constant AFR at 200 g/min. As
there is increase in TS and SOD as there is rise in angle of taper. Similarly, at lower
SOD and TS there is fall of angle of taper. Figure 5c shows that at higher TS and
lower of AFR there is rise in the angle of taper keeping the value constant of SOD
at 2 mm. Similarly, at the higher value of AFR and lower TS there is decrement in
angle of angle of taper.
4 RSM Optimization
(a) (b)
(c)
Table 6 Constraints of
Condition Goal Lower limit Upper limit
parameters for optimization
Traverse speed, A In range 80 160
Abrasive flow rate, B In range 100 300
Stand-off distance, C In range 1 3
Taper angle (degree) Minimum 1.41 1.70
of calculating weights and individual desirability for every response. Higher the value
of‘d’ implies higher desirables optimum values obtained for response parameter [6].
RSM optimization was performed for input parameters, multi-objective optimiza-
tion were used for response parameter. Figure 6 shows plot for response optimization.
It implies the composite desirability of 0.9195 for optimizing the taper angle. The
desirability value 0.9195 indicates acceptance of the response parameter. Optimal
values obtained TS 80 mm/min, AFR 300 gm/min, and SOD of 1 mm, so as to
achieve minimum value of angle of taper.
To Study and Optimize the Effects of Process Parameters … 991
5 Confirmation Analysis
Confirmation analysis is the analysis is used to analyze the accuracy of the model
performed by RSM. Table 7 is referred to plot the value of experiment and predicted
results. Percentage error is calculated between experimental and predicted results. It
is observed that percentage error is less than 6% within the limit of permissible. The
predicted values of RSM model show a high level of accuracy.
Figure 7 shows the difference between experimental results and modeling results.
It can be concluded to get low value of the angle of taper for the input parameters
such as TR, AFR and SOD the obtained RSM model is fulfilled.
6 Conclusions
The analysis reveals the impact of input parameters on angle of taper of stainless
steel material and thus we conclude as follows—
(i) The mathematical model is formulated for the taper angle showed the corre-
lation between the flow rate of abrasive, traverse speed and SOD. The most
significant parameter is AFR influencing on angle of taper significant statis-
tically followed by TS and SOD. The value of R-sq is 89.47% and R-sq (adj)
992 M. K. Gawade and V. S. Jatti
is 70.53% for taper angle implies accuracy and fitness of model by ANOVA
technique
(ii) Main plots and 3D surface plots were studied for the influence of control
parameters on angle of taper. AFR is comparably slighter influential on angle
of taper.
(iii) Optimization of process parameter using multi-objective RSM for minimum
taper angle which results the desirability of 0.9195. Optimal values obtained
To Study and Optimize the Effects of Process Parameters … 993
References
1. Srinivas S (2018) Machinability Studies on Stainless steel by abrasive water jet—Review. Mater
Today: Proc 57:2871–2876
2. Khan AA, Hague MM (2007) Performance of different abrasive material during abrasive water
jet machining of glass. J Mater Process Technol 41:404–407
3. Gupta V, Pandey PM, Garg M, Khannaand R, Batra NK (2014) Minimization of kerf taper angle
and kerf width using Taguchi’s method in abrasive water jet machining of marble. Proc Mater
Sci 6:140–149
4. Jagadish Bhowmik S, Ray A (2016) Prediction and optimization of process parameters of green
composites in AWJM process using response surface methodology. In: Int J Adv Manuf Technol,
Springer, London
5. Hascalik A, Aydas Guru UC (2007) Effect of traverse speed on abrasive water jet machining of
Ti–6Al–4 V alloy. Mater Des 18:1953–1957
6. Dumbhare P, Dubey S, Deshpande Y, Andhare A, Barve P (2018) Modelling and multi-objective
optimization of surface roughness and kerf taper angle in abrasive water jet machining of steel.
J Brazilian Soc Mech Sci Eng 40:48–55
Effect of Process Parameters on Response
Measures of Cartridge Brass Material
in Photo Chemical Machining
Abstract Now a days to produce stress free and burr free micro components, photo-
chemical machining is one of the emerging technologies. The micro level etching of
components of different materials is carried out by using chemical etching process.
The very precision parts such as microchannels, heat sinks, Fine mesh or screens
and printed circuit boards are developed by this process. In this work, the cartridge
brass was selected as a base metal which has good electrical and thermal conduc-
tivity. The process parameters are etching temperature and time and responses are
etching depth and surface roughness. The objective of this study is to achieve etching
depth on the Cartridge brass plate having thickness 0.5 mm. The etching depth and
surface roughness of material increases with increase in time and temperature. The
study of copper and brass were reported by many researchers. Very few literatures
are available on an alloy of copper and brass. The effect of process parameters is
changing by changing composition of material. This article is focused on the study
of etching depth and surface roughness of cartridge brass material at different time
and temperature. The cartridge brass has been widely used in industrial applications.
Keywords Photo chemical machining · Photo tool · Etch depth · Photo resist
1 Introduction
The method also named as chemical etching, acid etching, Photochemical milling.
The photoresist used PCM has two types given below.
• Positive photoresist—The part of UV exposure becomes soluble in developing
solution.
• Negative photoresist—The part of UV exposure becomes insoluble in developing
solution.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 995
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_99
996 Bandu. A. Kamble et al.
The method starts with generating a photo tool i.e. the design created by using
drawing softwares like AutoCAD, Catia etc. and printing it on trace paper. The nega-
tive photoresist solution is applied on a base metal at uniform thickness of coating.
The photo tool is kept on base metal and the surface is exposed by using ultraviolet
light. The transferred pattern of design will be visible after the developing process. In
the final step of method, the base metal is dipped into the etchant solution for etching
or machining. Davis et al. [1] developed a process to etch micro-fluidic channels
and other micro components. Wangikar et al. [2] optimized process parameters of
brass and German silver and concluded with brass has a better etching performance.
Agrawal et al. [3] creates microfluidic channels, micro holes by using the PCM
process. The manufacturing of such parts by using cutting tools are very challenging
work and many defects will be occurring. Jang et al. [4] and Sadaih et al. [5] devel-
oped a 3D system in the package (SIP) process. It has been considered as a superb
microelectronics packaging system. The selection of etchant for machining of parts
is depending on the surface topology and hardness of material. The surface finishing
of the etched part is varying with etchant to etchant.
The major steps involved on 3D PCM is given below
• Creation of photo tool
• Selection of material for etching
• Modelling of workpiece
• Film coating using photoresist
• Developing
• Etching
• Stripping and inspection.
Misal et al. [6, 7] shows the surface topography of Inconel 718 and responses
are measured at different conditions and checked effect of process parameters on
response measure for hard materials like Monel and Inconel 718. Saraf et al. [8]
presented a mathematical model of PCM and optimized conditions of PCM process.
Cakir et al. [9] etched Cartridge brass by cupric chloride and suggested a suitable
process for regenerated waste treatment. Allen D.M et al. [10] demonstrate the three-
dimensional etching process and gave characterisation of ferric chloride etchant.
the development of three-dimension components is possible by using lamination
techniques. Bruzzone et al. [11] generated simulation model of 2D photochemical
machining process. The simulation was carried out in six steps. An analysis of the
PCM process was carried out at differentconditions. Kamble et al. [12] shows variable
etching process for different colours. The depth of etching is changing by changing
UV light exposing at different intensity of light. The light passing intensity is different
for different colours due to this hardening of photoresist is different for different inten-
sity of light or colour. The more and less cross linking of photoresist results in uneven
etching of base metal. Jadhav et al. [13] have focused the etching depth variation
of brass material for different operating conditions. The parametric study of process
was reported. Mazarbhuiya et al. [14] studied PCM process for Aluminium material.
The effect of different parameters such as etching time, temperature and concentra-
tion were taken into consideration to measure responses like depth, roughness and
Effect of Process Parameters on Response Measures of Cartridge … 997
edge deviation. Manoharan et al. [15] compared the PCM process to binder jetting
to develop micro post arrays. PCM found the cheapest process for developing micro
arrays. Further, the study related to photochemical machining on copper, brass, etc.
for producing microchannels, micro textures in bearing, heat sinks, etc. were reported
by many researchers [16–25]. Very few literatures are available on an alloy of copper
and brass. The study of PCM process for copper and Brass alloys has wide scope. In
this work, the etching behaviour of Cartridge Brass was analysed.
The flowchart for PCM is presented in Fig. 1, which explains the different steps
in PCM. The cartridge brass was selected for study which has good electrical and
thermal conductivity and it is widely used in industrial applications. The selected
material for etching has been cut into a size of 40 mm × 40 mm × 0.6 mm (L × B×
T). The design was created by using drawing softwares like AutoCAD, Catia etc.
and printed it on trace paper. The 2400 × 2400 dpi printer was used to print photo
tool as shown in Fig. 2a. The accuracy of PCM is mainly depending on the quality
of printing images. The thickness of the photo tool is 80 microns.
The selected metal is cleaned by using paper or cleaning solution i.e. thinner
to remove all dust particles on it. After cleaning of base metal, in the next step
negative type photoresist solution applied on metal by using spin coating method.
The uniform coating thickness was maintained throughout the coating process. After
Cleaning
Photoresist coating
UV light exposing
Developing
Etching
Stripping
998 Bandu. A. Kamble et al.
a) b) c)
d)
that, the coating is baked for 2 min for better adhesion between photoresist and brass
material. The high intensity UV light is used to expose coating on material at 90 s. In
the exposing process, the UV light is passed through white color and restricts through
black color, the exposed area is hardened (white) and unexposed area becomes soft
(Black). After exposing, the sample is developed in a developing solution (Mixture of
sodium carbonate and water) at time 90 s. The hardened area of photoresist remains
on base metal and unexposed area is washed out during the development process.
The development process required to transfer required design onto a coating surface
which is open to the atmosphere. The parts open to the atmosphere are etched in
a chemical etching solution (mixture of ferric chloride and water. After the etching
process, the remaining photoresist is washed out by using a striping solution (sodium
hydroxide).
The etching process setup is shown in Fig. 2c.
3 Design of Experiments
The numbers of preliminary experiments were carried out to study the etching mech-
anism of Cartridge brass material. The process parameters are etching time, temper-
ature etc. and response parameters are etching depth and surface roughness. The
etching depth was measured by using 0.001 L.C Digital Micrometre. The concen-
tration of etchant is constant throughout the experiments which is 40 Baume scale
measured by hydrometer. The level of experiments is selected by studying the etching
Effect of Process Parameters on Response Measures of Cartridge … 999
The performance of photo chemical machining of Cartridge brass has been analyzed
at different time and temperature. The etching depth of each specimen was recorded
at three locations, and mean value is recorded. The etching depth of etching material
increases by changing process parameters such as etching temperature and etching
time. The depth of etching increases by increase in etching time and temperature. At
the low temperature, the depth of etch is low due to i.e. less molecules are available
to etching reaction i.e. less collision of molecules on the Cartridge brass surface. As
temperature increases the collisions in etchant molecules is increasing, it results in
higher depth of etch.
The depth of etching material is changed by changing process parameters such
as etching temperature and etching time. Figures 3 and 4 show the relation between
etching depth and process parameters. The depth of etching increases by increase in
1000 Bandu. A. Kamble et al.
300
250
200
150
100
50
0
3 6 9
Etching time (Min)
43 46 49
0.5
0.4
0.3
0.2
0.1
0
3 6 9
Time (Min)
43 46 49
etching time and temperature. The depth of etching and roughness at low temperature
i.e. for 43 °C is low at time 2 min and it increases with time its highest at time 6 min
which is 80 microns. Similarly, for other temperatures, the depth of etching is highest
at higher temperature and time.
Effect of Process Parameters on Response Measures of Cartridge … 1001
The surface roughness of Cartridge brass material is higher at highest etching temper-
ature and time. The value of roughness changes by change in process parameters, it
increases by increase in temperature and time. The below fig shows the roughness
of cartridge material over a time and temperature. The surface roughness increases
from 0.312 to 0.555 microns. The increase in roughness caused by uneven etching of
surface as time and temperature increases. The energy of etchant molecules depends
on concentration and temperature of etching solution.
The reaction between etchant species on the surface increases with time which
gives maximum etch rate and depth of etch. As time increases, the uneven collisions
of etchant species increases which results in higher roughness value. The reaction of
or etch rate of material is depending on the different properties of material such as
hardness, conductivity and composition of material. The less etch rate obtained for
harder materials. Depending on the characteristic of material, the chemical solution
will be prepared. Figure 4 shows the relationship between surface roughness, temper-
ature and etching time. The surface roughness is higher i.e. 555 µm at temperature
49 °C and time 9 min.
1002 Bandu. A. Kamble et al.
5 Conclusions
The present study evaluate etching mechanism of cartridge brass material at different
temperatures and time. In this study, the effect of process parameters like etching
temperature and time on an etching depth has been studied. The temperature has a
significant effect on etching depth and surface roughness. The higher temperature
results in more etching depth and surface roughness, due to the rate collision of
etchant species on the surface of Cartridge brass is more at higher temperature.
Similarly at low temperature, the etchant molecules are less reactive and less in
number which gives less depth of etch. Depth of etching and roughness goes on
increasing by increase in temperature and etching time. The more etching depth and
surface roughness obtained at 49 °C temperature i.e. 371 microns and 0.555 microns
in 9 min. The lowest depth and roughness obtained at 40 °C temperature i.e. 20
microns and 0.312 micronsin 3 min.
Acknowledgements Authors are extending thanks to the Principal Collaborator Dr. Balasubrama-
niam (BARC) for providing facilities for experimentation funded by BRNS (DAE).
References
1. Davis PJ, Overturf GE III (1986) Chemical machining as a precision material removal process.
Precis Eng 8(2):67–71
2. Wangikar SS, Patowari PK, Rahul DM (2017) Effect of process parameters and optimization for
photochemical machining of brass and german silver. Mater Manuf Process 32(15):1747–1755
3. Agrawal D, Kamble D (2019) Optimization of photochemical machining process parameters
for manufacturing microfluidic channel. Mater Manuf Process 34(1):1–7
4. Jang DM, Ryu C, Lee KY, Cho BH, Kim J, Oh TS, Yu J et al (May 2007) Development
and evaluation of 3-D SiP with vertically interconnected through silicon vias (TSV). In 2007
proceedings 57th electronic components and technology conference. IEEE, pp 847–852
5. Sadaiah M, Patil DH (8–12 Jun 2015) Some investigations on surface texturing on monel 400
using photochemical machining. In proceedings of the ASME 2015 international manufacturing
science and engineering conference MSEC2015, Charlotte, North Carolina, USA. https://doi.
org/10.1115/msec2015-9294
6. Misal ND, Sadaiah M (2017) “Investigation on surface roughness of inconel 718 in photo
chemical machining”. Adv Mater Sci Eng
7. Misal ND, Saraf AR, Sadaiah M (2017) Experimental investigation of surface topography in
photo chemical machining of Inconel 718. Mater Manuf Process 32(15):1756–1763
8. Saraf AR, Misal ND, Sadaiah M (2012) Mathematical modelling and optimization of photo
chemical machining. Adv Mater Res 548:617–622
9. Cakir O (2006) Copper etching with cupric chloride and regeneration of waste etchant. J Mater
Process Technol 175:63–68
10. Allen DM, Almond HJ (2004) Characterisation of aqueous ferric chloride etchants used in
industrial photochemical machining. J Mater Process Technol 149(1–3):238–245
11. Bruzzone AAG, Reverberi AP (2010) An experimental evaluation of an etching simulation
model for photochemical machining. CIRP Ann 59(1):255–258
12. Kamble B, Utpat A, Misal ND, Ronge BP (2019) 3D Photochemical machining of copper
by using colored phototools. Int J New Technol Res (IJNTR) ISSN: 2454–4116, July 2019,
5(7):28–32
Effect of Process Parameters on Response Measures of Cartridge … 1003
13. Jadhav Saurabh M, Karatkar Onkar V, Bangale Kamesh N, Choudhari Deepak B, Utpat AA,
Kamble Banduraj K (2019) Etching depth variation of brass material for different operating
conditions. Int J New Technol Res (IJNTR) ISSN: 2454–4116, April 2019, 5(4):93–96
14. Mazarbhuiya RM, Rahang M (2020) Parametric study of photochemical machining of
aluminium using taguchi approach. Advances in mechanical engineering. Springer, Singapore,
pp 497–504
15. Manoharan S, Summerville S, Freiberg L, Coblyn M, Touma JG, Jovanovic G, Paul BK
(2020) Manufacturing process design of a micro-scale liquid-liquid extractor and multi-phase
separator. J Manuf Process
16. Das SS, Tilekar SD, Wangikar SS, Patowari PK (2017) Numerical and experimental study of
passive fluids mixing in micro-channels of different configurations. Micros Techno 23:5977–
5988
17. Wangikar SS, Patowari PK, Misra RD (2018) Numerical and experimental investigations on
the performance of a serpentine microchannel with semicircular obstacles. Micros Techno
24:3307–3320
18. Gidde RR, Pawar PM, Ronge BP, Shinde AB, Misal ND, Wangikar SS (2019) Flow field analysis
of a passive wavy micromixer with CSAR and ESAR elements. Micros Techno 25:1017–1030
19. Wangikar SS, Patowari PK, Misra RD (2016) Parametric optimization for photochemical
machining of copper using grey relational method. In 1st techno-societal, international
conference on advanced technologies for societal applications. Springer, pp. 933–943
20. Wangikar SS, Patowari PK, Misra RD (2018) Parametric optimization for photochemical
machining of copper using overall evaluation criteria. Mater Tod Proceed 5:4736–4742
21. Wangikar SS, Patowari PK, Misra RD, Misal ND (2019) Photochemical machining: a less
explored non-conventional machining process. In: Non-conventional machining in modern
manufacturing systems, IGI Global, pp 188–201
22. Kulkarni HD, Rasal AB, Bidkar OH, Mali VH, Atkale SA, Wangikar SS, Shinde AB (2019)
Fabrication of micro-textures on conical shape hydrodynamic journal bearing. Int J Tre Engg
Techno 36:37–41
23. Raut MA, Kale SS, Pangavkar PV, Shinde SJ, Wangikar SS, Jadhav SV, Kashid DT (2019)
Fabrication of micro channel heat sink by using photo chemical machining. Int J New Techno
Res 5:72–75
24. Patil PK, Kulkarni AM, Bansode AA, Patil MK, Mulani AA, Wangikar SS (2020) Fabrication
of logos on copper material employing photochemical machining. Nov MIR Res J 5(6):70–73
25. Wangikar SS, Patowari PK, Misra RD, Gidde RR, Bhosale SB, Parkhe AK (2020) Optimization
of photochemical machining process for fabrication of microchannels with obstacles. Mater
Manuf Proces. https://doi.org/10.1080/10426914.2020.1843674
Thermal Performance of Two Phase
Closed Thermosyphon with Acetone
as Working Fluid
1 Introduction
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1005
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_100
1006 S. V. Pawar and A. K. Chandgude
to improve Thermal properties. Kamyar et al. [2] use Al2 O3 and TiSiO4 Nanoparti-
cles suspended in water. He concluded that thermal resistance of TPCT decreases,
evaporator wall temperature decreases and an increase in heat transfer coefficient.
The surface of evaporator and condenser modified to improve Thermal performance
of TPCT. Rahimi et al. [3] analyzes the effect of surface modification on the evapo-
rator and condenser surface. It decreases the thermal resistance of TPCT. Tsai et al.
[4] analyze the TPCT for electronic cooling. He analyzes the effect of parameters
such as fill ratio, evaporator surface structure, heating power on the performance of
TPCT. Brusly Solomon et al. [5] analyzed anodised and non anodised TPCT with
acetone as working fluid. Simple porous surface is prepared by anodising technique.
Anodised surface has a significant effect on evaporator’s thermal resistance and heat
transfer coefficient and the same is negligible on the condenser. Ong and Haider-
E-Alahi [6] filled Refrigerant R-134a in TPCT and analyze for fill ratio, coolant
mass flow rates, the mean temperature difference between evaporator and condenser.
Nano-fluid cause fouling on the internal surface of TPCT. Sarafraz et al. [7] analyzes
the effect of fouling caused by Nano-fluid on an internal surface of TPCT. Fouling
increases with an increase in the mass of Nano-fluid. Fouling causes deterioration in
thermal performance with time. Noie [8] analyzes the TPCT made up of copper tube
filled with distilled water. In his study, three parameters are considered: heat transfer
rates, working fluid filling ratio, evaporator length. Problem with Nanofluid is that it
causes fouling on the internal surface of TPCT. Sarafraz et al. [7] analyze Effect of
fouling [2]. Never mentioned anything about fouling and it’s an effect on heat trans-
mission. Dry out at higher is one of the limitations of TPCT at higher heat input. [4,
6, 8] analyze the effect of filling ratio on the performance of TPCT. Ong and Tong
[9] analyzes the effect filling ratio and inclination. Water is used as a working fluid.
Effect of filling ratio is determined. Jiao and Qui [10] analyzes vertical TPCT filled
with nitrogen. Parameters like filling ratio, heat input, operating pressure, geome-
tries are analyzed. Ordaz-Flores [11] analyze acetone and methanol filled two-phase
flat plate solar system. Experimental result shows that partial vacuumed condition at
starting to improve Thermal performance of phase change system.
From the literature survey, it is clear that different method implemented to improve
the thermal performance of TPCT. The researcher tries a different combination of
pipe material and working fluid. In this study, aluminium and acetone combination
is used and it’s thermal performance evaluated for 30% filling ratio.
2 Experimental Setup
unit are Pump, Control panel, Thermocouple, Rotameter, tilting bracket, heating coil.
TPCT pipe placed in two aluminium block at two ends called as evaporator block and
condenser block. Middle section made up of glass wool is called adiabatic section.
The evaporator is 100 mm block with a slot at the centre for TPCT pipe. Four
holes drilled for heating coils around TPCT slot. Three thermocouples are placed at
equidistance to measure the surface temperature of TPCT. The similar block is used
for a condenser with two openings for water inlet and outlet. Water is used to carry the
heat reject by TPCT pipe. Three thermocouples are placed over the condenser section.
The adiabatic section made up of glass wool so that no heat transfer should take place.
Three number of a thermocouple is placed over there. Two number of a thermocouple
is used for measuring the inlet temperature (T11) and outlet temperature (T12) of
water. One thermocouple measures atmosphere temperature (T10). Rotameter is used
to adjust the flow of cooling water (LPH). Control unit is used to vary the heat input.
Acetone is used as a working fluid. Most of working fluids used have lower latent
heat of vaporization and lower boiling point at atmospheric pressure. But in case
of acetone latent heat of vaporization and boiling point, both are high as compare
to refrigerant like R134a. Also, acetone shows good compatibility with aluminium
metal. Due to the lower amount of working fluid in the pipe, dry out situation may
arise. So to deal with this phenomenon 30/c/o of filling ratio is used in TPCT.
Properties of acetone
Boiling point 329.2 k
The heat of evaporation 534 kJ/kg k
Specific heat capacity, Cp 2.14 kJ/kg k
Specific heat capacity, CV 1.55 kJ/kg k
1008 S. V. Pawar and A. K. Chandgude
3 Results
The acetone filled thermosyphon is analyzed in this paper. The vertical thermosyphon
having 30% filling ratio analyze for increasing heat input.
Heat input can be calculated as,
Qin = VI
η = Qout /Qin
4 Conclusion
with the increase in heat input. The percentage Change in efficiency when heat input
increases from 50 W to 300 W is 17.67%. Thermal performance of TPCT is best at
high heat input.
References
10. Jiao B, Qui LM (2008) Investigation on the effect of filling ratio on steady-state heat transfer
performance of a vertical TPCT. Appl Therm Eng 28:1417–1426
11. Ordaz-Flores A (2012) Finding to improve the performance of a two-phase flat plate solar
system using acetone and methanol as working fluid. Solar Energy 86:1089–1098
Effect of Slip on Vortex Formation Near
Two-Part Cylinder with Same Sign Zeta
Potential in a Plane Microchannel
Abstract We investigate the effect of slip on the formation of recirculation zone near
the two-part cylinder with the same sign zeta potential placed in a microchannel. The
external electric field is used to actuate the electroosmotic flow (EOF). The governing
transport equations are solved using a finite element based numerical solver. The
vortex formation takes place near the upstream part of the cylinder. The strength of
the vortex is analyzed in terms of the maximum magnitude of reversed flow velocity
(UR ). It is found that the extent of the recirculation zone is smaller for the slip case as
compared to the no-slip case. The magnitude of UR increases with the slip coefficient
(β) for smaller values of β. Also there is a decrement in UR at larger values of slip
coefficient and the decrement is amplified at higher values of zeta potential. The flow
rate monotonically increases with the slip coefficient and zeta potential.
1 Introduction
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1013
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_101
1014 S. Pabi et al.
an effort to enhance and control mixing, different methods are employed such as
active and passive mixing techniques. In active mixers, an external energy other than
flow energy is applied to perturb the flow and creates vortices [9–11]. On the other
hand, a passive mixer uses confined geometries in the form of rivets, grooves in the
channel to create vortices and enhance its vorticity for proper mixing [12–14]. In
passive micromixers, no external energy source is applied and it requires only the
pumping energy for flow actuation. On the other hand, active micromixers are more
efficient in micromixing although quite difficult to fabricate [15]. Over the years
researchers have been working on the formation of vortices in microfluidic devices.
Wang et al. [6] investigated numerically the formation of vortices near a two-part
cylinder with the same sign of zeta potential but different in magnitudes under a
DC electric field. Mondal et al. [15] investigated numerically the mixing character-
istics in raccoon and serpentine micromixers. Soleymani et al. [16] investigated the
fluid mixing in T-type micromixers and showed that the formation of the vortices is
essential to enhance better mixing performances and the vortex formation strongly
depends on the geometrical parameters of the micromixer. Erickson and Li [17–19]
carried out a study to analyze the formation of vortices in a straight microchannel
wall under electroosmotic flow with different sign or polarity of zeta potential. Song
et al. [20] investigated both numerically and experimentally the flow field in a two-
section straight microchannel of the same polarity and different magnitude of zeta
potential. As in macroscopic level no-slip condition is valid, for microscopic flow
regime, studies exhibit the existence of slip boundary condition for hydrophobic
PDMS surfaces [21]. From the literature, it is seen that there is no work on the
analysis of the formation of vortices due to the same polarity of zeta potential and
interfacial slip. Therefore, we set our objective to investigate the effect of slip on the
vortex formation in a microchannel with the same sign zeta potential of the cylinder,
placing inside it. In the present study, the slip boundary conditions on the wall in the
microchannel [22–29] and the cylinder are considered in order to determine the flow
rate and recirculation reversed flow velocity near the cylinder surface accounting the
effect of vortices.
2 Theoretical Formulation
smaller than the radius of the cylinder (λD /r 1), therefore the curvature effect
on the charge distribution is neglected. Here, λD is the characteristic EDL thickness
and r is the radius of the cylinder. The dimensionless governing equations (Laplace,
Poisson-Boltzmann, Continuity and Momentum) can be written as [6, 20]:
∇2ϕ = 0 (1)
∇ 2 ψ = κ 2 sinh ψ (2)
∇ ·U=0 (3)
Re(U · ∇)U = −∇P + ∇ 2 U + κ 2 sinh ψ ∇ (ψ Ω) + ϕ (4)
Here,
ϕ is
the
dimensionless external electric potential, normalized with the scale
∗
ϕref = V L ; V is the external potential difference between the channel ends;
∗
ψ is the normalized EDL potential normalized with the scale ψref = kB To ze ;
kB , To , z and e are the Boltzmann constant,reference temperature,
valance
on ions
and charge on the electron, respectively; U = U î + V ĵ = uî + v ĵ uref is the
dimensionless velocity vector, u ref is the reference
velocity, ∇ = ∂ ∂X , ∂ ∂Y ;
∗
X = x H ; Y = y H ; uref = εo εr Eo ψref μ is the reference velocity; Eo =
V L; P is dimensionless
pressure scaled by μu ref H ; κ = H λD is the Debye
−1/2
parameter and λD = 2n◦ z 2 e2 εo εr kB T is the characteristic EDL thickness.
Here, n◦ is the nominal ionic bulk concentration, εr is the relative permittivity
of the
medium and εo is the absolute permittivity of the free space; Re = ρuref H μ is
1016 S. Pabi et al.
the Reynolds number; Ω is the ratio of applied potential difference to the reference
EDL potential and taken as 100. It corresponds to the external potential difference
of 25 V and reference EDL potential 0.025 V, respectively.
The following boundary conditions are imposed.
At the inlet
At the wall
U = β(∇U) (6b)
Here, ζU = −ζ and ζD = −ζ 5. ζ is the upstream side dimensionless zeta
∗
potential scaled by ψref , β = β ∗ H is the slip coefficient and β ∗ is the slip length.
At the outlet
ϕ = 0; n · (∇ψ) = 0; PG = 0 (7)
0.5
Q= Udy (8)
−0.5
Finite element based numerical technique is used to solve the governing equations
[10]. The relative convergence criterion taken in this simulation is set to 10−6 . An
extensive grid independence test is conducted and the selected mesh system is M3
with a relative difference in reverse flow velocity less than 1% (Table 1).
Validation is made with the results of Zhao et al. [30] as presented in Fig. 2 at
κ = 10. The close agreement of the present results with reported in [30] establish
the accuracy of the current results.
Effect of Slip on Vortex Formation … 1017
Fig. 2 Comparison of
dimensionless flow velocity
with the results of Zhao et al.
[30] for the electroosmotic
flow through a plane
microchannel at κ = 10
The objective of this work is to analyze the effect of interfacial slip on the forma-
tion of vortex near a two-part cylinder with zeta potential of same sign in a plane
microchannel. The values of Reynolds number (Re) and Debye parameter (κ) are
fixed as Re = 0.01, κ = 100, respectively [6, 20–22]. The dimensionless zeta poten-
tial (ζ ) is varied in the range of 1 ≤ ζ ≤ 4, which is equivalent to in the range
of 25–100 mV [6, 20–22]. The slip length varies in the order of 1 nm to 1 μm
[7], and therefore, the value of β is varied in the range of 0–0.1 for the microflu-
idic devices. The results are analyzed in terms of dimensionless flow velocity (U ),
streamlines, recirculation reversed
flow velocity
(Q), recirculation
(UR ), flow rate
reversed flow velocity ratio UR UR(β=0) and flow rate ratio Q Qβ=0 .
Figure 3a and b shows the streamlines and dimensionless flow velocity (U) at
β = 0 and β = 0.1, respectively. The vortex is formed near the upstream side of
the cylinder. It is because of the larger upstream zeta potential allows larger EOF
actuation along with the higher local electric field intensity near the channel wall and
upper part (or lower tip) of the cylinder. Therefore, to follow the mass conservation,
reversed flow is induced near the microchannel wall in upstream. It can be noted that
the effect of slip enhances the flow velocity and reversed flow velocity magnitude
(see Fig. 3a, b). The corresponding flow velocity profile is presented in Fig. 3c and d
1018 S. Pabi et al.
Fig. 3 Dimensionless flow velocity and streamlines contours at a β = 0 and b β = 0.1 with ζ = 4.
Variation of the dimensionless flow velocity profile at X = 4.7 for different zeta potential, near the
cylinder at c β = 0 and d β = 0.1
at different ζ for β = 0 and 0.1, respectively at the section X = 4.7. It can be noted
that the flow velocity and reversed flow velocity magnitude enhances with ζ and
β due to the increase in EOF actuation and flow resistance due to hydrophobicity,
respectively. It is also noted that the extent of the recirculation zone (vertical length
of negative flow velocity) is larger for the no-slip case as found Y = 0.156 and 0.119,
respectively for β = 0 and 0.1. It is because of the enhancement in primary flow for
the slip case and thus suppresses the extent of the flow reversal zone. It can be noted
that this length is fixed for any values of ζ .
Figure 4 shows the variation of reversed flow velocity (UR ) with β for different
ζ . The value of UR increases rapidly with β at its smaller values for all ζ . While
for higher values of β, a reduction in UR is observed, and the reduction is more
significant for higher values of ζ . It can be explained as follows. For smaller values
of β, UR enhances due to the presence of slip and for larger values of β and ζ , the
amplification in primary flow velocity suppresses the strength of UR . The percentage
increases in UR for slip case with β = 0.1 as compared to the no-slip case are found as
207.54%, 245.94%, 309.08%, and 394.56%, for ζ = 1,2, 3, 4, respectively.
Figure 5
shows the variation of reversed flow velocity ratio UR UR(β=0) with β at different
values of ζ . It can
be observed that this
ratio is greater than unity beyond the critical
slip coefficient βCri ≈ 2.97 × 10−4 . The value of this ratio increases with β for
Effect of Slip on Vortex Formation … 1019
Fig.5 Variation of
UR UR(β=0) with β at
different ζ
β > βCri up to β ∼ 0.08 and beyond that limit the reverse flow velocity ratio
decreases with β.
Figure 6 shows the variation of flow rate (Q) with β at different values of ζ .
The value of Q increases with β and ζ due to the increase
in primary flow velocity
(see Fig. 3c, d). Figure
7 shows the variation of Q Qβ=0 with β for different
ζ . The value of Q Qβ=0 is greater than unity beyond a critical slip coefficient
βCri ≈ 4.8 × 10−4 . The value of Q Qβ=0 monotonically enhances with β for β >
βCri at all values of ζ .
1020 S. Pabi et al.
Fig. 7 Variation of Q Qβ=0
with β at the different ζ
5 Conclusions
The present paper investigates the effect of slip on the formation of vortex near a
two-part cylinder with same sign of the zeta potential placed in a microchannel. The
results are presented in terms of flow velocity, flow rate, and reverse flow velocity
by varying the slip coefficient and zeta potential. The key findings of this study are
as follows:
• The extent of the recirculation zone is smaller for the slip case as compared to the
no-slip case.
Effect of Slip on Vortex Formation … 1021
• The value of characteristic reverse flow velocity (UR ) increases with the slip
coefficient (β) for its smaller values. While a decrement in UR is observed at larger
values of β. This decrement is amplified at the higher values of zeta potential.
• The flow rate monotonically increases with the slip coefficient and zeta potential.
References
1. Zhao G, Jian Y, Li F (2016) Streaming potential and heat transfer of nanofluids in microchannels
in the presence of magnetic field. J Magn Magn Mater 407:1–25
2. Ohno K, Tachikawa K, Manz A (2008) Microfluidics: applications for analytical purposes in
chemistry and biochemistry. Electrophoresis 29(22):4443–4453
3. Becker H, Gartner C (2000) Polymer microfabrication methods for microfluidic analytical
applications. Electrophoresis 21(1):12–26
4. Nandy K, Chaudhuri S, Ganguly R, Puri IK (2008) Analytical model for the magnetophoretic
capture of magnetic spheres in microfluidic devices. J Magn Magn Mater 320:1398–1405
5. Ng TN, Chen XQ, Yeung KL (2015) Direct manipulation of particle size and morphology of
ordered mesoporous silica by flow synthesis. RSC Adv 5(18):13331–13340
6. Wang C, Song Y, Pan X (2020) Electrokinetic-vortex formation near a two-part cylinder with
same-sign zeta potentials in a straight microchannel. Electrophoresis 00:1–9
7. Liu Y, Jian Y (2019) The effects of finite ionic sizes and wall slip on entropy generation in
electroosmotic flows in a soft nanochannel. J Heat Transf 141:1–11
8. Chakraborty J, Pati S, Som SK, Chakraborty S (2012) Consistent description of electrohydro-
dynamics in narrow fluidic confinements in presence of hydrophobic interactions. Phys Rev E
85:046305
9. Mehta SK, Pati S (2021) Thermo-hydraulic and entropy generation analysis for magnetohydro-
dynamic pressure driven flow of nanofluid through an asymmetric wavy channel. Int J Numer
Method Heat Fluid Flow 31(4):1190–1213
10. Mehta SK, Pati S (2019) Analysis of thermo-hydraulic performance and entropy generation
characteristics for laminar flow through triangular corrugated channel. J Therm Anal Calorim
136(1):49–62
11. Mehta SK, Pati S (2020) Numerical study of thermo-hydraulic characteristics for forced
convective flow through wavy channel at different Prandtl numbers. J Therm Anal Calorim
141:2429–2451
12. Nguyen NT, Wu Z (2005) Micromixers-a review. J Micromech Microeng 15(2):R1–R16
13. Maeng JS, Yoo K, Song S (2006) Modeling for fluid mixing in passive micromixers using the
vortex index. J Korean Phys Soc 48(5):902–907
14. Shang X, Huang X, Yang C (2015) Mixing enhancement by the vortex in a microfluidic mixer
with actuation. Exp Therm Fluid Sci 67:57–61
15. Mondal B, Mehta SK, Patowari PK, Pati S (2019) Numerical study of mixing in wavy
micromixers: comparison between raccoon and serpentine mixer. Chem Eng Process Process
Intensif 136:44–61
16. Soleymani A, Kolehmainen E, Turunen I (2008) Numerical and experimental investigations of
liquid mixing in T-type micromixers. Chem Eng J 135S:S219–S228
17. Erickson D, Li D (2003) Three-dimensional structure of electroosmotic flow over heteroge-
neous surfaces. J Phys Chem B 107(44):12212–12220
18. Erickson D, Li D (2002) Influence of surface heterogeneity on electrokinetically driven
microfluidic mixing. Langmuir 18(5):1883–1892
19. Erickson D, Li D (2002) Microchannel flow with patchwise and periodic surface heterogeneity.
Langmuir 18(23):8949–8959
1022 S. Pabi et al.
1 Introduction
Nimonic-75 (Ni-75) is an 80-20 Ni-Cr alloy with controlled additions of titanium and
carbon. It can be easily manufactured and welded. It possesses excellent corrosion
resistance, mechanical properties, and heat resistance. It finds extensive applications
in aerospace fasteners, heat treatment equipment, gas turbine engineering, nuclear
engineering, industrial furnace structural parts, etc.
There are many challenges faced during traditional machining of the Nimonic-75
alloy such as notching of tool nose, confrontation to forced deformation required for
chip formation, miniature cavities created on the rake face, dissolution and diffusion
wear of the rake face of the tool, abrasive wear of the tool, etc. [1, 2]. Wire electrical
discharge machining (WEDM) is one of the non-traditional machining processes
which can be effectively employed to machine such a challenging to cut alloys. It
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1023
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_102
1024 S. A. Sonawane and S. S. Wangikar
2 Experimentation
Table 2 Results of
Exp. No. MRR mm3 /min SR µm KW mm
orthogonal tests
1 33.30 1.82 0.285
2 27.60 1.68 0.269
3 26.60 1.6 0.265
4 31.16 1.76 0.279
5 28.43 1.7 0.27
6 26.64 1.61 0.266
7 32.25 1.79 0.28
8 25.67 1.58 0.26
9 24.53 1.52 0.254
10 55.00 2.32 0.307
11 48.63 2.19 0.297
12 33.70 1.85 0.286
13 52.37 2.28 0.305
14 32.92 1.79 0.285
15 37.62 2.01 0.289
16 34.00 1.95 0.287
17 39.20 2.06 0.293
18 36.60 1.97 0.288
19 58.70 2.51 0.327
20 38.90 2.04 0.291
21 56.34 2.41 0.313
22 44.37 2.1 0.295
23 55.20 2.36 0.309
24 50.90 2.25 0.299
25 57.60 2.5 0.322
26 54.50 2.31 0.305
27 29.20 1.3 0.275
Zij
Nij = √ (1)
(Zij)2
g
n
Yi = Nij − Nij (2)
j=1 j=g+1
g
n
Yi = WjNij − WjNij (3)
j=1 j=g+1
The AHP is derived from the knowledge acquired by its inventor, Saaty [13] while
steering research assignments in the US Arms Control and Disarmament organiza-
tion. The AHP presents a means of disintegrating the objective into a hierarchy of
associate difficulties which can more easily be figured out and subjectively estimated.
The subjective appraisals are renovated into numerical assessments and processed to
grade each choice on a numerical degree. The scheme of the AHP can be elucidated
in the subsequent steps:
Step 1 The hitch is decayed into a chain of command of objectives, criterion,
sub-criteria and options.
Step 2 Devise the pronouncement matrix.
The verdict matrix is created containing all the quality characteristics
indicated in quantitative or numerical value.
Step 3 Engender pair wise matrices
A couple wise association matrix is created using a level of comparative
importance as shown in Table 2. The oblique elements of the array are 1.
The measure in the ith row is better than the measure in the jth column if
the rate of the aspect (i, j) is in excess of 1; otherwise, the measure in the
jth line is superior than that in the ith chain. The (j, i) element of the array
is the mutual of the (i, j) constituent.
Step 4 Determine the eigenvectors using a geometric mean method.
Step 5 Calculate the weights by normalizing the eigenvector by the total sum of
eigenvectors.
By applying the above steps to the values of the output features as shown
in Table 2, the overall assessment value is calculated and is represented in
Table 3.
It is observed that Exp. No. 9 has the highest value of Yi. Thus, the optimum
combination of explanatory factors are Pon = 110 machine units (mu), Poff
= 51 mu, GV = 40 V, IP = 230 Amp., WV = 5 m/min and WT = 8 mu.
The evidence tests are conducted to validate the most favorable parameters
settings. The results of evidence tests are indicated in Table 4.
4 Conclusions
3. The optimum permutation of input factors obtained are Pon = 110 machine
units (mu), Poff = 51 mu, GV = 40 V, IP = 230 Amp., WV = 5 m/min and WT
= 8 mu.
References
1. Ezugwu EO, Wang ZM, Machado AR (1999) The machinability of nickel-based alloys: a
review. J Mater Process Technol 86(1–3):1–16
2. Chaudhary IA, El-Baradie MA (1998) Machinability of nickel-based super alloys: a general
review. J Mater Process Technol 77(1–3):278–284
3. Chakraborty S (2011) Applications of the MOORA method for decision making in manufac-
turing environment. Int J Adv Manuf Technol 54(9–12):1155–1166
4. Gadakh VS (2011) Application of MOORA method for parametric optimization of milling
process. Int J Appl Eng Res 1(4):743–758
5. Gadakh VS, Shinde VB, Khemnar NS (2013) Optimization of welding process parameters
using MOORA method. Int J Adv Manuf Techno 69(9–12):2031–2039
6. Rajesh S, Pethuraj M, Thirumalai Kumaran S, Uthaykumar M, Rajini N (2015) Some studies
on drilling of red mud reinforced aluminium composite. Proc Inst Mech Eng Part L J Mater
Des Appl 231(4):382–393
7. Singaravel B, Selvaraj T, Vinodh S (2016) Multi-objective optimization of turning parameters
using the combined MOORA and Entropy method. Trans Can Soc Mech Engg 40(1):101–111
8. Saha A, Mondal SC (2017) Machining optimization of nano-structured hardfaced tool insert in
WEDM using MOORA method. In: Interanational conference on research into design. Springer
Publication, pp 905–917
9. Kalirasu S, Rajini N, Rajesh S, Winowlin Jappes JT, Karuppasamy K (2017) AWJM perfor-
mance of Jute/Polyester composite using MOORA and analytical models. Mater Manuf Process
32(15):1730–1739
10. Majumdar H, Maity K (2017) Optimization of machining condition in WEDM for Titanium
Grade 6 using MOORA coupled with PCA-A multivariate hybrid approach. J Adv Manuf Syst
16(2):81–99
11. Anand babu K, Venkataramaiah P, Dileep P (2017) AHP-DENG’S similarity based optimization
of WEDM process parameters of Al/SiCp composite. Am J Mater Sci Technol 6(1):1–14
12. Mohapatra KD, Sahoo SK (2018) Optimization of single pitch error and MRR in a WEDM gear
cutting process. In: Precision product-process design and optimization. Springer Publication,
pp 285–312
13. Satty TL (2007) Fundamentals of decision making and priority theory with AHP. RWS
Publications, Pittsburg
Structural Analysis of Novel Mini
Horizontal Axis Wind Turbine
(NMHAWT)
Abstract The wind energy is most of the promising renewable energy source. In
wind turbine technology, the turbine blades play an important role as it directly
comes in contact with the wind. Many researchers have concentrated on improving
the aerodynamic performance of wind turbine blade through testing and theoretical
studies. In general, moderate to high-speed winds, typically from 5 m/s to about
25 m/s are considered favourable for most wind turbines in India. But in rural areas,
wind speed is near about 3–9 m/s. The present investigation aims is to compare the
performance of eight blade novel mini Horizontal Axis Wind Turbine (NMHAWT)
blades of novel airfoils over National Advisory Committee for Aeronautics (NACA)
eight blade profiles in terms of loads and performance. Therefore blade design of
well known series of NACA airfoil was selected from Q blade and Mat-Lab software.
Analytical calculations are help out for the selection of NACA 4418 is suitable for
the comparison. The objective of this paper is to decide the novel profile of the blade
for the development. The comparison is done through software for predicting the
performance of a novel mini horizontal axis wind turbine.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1031
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_103
1032 P. Magade et al.
1 Introduction
Renewable energy of wind is one of the cheapest sources for providing electrical
or mechanical power. It can be found everywhere and every time. But wind cannot
produce required power if it has not sufficient strong velocity. The blowing wind
possesses kinetic energy and strikes on wind turbine blades. By this action kinetic
energy are lost and transmit to mechanical energy of wind turbine shaft [1]. The
maximum power that can be extracted is 59.3% of the power in wind. Wind turbine
shaft is coupling with generator. Therefore, wind turbine blade is an energy converter.
Besides the turbine blades’ profile also plays an important role to capture the wind
energy efficiently and effectively [2]. Energy generated from the wind depends on
direction of the blade profiles, airstream and angle of attack [3]. The optimized airfoil
profile is chosen from maximum lift to drag ratio at design Reynolds number and
angle of attack [4]. In order to achieve the maximum power coefficient design of blade
positioned properly [5]. The performance of blade is simulated in detailed by using
analysis [6]. Experimental results has been carried out for various factors like wind
speed, numbers of blade and size of blade. The eight blades model with is appropriate
for designing and it gives around 40 W power output [7]. Above researches gave an
idea of comparison of traditional and novel blade.
2 Design of Blade
The fundamental theory of design and operation of wind turbines is derived based
on a first principles approach using conservation of mass and conservation of energy
in a wind stream. The diameter of the blade decided by the generator specification
and Beltz limits. The wind power equation is shown as follows.
1
FD = ρAv 2 Cd Therefore Drag Coefficient ∴ Cd = 0.043 (3)
2
(b) Lift to Drag ratio: In order to obtain high efficiency, it is therefore essential to
use airfoil shaped rotor blades with a very high lift to drag ratio,
cl
Lift to Drag Ratio ∴ = 36.816 (4)
Cd
(c) Reynolds Number
Low Re airfoils is suited for mini wind turbine applications must be designed
to avoid high leading edge suction peak and high adverse pressure gradients
that leads to flow separation.
D×V
Reynolds Number Re = ∴ Re = 500000 (5)
v
(d) Tip speed ratio is the ratio between the tangential speed of the tip of a blade
and the actual speed of the wind. Higher tip speeds result in higher noise levels
and require stronger blades due to larger centrifugal forces. Also calculated by
λ = 4Πn
from the wind energy math calculation n is no of blade.
∴ λ = 1.5. (7)
1034 P. Magade et al.
The optimum airfoil is selected from maximum lift to drag ratio at designed Reynold
number and angle of attack. Therefore Q blade software is used to select the suitable
airfoil based on calculations. For Cl/Cd = 36.819, Alpha = 15° and Reynold number
is 500000 as shown in Figs. 1 and 2.
The mat-lab programmed is used to calculate the position, chord length & twist angle
of blade. Input parameter require in mat-lab to calculate the chord length and twist
angle is number of blade’s, capacity of turbine, velocity of wind, Tip speed ratio,
Lift coefficient, Reynold number and Air density. Shown in Table 1.
Base on mat-lab calculation different chord length and twist angle are obtained
but at 15° angle of attack blade gives the maximum lift force. Therefore novel blade
and regular shape blade are designed for 15° angle of attack.
The design of structure which is necessary in turbine design. To find out the structural
analysis (von mises) stress and total deformation of 1200 mm diameter with eight
blade [7] geometry of two different blade profiles is compared with each other. The
analysis is done by using ANSYS workbench software.
The results of von mises stresses and total deformation of regular shape blade and
novel shape blade is compared by using ANSYS workbench software. The Von
mises stress is also known as equivalent stress. It is used in design work because any
arbitrary three-dimensional stress can be represented as a single positive stress value.
The Von mises stress is the part of the maximum equivalent stress failure theory used
to predict yielding in a ductile material. The von mises stress measurements is in
metric and to determine if the structure has started to yield at any point. In ANSYS
software material properties is necessary. Aluminium material is used for design
of blade and its Density is 2.77 × 10−6 kg /mm3 , Compressive yield strength (Syc )
is 280 MPa, Tensile yield strength (Syt ) is 280 MPa, Ultimate tensile strength is
310 MPa, Modulus of elasticity is 69 GPa and Poisons ratio is 0.34.
Now find out the von mises stress and total deformation by using ANSYS workbench
software with the load of 20 N. The following steps demonstrate how to create the
model geometry and find a solution by using ANSYS workbench.
1. Conduct the pre-processing by specifying the element type such as material
properties and meshing. The accuracy that can be obtained from any FEA model
is directly related to the finite element mesh that is used. The finite element mesh
is used to subdivide the CAD model into smaller domains called elements, over
which a set of equations are solved. These equations approx. represent the
governing equation of interest via a set of polynomial functions defined over
each element. As these elements are made smaller and smaller, as the mesh
is refined, the computed solution will approach the true solution. This process
of mesh refinement is a key step in validating any finite element model and
gaining confidence in the software, the model, and the results. The following
Fig. 3 shows the meshing of the blade.
1036 P. Magade et al.
2. Specify the constraint and loads. The fixed support of the in middle side and
load are applied to the blade is 20 N. Following figure shows the stresses occurs
in NACA 4418 and NMHAWT.
The maximum amounts of stresses are developed in the root of blade as shown in
Fig. 4. For the load of 20 N NACA profile blade creates 26.9 MPa stress and for
novel profile it will reach up to 17.059 Mpa. Stresses occur in novel blade profile
is less than NACA blade profile in connection with this the ultimate strength of
aluminum is 40–50 Mpa, Hence the maximum von-Mises stress found in both
the blade is below the ultimate strength of aluminum therefore blade will not
fails under 20 N defined load. It can be conclude from the results that blade
design is safe for 20 N load.
3. Deformation is the change in the physical properties of a material. It can be clas-
sified into two categories i.e. plastic and elastic deformation. The deformation
that is permanent or which remains even after the removal of stress is known as
plastic deformation whereas the deformation that remains only until the force
is applied is known as elastic deformation. The following result shows the total
deformation at different load conditions.
Figure 5 shows the deformation occurs in NACA 4418 and NMHAWT for a load
of 20 N.
The maximum amounts of deformation are developed beyond swept area or tip
side of both the blade as shown in Fig. 5. For the load of 20 N NACA profile blade
deform 2.29 mm and for novel profile is up to 2.52 mm. Deformation occur in novel
blade profile is quite less in NACA blade profile.
The present paper describes the static structural analysis by ANSYS for uniformly
distributed loading. A parameter study is carried out where advances in this field
have allowed the price of wind energy to be competitive with more conventional
sources. The blades are an important component of wind turbines and much research
has been done in this area. Optimization of blade consider for analysis where wind
load of 20 N. Identify its suitability for its application on wind turbine blades and
good agreement is made between results. In two-dimensional aerofoil modeling, the
comparison between the two profiles of blades in structural analysis by ANSYS
results is a reliable benchmark. Stress and deformation of the blade performances
of different aerofoils have been compared. An airfoil of different profile blade for
a horizontal axis wind turbine is designed based on Structure, it is concluded that
the structural analysis is efficient in predicting rotor aerodynamic characteristics.
The structural analysis is performed to evaluate the proposed novel blade design
configuration is the suitable for further development. In this study,
1. Static structural analysis of three, five and eight blade with 1200 mm diameter
NACA4418 and Novel HAWT is checked under various uniform load conditions
as like 10 N, 15 N and 20 N.
2. The amount of stresses and deformation occurs in the three blades NACA 4418
for the load of 10 N is 37.34 Mpa and 4.92 mm.
1038 P. Magade et al.
3. The amount of stresses and deformation occurs in the five blades NACA 4418
for the load of 15 N is 22.40 Mpa and 2.98 mm.
4. The amount of stresses and deformation occurs in the eight blades NACA 4418
for the load of 20 N is 26.98 Mpa and 2.29 mm.
5. The amount of stresses and deformation occurs in the three blades NMHAWT
for the load of 10 N is 20.37 Mpa and 5.86 mm.
6. The amount of stresses and deformation occurs in the five blades NMHAWT
for the load of 15 N is 14.46 Mpa and 2.72 mm.
7. The amount of stresses and deformation occurs in the eight blades NMHAWT
for the load of 20 N is 17.05 Mpa and 2.52 mm.
8. It was observed that NMHAWT gives very less amount of stress generation
at the root of the blade and negligible amount of deformation of the blade as
compared to NACA 4418.
9. Newly developed NMHAWT gives better results as compared to conventional
blade profile as shown in Table 2. So NMHAWT can be proposed for further
development and testing.
References
1. Guo Y, Cao X, Wang N (2019) A hybridized design of pinwheel by coupling tribo electrification
and electromagnetic induction effects. Elsevier Nano Energy, vol 60, pp 641–648
2. Riyanto,Pambudi NA (2019) The performance of shrouded wind turbine at low wind speed
condition, (ICAE 2018), Hong Kong, china, pp 260–265
3. Tian W, Yang Z, Zhang Q (2017) Bionic design of wind turbine blade based on long-eared owls
airfoil
4. Thu AM, Aung YYH (2016) Design and performance testing of a small-scaled horizontal axis
wind turbine for low wind speed regions
5. Sinha DK, Tummala A (2016) A review on small scale wind turbines. Science direct
6. Elmagid WMA, Mekhali TA (2015) Experimental and CFD of designed small wind turbine
7. Magade Pramod B, Chavan Srirang P, Magade Sushilkumar B (2019) Experimentation on design
and development of mini wind turbine. Int J Innov Technol Explor Eng (IJITEE) 8(11). ISSN:
2278-3075
Fabrication of Tree Type Micro-Mixer
with Circular Baffles Using CO2 Laser
Machining
Abstract Micro-mixer is a device which is used to carry the mixing of two or more
than two fluids. Any one dimension of it is in micrometer. Due to low Reynolds’
Number Micro-mixer design is challenge for the designers. To address this chal-
lenge, novel methods of mixing enhancement within micro-fluidic devices have been
explored for a variety of applications. For Passive Micro-mixer mixing will be carried
out with the help of geometry of the micro-mixer. In this study fabrication of a tree
type Micro-mixer with Circular baffles using CO2 Laser machining is discussed. The
Circular baffles are responsible for making diversion of flow of the fluid in Conver-
gent & Divergent manner due to which proper mixing of fluids become possible. This
type of micro-mixer is suitable for Bio-medical applications such as Urine testing,
Blood testing, in preparation & in testing of Drugs. This paper focuses on design
and fabrication of tree type Micro-mixer with Circular Baffles. The parametric study
for CO2 laser machining by considering the speed and power as the control param-
eters and the depth as a performance measure for engraving different width micro
channels is presented in this paper. The tree type Micro-mixer with Circular Baffles
fabricated by CO2 laser machining is found to be suitable for master molds used in
soft lithography process.
1 Introduction
A device having the size in microns which is employed for mixing the fluids is called
as Micro-mixer. parts used to mix fluids. This device represents a significant tech-
nology to fields like pharmaceutical industry, chemical industry, analytical chemistry,
high-throughput synthesis, and biochemical analysis, as it employs miniaturized parts
in turn the reduced quantities involved in the chemical and/or biochemical processes.
The Active and passive micro-mixers are the main to types of it. The external energy
source like electric or magnetic source is used by active micro-mixers for performing
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2021 1039
P. M. Pawar et al. (eds.), Techno-Societal 2020,
https://doi.org/10.1007/978-3-030-69921-5_104
1040 S. R. Gavali et al.
the mixing of the fluids. There is no any power source utilized in the passive micro-
mixers and the pressure guides the flow of fluids. Mixing in micro-fluidic devices
presents a crucial challenge [1–5]. A lot of research has been done on micro-channels
with different configurations, both experimentally and numerically. Wong et al. [2]
fabricated micro-T-mixersfrom silicon substrate and tested it for investigating their
probability as quick mixing micro-mixers. Orsi et al. [3] has performed numer-
ical simulations for studying the mixing dynamics of two miscible liquids within
a T shaped micro-mixer by comparing the case where the two inlet fluids are both
water (W-W). Tsai et al. [4] investigated the mixing phenomena between ferro-nano
fluids (a water solution with suspended Fe3 O4 nano particles) and water in a Y-type
semi-active micro-mixer. Further, many researchers have reported the performance
analysis of different types of microchannels using various approaches [3–13]. The
photochemical machining process was also used by various researchers for fabrica-
tion of micro-mixers and molds for micro-mixers employing soft lithography process
[14–21]. The CO2 laser machining was employed for fabrication of micro features
and some straight channels by Chavan et al. [22].
In this study, fabrication of a tree type Micro-mixer with Circular baffles is
performed by CO2 laser machining. The Circular baffles are responsible for making
diversion of flow of the fluid in Convergent & Divergent manner due to which proper
mixing of fluids become possible. The design is suitable for not only two fluids but
also for four fluids. This type of micro-mixer is suitable for Bio-medical applications
such as Urine testing, Blood testing, in preparation & in testing of Drugs. This design
is also helpful in testing of liquids used for Food Processing, testing of Milk, alcohol,
wines etc. It is also suitable for Soil testing purposes. The parametric study for CO2
laser machining by considering the speed and power as the control parameters and
the depth as a performance measure for engraving different width microchannels is
presented in this paper.
Acrylic is extensively known for its significance to the industry. It is employed for
making the different products like bath enclosures, shower doors, skylights, and
windows. It is preferred over the glass for different reasons. It is stronger than glass
which makes it impact resistant and hence the safer. Another, big advantage is that
acrylic is only half as heavy as glass which makes working with acrylic much easier
[5].
Fabrication of Tree Type Micro-Mixer … 1041
2.2 Methodology
Fig. 1 Micro-mixer
(without circular baffles)
1042 S. R. Gavali et al.
The micro-mixer without baffles and the micro-mixer with circular baffles have been
fabricated on acrylic material using CO2 laser machining and depicted in Figs. 4 and
5, respectively.
4 Conclusion
0.551 mm and 0.517 mm. The influence of the power is observed to be prominent
and the depth is noted to be increasing with increase in laser power. This study shows
the suitability of CO2 laser cut machining for fabrication of micro-mixers or molds
for micro-mixers used in soft lithography process.
References
1. Klank H, Kutter JP, Geschke O (2000) CO2-Laser micromachining and back-end processing
for rapid Production of PMMA-based microfluidic systems. La Chi 2:242–246
2. Das SS, Tilekar SD, Wangikar SS, Patowari PK (2017) Numerical and experimental study of
passive fluids mixing in micro-channels of different configurations. Micros Technol 23:5977–
5988
3. Seck HW, Patrick B, Michael W, Christopher W (2013) Investigation of mixing in a cross-
shaped micromixer with static mixing elements for reaction kinetics studies. Sens Actua B
95:414–424
4. Gianni O, Mina R, Elisabetta B, Chiara G, Roberto M (2013) Water-ethanol mixing in T-shaped
microdevices. Chem Eng Sci 95:174–183
5. Tsung-Han T, Dar-Sun L, Long-Sheng K, Ping-Hei C (2009) Rapid mixing between ferro-
nanofluid and water in a semi-active Y-type micromixer. Sens Actua A 153:267–273
6. Wangikar SS, Patowari PK, Misra RD (2018) Numerical and experimental investigations on
the performance of a serpentine microchannel with semicircular obstacles. Micros Technol
24:3307–3320
7. Gidde RR, Pawar PM, Ronge BP, Misal ND, Kapurkar RB, Parkhe AK (2018) Evaluation of the
mixing performance in a planar passive micromixer with circular and square mixing chambers.
Micros Technol 24:2599–2610
8. Zonghuan L, Jay McMahon HM, Barnard D, Shaikh TR, Mannella CA, Wagenknecht T, Toh-
Ming L (2010) Passive microfluidic device for submilli second mixing. Sens Actua B 144:301–
309
9. Gidde RR, Shinde AB, Pawar PM, Ronge BP (2018) Design optimization of a rectangular
wave micromixer (RWM) using Taguchi based grey relational analysis (GRA). Micros Technol
24:3651–3666
10. Gidde RR, Pawar PM, Ronge BP, Shinde AB, Misal ND, Wangikar SS (2019) Flow field analysis
of a passive wavy micromixer with CSAR and ESAR elements. Micros Technol 25:1017–1030
11. Gerlach A, Knebel G, Guber AE, Heckele M, Herrmann D, Muslija AS (2002) Microfabrication
of single-use plastic microfluidic devices for high-throughput screening and DNA analysis.
Micros Technol 7:265–268
12. Arshad A, Kwang-Yong K (2012) Passive split and recombination micromixer with convergent-
divergent walls. Chem Eng J 203:182–192
13. Rotting O, Ropke W, Becker H, Gartner C (2010) Polymer microfabrication technologies.
Micros Technol 8:32–36
14. Wangikar SS, Patowari PK, Misra RD (2017) Effect of process parameters and optimization
for photochemical machining of brass and german silver. Mater Manuf Process 32:1747–1755
15. Wangikar SS, Patowari PK, Misra RD (2016) Parametric optimization for photochemical
machining of copper using grey relational method. In: 1st Techno-societal, international
conference on advanced technologies for societal applications. Springer, pp 933–943
16. Wangikar SS, Patowari PK, Misra RD (2018) Parametric optimization for photochemical
machining of copper using overall evaluation criteria. Mater Tod Proc 5:4736–4742
17. Wangikar SS, Patowari PK, Misra RD, Misal ND (2019) Photochemical machining: a less
explored non-conventional machining process. In: Non-conventional machining in modern
manufacturing systems. IGI Global, pp 188–201
1046 S. R. Gavali et al.
18. Kulkarni HD, Rasal AB, Bidkar OH, Mali VH, Atkale SA, Wangikar SS, Shinde AB (2019)
Fabrication of micro-textures on conical shape hydrodynamic journal bearing. Int J Tre Eng
Technol 36:37–41
19. Raut MA, Kale SS, Pangavkar PV, Shinde SJ, Wangikar SS, Jadhav SV, Kashid DT (2019)
Fabrication of micro channel heat sink by using photo chemical machining. Int J New Technol
Res 5:72–75
20. Patil PK, Kulkarni AM, Bansode AA, Patil MK, Mulani AA, Wangikar SS (2020) Fabrication
of logos on copper material employing photochemical machining. Nov MIR Res J 5(6):70–73
21. Wangikar SS, Patowari PK, Misra RD, Gidde RR, Bhosale SB, Parkhe AK (2020) Optimization
of photochemical machining process for fabrication of microchannels with obstacles. Mater
Manuf Process. https://doi.org/10.1080/10426914.2020.1843674
22. Chavan NV, Bhagwat RM, Gaikwad SS, Shete SS, Kashid DT, Wangikar SS (2019) Fabrication
& characterization of microfeatures on PMMA using CO2 laser machining. Int J Tre Eng
Technol 36:29–32