International Conference On Artificial Intelligence For Smart Community
International Conference On Artificial Intelligence For Smart Community
Rosdiazli Ibrahim ·
K. Porkumaran · Ramani Kannan ·
Nursyarizal Mohd Nor ·
S. Prabakar Editors
International
Conference on Artificial
Intelligence for Smart
Community
AISC 2020, 17–18 December, Universiti
Teknologi Petronas, Malaysia
Lecture Notes in Electrical Engineering
Volume 758
Series Editors
Leopoldo Angrisani, Department of Electrical and Information Technologies Engineering, University of Napoli
Federico II, Naples, Italy
Marco Arteaga, Departament de Control y Robótica, Universidad Nacional Autónoma de México, Coyoacán,
Mexico
Bijaya Ketan Panigrahi, Electrical Engineering, Indian Institute of Technology Delhi, New Delhi, Delhi, India
Samarjit Chakraborty, Fakultät für Elektrotechnik und Informationstechnik, TU München, Munich, Germany
Jiming Chen, Zhejiang University, Hangzhou, Zhejiang, China
Shanben Chen, Materials Science and Engineering, Shanghai Jiao Tong University, Shanghai, China
Tan Kay Chen, Department of Electrical and Computer Engineering, National University of Singapore,
Singapore, Singapore
Rüdiger Dillmann, Humanoids and Intelligent Systems Laboratory, Karlsruhe Institute for Technology,
Karlsruhe, Germany
Haibin Duan, Beijing University of Aeronautics and Astronautics, Beijing, China
Gianluigi Ferrari, Università di Parma, Parma, Italy
Manuel Ferre, Centre for Automation and Robotics CAR (UPM-CSIC), Universidad Politécnica de Madrid,
Madrid, Spain
Sandra Hirche, Department of Electrical Engineering and Information Science, Technische Universität
München, Munich, Germany
Faryar Jabbari, Department of Mechanical and Aerospace Engineering, University of California, Irvine, CA,
USA
Limin Jia, State Key Laboratory of Rail Traffic Control and Safety, Beijing Jiaotong University, Beijing, China
Janusz Kacprzyk, Systems Research Institute, Polish Academy of Sciences, Warsaw, Poland
Alaa Khamis, German University in Egypt El Tagamoa El Khames, New Cairo City, Egypt
Torsten Kroeger, Stanford University, Stanford, CA, USA
Yong Li, Hunan University, Changsha, Hunan, China
Qilian Liang, Department of Electrical Engineering, University of Texas at Arlington, Arlington, TX, USA
Ferran Martín, Departament d’Enginyeria Electrònica, Universitat Autònoma de Barcelona, Bellaterra,
Barcelona, Spain
Tan Cher Ming, College of Engineering, Nanyang Technological University, Singapore, Singapore
Wolfgang Minker, Institute of Information Technology, University of Ulm, Ulm, Germany
Pradeep Misra, Department of Electrical Engineering, Wright State University, Dayton, OH, USA
Sebastian Möller, Quality and Usability Laboratory, TU Berlin, Berlin, Germany
Subhas Mukhopadhyay, School of Engineering & Advanced Technology, Massey University,
Palmerston North, Manawatu-Wanganui, New Zealand
Cun-Zheng Ning, Electrical Engineering, Arizona State University, Tempe, AZ, USA
Toyoaki Nishida, Graduate School of Informatics, Kyoto University, Kyoto, Japan
Federica Pascucci, Dipartimento di Ingegneria, Università degli Studi “Roma Tre”, Rome, Italy
Yong Qin, State Key Laboratory of Rail Traffic Control and Safety, Beijing Jiaotong University, Beijing, China
Gan Woon Seng, School of Electrical & Electronic Engineering, Nanyang Technological University,
Singapore, Singapore
Joachim Speidel, Institute of Telecommunications, Universität Stuttgart, Stuttgart, Germany
Germano Veiga, Campus da FEUP, INESC Porto, Porto, Portugal
Haitao Wu, Academy of Opto-electronics, Chinese Academy of Sciences, Beijing, China
Walter Zamboni, DIEM - Università degli studi di Salerno, Fisciano, Salerno, Italy
Junjie James Zhang, Charlotte, NC, USA
The book series Lecture Notes in Electrical Engineering (LNEE) publishes the
latest developments in Electrical Engineering—quickly, informally and in high
quality. While original research reported in proceedings and monographs has
traditionally formed the core of LNEE, we also encourage authors to submit books
devoted to supporting student education and professional training in the various
fields and applications areas of electrical engineering. The series cover classical and
emerging topics concerning:
• Communication Engineering, Information Theory and Networks
• Electronics Engineering and Microelectronics
• Signal, Image and Speech Processing
• Wireless and Mobile Communication
• Circuits and Systems
• Energy Systems, Power Electronics and Electrical Machines
• Electro-optical Engineering
• Instrumentation Engineering
• Avionics Engineering
• Control Systems
• Internet-of-Things and Cybersecurity
• Biomedical Devices, MEMS and NEMS
For general information about this book series, comments or suggestions, please
contact [email protected].
To submit a proposal or request further information, please contact the Publishing
Editor in your country:
China
Jasmine Dou, Editor ([email protected])
India, Japan, Rest of Asia
Swati Meherishi, Editorial Director ([email protected])
Southeast Asia, Australia, New Zealand
Ramesh Nath Premnath, Editor ([email protected])
USA, Canada
Michael Luby, Senior Editor ([email protected])
All other Countries
Leontina Di Cecco, Senior Editor ([email protected])
** This series is indexed by EI Compendex and Scopus databases. **
Rosdiazli Ibrahim · K. Porkumaran ·
Ramani Kannan · Nursyarizal Mohd Nor ·
S. Prabakar
Editors
International Conference
on Artificial Intelligence
for Smart Community
AISC 2020, 17–18 December, Universiti
Teknologi Petronas, Malaysia
Editors
Rosdiazli Ibrahim K. Porkumaran
Department of Electrical and Electronics Department of Electrical and Electronics
Engineering Engineering
Universiti Teknologi PETRONAS Sri Sairam Engineering College
Seri Iskandar, Perak, Malaysia Chennai, India
S. Prabakar
Department of Biomedical Engineering
Sona College of Technology
Salem, India
© The Editor(s) (if applicable) and The Author(s), under exclusive license to Springer Nature
Singapore Pte Ltd. 2022
This work is subject to copyright. All rights are solely and exclusively licensed by the Publisher, whether
the whole or part of the material is concerned, specifically the rights of translation, reprinting, reuse
of illustrations, recitation, broadcasting, reproduction on microfilms or in any other physical way, and
transmission or information storage and retrieval, electronic adaptation, computer software, or by similar
or dissimilar methodology now known or hereafter developed.
The use of general descriptive names, registered names, trademarks, service marks, etc. in this publication
does not imply, even in the absence of a specific statement, that such names are exempt from the relevant
protective laws and regulations and therefore free for general use.
The publisher, the authors, and the editors are safe to assume that the advice and information in this book
are believed to be true and accurate at the date of publication. Neither the publisher nor the authors or
the editors give a warranty, expressed or implied, with respect to the material contained herein or for any
errors or omissions that may have been made. The publisher remains neutral with regard to jurisdictional
claims in published maps and institutional affiliations.
This Springer imprint is published by the registered company Springer Nature Singapore Pte Ltd.
The registered company address is: 152 Beach Road, #21-01/04 Gateway East, Singapore 189721,
Singapore
Contents
v
vi Contents
1 Introduction
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 1
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_1
2 A. Parvathi et al.
signal which is the difference between the desired and actual output of the processes
is an input for the PID controller.
It offers numerous options to the researchers for tuning. The earlier tuning princi-
ples presented by Ziegler and Nichols [1]. They described many tuning procedures
for different types of systems. This tuning approach becomes complicated when the
process having the integrating terms and zeros on the right half of the s-plane. The
intricacy becomes more significant when the time delay (dead time) is related to
the procedure. If the process has large time delay in their model then this type of
design procedure may not give the accurate desired performance of the processes
considered. Therefore a new developments or modification in the basic structure of
PID controller/tuning plan is required.
Procedures with inverse reaction attributes are regularly experienced in industries,
for example, the degree of the drum kettle in a refining segment, the left temperature
of a cylindrical exothermic reactor. Integrating procedures with inverse reaction can
likewise be knowledgeable about mechanical applications, for example, level control
of a kettle steam drum. The model of this type of process includes the zero of transfer
function on the RHS.
It is noted that PID-type controllers bring about unacceptable performance exhi-
bitions in controlling the processes like integrating type. Numerous researchers have
projected different approaches for tuning this type of process models as: Jin and Liu
[2] proposed an upgraded inside the model controller and determined scientific PID
tuning rules for incorporating forms. Anil and Sree [3] planned a PID controller for
different types of industrial processes with time deferring based on directsynthesis
approach. Ajmeri and Ali [4] utilized a direct way to combine and deal with tuning
PID controllers for complete integrating, and twofold request integrating procedures
with huge time delays.
Then again, there are very few works tending to the control of integrating proce-
dures with reverse/inverse reaction and dead time. Luyben [5] proposed a PI and
additionally a PID controller in the traditional single-input single-yield (SISO) frame-
work to control incorporating forms with converse reaction. Pai et al. [6] have used
PID for inverse operating integrating process models.
In recent years, Kaya [7] has projected a PI–PD form of controller to improve
the performance of such (integrating with inverse operating) processes using Smith
indicator plan.
The target of this work is to structure an effective control methodology for the
stable-integrating processes of first order with delay in operating time. The goal is
cultivated by utilizing a PID controller along filter is designed. The controller param-
eters are inferred as far as procedure parameters with the assistance of polynomial
strategy.
The remaining portion of paper is structured as follows: The projected control
scheme is explained in part II, control system design in part III, and simulation
studies in part IV. Finally conclusion is presented.
Design of PID Controller for Integrating Processes … 3
The projected control loop structure for (capacity type) integrating processes is
given in Fig. 1. This control loop consist of controller—Gc (s), disturbance signal—
d, process with inverse response—Gp (s) and set point filter—F(s).In this structure
r—denotes the reference and Y —denotes the output response.
In this study the process considered is (capacity type) integrating type which
has one pole (type-1) at origin of s-plane. The form of integrating (capacity type)
processes with inverse response is given as:
K p (−αs + 1) −θs
G p (s) = e α > 0; τ, θ > 0 (1)
s(τ s + 1)
where, Kp —is the steady-state gain, τ—is time constant, and θ—is the time delay.
The servo and regulatory responses are derived as presented in Eqs. (2) and (3)
respectively.
k(−αs + 1) −θs
G p (s) = e (4)
s(τ s + 1)
k(αs + 1)(−αs + 1) −θs
G p (s) = e
s(αs + 1)(τ s + 1)
These sections explains the steps to intend the overall controller which includes
set point filter F(s), tuning rules, selecting λ value and performance measurement.
It is clear from the controller presents zeroes in the servo reaction. The zeroes are the
main reason for undesired over shoot in the servo reaction prompting motions and
the settling time is very high. The presented strategy utilized a set point channel as
appeared in Eq. (6) so as to defeat the impact of zeroes presented by the controller
in servo reaction [8].
1
F(s) = (6)
kd τi s 2 + k p τi s + 1
The suggested scope of MS is 1.2–2, which is a trade off between the speed of reaction
and powerful stability. Be that as it may, numerous multiple times, researchers think
about qualities past the most extreme point on account of precarious and incorporating
forms in order to accomplish desired speed of reaction. For the projected technique,
the loop transfer function (L) is,
L = G c (s)G p (s)
−1
Ms = max 1 + G p (s)G c (s) (8)
The performance indexes considered in the current study are integral of absolute
error (IAE), integral of the square error (ISE). Lessening the IAE and ISE is the goal
of present paper.
These are expressed as:
∞
IAE = (|E(t)|)dt (9)
0
∞
2
ISE = E (t) dt (10)
0
0.547(−0.418s + 1)
G p (s) = (12)
s(1.06s + 1)
6 A. Parvathi et al.
0.0548s 2 + 0.468s + 1
f P I D (s) = (13)
0.0433s 2 + 0.5215s + 1
The set point filter of the process is calculated as given in Eq. (6).
1
F(s) = (14)
2.9216s 2 + 3.815s + 1
0.5(−0.5s + 1)
G p (s) = e−0.7s (15)
s(0.5s + 1)(0.4s + 1)(0.1s + 1)
0.5183(−0.4699s + 1) −0.81s
G p (s) = e (16)
s(1.1609s + 1)
0.1914s 2 + 0.875s + 1
f P I D (s) = (17)
0.0974s 2 + 0.677s + 1
The set point filter of the process is calculated as given in Eq. (6).
1
F(s) = (18)
6.1703s 2 + 6.475s + 1
This procedure move capacity has a place with a mechanical contextual investi-
gation of heater steam drum and the level is constrained by controlling evaporator.
8 A. Parvathi et al.
The PI-PD controller parameter utilized by Kaya [7] is used for comparison
purpose. Controller parameters for the current strategy recommended are Kc =
1.1292, ki = 0.1744 and kd = 1.0761. In this it is used that λ = 1.500.
Figure 4 shows reactions for all plan techniques to both a change in reference and
a stage upsets with size of −0.5 at time instant(t) = 25 s. The corresponding control
signal has been shown in Fig. 5.
Execution particulars for all structure techniques are condensed in Table 2. In all
methods, for both nominal and distributed system the method that present in this
work has low values of ISE, IAE and TV.
(−s + 1) −7s
G p (s) = e (19)
s(10s + 1)
0.218s 2 + 0.112s + 1
f P I D (s) = (20)
0.084s 2 + 0.571s + 1
The set point filter of the process is calculated as given in Eq. (6).
1
F(s) = (21)
1.4285s 2 + 3.2075s + 1
Design of PID Controller for Integrating Processes … 9
The PI-PD controller parameter utilized by Kaya [7] is used for comparison
purpose. Controller parameters for the current strategy recommended are Kc = 0.119,
ki = 0.0371 and kd = 0.053. In this it is used that λ = 1.500.
Figure 6 shows reactions for all plan techniques to both one unit reference change
and a stage upsets with size of −0.5, at time instant t = 250 s. Execution particulars
for all structure techniques are condensed in Table 3. In all design methods, for both
nominal and distributed system the method that present in this work has low values
of ISE, IAE and TV (Fig. 7).
4 Conclusions
This paper portrayed the controller plan for integrating procedure with backwards
reaction and pole at RHS. The PID controller has been planned in an adjusted structure
with arrangement of forward path. The set point channel/filter additionally used to
improve the reaction of process because of step changes in reference signal. The
proposed technique is connected to the two distinctive relative investigations and their
reactions have been appeared in Figs. 2 and 3. The reaction obviously clears that the
10 A. Parvathi et al.
proposed strategy has improved reaction than the strategies utilized for comparison.
The proposed design plan is additionally simpler than other related methodologies
for integrating/capacitive type processes.
References
1. Ziegler JG, Nichols NB (1942) Optimum settings for automatic controllers. Am Soc Mech Eng
64:759–768
2. Jin QB, Liu Q (2014) Analytical IMC-PID design in term of performance/robustness tradeoff
for integrating processes: from 2-Dof to1-Dof. J Process Control 24:22–32
3. Ch, Anil, Sree RP (2015) Tuning of PID controllers for integrating processes using direct
synthesis method. ISA Trans 57:211–219
4. Ajmeri M, Ali A (2015) Simple tuning rules for integrating processes with large time delay.
Asian J Control 17:2033–2040
5. Uma S, Chidambaram M, Rao AS (2010) Set point weighted modified Smith predictor with
filter controllers for non-minimum-phase (NMP) integrating processes. Chem Eng Res Des
88:592–601
6. Pai NS, Chang SC, Huang CT (2010) Tuning PI/PID controllers for integrating processes with
dead-time and inverse response by simple calculations. J Process Control 20:726–733
7. Kaya I (2018) Controller design for integrating processes with inverse response and dead time
based on standard forms. Electr Eng 100(3):2011–2022
8. Kumar, Praveen M, Manimozhi M (2018) A new control scheme for integrating processes with
inverse response and time delay. Chem Product Process Modeling 13(4)
Environmental Feasibility Survey
of Solar Photovoltaic Cells
Sai Goutm Golive, B. Vijaya Krishna, B. Parama Sivam,
and K. Ramash Kumar
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 13
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_2
14 S. G. Golive et al.
Keywords Solar Photovoltaic Cells · Life cycle inventory (LCI) · Life cycle
assessment (LCA) · ECO-indicator · Renewable energy Systems (RES)
1 Introduction
compensation periods because of restricted sun based radiation [10]. Counting the
current examinations on life assessment investigation of PV, this examination thinks
about the cycle energy and typified energy of ascend constructs in assessing SPV
framework maintainability angles. Dissimilar to the vast majority of the investiga-
tions an incorporated way to deal with derive stage savvy discharges at all steps and
parts of SPV framework is concocted with this examination. The extent of its investi-
gation was limited to creation phase of PV frameworks barring the capacity compo-
nents similar to batteries and power modules. To measure outflows underway period
of SPV ECO-indicator instrument was utilized and useful unit is huge amounts of
CO2 per 1 kWp SPV framework creation. Rest of the paper is sorted out as includes:
Methodology received to assess life cycle emanations is portrayed in segment II.
Segment III diagrams the portrayal of 1 kWp framework for various kinds of PV
boards, body of construct and ascend constructs. The segment IV the outcomes are
introduced and examined as for generally speaking framework just as every segment.
2 Methodology Adopted
CO2 counterparts. The ReCiPe (RCP) esteems are unit results which mean the seri-
ousness to outflows with choosen it emor movement. Large data, high the rapid
articulated on effect on climate. The ECO-it device comprises off our areas/phases
of assessment. The initial stage gives data about the framework and any definitions
with respect to the frame work, practical element can be characterized to assist an
ce the professional in evaluation. The subsequent stage was the creation period of
the item/administration whose step and cycles for the frame work can be compre-
hended individually to assess coming about emanations. The ECO-it was encouraged
to incorporate various components like mass, zone and energy indifferent decimal
standards. Any product, energy and cycle in flux into the frame work or item can
be determined via five classifications of LCI gave i.e., product, energy, relocate,
handling and administration/framework (Fig. 2). When the information from LCI is
began coming about emanations can be seen as far as CO2 counter parts or ECO-it
scores. The score of each cycle or part can be utilized to assess the effect of specific
segment/measure in the entire life (Fig. 3).
3 Model Representation
In this complete scenario, life rotation discharges for a 1 kWp Solar framework is
dissected as four kinds of boards, three sorts of ascend cohesive and regularly utilized
BOS. The virtual and specialized attributes of a wide range of Solar boards utilized
Environmental Feasibility Survey of Solar Photovoltaic Cells 17
in this examination are specified in Table 1. The average LCI of SPV boards per 1m2
zone of board is embraced [9] in discussion with zone needed for 1 kWp for different
sorts of SPV boards dissected in this investigation. For the ascend framework, three
kinds of ascend courses of action are examined in particular, in-rooftop, on-rooftop
and ground surrounded constructs. As the products and zone necessity differs amid
the diverse ascend surrounded considered, the in general emanations may likewise
shift. The boundaries of the ascend constructs are specified in Table 2. Since the
existence cycle stock of three ascend construction it tends to be seen that, ground
ascend constructs are material serious and in rooftop framework can spare material
contribution surrounded rooftop tile sparing. Be that as it may, the energy repay of in
rooftop frameworks should be lower contrasted with that of on rooftop frameworks
because of skewed environmental and air refrigeration on lower edge of PV boards
[12]. To examine outflows owed to BOS of SPV framework, the framework is larger
than usual by half [11]. Thusly, outflow investigation for 1 kWp will in the long run
outcome in discharges for 1.5 KVA of inverter limit. In the association part, LCI
of Direct wiring is additionally remembered for BOS itself instead of take up as
18 S. G. Golive et al.
The 1 kWp SPV framework is planned from support to entryway regarding emana-
tions during assembling/creation phase of cell, ascending and body system of the
framework. The framework limit for module is found from wafer manufacturing
leaving Si raw material product and energy inputs.
The vitality and material concentrated assembling measure of SPV module applies
impressive test in prepare SPV a practical energy reaping innovation. The mono
translucent innovation is product just as energy serious there by bringing about most
noteworthy discharges of 1257 kg of CO2 for every 1 kWp of creation. While, strip
Si with low product and energy utilization fueled by present day creation innovation
brought about best natural alternative with 311 kg of CO2 per 1 kWp of creation
(Fig. 4).
1400
1200
EMISSIONS (Kg of
1000
800
CO2)
600
400
200
0
Mono Si Poly Si Thin film Ribbon cell
creation outflows of mono Si, emanations in wafer creation phase are simple 46.7%
and module bundling had brought about most noteworthy discharges of over 48%
of complete poly Si creation discharges. The principle distinction in the creation
measure is in to wafer phase of creation. As an aggregate decrease of 530 kg of CO2
for each 1 kWp wafer creation is seen which was proportionate to 46.7% contrasted
with 69.2% of mono Si. In this manner, poly Si wafer creation measure is naturally
22.4% suggested over mono Si wafer creation. Additionally, when contrasted with
mono Si the emanations from wafer creation measure are decreased by 7.3%, 58.7%,
7.27% and 22.88% in the event of solids, fluids, gases and power individually (Fig. 6).
(3) LCA of Thin Film CdTe:
Additionally, the slight film modules had came up long way from their presentation
in 1975. The record effectiveness of slight film CdTe modules is accounted for as
21.5% [18]. The primary bit of leeway of the flimsy film modules is the product
adequacy and wide scope of activity in both diffused and direct retention range
[19]. Notwithstanding, the effectiveness is lower contrasted with mono and poly Si
partners, there with heightening geological necessity to tackle same sum of energy
contrasted with previous PV advancements. Creation of 1 kWp of flimsy film PV
module brought about 475 kg-eq of CO2 which is 62%, 34.6% lower contrasted with
mono and poly Si advances separately (Fig. 1). The discharges from creation phase
of Cd-Te innovation are partitioned into three classes to be specific module measure,
plant compound utilization and plant equipment use. The module creation measure
involve 98% of absolute discharges while plant compound utilization and equipment
use represented 1.4%, 0.5% individually (Fig. 7). In the module cycle, material and
energy inputs involve emanation portions of 60, 40% individually.
500 465
400
300
200
100
6.6 2.8
0
CdTe PV module Plant chemical usage Plant hardware usage
300
250
Kg CO2 Eq
200
150
100
50
0
Material Electricity Emission to air
When the PV producing was finished, establishment measure is the key territory
which devours impressive measure of materials through ascend constructs [9]. The
ordinary LCI of three ascend constructs used in this examination is represented in
22 S. G. Golive et al.
160
140
120
Kg CO2 Eq
100
80
60
40
20
0
On roof Inroof Ground mounted
Table 2. In light of the undertaking type mounting constructs are separated into three
sorts specifically on-rooftop constructs which are sent in average rooftop top PV
frameworks, in rooftop constructs in which boards are coordinated to rooftop top
supplanting the tiles there by offering material reserve funds in building develop-
ment and the third sort is ground ascend constructs which require extra material as
concrete utilized for readiness of establishments. The outflows during establishment
measure are thought to be exclusively from exemplified energy of product utilized
for ascending constructs. The discharges three kinds of discharges are represented in
Fig. 9. Ground layed constructs brought about most elevated emanations of 135 kg
of CO2 –eq. thought about to 61 and 13 kg for on rooftop and in rooftop partners. The
contrast in emanations if there should be an occurrence of in-rooftop construct can
be credited to the way that, 1 m2 of PV boards conceivably can supplant 40 rooftop
tile there by decreasing epitomized energy of in general framework (Fig. 10).
LCA of generally framework To comprehend the portion of various framework
parts in emanations of entire framework consisting casing and BOS, LCA of entire
framework is to be performed. Here, LCA of entire framework is assessed for a
1 kWp framework including ascending and BOS prerequisites. The emanations of
1 kWp framework are investigated with in rooftop framework ascending for all PV
advances as appeared in Fig. 11. The emanation portion of PV module with the
entire framework is diminished from 96.3% for mono Si to 86.8% for lace Si wafer
innovation. While emanations with poly and slight film advances remained at 94%,
90% separately.
Chart Title
150
100
50
Kg CO2 Eq
0
Steel Sell convert Aluminium Concrete Roof Ɵle
-50
-100
-150
5 Conclusion
References
1. The Emissions Gap Report 2012. United Nations Environment Programme (UNEP), Nairobi.
Advancing State Clean Energy Funds Options for Administration and Funding, U.S. Environ-
mental Protection Agency’s Climate Protection Partnerships Division, May 2008
2. Available [online] http://www.epa.gov/greenpower/documents/purchasingguideforweb.pdf
3. Guide to Purchasing Green Power Renewable Electricity, Renewable Energy Certificates, and
On-Site Renewable Generation
4. Environmental management—Life Cycle Assessment—Principles and framework, ANSI/ISO
14040, American National Standard (1997)
5. Kreith F, Norton P, Brown D (1990) A comparison of CO2 emissions from fossil and solar
power plants in the United States. Energy 15:1181–1198
6. Kannan R, Leong KC, Osman R, Ho HK, Tso (2006) CPO Life Cycle assessment study of
solar PV systems: an example of a 27 kWp distributed solar PV system in Singapore. Solar
Energy 80:55–63
7. Nawaz I, Tiwari GN (2006) Embodied energy analysis of photovoltaic (PV) system based on
macro- and micro-level. Energy Policy 34:3144–3152
8. Raghu N, Kshatri SS, Boyina VK (2020) Impact of temperature and irradiance on life time of
Pv inverters. Int J Res 2:283–291
9. Knapp K, Jester T (2001) Empirical investigation of the energy payback time for photovoltaic
modules. Sol Energy 71:165–172
10. Fthenakis V, Kim HC, Frischknecht R, Raugei M, Sinha P, Stucki M (2011) Life cycle inven-
tories and life cycle assessments of photovoltaic systems. In: International energy agency
photovoltaic power systems programme, October 2011
11. Eco-indicator 99 manual for designers, A damage oriented method for life cycle impact assess-
ment, Ministry of Housing, Spatial planning and environment, The Netherlands, October 2000
(Online): www.pre-sustainability.com/download/EI99Manual.pdf
24 S. G. Golive et al.
12. Vijaya Krishna B, Venkata Prasanth B, Sujatha P (2018) Design and implementation of DPFC
for multi-bus Power system. Int J Eng Technol 7(2.8):18–29
13. Sunmodule SW 250 POLY (33mm frame). (Online): http://www.solarworld-usa.com//media/
www/files/datasheets/sunmodule-pro-series/33mm-frame/sunmodule-solar-panel-pro-series-
poly-33mm-frame-ds.pdf
14. Vijaya Krishna B, Venkata Prasanth B, Sujatha P (2018) Enhancement of power quality in
distribution system using hybrid seven level H-bridge inverter based Dpfc. J Electrical Eng
2:61–71
15. Vijaya Krishna B, Venkata Prasanth B, Sujatha P (2018) MA TLAB/Simulink study of of multi-
level inverter topologies using minimized quantity of switches. Int J Eng Technol 7(1.5):209–
216
16. Evergreensolar, ES-A Series Photovoltaic panels. [Online] http://www.evergreensolar.com/upl
oad2/InstallManuals/2011-JANES-AIMWETLSM-0029-USRevE
Intelligent Control Techniques
for Parameter Tuning of PID Controller
for LFC with Emphasis on Genetic
Algorithm
Abstract The theme of this work is carried out to reduce the deviations involved in
LFC of a developed system by via GA technique (intelligent controller) deployed PID
controller. In MATLAB environment, the system is modelled and as well as simu-
lated for conventional controller, Fuzzy, PSO and GA based PID tuning parameters.
Simulation results shows that there is an improved dynamic response of the system
when subjected to different step perturbations with proposed GA-PID controller.
Keywords Load Frequency Control · Integral Time Absolute Error · Fuzzy Logic
Controller · Particle Swarm Optimization · Genetic Algorithm · PID controller
1 Introduction
The goal of LFC during transient conditions is to maintain constant frequency, power
exchanges and tackling system complex model as well as variations involved [1–3].
In general case, the habitual apparent power produced by generators must be met in
accordance with the load power variations. The principal task before a power system
engineers is to provide good quality of power supply generated by a mixture of
renewable and non-renewable sources of energy to the utility customers without any
distortions. Load frequency control to be referred as by maintaining frequency within
permissible limits by regulating both wattage input power generation and demand
B. Venkata Prasanth
QIS College of Engineering & Technology, Ongole, India
R. Sathish Kumar
Keshav Memorial Institute of Technology, Hyderabad, India
V. K. Boyina (B)
Bapatla Engineering College, Bapatla, India
e-mail: [email protected]
K. Ramash Kumar
Department of Electrical and Electronics Engineering, Dr.N.G.P. Institute of Technology,
Coimbatore 641048, India
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 25
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_3
26 B. Venkata Prasanth et al.
[4]. The variables (f and Ptie ) effects during power transient load power condi-
tions. Automatic control is more efficient method of load frequency control because
manual controls are sluggish and involve inherent human time lags. The problem of
ALFC resolves into not only measuring f but also analysing the measured change
from a reference measurement value. Keep view on this correction is initiated by
using control strategy to keep the system original measured value. The risk of the
interconnected system increases because of system design deals with more time
varying parameters. Valid assumptions are to be incorporated in design of a system
controller to enhance overall performance. Tyreus-Luyben, Cohen-Coon, Fertik, Z-
N and Integral Control Methods, are considered for controlling the LFC of isolated
single area system [3, 5, 6]. Integral gain of conventional controller restricts the
dynamic performance of the system. An optimal control scheme based PSO which
enhances the gain of controller for addressing LFC in power systems of single area
or multi area type power systems. The authors [7, 8] proposed the AGC by using
a optimal controller for two area power system as well as compared the results.
Intelligent control technique needs to be deployed to achieve further improvement
in power system dynamics [9–12]. Among them GA has been used to deal complex
optimization problems which uses H2/H∞ Controller [6–8].
The power plant model shown in Fig. 1 consists of governor, turbine, a generator and
load with speed regulation as feedback etc., are characterized by transfer functions.
The values of gain and time constants chosen for the model are listed out in Table 1.
The state variables are listed below for the model shown above in Fig. 1.
ẋ2 = x6
1 −1
ẋ3 = x2 + x3
Tt Tt
ẋ4 = x5
−K P KP −1 KP
ẋ5 = x2 + x3 + x5 + x6
TP Tt TP Tt2 Tg TP Tt
ki k p R−1 kd −1
ẋ6 = x1 + x4 + x5 + x6
Tg Tg R Tg Tg
Genetic Algorithm:
To solve more complex problems GA isused and the process involved represented
by flow chart which is shown in Fig. 5. The procedure for minimization of perfor-
mance index which is integral time absolute error (ITAE) = ∫ t e(t)dt and frequency
deviation is an error function is obtained by reproduction, crossover and mutation.
3 Simulation Results
The dynamic response of the system is studied by applying step change in load
perturbation (PD = 0.01, 0.02, 0.03, 0.04 and 0.05) to the generator. The dynamics
are studied in terms of time domain specifications with GA, PSO and Fuzzy Logic
PID Controllers as well as concluded from the responses shown in Figs. 6, 7, 8,
9 and 10 that GA-PID controller perks up better transient behavior among all the
mentioned controllers (Table 3).
30 B. Venkata Prasanth et al.
Fig. 6 Dynamic response of isolated thermal system with a disturbance of PD = 0.01
Intelligent Control Techniques for Parameter Tuning of PID … 31
Fig. 7 Dynamic response of isolated thermal system with a disturbance of PD = 0.02
Fig. 8 Dynamic response of isolated thermal system with a disturbance of PD = 0.03
Fig. 9 Dynamic response of isolated thermal system with a disturbance of PD = 0.04
32 B. Venkata Prasanth et al.
Fig. 10 Dynamic response of isolated thermal system with a disturbance of PD = 0.05
Table 3 (continued)
Disturbance with PD = 0.01
Type of Controller Undershoot Static Error Settling Time
Disturbance with PD = 0.04
Type of Controller Undershoot Static Error Settling Time
Uncontrolled −0.12201 −0.09405 5.045
PID controlled −0.11560 0 25.310
Fuzzy-PID controlled −0.10250 0 11.49
PSO-PID controlled −0.05332 0 4.566
GA-PID controlled −0.05228 0 4.255
Disturbance with PD = 0.05
Type of Controller Undershoot Static Error Settling Time
Uncontrolled −0.15340 −0.11770 5.685
PID controlled −0.14590 0 26.451
Fuzzy-PID controlled −0.12780 0 12.760
PSO-PID controlled −0.06599 0 5.349
GA-PID controlled −0.06579 0 5.021
4 Conclusion
A single area Load frequency control problem with PID controller is modelled and
state variables are derived. Conventional controller, Fuzzy based, PSO based and GA
based PID controllers are applied to the generator to get dead beat response from
the system. These controllers are used to minimize the system transient behaviour
pertaining to step load disturbances i.e., PD (p.u.) = 0.01 to 0.05 in steps of 0.01. GA
based PID tuned controller improves the system transient behaviour in contrast with
other controllers implemented. The proposed GA controller is more robust in nature
to minimize the system transient behaviour. There is a scope for further research
to where the robustness of power system can be improved by new soft computing
techniques which deals optimization of complex problems.
References
1. Sathish Kumar R (2018) LFC problem by using improved genetic algorithm tuning PID
controller. Int J Pure Appl Math 120:7899–7908
2. Prasanth BV (2018) Tuning of PID controller by improved GA technique for an Isolated Nuclear
Power System. Int J Pure Appl Math 120:8011–8020
3. Fagna R (2017) Load frequency control of single area thermal power plant using type 1 fuzzy
logic controller. Sci J Circuits Syst Signal Process 6: 50–56
4. Vijaya Krishna B (2018) Enhancement of power quality in distribution system using hybrid
seven level H-bridge inverter based DPFC. J Electr Eng 2:61–71
34 B. Venkata Prasanth et al.
5. Prasanth BV (2015) PSO-PID tuning technique for load frequency control problem. Int J Global
Innov 3:204–209
6. Rajani C (2015) Multi-objective design of load frequency problem using genetic algorithm
based PID controller. Int J Adv Eng Global Technol 3:307–318
7. Kumar RS (2014) Load frequency control of two area power system with fuzzy logic and ANN
control. Int J Adv Eng Global Technol 2:143–147
8. Prasanth BV (2013) New control strategy for improvement of transient behavior in multi area
interconnected power system with emphasis on robust genetic algorithm. Int J Eng ResTechnol
(IJERT) 2:3989–3993
9. Subramanyam B (2009) Comparative study of different control strategies for a load frequency
problem with emphasis on new fuzzy logic controller. IJEEE Spring Edn 1:25–33
10. Vijaya Krishna B (2018) MATLAB/simulink study of multi-level inverter topologies using
minimized quantity of switches. Int J Eng Technol 7:201–216
11. Vijaya Krishna B (2018) Design and implementation of DPFC for multi-bus power system. Int
J Eng Technol 7:18–29
12. Vijaya Krishna B (2016) Designing of multilevel DPFC to improve power quality. In: Inter-
national Conference on Electrical, Electronics and Optimization Techniques (ICEEOT) @
IEEE
13. Wood AJ, Wollenberg BF (1996) Power generation operation and control. Wiley, New Jersey
14. Kennedy J (1995) Particle swarm optimization. In: Proc IEEE Int Conf Neural Netw 4:1942–
1948
Simplified Decoupler Based Fractional
Order PID Controller for Two Variable
Fractional Order Process
Abstract In recent years, more research work has been proved the advantages of
using fractional order modeling and control techniques. This paper describes the
design of fractional order proportional integral derivative (FO-PID) controller along
with simplified decoupling method of two variable fractional order processes. The
structure of two variable processes is different from single variable process. The
interaction effects occur for two variable fractional order processes. So to reduce the
interaction effect, the process is decoupled by simplified decoupling method. The
maximum sensitivity based frequency domain strategy is suitable for parameters
tuning of FO-PID controller. The described controller yield better execution for
set point-tracking. Disturbances like white noise are included in the system so as
to exhibit the FO-PID method shows better result in rejection of disturbance. The
parameter uncertainties are added to the process. The course of action is more robust
to such dissimilarities. Simulation results will exhibit the better achievement of the
proposed method.
1 Introduction
One of the oldest control method proposed in 1920s is PID control method. Mostly it
is utilized in several industrial fields because of good performance, simple design and
small settling time [1, 2]. Many of the PID control methods are modelled as integer
order system. In some cases the systems should be modelled as fractional order,
for the systems the fractional PID controller is employed. Many scientists employ
fractional order controllers for many applications to obtain desired performance of the
systems [3–5]. In 1994 I. Podlubny [6] introduced a PID controller for a fractional
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 35
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_4
36 A. Divya Teja et al.
t
−v 1
a Dt f(t) = (t − τ )v−1 f(τ )dτ (1)
(γ )
a
The fractional calculus operator a Dt−v represents fractional derivative and f(t) is
an function of integer. G(γ) defines gamma function.
Simplified Decoupler Based Fractional Order PID Controller … 37
Podulbuny has first described the FOPID controller. It was an expansion of conven-
tional PID controller using fractional calculus. The fractional order PID based control
method is presented in Fig. 1. Gp (s) is the process, Gc (s) is the FOPID controller.
By adding three control actions, the expression for the control method is obtained
and it is described in time domain below:
de(t)
u(t) = Kp e(t) + Ki e(t)dt + (2)
dt
u(S) 1
Gc (s) = Kp + Ki + Kp s (3)
e(s) s
A two variable fractional order linear system is represented by the following fractional
order differential equations:
γ
0 Dt x(t) = Ax(t) + Bu(t), (4)
Assuming initial conditions are zero and by applying Laplace transform, the
representation of state space equations can be transformed as
By substituting (4) and (5) in (6), the following equation can be obtained as:
38 A. Divya Teja et al.
Gp11 (s) G p12 (s)
−1
G p (s) = C(I−A) B = (7)
Gp21 (s) G p22 (s)
where each Gi j (s) is a fractional order transfer function and defined as follows:
where βij and αij are the fractional order, aij and bij are the coefficients. The
structure of fractional order two variable process is introduced in Fig. 1.
G12 (s)
Gl1 (s) = − (9)
G11 (s)
G21 (s)
Gl2 (s) = − (10)
G22 (s)
The output of c1 , c2 and the input u1 , u2 of G(s) have the relationship as:
G12 (s)
u1 = c1 + c2 (11)
G11 (s)
G21 (s)
u2 = c2 + c1 (12)
G22 (s)
The relationship between the inputs u1, u2 of G(s) and the output is:
A fractional order two variable processes is considered. After applying the simplified
decoupling to the two variable process, the interactions are present between two
control loops has been eliminated. The FOPID control method discussed in Sect. 2
obtained the following expressions:
1
C11 (s) = Kp1 + Ki1 + Kd1 s μ1 (20)
s λ1
1
C22 (s) = Kp2 + Ki2 Kd2 s μ2 (21)
s λ2
In controller design, parameter tuning is of great importance because of changes in
operating conditions. To solve this problem tuning the parameters based on maximum
sensitivity approach, the controller values are obtained. It is defined as
1
Ms = max (22)
ω∈[0,+∞] 1 + C(jω)G(jω)
4 Simulation Study
In this section, the TITO-FOIMC method will compare with the proposed method.
The model of simulink diagram is shown in Fig. 2.
4.1 Example 1
2s0.5 + 1
Gl12 (s) = − (24)
2 1 + 3s0.7
3s0.6 + 1
Gl21 (s) = − (25)
3 1 + s0.8
Simplified Decoupler Based Fractional Order PID Controller … 41
The TITO-FOIMC method is compare with the proposed method in this part. A
maximum sensitivity based parameter tuning method is used to design the FOIMC
controller. The maximum sensitivity value of loop1 is equal to 1, the value of Ms2 for
loop 2 is equal to 1. The transfer functions of TITO-IMC controller are given below:
2s0.5
CIMC−1 (s) = (26)
1.2(1 + 0.667s)
3s0.6 + 1
CIMC−2 (s) = (27)
1.5(1 + 0.5843s)
33.307
C11 (s) = 10.7613+ (28)
s0.8143
30.723
C22 (s) = 10.1953+ (29)
s0.7573
The output behaviour of the FO-PID method is obtained in Fig. 3. At two instants
at t = 0 s and at t = 4 s, a unit step input is applied and also for the two inputs, a
disturbance of the magnitude 0.1 are added at t = 10 s. From Fig. 3 the simplified
2.5
1.5
Output
0.5
y1 of the proposed method
y1 of the TITO FO-IMC method
0 y2 of the proposed method
y2 of the TITO FO-IMC method
-0.5
0 2 4 6 8 10 12 14 16 18 20
Time (s)
decoupling of the FO-PID controller exhibit less overshoot when compared to FO-
IMC method. The output response of the proposed method reaches the steady state
from the disturbance faster than FO-IMC method.
The time domain specifications are listed in Table 1. The values of rise time,
settling time and peak overshoot have decreased when compare to the FO-IMC
method. The presented method having quicker settling time and shows better results.
The parameter uncertainties (50%) are added to the process to test the robustness
of the system. The following results are obtained in Figs. 4 and 5. In the case of
model mismatch with parameter uncertainties, the presented method is more robust
than FO-IMC method. The output response of the proposed method will not undergo
any change for parameter uncertainties.
Sometimes external noises are included in the system while operation of process.
In order to reject noise in the system and to exhibit the better results of the presented
design, a Gaussian white noise is included to the step input. The result is obtained in
Fig. 6. From the figure, the proposed method performs well in noise rejection.
1.5
Output
y1
y1 with 50% parameter perturbation
0.5
y2
y2 with 50% parameter perturbation
0
0 2 4 6 8 10 12 14 16 18 20
Time (s)
Fig. 4 Step response of the proposed method with 50% parameter perturbation
Simplified Decoupler Based Fractional Order PID Controller … 43
2.5
1.5
Output
0.5 y1
y1 with 50% parameter perturbation
0 y2
y2 with 50% paramter perturbation
-0.5
0 2 4 6 8 10 12 14 16 18 20
Time (s)
2.5
1.5
Output
0.5
y1 of the proposed method
y1 of the TITO-FOIMC method
0 y2 of the proposed method
y2 of the TITO-FOiMC method
-0.5
0 2 4 6 8 10 12 14 16 18 20
Time (s)
Fig. 6 Tracking and disturbance rejection performance of the proposed method with white noisy
4.2 Example 2
2s0.5 + 1 −0.1s
Gl12 (s) = e (31)
2 1 + 3s0.7
3s0.6 + 1 −0.1s
Gl21 (s) = − e (32)
3 1 + s0.8
The TITO-FOIMC method is compare with the proposed method in this part. A
maximum sensitivity based parameter tuning method is used to design the FOIMC
controller. The Maximum sensitivity value of loop1 is equal to 1.2785. The value of
Ms2 for loop 2 is equal to 1.2557.
The transfer functions of TITO-IMC controller are given below:
2s0.5 + 1
CIMC−1 (s) = (33)
1.2(1 + 0.6885s)
3s0.6 + 1
CIMC−2 (s) = (34)
1.5(1 + 0.6521s)
7.35
C11 (s) = 1.2 + (35)
s0.93
5.12
C22 (s) = 1 + (36)
s0.81
The output behaviour of the FO-PID method is obtained in Fig. 7. At two instants
at t = 0 s and at t = 4 s a unit step input is applied and also for the two inputs,
a disturbance of the maitude 0.1 are added at t = 10 s. From Fig. 7 the simplified
decoupling of the FO-PID controller exhibit less overshoot when compared to FO-
IMC method. The output response of the presented method has attained steady state
from the disturbance faster than FO-IMC method. The time domain specifications
are listed in Table 2. The values of rise time, settling time and peak overshoot have
decreased when compare to the FO-IMC method. The presented method having
quicker settling time and shows better results.
The parameter uncertainties (50%) are added to the process to test the robustness
of the system. The following results are obtained in Figs. 8 and 9. In the case of
model mismatch with parameter uncertainties, the presented method is more robust
than FO-IMC method. The output response of the proposed method will not undergo
any change for parameter uncertainties.
Sometimes external noises are included in the system while operation of process.
In order to reject noise in the system and to exhibit the better results of the presented
Simplified Decoupler Based Fractional Order PID Controller … 45
2.5
1.5
Output
0.5
y1 of the proposed method
y1 of the TITO-FOIMC method
0 y2 of the TITO-FOIMC method
y2 of the proposed method
-0.5
0 2 4 6 8 10 12 14 16 18 20
Time (s)
Table 2 Time domain specifications of proposed and FO-IMC method for time delay process
S. No Control method Rise time Peak time Settling time Overshoot (Mp )
(t r ) (t p ) (t s )
1 FO-IMC Y1 1.1056 6 15.9399 9.6176
method Y2 0.7162 16.5009 15.6956 4.6633
2 Proposed Y1 0.4491 2.2389 15.6025 3.5137
method Y2 0.5921 8.0387 9.9982 0.3095
design, a Gaussian white noise is included to the step input. The result is obtained in
Fig. 10. From the figure, the proposed method performs well in noise rejection.
5 Conclusion
2.5
1.5
Output
0.5 y1
y1 with 50% parameter perturbation
0 y2
y2 with 50% parameter perturbation
-0.5
0 2 4 6 8 10 12 14 16 18 20
Time (s)
Fig. 8 Step response of the proposed method with 50% parameter perturbation with time delay
2.5
1.5
output
0.5
y1
y1 with 50% parameter perturbation
0 y2
y2 with 50% parameter perturbation
-0.5
0 2 4 6 8 10 12 14 16 18 20
Time (s)
Fig. 9 Step response of the TITO-FOIMC method with 50% parameter perturbation with time
delay
better than FO-IMC method. In few words, the FO-PID controller presented in this
paper has produced robust performance for both set point-tracking and rejection in
disturbance. The proposed method will extend to three variable FO-TITO process
for future work.
Simplified Decoupler Based Fractional Order PID Controller … 47
2.5
1.5
Output
0.5
y1 of the proposed method
y1 of the TITO-FOIMC method
0 y2 of the TITO-FOIMC method
y2 of the proposed method
-0.5
0 2 4 6 8 10 12 14 16 18 20
Time (s)
Fig. 10 Tracking and disturbance rejection performance of the proposed method with white noisy
References
1. Xue Y, Zhao H, Yang Q (2008) Self-tuning of PID parameters based on the modified particle
swarm optimization, IEEE
2. Maiti D, Acharya A, Chakraborty M, Konar A (2005) Tuning PID and PIDδ Controllers using
the integral time absolute error criterion, IEEE, Aug 2005
3. Monje CA, Vinagre BM, Feliu V, Chen Y (2008) Tuning and auto-tuning of fractional order
controllers for industry applications. Control Eng Prac 16(7):798–812
4. Chen Y, Bhaskaran T, Xue D (2008) Practical tuning rule development for fractional order
proportional and integral controllers. J Computat Nonlinear Dyn 3(2):021403
5. Zhao C, Xue D, Chen Y (2005) A fractional order PID tuning algorithm for a class of fractional
order plants. In: 2005 IEEE international conference mechatronics and automation, vol 1. IEEE,
pp 216–221
6. Podlubny I (1994) Fractional-order systems and fractional-order controllers. Inst Exp Phys
Slovak Acad Sci, Kosice
7. Polubny (1999) Fractional-order systems and PIλ Dμ controller. IEEE Trans Automatic Control
44:208–214
8. Monje CA, Vinagre BM, Feliu V (2008) Tuning andauto-tuning of fractional order controllers
for industry applications. Control Eng Pract 16(7):798–812
9. Samko SG, Kilbas AA, Marichev OI (1993) Fractional integrals and derivatives. Theory and
applications, Gordon and Breach, Yverdon
10. Li Z, Chen YQ (2014) Ideal, simplified and inverted decoupling of fractional order TITO
processes. In: IFAC Proceedings Volumes 2014, vol 47, no 3, pp 2897–2902
Improved Centralized PID Controller
with Disturbance Rejection for LTI
Multivariable Processes
S. Anitha, R. Kiranmayi, K. Nagabhushanam,
and K. Firdose Kowser Ahamadia
1 Introduction
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 49
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_5
50 S. Anitha et al.
and Multi Output (MIMO) processes. Because of the presence of associations and
coupling in MIMO processes, there are a few problems to increase the difficulty of
designing use of SISO structures into design of MIMO structures directly [1].
In multiple loop control approach, the MIMO techniques are treated as combi-
nations of multiple-single controlling loops, and a main controller is organized and
designed on each controlling loop by in view of loop variable associations [2]. Due
to their reasonable sensitivity, and structure straight and direct forwardness, multiple
loop controlling has been a lot of acknowledge by controlling loop control industry
and broad attempt has been determined to get better the introduction of multiple-loop
PI/PID controllers. Different control structure, such as, detuning factor procedures,
chronological loop closing systems, self-governing control methods and decentral-
ized control methodologies, etc., have been proposed consistently. Exactly when
the associations along with different variables are nominal a multiple-loop PID/PI
controller is conventionally enough, in such case, the off diagonal- decentralized
control approach procedure may not succeed to give acceptable reaction if there be
present genuine circle interchanges [3].
Decoupling (Off-diagonal) control is a popular method for MIMO processes to
take out the impacts of process variable collaborations. This at last builds the diffi-
culty in designing a controller. In decoupling (Off-diagonal) PID-like controller have
intended for the MIMO systems directly with straightforwardness procedure as in
case of SISO systems. The methodology describes how to decouple (off-diagonal
elements) and control approach of the systems in a single step. Despite the fact that
decoupling control is a viable approach in favour of MIMO type processes, outer
disturbances actually have consequences for control execution of the structure. To
acquire a superior in the interruption limination, disturbance observers (DOB) be
presented [1]. DOBs have been recognized for their capacities in minimizing the
effect of disturbance.
Accordingly, various logical decentralized control schemes are used to tune and to
make sure system steadiness, which basis design procedure more simple and reduces
costs.
For the MIMO structures with extraordinary loop variable associations, the decou-
pling (off-diagonal) control format shave been regularly utilized [4]. The decoupling
(off diagonal) control as a rule necessitates two phases: (1) arrangement of decoupler
off diagonal elements to restrict the interrelation among control loops, (2) preparation
of the standard loop regulators for as a rule for carrying out control algorithm.
The important purposes of the decoupling (Off-diagonal) control approach are,
(1) it allows the usage of SISO control strategies, (2) if there ought to be an event
of sensor issues, it is decently effortless to adjust loop physically, adjust individual
loop is reasonably exaggerated by the breakdown. Regardless, such an arrangement
may outcome especially stable control composition, especially, when the structure
estimation is high. Thus, the large numbers of investigators are generally focused on
TITO structures.
For more number of control variables methods, Wang et al. [5]projected certain
procedures for fully coupled PID or other than PID regulators with deliberate arrange-
ment strategy and for the most part system structures [6]. Regardless, owing to
Improved Centralized PID Controller with Disturbance … 51
the mathematic assessment of the decouplers (off diagonal) for large number vari-
able system, the vitality of like a control structure can’t be definite. Subsequently,
stationary decoupling come close to which can in a general sense improve control
shows with their robustness like so as to of multiple-loop control structures are
obtained beside the production to control loop relationship for large variable MIMO
structures.
Decoupling systems only may not give the guaranteed stability in a control
strategy. In spite of the way that static off diagonal element is simple but difficult
to design and complete, they may possibly not commonly give sensible robustness
of loop but there subsist significant interactions. However, a combination of decou-
pler with diagonal PID will give a centralized control scheme and will improve the
response of the system.
In this projection, another off diagonal control procedure [7] is along with
frequency response based PID controller is anticipated. The primary objective of
this direct methodology is to get better the trouble dismissal when arrangement has
serious outside disturbance influences, and to simplify the control algorithm.
The organization of the proposal is structured as: Section 2 discusses the illus-
tration of IMC based centralized control system. Section 3 depicts the centralized
control arrangement design using frequency response approach, Section 4 presents
the simulation of case studies and projected algorithm justification, and summary
are depicted in Sect. 5.
A graphic description of broad and projected closed loop IMC control arrangement of
multiple control variable strategies are showed up in Figs. 1 and 2 independently. In
projected approach, the diagonal regulator is used for excusal of loop internal variable
combinability and desirable set value/reference feed is for next the set value. The
parts exist in projected control system are decentralized- main loop controller Gc(s),
Gm(s) is inside model and process Gp(s).
For ease, it is acknowledged that the number of restricted variables in multiple
variable techniques is comparable to numbered controlled components. This allows
mixing of single (main) variable control with a single variable in the course of an
investigation of controller.
In common the basic exhibition of the transfer function of the 2 × 2 (TITO)
process is:
Gp11 (s) Gp12 (s)
Gp (s) = (1)
Gp21 (s) Gp22 (s)
Where, the process G pi j (s) is embedded as first order process with delay time
(FOPDT), i.e.,
K
Gpij (s) = e - θs (2)
τs + 1
The RGA of the Eq. (1) can be written as:
y1 (s) u 1 (s)
Y (s) = , U (s) = (6)
y2 (s) u 2 (s)
Here, Y(s), U(s) are specifies an output, input vectors of the process correspond-
ingly. The large variable processing out variable association for the process with two
variables is able to be given as:
In the TITO (two variable) structure, while the resulting loop is off (closed), the
commitment from ui to yi has two different ways of sign transmission is feasible.
The grouping of two conduction ways is measured as productive open control loop
components. In case the later analysis controller would be in modified configuration,
with ys2 = 0, by then the as a rule closed loop model everywhere in the scope of y1
and u1 is known by
Here, G11 and G22 are the incredible open-loop effective transfer function (EOTF).
These EOTFs are confounded models, and it is difficult to clearly use them for the
controller structure. The decouplers are resolved for the off diagonal positions as:
G p,12 (s)
D12 (s) = − (14)
G p,11 (s)
G p,21 (s)
D21 (s) = − (15)
G p,22 (s)
The diagonal controller and off diagonal decouplers together forms a centralized
control system.
The most of the industrial processes are represented by second order plus delay model
as,
b0,i j
G i j (s) = e−θi j s (16)
a2 i j s2 + a1 i j s + 1
G i j (0)
Ĝ i j (s) = e−θ̂i j s (17)
a2 i j s 2 + a1 i j s + 1
ki,i
G c,i (s) = k p,i + + kd,i s (18)
s
The controller preserved as,
Ĝ i j (0) −θ̂i j s
G c,i (s)Ĝ i j (s) = k e (20)
s
Abbreviating the gain (GM) and phase margin (PM) provisions as Am,i and ψm,i ,
and their intersect frequencies as ωg,i and ω p,i , correspondingly.
gc,i ( jω p,i )ĝii ( jω p,i ) = 1 (23)
By simplification we obtain,
π ωg,i
ωg,i θ̂i j = Am,i =
2 k Ĝ i j (0)
π
k Ĝ ii (0) = ω p,i ψm,i = − ω p,i θ̂i j
2
Which results,
π 1 π
ψm,i = (1 − ), k=
2 Am,i 2 Am,i θ̂i j Ĝ i j (0)
56 S. Anitha et al.
As a result of this plan, the increase and phase (PM) and gain margins (GM)
are consistent to one another, some conceivable phase and gain margin choices are
provided in Table 1.
The controller considerations are obtained by,
⎡ ⎤ ⎡ ⎤
k p,i a1,i j
⎣ ki,i ⎦ = π ⎣ 1 ⎦ (25)
kd,i 2 Am,i θ̂i j Ĝ i j (0) a
2,i j
The complete design steps of proposed centralized control system are given below.
Step 1: Choose the minimum phase multivariable process and get the mathemat-
ical model of the process Gp (s).
Step 2: Determine the control loops by using relative gains of the variables in the
process as in Eq. (3).
Step 3: Approximate the mathematical model of equivalent processes and
determine the decoupler elements in off diagonal positions as in Eq. (16).
Step 4: Design the PID controllers of main diagonal using the frequency response
tuning method as given in Eq. (25).
Step 5: An exhibition of the projected unified control is assessed regarding indis-
pensable of integral over time of working time, i.e., necessary of IAE (Integral Abso-
lute Error), ISE (Integral Square Error) as given in Eqs. (25) and (26). The lesser the
estimations of these IAE and ISE shows better exhibition of control algorithm.
The expressions for two variable-TITO process are:
∞
4 Simulation Results
12.8e−3s −18.9e−3s
G p (s) = 16.7s+1 21s+1
6.6e−7s −19.4e−3s (28)
10.9s+1 14.4s+1
Since it is 2 × 2 process, there two loops should be controlled. The control loops
are determined by RGA method. Based on the outcome of RGA the pairing of control
loops have selected. The RGA of the given process is,
2.0094 −1.0094
RG A = (29)
−1.0094 2.0094
From the RGA [8, 9], it is concluded that the diagonal loops are need to have
controller for improving the performance of the process. The effective models of the
corresponding processes are:
1.3046s+0.0781
0
G c, pr oposed (s) = 10s+1
−0.7422s−0.0515 (32)
0 6s+1
(189s+18.9)e−2s
1 (21s+1)
D(s)inver ted = (−39.66s−6.6)e−6s (33)
(10.9s+1)
1
enhanced the whole performance of the two variable process by attenuating influence
of disturbance. The proposed method shows IAE for set point is 4.267 and 5.892
respectively for two channels. Similarly IAE for disturbance is 10.528 and 46.251
for the two channels respectively. These values show improvement as compared to
existing method used for comparisons.
Example 2: The transfer function of the Shell Heavy Oil process (HOP) by [1] is
known as:
4.05e−27s 1.77e−28s
G p (s) = 27s+1 21s+1
5.39e−18s 5.72e−14s (34)
50s+1 60s+1
Since it is 2 × 2 process, there two loops should be controlled. The control loops
are determined by RGA method. Based on the outcome of RGA the pairing of control
loops have selected. The RGA [9] of the given process is,
1.7002 −0.7002
RG A = (35)
−0.7002 1.7002
From the RGA, it is concluded that the diagonal loops are need to have
controller for improving the performance of the process. The effective models of
the corresponding processes are:
0.6663s+0.2469
0
G c, pr oposed (s) = 19s+1
10.488s+0.1748 (38)
0 26s+1
Improved Centralized PID Controller with Disturbance … 61
(−33.63s−1.77)e−2s
1 (60s+1)
D(s)inver ted = (−140.14s−5.39)e−4s (39)
(50s+1)
1
The proposed method shows IAE for set point is 31.159 and 29.357 respectively
for two channels. Similarly IAE for disturbance is 135.226 and 198.498 for the two
channels respectively.
5 Conclusions
The present paper discusses the design of controller for multi variable process. The
proposed method improves the set point response as well as disturbance response of
the processes. The pairing of multivariable systems has been decided by relative gains
of corresponding transfer function models of the given multivariable systems. The
64 S. Anitha et al.
References
1. Hao Y et al (2020) Decoupling controller with disturbance observer for LTI MIMO systems.
Asian J Control 22(2): 831–840
2. Wood RK, Berry MW (1973) Terminal composition control of a binary distillation column.
Chem Eng Sci 28(9):1707–1717
3. Zhang W (2011) Quantitative process control theory, vol 45. CRC Press, Boca Raton
4. Xiong Q, Cai W-J, He M-J (2005) A practical loop pairing criterion for multivariable processes.
J Process Control 15:741–747
5. Wang Q-G, Huang B, Guo X (2000) Auto-tuning of TITO decoupling controllers from step
tests. ISA Trans 39(4): 407–418
6. Vu TNL, Lee M (2009) Independent design of Multiloop PI/PID controllers for Multi-delay
processes, World Academy Science, Engineering and Technology (60):703–708
7. Tavakoli S, Griffin I, Fleming PJ (2006) Tuning of decentralised PI (PID) controllers for TITO
processes. Control Eng Pract 14(9):1069–1080
8. Seborg DE et al (2010) Process dynamics and control. Wiley
9. Hanuma NR, Ashok Kumar DV, Anjaneyulu KSR (2014) Control configuration selection and
controller design for multivariable processes using normalized gain. World Acad Sci Eng
Technol Int J Electr Comput Electron Commun Eng 8(10)
10. Shen Y, Sun Y, Li S (2012) Adjoint transfer matrix based decoupling control for multivariable
processes. Ind Eng Chem Res 51(50):16419–16426
Tuning of PID Controller Using SIMC
Method for Systems with Time Delay
and RHP Poles
P. Keerthi, R. Kiranmayi, K. Nagabhushanam,
and K. Firdose Kowser Ahamadia
Abstract To control a class of system with RHP poles and time delay, a revised
cascade control model is proposed in this work. This revised cascade control model
aims to enhance both regulatory and servo performance of the system. Structurally,
this proposed revised cascade control scheme has three controllers. To minimize the
set point variations, an auxiliary controller is also used. An analytical method based
on IMC design principle is used in designing the controller for set point tracing.
It is developed using H2 optimization and Internal Model Control theory. Distur-
bance rejecter is developed as P + I + D controller. To make the set point tracking
controller physically realizable, a low pass filter is cascaded with it. Simplified
Internal Modal Control (SIMC) with only one adjustable parameter ‘Tc ’ is used
for tuning the controllers. The simulation studies for proposed approach are done
using MATLAB/SIMULINK to demonstrate its effectiveness. This proposed revised
cascade control scheme depicts 3-Degree of Freedom Control Structure.
1 Introduction
Processes with RHP poles are unstable in nature and are very difficult to control
when compared to stable systems. This RHP pole may lead to large setting time
and increased overshoot. In many systems, this unstable pole may also create an
imbalance between the input and the output due to some external disturbance [1].
Therefore the study of the processes with RHP poles and time delay are carried out
by some authors in past. In the initial stage, studies were focused mainly on tuning
methods [2]. Chidambaram along with his co-workers proposed several techniques
for systems having delay and RHP poles. They developed P + I controller using
Ziegler-Nichols tuning formulae [3]. A formula for P and P + I controllers using
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 65
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_6
66 P. Keerthi et al.
IMC principle is explained in [4]. To control a class of linear and non-linear systems,
an Internal Model Controller cascaded with first order filter was suggested [5]. Anvil
[6] developed a P + I + D controller using Internal Model Control theory and
H2 optimization technique for process with RHP poles. In [7, 8] revised form of
SP structure based on 2-DOF control scheme is shown. But in general, one loop
control method offers a satisfied output for systems with delay in few cases. The
cascade control technique has an ability to reject the load disturbances very fast
which enhances the closed loop performance of the system. Generally, there are two
control loops in a cascade structure, one is secondary and the other is primary loop.
Reducing the disturbances in the inner loop itself and not letting them enter into the
primary loop is the main concept behind the cascade control technique [9]. A general
cascade control scheme which was derived from Internal Model Control P + I + D
method was proposed by Lee and Oh [10].
Cascade control structure alone might not give an effective regulatory control for
systems with RHP poles whose control quality should be held at a peek level. Smith
predictor (SP) can compensate the adverse effects of time delay. So combinations of
cascade control scheme and SP have the ability to achieve a better control perfor-
mance [11]. A cascade control scheme is derived from optimal Internal Model Control
technique for control of systems with RHP poles with combination of controllers
and filter was proposed by Dasari et al. [12]. In [13] the disturbance rejecter and set
point tracking were developed using Internal Model Control. The main motive of
the present study is to enhance both disturbance rejection and set point tracking. To
achieve this we are using SIMC tuning method with only one adjustable parameter
‘Tc ’. This SIMC tuning method reduces recovery time and enhances the regulatory
performance of the system.
The proposed revised cascade control scheme for systems consisting RHP poles is
displayed in Fig. 1. In Fig. 1, P2m and P1m are the process models of the primary
unstable systems and intermediate stable system, P1 and P2 respectively. P1mo and
P2mo are the perfect transfer function models of P1m and P2m respectively without
any time delay; i.e., P1m =P1mo e−θ1 mS and P2m =P2mo e−θ2 mS
Proposed revised cascade control model has three controllers namely, the set point
tracker C, inner loop load disturbance rejector F1 , and the outer loop load disturbance
rejector F2 . To minimize the set point output variations an auxiliary controller set Pc
is installed. F1 rejects the disturbances that are introduced in the intermediate system
and F2 rejects the disturbances that are introduced in the primary systems. Note that,
the auxiliary controller Pc has no effect on the controllers C, F1 and F2 . This revised
cascade control technique depicts the 3-DOF control scheme. This method offers a
better control in the model parameter error case.
The intermediate system is stable in nature,
Tuning of PID Controller Using SIMC Method … 67
Considering the ideal case, i.e.,P1m and P2m are exact process models without any
time delay. Therefore, set point transfer function is simplified
C P1 P2
Hr =
1 + Pc P1mo P2mo
K 1 K 2 Ce−(θ1 +θ2 )S
= (4)
T1 T2 S2 + (T2 − T1 )S + K 1 K 2 PC − 1
Case 1: T1 < T2 , consider Pc = K c and (K c > 1/K 1 K 2 ) Hence, C.E of set point
transfer function becomes T1 T2 S 2 + (T2 − T1 )S + K 1 K 2 K c = 0, is clearly stable as
per RH Criterion for stability.
Case 2: T1 >= T2 , consider Pc = K c + K d S, (K c > 1/K 1 K 2 ) K d >
(T1 − T2 )/(K 1 K 2 ). Hence, C.E of set point transfer function becomes T1 T2 S 2 +
(K 1 K 2 K d + T2 − T1 )S + K 1 K 2 K c − 1 = 0 which is stable as per RH criterion for
stability.
Note that,Pc can designed as a regular P + I + D controller but it may result in
complex tuning procedure. So, it is not suggestible.
If the set point tracker C is developed as stable and rational then there exist zero dead
time element in the denominator of Eq. 4,which
may yield a better set point response.
We use H2 performance measure min e22 i.e., the integral square error (ISE) in
designing optimal set point tracker C, which should meet the min W (1 − Hr )22
condition.
Here W represents set point input weight function.
Using Pade’s approximation for delay term in Eq. 4 gives
where,
n
(2n − j)!n!
Q nn [(θ1 + θ2 )S] = ∗ [(θ1 + θ2 )S] j
j=0
(2n)! j!(n − j)!
T1 T2 S 2 + (T2 − T1 )S + K 1 K 2 K c − 1
C1m(s) =
K1 K2
T1 T2 S 2 + (T2 − T1 )S + K 1 K 2 K C − 1
C(S) = (5)
K 1 K 2 (TC S + 1)2
T1 T2 S 2 + (K 1 K 2 K d + T2 − T1 )S + K 1 K 2 K c − 1
C(S) = (6)
K 1 K 2 (TcS + 1)2
Remark 1: Substitute Eq. 5 or 6, (if T1 >= T2 ) in Eq. 4, then practical form of set
point transfer function is derived as,
1
Hr (S) = e−(θ1 +θ2 )S (7)
(Tcs + 1)2
0 , t ≤ θ1 + θ2
Yr (t) = −(t−θ1 −θ2 )/Tc (8)
1− 1+ t
Tc
e , t > θ1 + θ2
There exist zero overshoot in ideal set point output and by tuning Tc quantitative
time domain specifications are achieved.
From Eq. 8, tuning formula can be figured as tr = 3.8897Tc + θ1 + θ2 where tr
is the rising time.
From Eqs. 7 and 8, it is clear and obvious that Pc does not affect the set point tracker
C. In ideal case, i.e., when Tc is adjusted to zero, the set point response Hr (S) =
70 P. Keerthi et al.
e−(θ1 +θ2 )S i.e., system output reaches the desired value for set point immediately after
the overall systems’ delay time. This single adjustable parameter Tc gives the trade off
among the ideal performance of set point output and the output capacity of set point
tracker C. Therefore, when Tc is adjusted to small value, the setpoint tracing becomes
faster and the output of C is larger resulting in more aggressive dynamic behaviour.
When Tc is adjusted to a higher value, the set point tracing becomes slow, and the
output of C is small resulting in less aggressive dynamic behaviour. Therefore, it is
suggested to adjust ‘T C ’ to a value which approximately equals the system delay
initially. If tuning is unsatisfactory, then decrease/increase Tc to achieve required set
point response.
In order to make the design procedure simpler, leave the primary outer loop and
consider the inner loop load disturbance output independently. This results in degra-
dation of actual system performance, due to load disturbance response oscillation
may seep into the inner loop.
From Fig. 1, the ideal load disturbance process models of inner loop are given as
y1 P1 (S)
Hd1 (S) = = (9)
d1 1 + F1 (S)P1 (S)
y1 1
Hd2 (S) = = (10)
d2 1 + F1 (S)P1 (S)
f1 F1 (S)P1 (S)
Td−inner = = (11)
d1 1 + F1 (S)P1 (S)
1 (−θ1s )
Td−inner (S) = e (12)
T f1 s +1
Td−inner (S) 1
F1 (S) = (13)
1 − Td−inner (S) P1 (S)
y2 1
Hd3 (S) = = (14)
d3 1 + F2 (S)P1 (S)P2 (S)
aS + 1 −(θ1 +θ2 )S
Td− outer (S) = 3 e (16)
T f2 S + 1
To reject the step load disturbances, the following asymptotic constraint is required
3
T f2 (θ1 +θ2 )/T2
a = T2 +1 e −1 (18)
T2
from Eqs. 15, 16 and 18, the required load disturbance is obtained as,
At S = 1/T2 , Eq. 19 has a zero-pole cancellation on Right Hand Plane, this may
result in unreliable working of F2 .
So, Maclaurin expansion formulae are used to regenerate the ideal disturbance
rejector.
Say F2 (S) = M(S)/S,
1 M (0) 2 M i (0) i
F2 (S) = M(0) + M (0)S + S +.+ S (20)
m 2! i!
1
F2−P I D (S) = K F + + TD S (21)
T1 S
d3 S 2 + d2 S + d1 d0
F2−3/3 (S) = + (22)
C2 S + C1 S + 1
2 S C2 S + C1 S + 1
2
The 1st part is a lead-lag controller of second order, whereas the 2nd part represents
an integrator cascaded with second order low pass filter.
m 1 (S) = e−(θ1 )S − 1
Substitute Eqs. 16 and 18 in Eq. 23 which gives the robust stability constraint.
3
T2 T2
+1 e (θ1 +θ2 )/(T2 −1)S
+ 1
T2 1
3 < (27)
m 2 (S)∞
T f2 S + 1
∞
5 Simulation Examples
Three unstable cascade processes are simulated. For numerical comparison, perfor-
mance measure Total Variation (TV), Integral Absolute Error (IAE) and Peak Value
(PV) of the system output are considered.
∞
IAE = |e(t)|dt
0
74 P. Keerthi et al.
N
TV = |u n+1 − u n |
n=0
Considering the above controller values, the performance of the two techniques
are simulated by enabling unit step change at t = 1 s in the set point and applying
negative step change at t = 150 s in disturbance d1 and negative step change at t =
350 s in disturbance d2 respectively. System output and control signals are depicted
in Figs. 2 and 3 for perfect model case.
To analyse the system robustness, assume a 20% raise in intermediate process
gain along with 20% raise in intermediate delay time. The system output is given in
Fig. 4 and control action response is given in Fig. 5. On observing Figs. 4 and 5, the
proposed technique provides gives better closed loop response and the control action
undergoes several oscillations at first which later becomes smoother obviously. The
performance measures for proposed method and other method are tabulated below
for both perfect and perturbed cases. From Table 1 the proposed method has lower
IAE and TV than Cheng-Qiang technique. Therefore, the proposed technique gives
an enhanced performance than the Cheng-Qiang technique.
Table 1 Performance
Method IAE PV TV
measures of Example 1
Perfect model parameter
Cheng-Qiang 47.4591 1.01 39.65
Proposed 25.1821 1.26 31.45
Error model parameter
Cheng-Qiang 55.81 1.52 62.74
Proposed 27.6376 1.0 59.28
76 P. Keerthi et al.
Table 2 Performance
Technique IAE PV TV
characteristics of Example 2
Perfect model parameter
Cheng-Qiang 28.3552 1.05 23.91
Proposed 24.2949 1.06 18.26
Error model parameter
Cheng-Qiang 31.10 1.16 33.53
Proposed 28.3028 1.0 27.83
Table 3 Performance
Technique IAE PV TV
characteristics of Example 3
Perfect model parameter
Cheng-Qiang 4.1129 1.05 50.9
Proposed 3.4976 1.0 43.73
Error model parameter
Cheng-Qiang 4.14 1.05 76.23
Proposed 3.7929 1.0 64.154
6 Conclusion
This paper work proposes a revised cascade control method for systems with RHP
poles and time delay. The set point tracker C is developed derived from H2 optimiza-
tion and Internal Model Control theory. The disturbance rejector F2 is developed as a
P + I + D controller. A filter is cascaded in series with controller C to make it phys-
ically realizable. SIMC tuning method is used to tune the controllers. SIMC has ‘Tc ’
as variable measure which gives the trade-off among robustness and performance.
MATLAB simulated results depicts that suggested technique gives enhanced outputs
for both rejection of disturbance and also set point tracing. Thus, both regulatory and
servo performances are enhanced. This method also yields lower IAE, PV and TV
values.
References
1. Liu T, Gao FR (2011) Enhanced IMC design of load disturbance rejection for integrating and
unstable processes with slow dynamics. ISA Trans 50(2):239–248
2. Chidambaram M (1997) Control of unstable systems: a review. J Energy Heat Mass Transf
19:49–56
3. Kavdia M, Chidambaram M (1996) On-line controller tuning for unstable systems. Comput
Chem Eng 20(3):301–305
4. Jacob EF, Chidambaram M (1996) Design of controllers for unstable first-order plus time delay
systems. Comput Chem Eng 20(5):579–584
5. Saxena S, Hote YV (2017) Internal model control based P+I+D tuning using first-order filter.
Int J Control Autom Syst 15(1):149–159
6. Vanavil B, Anusha AVNL, Perumalsay M, RaoAS (2014) Enhanced IMC-P+I+D controller
design with lead lag filter for unstable and integrating processes with time delay. Chem Eng
Commun 201(11):1468–1496
7. Zhang WD, Gu DY, Wang W, Xu XM (2004) Quantitative performance design of a revised
Smith predictor for unstable processes with time delay. Ind Eng Chem Res 43(43):56–62
8. Rao AS, Rao VSR, Chidambaram M (2007) Simple analytical design of revised Smith predictor
with enhanced performance for unstable first-order plus time delay (FOPTD) processes. Ind
Eng Chem Res 46(13):4561–4571
9. Vijaya Krishna B, Venkata Prasanth B, Sujatha P (2018) Design and implementation of DPFC
for multi-bus Power system. Int J Eng Technol 7(2.8):18–29
10. Lee YH, Oh SG (2002) Enhanced control with a general cascade control structure. Ind Eng
Chem Res 41(11):2679–2688
80 P. Keerthi et al.
11. Vijaya Krishna B, Venkata Prasanth B, Sujatha P (2018) Enhancement of power quality in
distribution system using hybrid seven Level H-bridge inverter based Dpfc. J Electrical Eng
2:61–71
12. Vijaya Krishna B, Venkata Prasanth B, Sujatha P (2018) MATLAB/Simulink study of multi-
level inverter topologies using minimized quantity of switches. Int J Eng Technol 7(1.5):209–
216
13. Yin C-Q, Wang H-T, Sun Q, Zhao L. Enhanced cascade control system for a class of unstable
processes with time delay
14. Liu T, Zhang W, Gu D (2005) IMC-based control strategy for open-loop unstable cascade
processes
Independent Controller Design
for Non-minimum Phase Two Variable
Process with Time Delay
Abstract Non-minimum phase (NMP) second order with time delay Process
(SOPTDP) are stabilized by tuning process. Certain restrictions on the attainable
performance when system have right hand side zeros of s-plane. The PID controller
had been designed by Enhanced IMC method. The proposed Proportional Integral
and Derivative controller (PID) are implemented to non-minimum phase systems like
Fermenter, Isothermal CSTR. The performance was analysed under model uncer-
tainty and considering perturbations in one model parameter. The performance of
the system under recommend method compare with the existing method in the liter-
ature. From the performance detected as present Approach effectively stabilize and
advances the performance of process with disturbances. From simulated result of
three examples demonstrates present work offers better-quality presentation of the
set-point track and Hload rejection than existing method through better time domain
specifications.
Keywords Enhanced IMC · NMP · Second Order plus Time Delay systems
(SOPTDS) · Fermenter and Isothermal CSTR
1 Introduction
PID controller was a naturally uses controller amongst industries. Valuation of PID
controllers justifies conversation now. Normal PID controllers were exchanged by
electronic controller. These offers various improvements than earlier controller.
Thereafter, electronic controller is substituted with computer programs. presently
PID controller are software programs in place of physical controllers. Dealing
unstable processes were at all times difficulty with comparison to the stable processes.
Regrettably, various chemical processes are unstable processes. Some samples were
bottom level control of distillation column, chemical reactors, bioreactors, etc. The
difficulty controlling becomes higher though the processes have time delay. Time
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 81
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_7
82 C. Venkatasuresh et al.
delay may be due to process lag. The goal has to develop a simple controller for
stable/unstable SOPTD systems [1–5]. A tuning parameter selection depends on
maximum sensitivity (Ms) function. Set point filter was used to decrease overshoots
in servo problem. “Many Approaches are used to construction of PI/PID controller
for stable NMP Processes are stated in literature as Internal Model Control approach
(IMC) [6, 7], “PM and GM” [8], “optimization approach” [9, 10], control structure
approach [11–13] etc.
Here numerous approaches existing in literature to construct PI/PID controllers
to NMP FOPTD systems. But a very few approaches were offered to NMP SOPTDS
processes.
2 Proposed Method
Therefore, in this study, a SOPDS process model G p (s) given by Eq. 1 is considered
for the construct of PID controller G c (s)
k p (u − ts) −ds
G p (s) = e (1)
ps 2 + qs + r
C(s) e−ds
= Z (s) = (3)
R(s) (1 + αs)2
From the closed loop transfer function controller transfer function derived as
1 Z (s)
G c (s) = (4)
G p (s) 1 − Z (s)
ps 2 + qs + r e−ds
G c (s) = (5)
k p (1 − ts) (1 + αs)2 − e−ds
A(s)
G c (s) = (8)
s
For the ideal form of PID apply Maclaurin series expansion theorem then
1 A(0) 2
G c (s) = A(0) + A(0) s + s + ... (9)
s 2
1
G c (s) = kc 1 + + τD s (10)
τI s
The controller parameters are assessed by comparing the constants of ‘s’ of. The
following terms for controller parameters are gotten and written as
kc A(0)
kc = A(0) , τ I = , τD = (11)
A(0) 2kc
3 Selection of α
For the controller design the tuning parameter α is required. A suitable tuning param-
eter α is chosen by the use of maximum sensitivity function. The chosen tuning
parameter for attained controller should offer better performance and robustness.
For smaller α value of gives a faster response and displays improved results in the
case of load disturbance for the processes. Ms is defined as
1
Ms = (12)
1 + G p (s)G c (s)
84 C. Venkatasuresh et al.
Ms value uses to determine a range for the Gain Margin (GM) and Phase Margin
(PM) and relation between Ms and GM and Ms and PM is given below.
Ms 1
GM ≥ ; P M ≥ 2 sin−1 (13)
Ms − 1 2Ms
4 Simulation Studies
Three cases of NMP SOPDT Process were taken to the Valuation of present approach.
Simulation approach has been made to achieve the best value of α. For finding
optimum PID values, the original PID values were essentially varied frequently using
α through computer simulation till the lowest MS index and the required closed loop
performance reached. Effectiveness of the controllers stands measure in the form of
errors, total variation (TV) of operated variables and time domain specification.
Case study-1 (fermenter).
Fermenter process exact model has given below. In this the microbial growth was
supposed to following Monod kinetics has specified by the following equations.
dZ
= (μ−K d )Z−Zu (14)
dt
dS μ
= ( − m)Z − (S−St )u (15)
dt Y
μmax S
μ= (16)
km + S
Fig. 1 Servo performance of process with the controller parameters obtained from the enhanced
IMC, proposed method
Fig. 2 Regulatory performance to fermenter process with controller parameters attained through
enhanced IMC, proposed Approach
86 C. Venkatasuresh et al.
time (ts) were less than the available approach and these are recorded in Table 1.
From both responses of current and existing method observes the current method
accomplishes a smoother undershoot than the existing method. The current method
expresses important development in the controller through reduce of ITAE 11% than
enhanced IMC method. The controller actions of both servo and regulatory problems
specified in Figs. 3 and 4 correspondingly.
Case study 2 (isothermal CSTR).
Van De Vesse isothermal CSTR [13] is taken now.
P1 P2
X−
→ 3Y −
→Z
P3
2X −
→W
dC X F
= C X, f − C X − P1 C X − P3 C X2 (18)
dt V
dC X F
= − C B + P1 C X − P2 CY (19)
dt V
From solving above equations resulted function is as below
2 Enhanced IMC 1.068 0.533 0.651 1.103 26.654 1.748 3.58 8.24
Proposed 0.911 0.496 0.626 1.001 23.256 1.018 3.46 6.289
3 Enhanced IMC 36.423 267.67 3.575 57.261 148.81 6.347 11.3 26.7
Proposed 35.7 225.36 2.92 50.215 139 5.983 10.86 25.9
87
88 C. Venkatasuresh et al.
Fig. 3 Control action attained through servo response with PID parameters for enhanced IMC,
proposed Approach
Fig. 4 Control action attained through regulatory response with PID parameters for enhanced IMC,
proposed Approach
Independent Controller Design for Non-minimum … 89
Fig. 5 Servo performance of isothermal process with the controller parameters attained from the
enhanced IMC, proposed approach
Fig. 6 Regulatory performance to isothermal process with controller parameters attained through
enhanced IMC, proposed Approach
90 C. Venkatasuresh et al.
Fig. 7 Control action attained through servo response with PID parameters for enhanced IMC,
proposed Approach
Fig. 8 Control action attained through regulatory response with PID parameters for enhanced IMC,
proposed Approach
Independent Controller Design for Non-minimum … 91
Fig. 9 Servo performance of fermenter process with the controller parameters attained from the
enhanced IMC, proposed approach
5 Conclusion
A tuning process was used to develop the PID for the NMP SOPDT systems. The
proposed method was associated with the method existing method in literature such
as Enhanced IMC. Analysed that from simulation results, proposed method gives
important development in the closed loop performance has detected in three cases
related to the current method in literature. The current method expresses improved
performance indices IAE and the ITAE. smother control action was attained by the
92 C. Venkatasuresh et al.
Fig. 10 Regulatory performance to fermenter process with controller parameters attained through
enhanced IMC, proposed Approach
Fig. 11 Control action attained through servo response with PID parameters for enhanced IMC,
proposed Approach
Independent Controller Design for Non-minimum … 93
Fig. 12 Control action attained through regulatory response with PID parameters for enhanced
IMC, proposed Approach
present Approach. The performance was enhanced in present approach than Existing
approach. This could be exposed that the present PID is able to control NMP SOPTD
systems.
References
1. Sankar Rao C (2019) Enhanced PID controller for non-minimum phase second order plus time
delay system. In: Chemical product and process modeling, 20180059
2. Normey-Rico JE, Camacho EF (2008) Simple robust dead-time compensator for first-order
plus dead-time unstable processes. Ind Eng Chem Res 47:4784–4790
3. Uma S, Rao AS (2016) Enhanced modified Smith predictor for second-order non-minimum
phase unstable processes. Int J Syst Sci 47:966–981
4. Begum KG, Rao AS, Radhakrishnan TK (2017) Enhanced IMC based PID controller design
for non-minimum phase (NMP) integrating processes with time delays. ISA T 68:223–234
5. Hauksdottir A, Sigurosson S (2011) The matching coefficients PID controller. In: American
control conference, San Francisco, CA, USA
6. Kuhlmann A, Bogle ID (2001) Controllability evaluation for non-minimum phase-processes
with multiplicity. AIChE J 47:2627–2632
7. Seaborg DE, Edgar TF, Mellichamp DA (2004) Process dynamics and control. Wiley, USA
8. Skogestad S (2003) Simple analytic rules for model reduction and PID controller tuning. J
Process Contr 13:291–309
9. Simi S, Chidambaram M (2013) A simple method of tuning series cascade controllers for
unstable systems. J Control Theory Appl 11:661
10. Sri RP, Chidambaram M (2003) A simple method of tuning PID controllers for integrator/dead-
time processes. Comput Chem Eng 27:211–215
94 C. Venkatasuresh et al.
11. Vijaya Krishna B (2018) MATLAB/simulink study of multi-level inverter topologies using
minimized quantity of switches. Int J Eng Technol 7:201–216
12. Vijaya Krishna B (2018) Design and implementation of DPFC for multi-bus power system. Int
J Eng Technol 7:18–29
13. Vijaya Krishna B (2018) Enhancement of power quality in distribution system using hybrid
seven level H-bridge inverter based DPFC. J Electr Eng 2:61–71
Control of DC Link Voltage and Load
Voltage Variations in a Pitch Angle
Controlled PMSG Based Wind Energy
Conversion System
Abstract The world is looking for a better alternative energy source to generate
electricity. Wind Energy Conversion Systems stands first in the choice of researchers
in terms of renewable energy. Permanent Magnet Synchronous Generator (PMSG)
plays an important role in wind energy generation system design. They exhibit a
character of generating electrical energy even at a low speed of operation. Also, the
control strategies of PMSG based WECS are quite easy compared to other generators
used in wind energy generators. A pitch angle control technique along with load
voltage variation control is proposed in this article. A MATLAB/Simulink based
simulation work has been proposed here and the test results are quite satisfactory.
1 Introduction
Wind power conversion system is one of the cleanest sources of electrical power
generation systems. The power produced per unit area is quite impressive compared
to all other power generation techniques. Variable speed WECS are most popular in
wind power technologies. They can deliver maximum power point tracking as well.
Their efficiency is also compared to fixed speed and limited variable speed WECS.
The first choice of electrical generator now a days is doubly fed induction generator
(DFIG). In this article PMSG based WECS have been proposed. This generator
improves the reliability of the entire system. The features of PMSG are self-excitation
A. Nagarajan · G. Sureshkumaar
Karpagam College of Engineering, Coimbatore, India
e-mail: [email protected]
G. Sureshkumaar
e-mail: [email protected]
K. Ramash Kumar (B)
Dr. N.G.P Institute of Technology, Coimbatore, India
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 95
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_8
96 A. Nagarajan et al.
capability, leading to a high power factor and high efficient operation. A standalone
PMSG based WECS can be used as a prime source of energy to the places where grid
integration is impossible. The reason is PMSG based WECS can generate electricity
even at a low speed of wind flow.
The main challenge of standalone power system is the feeding the end users with
a balanced power supply with proper control of power converters. In wind energy
conversion systems some power quality issues like voltage imbalalnce, generation of
harmonics, flickering and load variation will occur very often. The change in voltage
occurs often due to variation in the load. The oscillating voltage of WECS is mainly
due to change in their wind speed. Also, the harmonics are produced due the high
frequency operating power converters of WECS. Those power quality problem could
not tolerated by the consumers and so, require mitigation methods [1]. Controllers
for power converters were detailed in [2–9].
Figure 1 shows the schematic diagram of permanent magnet synchronous gener-
ator based wind energy conversion system. When load varies with in a short period,
that causes a huge variation in voltage at common point of coupling. This will result
in torque ripple sin the shaft which may affect the life of the machine,
A standalone small scale PMSG based WECS has been explained in this article.
The major considerations are,
1 Implementation of pitch angle control under heavy wind turbulences.
2 Maintaining the constant voltage in the DC link to minimize ripple in the DC
input to the generator side converter.
The energy extracted from the wind is directly proportional to velocity of the wind.
The power traced from the wind energy is expressed as below,
where
ρ—air density (kg/m3 );
Control of DC Link Voltage and Load Voltage … 97
where
Cp—power coefficient,
β—pitch angle in degrees.
λ—tip speed ratio.
The tip speed ratio is defined as the ratio of the speed at which the tip of blade
rotates to the speed of the wind (wind velocity) Vw,
R
λ= (3)
vw
where
—speed of the turbine rotor.
R—radius of turbine rotor tip swept area.
Figure 2 shows the relationship between speed of the turbine rotor and the mechan-
ical power developed in the turbine. It is clear from the picture that is the mechanical
power developed in the rotor is directly proportional to the speed of rotor also the
developed power reaches maximum when at a particular point for various wind
speeds.
The typical speed vs power extraction plot is shown in Fig. 3. It shows that all the
WECS starts to produce electricity at a particular speed of wind. Till that the wind
generator will be acting as a motor that is receiving power from the system. Once the
cut-in velocity is attained the machine starts to generate electricity that is feeding the
power to the power system. One major advantage of using PMSG base WECS is that
it can produce electricity from a minimum wind velocity compared to all other types
of wind energy conversion systems. From the Fig. 4 it is evident that that extracted
power is maximum from the speed 12–25 m/s. if speed exceeds beyond this limit it
can be controlled either mechanically by using pitch angle control technique or yaw
mechanism or electrically using the power converters.
Pitch Angle Control (PAC) is defined as the angle at which the wind strikes the
turbine blades with respect to the angle of rotation. When the speed of wind exceeds
the wind blades experiences more torque as shown in the above figure. The figure
says that the power generated beyond the cut off velocity is zero. That is when the
speed of wind exceeding the maximum value of rated velocity the mechanical control
comes into action.
Control of DC Link Voltage and Load Voltage … 99
There are two ways to control the wind turbine mechanically which are
1. Pitch angle control
2. Yaw mechanism
Since the power extracted from the wind is depending mainly on pitch angle and
tip speed ratio the adjustment of pitch angle should be optimum below the rated
speed of the wind. The need of pitch angle control is as below;
(1) Maximizing the obtained power of the wind turbine, below the cut off speed
the angle must be set to derive maximum power from the system.
(2) Regulating the power generation within a safe value. Ensuring the machine
operates with in the SOA (safe operating area).
Following is the simulation model of the pitch angle control of wind energy conver-
sion system. Here the actual wind speed is measured and compared with the reference
value of wind flow. The error is amplified with help of a gain. After the gain point a
rate limiter is introduced which limits the speed of the rotor within a safe value. In
practical case a servo mechanism is employed to adjust the pitch angle. When the
wind velocity exceeds a particular value the servo motor turns the wind blade to a
particular angle. The feedback is given via the sensors which collect the speed of the
rotor of the turbine.
It is not necessary that the load is to be constant always. Due to variation in load
connected to the inverter the current in all the phases will not be uniform. In distribu-
tion network this is a major issue. This sudden change in load may lead to unbalance
voltage at the point of common coupling. There for compensation of the load voltage
variation is more important in the design of converters. A voltage control technique
using the PWM inverter is employed in this proposal. PI controllers have been used
to control the pulse generation. Here the actual value is compared with the reference
value, based on the variation between actual and reference current magnitudes the
PWM pulses are generated with different modulation index values (Fig. 5).
100 A. Nagarajan et al.
Fig. 5 MATLAB/Simulink model of PWM inverter controller for unbalanced load compensation
The following figure shows the simulation model of the power converter used in the
proposed system. The results are compared with and without the pitch angle control
strategy (Fig. 6).
The speed of wind is set to 12 m per second for first 3 s and it is set to 25 m/s up to
7 s the again 12 m/s up to 10 s. in this the base speed is set to 12 m/s. From Fig. 7 the
generator speed crosses more than 1.5 p.u which is 2550 rpm. In Fig. 8 it is shown
that the generator speed is controlled to 1.2p.u at the excess wind speed (25 m/s).
And parameters like generator torque and output voltage are also controlled.
Figure 9 shows the DC link voltage control. The disturbance occurs at 0.25 s of
the simulation period. But the system is providing a constant voltage over the entire
simulation period because of the control of the rectifier as explained in Sect. 5.
The load is varied at 0.25 s by using breaker in the Simulink model. In the above
Fig. 10, the voltage output is kept constant throughout the run time. The disturbance
is tolerated by the pulse generation. The magnitude of the modulating signal is shown
in below plot.
Control of DC Link Voltage and Load Voltage … 101
Fig. 9 Simulated responses of dc link voltage and magnitude of reference signal for PWM
generation of rectifier
Fig. 10 Simulated responses of output voltage of inverter and magnitude of reference signal for
pwm generation of inverter
104 A. Nagarajan et al.
7 Conclusion
It is evident that under mechanical discrepancy conditions like higher wind speed
and multi directional wind power mechanical control of the systems works better
compared to electrical control instantly. i.e. pitch angle control result is quick as
compared to the electrical control. Also controlling the DC link voltage with in a
safe limit saves the generator as well as the load. PI controller based control is giving
better results as shown in the result. Modern controllers may give better result.
Appendix
See Table 1.
References
8. Gayathri Monicka J, Guna Sekhar NO, Ramash Kumar K (2011) Performance evaluation of
membership functions on fuzzy logic controlled AC voltage controller for speed control of
induction motor drive. Int J Comput Appl 13(5)
9. Ramash Kumar K, Arunkumar N, Sivakumaran TS (2019) Implementation of non-linear
controller for contemporary DC-DC converters. Int J Electrical Eng Informatics 11(4):622–637
A Novel Approach of Wind MPPT Using
Fuzzy Logic
Nesar Uddin, Md Saiful Islam, Jewel Sikder Joy, and Ishaque Mozumder
Abstract The key motive of this article is to carry out the optimum power as well
as supreme power from wind turbine using fuzzy control technique. The wind power
system largely relies on wind velocity and pitch angle, so Fuzzy control technique is
used to extract supreme power in an efficient method which minimizes the dislocation
in the system reaction. For developing the fuzzy-based wind MPPT system, a DC/DC
step-up power converter is constituted to trigger the gate of its MOS-FET switch by
applying Fuzzy rules for regulating rotor acceleration of wind generator. The fuzzy
controller provides a signal as well as duty cycle as output through pulse width
modulation (PWM) that switches the gate of DC/DC converter. Simulation results
show that the developed fuzzy-based wind MPPT structure can perform nonlinear
functions of random difficulty and cover an extensive range of working conditions.
1 Introduction
N. Uddin (B)
Department of EEE, Chittagong University of Engineering and Technology, Chattogram 4349,
Bangladesh
M. Saiful Islam
Department of ETE, Chittagong University of Engineering and Technology, Chattogram 4349,
Bangladesh
J. Sikder Joy · I. Mozumder
Department of IET, Chittagong University of Engineering and Technology, Chattogram 4349,
Bangladesh
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 107
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_9
108 N. Uddin et al.
In 2019, be close to 19 TW fossil fuels have been used [3]. It also increases air pollu-
tion and damages the environment. To surpass the problem, the knowledge of new
technology of sustainable energy from solar irradiation, hydro source, wind resource,
geothermal store and biogas supply, are a more effective and reliable method in the
creation [4].
Several regions in Bangladesh are also probable for wind resource as well as wind
power generation via Patenga, Cox’s Bazar, around Teknaf area, Kuakhata and Kutub-
diya beach [5]. Generally, a wind turbine cannot generate extreme power by tradi-
tional control strategy [6]. To surpass these difficulties, MPPT methods are planned
to extract supreme power by applying FLC. Howsoever, Fuzzy control technique is
an expert intelligence thinking method that has beaten fruition in numerous cases to
take effective decisions like human beings by adjusting themselves for present and
future situations automatically [7]. Therefore, to extract supreme power from the
wind turbine, the fuzzy control technique has been used in this paper. The recom-
mended Fuzzy control technique can run an input/output planning by depending on
the human sense which is a more effectual and optimum way.
2 Methodology
Maximum power point tracking (MPPT) is a method that is used to bring out supreme
power from renewable foundations. The suggested wind MPPT is designed for
extracting efficient power that stands on wind turbine, DC/DC step-up converter,
fuzzy based control method and AC load as well as the electric appliance. The gener-
ated electricity from wind turbine relies on the atmospheric situation as well as wind
speed and direction of pitch that is distributed through DC/DC step-up converter.
The step-up converter is coupled with the fuzzy controller as well as the control unit.
The control division acts according to fuzzy rules which give duty cycle as output.
Subsequently, this output signal switches the gate of MOS-FET device through PWM
incessantly to get the extreme power. Control units have double inputs viz; error signal
and change in the error signal. The error signal is denoted as ‘e’ and change in the
error signal is denoted as ‘de’. Again ‘e’ and ‘de’ relies on a ratio of change in power
and change in voltage. Change in power is denoted as ‘P’ and change in voltage is
denoted as ‘V’. Overall block diagram has been presented in Fig. 1.
Generally, wind power conversion method, the output power of the wind turbine is
not linear because of unstable weather conditions. The fuzzy-logic-based controller
A Novel Approach of Wind MPPT Using Fuzzy Logic 109
The rotor velocity is the main dependence of the voltage of the wind turbine. As
a result, the FLC controller is introduced here to be in command of rotor speed for
higher power generation. The fuzzy formulas are précised in Table 1. The maximum
output is picked up using fuzzy control technique. DC/DC step-up converter is applied
here to position the duty ratio in the course of PWM. The error signal is a ratio of
power deviation and voltage deviation with respect to time. A mathematical model
of the control unit of MPPT system has been given in Fig. 6.
P P(t) − P(t − 1)
e= = (1)
V V (t) − V (t − 1)
The FLC is associated with double input (e and de) and single output (Duty Cycle)
that rely on IF–THEN affirmations which is named as fuzzy rules where IF indicates
the input and THEN indicates the outputs parts. Overall, 15 rules have been set
to produce supreme power from the scheme where various working conditions are
described that is given in Table 1.
The ultimate schematic illustration of MPPT system is exposed in Fig. 7 where the
arrangement is constructed with a wind turbine, DC/DC step-up converter, electric
112 N. Uddin et al.
load and other system’s controlling unit. The outputs have been culled by varying
the wind speed from the range of (3–8) meter per second.
The 3-D plane inspection of FLC is represented in Fig. 8 where X-axis is representing
the error signal (e) and Y-axis is indicating a change in error signal (de). The value
of ‘e’ is measured from negative 10 to positive 10 that rely on the fluctuation of
power and voltage. The value of ‘de’ is calculated as -3 to 3 which is précised for
the interruption period to bring the utmost power. The outcome signal of the FLC is
the duty cycle that varies with reference to ‘e’ and ‘de’.
The value of the error signal has been considered negative ten to positive ten
because of standard power as well as voltage of wind turbine. The standard power
and voltage of WES5 tulipo wind turbine are 2.5 kW and 400 V. The standard voltage
is converted to 220 V by 1-phase rectifier that is shown in Fig. 7. Therefore, we can
get a ratio between power and voltage which is approximately 10. This value can vary
negative ten to positive ten by depending on wind velocity. But the range of change
in the error signal does not change gradually. So, this value has been considered
negative three to positive three which is lower than the value of the error signal.
Case study—A
According to Fig. 9, when the value of the error signal is very negative and the
value of change in the error signal is negative, then the duty cycle is very positive
to get supreme power. Generally, it is not possible to get supreme power when the
value of the error signal is very negative and the value of change in the error signal is
negative. At any time, this situation may occur because of unavailable wind velocity.
Therefore, it is impossible to get maximum power from the turbine. So, the duty
cycle has been increased to bring out supreme power.
Case study—B
In this situations, when error signal is negative and change in error signal is
considered as the value of zero’ side, then duty cycle is assumed zero’s side. That
means the variation of change in error signal is not high by following error signal.
So, it is not necessary to change the duty cycle (Fig. 10).
Case study—C
In this case, the value of the error signal and change in error signal are considered
as positive and negative. So, the value of the duty cycle is assumed zero’s side because
of its operating condition. Error signal positive means the high operating condition
of wind turbine and change in error signal negative means it does not influence on
the operating system. So, the duty cycle has remained beside zero (Fig. 11).
Power difference of Fuzzy based wind MPPT system and without MPPT system
is given in Table 2. From the Table, it is clear that Fuzzy based wind MPPT system
is more efficient than a conventional system.
4 Conclusion
The suggested fuzzy based wind MPPT system that has diverse benefits over previ-
ously accessible systems. Fuzzy system is used which depends on nonlinear func-
tions which is very modern and reliable. The petition of Fuzzy technique is swiftly
mounting due to its strong performance regardless of complex situations and elas-
ticity. It is possible to increase the efficiency by adapting supplementary intellectual
fuzzy rules. A problem of the proposed structure is their remittent environment of
A Novel Approach of Wind MPPT Using Fuzzy Logic 115
wind speed that is dependent on weather variations. Fuzzy technique has been applied
to extract extreme power by two input parameters via Wind velocity & pitch angle.
The results of the Fuzzy based wind MPPT model has been compared with the
conventional system under a similar climate situation. Simulation result provides a
grand performance of Fuzzy based wind MPPT model and less error. The applica-
tion of renewable sources minimizes various gas discharges in the system. Besides,
the system is environment kindly, favorable. Developed Fuzzy based wind MPPT
structure more efficient than any other operable system as well as a grid-connected
system that must reduce power loss.
References
1. Soedibyo T, Hadi A, Putra HP, Suryoatmojo H, Ashari M (2016) Integration of hybrid PV/wind
generation system using fuzzy MPPT in grid connected system for remote area. In: MATEC web
of conferences, International Conference on Frontiers of Sensors Technologies (ICFST 2016),
vol 59, pp 1–2
2. Gowri M, Vinothkumar V, Punitha K (2015) ANFIS based MPPT and droop controller for power
sharing in interline power system. Int J Sci Res Dev 3(04)
3. Khalil GA, Lee DC, Seok jk (2014) Variable speed, wind power generation system based on
fuzzy logic control for maximum output power tracking. In: 35th annual IEEE power electronics
specialists conference Aachen, Germany, pp 5–6
4. Uddin N, Islam MS (2018) Optimal fuzzy logic based smart energy management system for real
time application integrating RES, grid and battery. In: 4th international conference on Electrical
Engineering and Information & Communication Technology, IEEE
5. Islam MR, Rahman MN, Mannan MA (2016) Study of wind power in different parts of
Bangladesh. Int Res J Eng Technol (IRJET) 03(09): 1290. ISSN: 2395-0056
6. Uddin N, Islam MS (2018) Optimization of PV energy generation based on ANFIS. In:
International conference on Innovations in Science, Engineering & Technology, IEEE
7. Hailemariam MM, Mekonnen T, Sudheendra H (2015) Novel approach to fuzzy logic controller
based hybrid solar/micro hydro/bio-mass generation, A real time analysis (Barsoma Village,
Ethiopia). Int J Innov Res Eng Multidisciplinary Phys Sci (IJIRMPS) 3(1). ISSN: 2349-7300
ISO 9001
8. Minh HQ, Cuong NC, Chau TN (2014) A fuzzy-logic based MPPT method for stand-alone wind
turbine system. Am J Eng Res (AJER) 3(9): 177–184. e-ISSN: 2320-0847 p-ISSN: 2320-0936
Shale Gas Productive Volume
Optimization
1 Introduction
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 117
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_10
118 A. Umar Ibrahim and B. Mamo Negash
are applied to predict productivity. These models are basically: analytical transient,
numerical and empirical models [3, 4]. Recently, a simple model was presented,
which successfully predict productivity with less than 3% over prediction. The analyt-
ical model is applied in pseudo-steady state flow conditions. However, the fracture
width is assumed and computed as an average in the model [4]. Therefore, there
is a need to find a more reliable way of obtaining a fracture width that will close
the overprediction gap of the model. An accurate fracture width is vital to proper
field performance prediction, fracture completion design and adequate placement of
proppants in shale gas wells [5]. Also, this will give rise to designing a fracturing
treatment that will propagate in a more optimal direction and avoid unnecessary
damage to the subsurface environment.
This paper presents a simple and more accurate method using a pseudo-3D model
to determine the fracture width in shale gas reservoirs. The model gives the visual
result of the fracture geometry using MATLAB software. So, the obtained fracture
width can be incorporated with the latest analytical model to forecast the long-term
productivity of multistage fractured shale gas wells.
Multistage hydraulic fracture in horizontal wells initiates many fractures from each
perforation cluster. These change the formation stress field and affect the propagation
of hydraulic fracture. Also, the low-viscosity fracturing fluid will leak off from the
surface of hydraulic fracture into the reservoir, therefore, increasing its pressure.
Consequently, the natural fractures will be stimulated by the induced stress. If the
shear stress is greater than the shear strength, shear failure occurs [6]. Likewise, if
the natural fractures inner pressure is greater than the tensile strength, tensile failure
happens [7, 8]. Therefore, these increase the shale reservoir permeability. The fracture
geometry, which has the fracture height, length, and width, directly control most of
the productivity of the well [9, 10]. Proper design of these parameters aid in gas
production optimization during fracturing treatment design.
Furthermore, it is good to visualize the fracture geometry using the pseudo-3D
model in MATLAB software to determine the fracture width. The P3D model is
compared with the well-known KGD and PKN models (Table 1).
Where, Qg = production rate of the gas (MScf/D); nf = number of fractures; μ =
viscosity of the fluid (cp); T = temperature of the formation (°R); km = permeability
of matrix (md); h = reservoir thickness (ft); p = average reservoir pressure (psia); pw
= wellbore pressure (psia); e = exponential function; xf = fracture half-length (ft);
kf = permeability of fracture (md); w = average fracture width (in); Sf = fracture
spacing (ft).
Shale Gas Productive Volume Optimization 119
The single hydraulic fracture propagation is first shown to compare a P3D model with
the conventional KGD and PKN model. The visualized fracture width provides the
quantitative result of the width, which makes it easy to be applied in the productivity
model.
4 Discussion
Fig. 4 KGD in 3D
Shale Gas Productive Volume Optimization 121
in proportional to t1/ 3. These models show a curvy tip which indicates that in the
KGD model, the effect of the tip of the fracture is not large. Both KGD and PKN
showed an approximation the width of the fracture and length. The pseudo-3D model
forecasts a complete ellipsoid in 3D space at a time which can be seen in Figs. 9,
10 and 11. Therefore, the P3D model is good for fracture propagation modelling.
The model can display the variation of the fracture height with length as the fracture
propagates.
The fracture width is directly taken from the visualized fracture geometry and
subsequently inputted in the multi fractured shale gas wells. This method is more
dependable to determine the fracture width of a fractured shale gas reservoir to
determine the productivity of gas wells. Also, it can be applied in order to optimize
production by adequate fracture spacing and proppant placement.
References
1. Abdelaziz A et al (2019) Unconventional shale hydraulic fracturing under true triaxial labora-
tory conditions, the value of understanding your reservoir. In: SPE annual technical conference
and exhibition, Society of Petroleum Engineers
2. Quainoo AK, Negash BM, Bavoh CB, Ganat TO, Tackie-Otoo BN (2020) A perspective on
the potential application of bio-inhibitors for shale stabilization during drilling and hydraulic
fracturing processes (in English). J Natural Gas Sci Eng 79:103380
3. Mahmoodi S, Abbasi M, Sharifi M (2019) New fluid flow model for hydraulic fractured wells
with non-uniform fracture geometry and permeability. J Natural Gas Sci Eng, Article 68, Art
no 102914
4. Li G, Guo B, Li J, Wang MJSD, and Completion (2019) A mathematical model for predicting
long-term productivity of modern multifractured shale-gas/oil wells, vol 34, no 02, pp 114–127
5. Belyadi H, Fathi E, Belyadi F (2019) Hydraulic fracturing in unconventional reservoirs:
theories, operations, and economic analysis. Gulf Professional Publishing
6. Wu Y-S (2018) Hydraulic fracture modeling. Gulf Professional Publishing
7. Ren L, Lin R, Zhao JZ (2018) Stimulated reservoir volume estimation and analysis of hydraulic
fracturing in shale gas reservoir, (in English). Arab J Sci Eng 43(11):6429–6444
8. Ren L, Lin R, Zhao JZ, Rasouli V, Zhao JY, Yang H (2018) Stimulated reservoir volume
estimation for shale gas fracturing: mechanism and modeling approach, (in English). J Petrol
Sci Eng 166:290–304
9. Ahn CH, Dilmore R, Wang JY (2017) Modeling of hydraulic fracture propagation in
shale gas reservoirs: a three-dimensional, two-phase model. J Energy Resources Technology
139(1):012903
10. Berawala DS, Andersen PO, Ursin JR (2019) Controlling parameters during continuum flow
in shale-gas production: a fracture/matrix-modeling approach, (in English). SPE J Conf Paper
24(3):1378–1394
11. Yew CH, Weng X (2104) Mechanics of hydraulic fracturing. Gulf Professional Publishing
A Comparative Study and Validation
of Kinematic Analysis of a Crank Rocker
Engine Prototype Using MATLAB
and ADAMS
1 Introduction
A new engine called Crank-Rocker engine (CR) was patented in 2018 by engineers
of Universiti Teknologi Petronas, Malaysia. Despite showing promising results, the
engine is still considered belonging to early development stage. Of these studies
which has been performed on crank-rocker engine and used as base reference for
this paper and any further improvements on this engine, work in researches [1–4].
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 125
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_11
126 A. M. Albaghdadi et al.
Since a MATLAB code is utilized to perform analytical study on this system [7–
9], it is required to include related mathematical equations that represent system
physics. Hence this section introduces kinematic fundamentals which are involved
in analyzing the system. Once the mechanism geometry and other conditions such
as Grashof’s law and initial conditions are identified, mathematical equations are
constructed. This study adopted the usage of vector-loop method for a crank-rocker
mechanism kinematic analysis. Figure 1, shows the crank rocker basic geometry
which consists of four main linkages namely ground R1 , Crank R2 , Coupler R3 and
Rocker R4 . Also, each respected link angular position with reference to the ground
is set to θ1 , θ2 , θ3 and θ4 , which are (ground linkage angle, crank angle, coupler
A Comparative Study and Validation of Kinematic … 127
angle and rocker angle) respectively. The angle between the coupler and rocker is
called the transmission angle γ. Another angle of concern in this design process is
the throw angle ϕ, which represents maximum rocker displacement, in other words
it is obtained by subtracting maximum and minimum crank rocker angle (θ4 ).
In this case, the vector loop equation is represented in Eq. (1):
R2 + R3 = R1 + R4 (1)
R2 + R3 − R1 − R4 = 0 (2)
Arbitrary link position relative to the X–Y axis, for each components of Eq. (2)
can be rewritten as:
Equation (3) represents system component position in the X-axis direction, while
Eq. (4) represents system component position in the Y-axis direction. When applying
first time derivative on Eqs. (3) and (4), also since is ground link is fixed (ω1 = 0),
we get angular velocity Eqs. (5) and (6):
Also, α2 , α3 , and α4 above represent the angular accelerations of links R2, R3,
and R4 respectively. Linear velocity for joints between linkages, in this case point A
and B as illustrated in Fig. 1, basically can be found by Eq. (9):
υ =ω×R (9)
where, R is linkage length and ω is the angular speed. Likewise, linear acceleration
for both points can be found by Eq. (10):
a = at + an (10)
where, at is the tangential acceleration of this point and an is the normal acceleration
of the same point, and both can be found by:
at = α × R (11)
v2
an = (12)
R
Here, α is the angular acceleration of link R, and v is the linear velocity at the
measured point. Using all of the above equations with known crank position (0° ≤ θ2
≤ 360°) and constant rotational speed ω2 = C, MATLAB codes are programmed for
system analysis. All remaining parameters are calculated and translated into system
simulation analysis.
Unlike programming with softwares such as MATLAB [6, 10, 11], ADAMS has the
advantage of less need to write mathematical equations to identify system physics.
The required task is only to represent the geometry of this system into the application
interface, and then ADAMS has the ability to translate this system into a model for
simulation and analysis. ADAMS also allows users to perform static, kinematic and
dynamic study on different system conditions, in addition to the ability to simulate
and predict the behavior of each case. However, other procedures are required to
A Comparative Study and Validation of Kinematic … 129
accomplish valid simulation such as identifying all links and joints dimensions.
Also, it is desired to create proper measurements and variables of each object to
be included and calculated during the simulation process. After model set-up is
finished, the simulation process can be performed, and a full evaluation of system
characteristics is created. Figure 2, shows ADAMS workspace environment, the
crank rocker mechanism is drawn with corresponding system geometry. Also, joints
between linkages are identified with crank joint having a motion function for system
initiation.
Figure 3, shows the complete rotation of the crank link (green), and the corre-
sponding rocker path motion (red). After analyzing this system path behavior, a
simulation was conducted in both softwares for studying the kinematic performance
of this mechanism. The outcome results were plotted in overlayed graphs for better
illustration and verification purpose. The plots of angular position, velocity and accel-
eration versus crank position (θ2 ), obtained from MATLAB and ADAMS simulations
are compared and shown in Figs. 4, 5, 6, 7 and 8.
Figures 4, 5, 6, 7 and 8, shows overlaying graphs of the different kinematic anal-
ysis for the crank-rocker mechanism by MATLAB and ADAMS. These data are
representing mechanism various positions, angular and linear velocities, angular
Fig. 4 Kinematic simulation comparison results between MATLAB and ADAMS, Mechanism
angular position
A Comparative Study and Validation of Kinematic … 131
and linear accelerations with respect to the crank position. When examining and
comparing results by both softwares, it can be noticed that both models are showing
almost same readings, which gives an indication that both results are verifying each
other. Hence, these results can be considered and applied for further system analysis
and enhancement.
5 Conclusion
References
1. Mohammed SE, Baharom MB, Rashid A, Aziz A (2016) Thermodynamic analysis of a single
cylider crank-rocker engine. ARPN J Eng Appl Sci 11:12239–12246
2. Mohammed SE, Baharom MB, Aziz ARA (2017) Performance and combustion characteristics
of a novel crank-rocker engine. J Mech Sci Technol 31:3563–3571
3. Hassan SEM (2018) Development, performance and emission assessment of a single-cylinder
air-cooled gasoline Crank-Rocker engine. Universiti Teknologi Petronas
4. Mohammed SE, Baharom MB (2019) Modelling of combustion characteristics of a single
curved-cylinder spark-ignition crank-rocker engine. Energies 1–14
A Comparative Study and Validation of Kinematic … 133
5. Verma SK, Kumar R, Chittawadigi RG, Saha SK (2016) Kinematic analysis of mechanisms
using velocity and acceleration diagrams (VAD) module in mechanalyzer software. In: The
proceedings of the Asian conference on multibody dynamics, p. 01_771082
6. Hroncová D, Delyová I, Frankovský P (2014) Kinematic analysis of mechanisms using MSC
Adams. Appl Mech Mater 611:83–89
7. Norton RL (2012) Design of machinery: an introduction to the synthesis and analysis of
mechanisms and machines, McGraw-Hill, New York
8. Mohammadzadeh A (2007) Analytical synthesis and analysis of mechanisms using MATLAB
and SIMULINK. Am Soc Eng Educ 12(242):1
9. Myszka DH, Celik A (2012) Machines and Mechanism. Prentice Hall
10. Hroncová D, Binda M, Šarga P, Kičák F (2012) Kinematical analysis of crank slider mechanism
using MSC ADAMS/View. Procedia Eng 48:213–22
11. Hroncová D, Frankovský P, Virgala I, Delyová I (2014) Kinematic analysis of the press
mechanism using MSC Adams. Am J Mech Eng 2:312–315
12. Albaghdadi AM, Baharom MB, bin Sulaiman SA (2021) Parameter design optimization of
the crank-rocker engine using the FMINCON function in MATLAB. IOP Conference Series:
Materials Science and Engineering 1088(1):012072. https://doi.org/10.1088/1757-899X/1088/
1/012072
Application of Machine Learning Models
in Gas Hydrate Mitigation
Sachin Dev Suresh, Bhajan Lal , Ali Qasim , Khor Siak Foo,
and Jega Divan Sundramoorthy
Abstract The production of oil and natural gas contributes to a significant amount of
revenue generation in Malaysia thereby strengthen the country’s economy. The flow
assurance industry is faced with impediments during smooth operation of the trans-
mission pipeline in which gas hydrate formation is the most important. It affects the
normal operation of the pipeline by plugging it. Gas hydrate is a crystalline struc-
ture composed of a network of hydrogen bonds between host molecules of water
and guest molecules of the incoming gases under high pressure and low temper-
ature conditions. Industry uses different types of chemical inhibitors in pipeline
to suppress hydrate formation. To overcome this problem, machine learning algo-
rithm has been introduced as part of risk management strategies. The objective of
this research is to evaluate the various types of machine learning models used to
predict the gas hydrate formation where the input parameters are gas composition,
pressure and concentration of inhibitor and the output parameter is hydrate deposi-
tion/formation temperature (HDFT). Three machine learning models are compared:
Artificial Neural Network (ANN), Least Square version of Support Vector Machine
(LSSVM), and Extremely Randomized Trees (Extra Trees). Comparison of the three
different machine learning models is based on the correlation coefficient, R2 . The
best choice of machine learning model that has highest R2 is obtained by Extra Trees
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 135
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_12
136 S. D. Suresh et al.
model of 0.9991 compared to other two machine learning models which predicted
R2 value greater than 0.96.
1 Introduction
In the 1930s, gas hydrate deposition was discovered in pipelines where several
attempts were made to understand the clathrate hydrate in terms of compound,
composition, equilibrium state and manner of handling them [1]. Gas hydrate, mostly
known as clathrate hydrates exist when sufficient amount of water molecules in oil
and hydrocarbon vapour form crystallize during transient flow condition [2, 3]. Gas
hydrates are developed in the crude oil and natural gas industries’ output, refining and
transmission facilities. [4–8]. Depending on the type and size of guest molecules,
the gas hydrates are produced in 3 structures, which are I, II and H [7, 9]. The
unit cells of Structure I had eight cavities (two smalls and six large ones) and were
formed by hydrogen-bound 46 water molecules. Structure II unit cells, by contrast,
contained 136 molecules of water and were enclosed in 24 cavities, including 16
small and 8 wide cavities. Whereas structure H is a mixture of methane, propane,
isobutene, CO2 , H2 S, N2 , and stabilises with one of the two hydrate structures forming
micro cavities[10, 11]. According to Qin et al. [12] and Olajire [13], at high pres-
sure and low temperatures as well as long distances, gas hydrates are produced,
often contained in subsea oil and gas pipelines. Although gas hydrate production
contributes to critical flow assurance problem, it is therefore important to address
this issue by implementing effective preventive and predictive strategies [4].
There are several risk management strategies that has been used to prevent the
growth of hydrates which are dehydration of natural gas, usage of inhibitors such as
thermodynamic hydrate inhibitors (THIs) and low dosage hydrate inhibitors (LDHIs)
[12–16]. Methanol (MeOH) and monoethylene glycol (MEG) are two commercially
used Thermodynamic hydrate inhibition compounds [17]. The goal of these THIs is to
interrupt the process in which hydrates form, often like water–ice antifreeze, to ensure
safety under the most severe hydrate forming conditions. However, the downside of
these THIs is that the hydrate condition gets extreme as the subcooling process is of
better consistency, leading to more inhibitor requirements. This leads to a rate limit
of methanol supply due to availability, storage and injection restrictions, leading to
an adverse effect on capital costs, operating costs, non-optimum performance, as
well as an increased risk of hydrate plug formation [1].
In contrast with THIs, LDHIs play a crucial position as an alternative to THIs in
terms of lower doses of inhibitors used and dosage levels that cause lower loss of
inhibitors due to evaporation. Other than that, LDHIs would decrease the chemical
storage and injection needs, which will have a beneficial impact by minimising
CAPEX and OPEX [1]. Two kinds of LDHIs, such as kinetic hydrate inhibitors (KHI),
are used to slow hydrate crystallisation and anti-agglomerants (AAs), which facilitate
Application of Machine Learning Models in Gas Hydrate Mitigation 137
hydrate molecule distortion [12]. For simulation purposes, computer intelligence can
be used, such as machine learning (ML) models, to test the accuracy of the inhibitor
used, composition of gas molecules, gas hydrate temperature and pressure regulation
[4].
Usually, ML is called a universal toolbox, ready to be used for classification issues,
defining an appropriate group for a new collection of findings, and regression tasks
that approximate the relationship between groups of data provided. ML is fundamen-
tally a numerical representation of a phenomenon, provided a certain value and based
on a certain setting, aimed at performing a task. ML techniques may be classified
based on learning activity priorities that define patterns for detection and estima-
tion, action learning, or inductive methods of learning. The algorithms can be further
grouped into three distinct families of learning, known as supervised learning, unsu-
pervised learning and reinforcement learning [18]. The main objective of this paper is
to compare numerous forms of ML models that will be used, such as Aritifical Neural
Network (ANN), Least Square Support Vector Machine (LSSVM) and Extremely
Randomized Trees (Extra Trees), which are used to help create advanced gas hydrate
control techniques [12]. In this article, theory and method employed for different ML
models have been discussed. Afterwards, a comparison among the models has been
performed and the their effectiveness has been assesed and identified. Finally, on
these basis, some important conclusion have been drawn and discussed.
Authors like Rebai et al. [19], stated that machine learning techniques such as ANN,
LSSVM and Extra Trees are considered eligible to reproduce any form a function and
have been successful in predicting the required output of the gas hydrate formation
which is hydrate formation/deposition temperature (HDT/HFT). The input parame-
ters that has been used in this 3 ML models were concentration of the inhibitors, gas
composition and pressure. Each ML models theory and method will be explained
below in details.
where ., . represent dot product; (x) denotes the nonlinear function that performs
regression; b and w are bias terms and weight vector, respectively. The LSSVM model
structure can be seen in Fig. 1 where the K (xi , xj ) is defined as kernel function [20].
Tree models were identified by the assembly method named as Extra Trees, in which
this process produces multiple trees from the complete testing examples in terms
of regression or decision trees. Extra Trees is a top-down procedure where it forms
a joint of untrimmed choice or regression tress where the act of splitting nodes by
selecting action points randomly completes and uses the entire experiment model to
grow these trees, distinguishing it from other tree-based ensemble approaches [4].
The growth of the tree is in the top-down order where the numbering is from left
to right, as can be seen from Fig. 2. The value of a parameter is compared with the
Application of Machine Learning Models in Gas Hydrate Mitigation 139
threshold value of each internal node. If the shown inequality above each node is
satisfied, the path is either the left branch or the path is not the right branch [4].
Artificial neural networks (ANN) are a method of calculation based on how the brain
conducts calculations [20, 21]. According to Samsudin et al. [20], this model has
the potential to carry out extensive input–output mapping and can create a network
that approximates nonlinear functionality. Figure 3 shows ANN feed-forward back-
ward propagation consists of three layers; data transmitted is the input (first layer),
interpreted data is hidden layer, and result provided is output (last layer) [20].
Output of the ANN assuming a linear output neuron j, a single hidden layer with
h sigmoid hidden nodes and the output variable (xt ) is given by [20]:
⎛ ⎞
h
xt = g ⎝ w j f s j + bk ⎠ (2)
j=1
where g(.) is the linear transfer function of the output neuron k and bk is its bias, wj
is the connection weights between hidden layers and output units, f (.) is the transfer
function of the hidden layer.
Below Fig. 4, shows the exact prediction of the distribution of temperature in the
gas-dominated system, with due regard to the exothermic effect of the composition
140 S. D. Suresh et al.
of hydrate and the resistive effect of the hydrate deposition layer on the rate of heat
transfer in the pipeline. The number on the Fig. 4 is decoded by (1) supplying hydrate
forming materials for deposition, (1’) reflecting the indirect effect by temperature,
pressure, and effective inner diameter variations. Whereas (2) signifies heat release
from hydrate formation by interaction of free water and gas; (2’) indicates the effect
of temperature differences on the hydrate formation sub-cooling. The heat tolerance
effect denoted by (3) of the hydrate layer on the heat transfer while (3’) indicates the
indirect effect influencing the rate of hydrate formation [13].
Application of Machine Learning Models in Gas Hydrate Mitigation 141
It is assumed that the accuracy and credibility of every statistical method is linked to
the universality and validity of the database used for its foundation. To compare the
results based on the several factors such as correlation coefficient, R2 , comparison of
the three ML models must be represented in terms of hydrate dissociation/ formation
temperature (HDT/HFT). All the data and information taken and referred from [4,
22] in order to obtain effective comparison and accuracy.
Based on article Yarveicy and Ghiasi [4], in the first step,1800 experimental data
points of different gas systems for liquid water-hydrate-vapor (LHV) or ice-hydrate-
vapor (IHV) balance were collected and divided into two classes called “training set”
and “test set” for Extra Trees and LSSVM models. 90% were used as training proce-
dures from the data point and the remaining 10% were used as testing procedures.
The primary goal is to approximate the HDT / HDFT of gases where electrolysis
and alcohol(s) are present in pure water or aqueous solutions, where it is corre-
lated as function of concentration of inhibitor(s) (C additive ), pressure (P), and the gas
composition (Zi ):
H D F T = f (Cadditive P, Z i ) (3)
Table 1 Error evaluation outcomes for the presented Extra Trees, LSSVM and ANN models
Parameter/model Extra trees LSSVM ANN
Train Test Total Train Test Total Train Test Total
R2 % 99.91 97.46 99.7 97.54 92.4 96.99 99.47 98.82 99.28
AARD% 0.01 0.35 0.04 0.28 0.61 0.32 0.1331 0.2137 0.1509
determining the reliability and efficiency of the built Extra Trees, LSSVM and ANN
models.
n
− ti )2
i (oi
R 2 % = 100 1 − n (4)
i (oi − average(ti ))
2
n
oi − ti
A A R D% = 100 /n (5)
i=1
ti
where n denotes the number of samples; oi and ti are the predictions of the model
and corresponding targets, respectively.
The error analysis results for the tabulated in Table 1 shows an excellent compli-
ance with the experimental hydrate deposition / forming temperature (HDT / HDF)
for the three distinct types of ML models that are acceptable outcomes [4]. Neverthe-
less, as the AARD percent is 0.04 relative to the other models with AARD percent of
0.32 and 0.1509 for LSSVM and ANN, respectively, the suggested tree-based model
has stronger prediction.
4 Conclusion
References
1. Bavoh CB, Lal B, Osei H, Sabil KM, Mukhtar H (2019) A review on the role of amino acids
in gas hydrate inhibition, CO2 capture and sequestration, and natural gas storage. J Nat Gas
Sci Eng 64:52–71
Application of Machine Learning Models in Gas Hydrate Mitigation 143
Abstract Overviewing the past the past period of 20 years, it is evident that in
certain geographical regions, the spectrum, lay unutilized whereas other areas face
spectrum crunch. To boost the spectral efficiency, unlicensed users (secondary users)
will be allowed to access unutilized parts of the spectrum and adapt their communi-
cations. It assists to utilize them while minimizing the interference on licensed users.
Such ability is the predominant feature of cognitive radio (CR) nodes, which needs,
algorithms and protocols for spectrum sensing, sharing, and management. The ratio-
nale is to facilitate optimization of the spectrum and make a conducive environment
between licensed and unlicensed users. The seamless method of spectrum sensing
is Energy Detection method, since it does not require to have priory knowledge of
Primary User (PU).In this paper, sequential sensing event at the Energy Detector is
used to decide on the optimal adaptive threshold value over Rayleigh fading channel.
1 Introduction
The year 2020 which is not yet over, went through health and economic crisis world-
wide due to Covid-19, faced lockdowns of months. All people during this pandemic
situation were connected and jointly fight the Covid-19 situation using digital tech-
nologies, which completely depends on mobile wireless network. At the outburst
of pandemic situation, an emergency meeting was called in starting week of April
and ITU emphasized on the importance of communication technologies in dissem-
inating timely critical information, supporting e-learning for more than 1.5 billion
B. Rose (B)
Department of IT, Dr. N.G.P. Institute of Technology, Coimbatore, Tamil Nadu, India
e-mail: [email protected]
B. A. Devi
Department of ECE, Dr. N.G.P. Institute of Technology, Coimbatore, Tamil Nadu, India
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 145
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_13
146 B. Rose and B. A. Devi
2 Related Works
In the case of relatively low SNR,when the difference between noise and signal
energy is not distinct, single-threshold detection methods are easy to cause SUs to
perceive the existence of a PU incorrectly, which will reduce the detection probability
of the entire system. To improve the detection the single threshold is replaced over
the years by adaptive or dynamic threshold.
There are methods, named three-event ED, which makes decision considering the
event immediately before and the one immediately after current sensing event. The
primary users (PU) activity duty cycle value was exploited for tracking the changes
of the PU state [11–13].
Furthermore, an adaptive double threshold based on history energy detection is
proposed for Cooperative Spectrum Sensing. The weighting coefficient of thresholds
according to the SNR of all cognitive nodes is calculated during sensing period; thus
adjusting the upper and lower thresholds, if the average history energy is still between
two thresholds, the single-threshold method will be used for the end decision. Finally,
the fusion center aggregates the detection results of each node and obtains the final
cooperative conclusion through “or” criteria [14, 15].
However, the double threshold methods mentioned above are all no-decision
when the energy is between the high and low thresholds. The multilevel quanti-
zation method is adopted in the uncertain region of the traditional double-threshold
energy detection to solve the problem that the sensing information is ignored between
the two determined thresholds [16, 17].
In this method, an adaptive threshold method to overcome sensing failure at very
low SNR with uncertain noise power using a check parameter and double threshold
concept is proposed. It was noted that this method improved the detection probability
at a low SNR [18].
Nowadays, many Machine Learning algorithms are put forward for Optimization
of Adaptive Threshold in energy detection. The machine learning algorithm looks
for the optimum decision threshold, which is the most important parameter to decide
the presence or absence of the primary user, using historical detection data [19].
3 Methodology
Let the complex signal received at time t by the secondary user in the desired band
be [20]
n(t) : H0
y(t) = (1)
h(t)s(t) + n(t) : H1
where n(t) is an additive complex white Gaussian noise process, h(t) represents a
fading process (e.g., nonfading means h(t) = 1), and s(t) is a signal transmitted by
148 B. Rose and B. A. Devi
the primary node [20]. The sensing decisions are made using N > > 1 samples of
y(t). A decision statistic T(y) is used for hypothesis testing, given by
T (y) < λaccept : H0
T (y) > λaccept : H1
where 0 < λ < ∞ is called the decision threshold. The reliability associated with the
decision rule (2) can be characterized by Probability of Detection Pd and Probability
of False Alarm Pf is shown in Fig. 1. Mathematically, it is a conditional probability
given as below:
where γ is the SNR and Q is the standard Gaussian upper tail probability function
∞ −x 2
given as Q(u) = 2π1
u exp 2
d x.
The probability of missed detection is given as Pm
Pm = 1 − Pd (6)
Analysis of Signal Sensing with Adaptive … 149
Pe = P f + Pm (7)
Pe = P f + (1 − Pd ) (8)
From Eqs. (4) and (5) suitable sample size N can be calculated from Pf and Pd
as,
2
N = γ −1 Q −1 P f − 1 + γ −1 Q −1 (Pd ) (9)
For constant false alarm rate (CFAR) the normalized threshold λf (optimum
threshold) can be calculated as,
√ √
λ f = 2(1 + γ ) N Q −1 P f + N (11)
Optimum Threshold is chosen to make Pd large (Pd ≥ 0.9) and Pf small (Pf ≤
0.1).CFAR and CDR is used for the selection of threshold [21]. It has been noted
that CFAR improves the throughput [22], but CFAR fails to give enough protec-
tion to the PU as compared to that of the CDR approach. To improvise the overall
throughput at low SNR, the combination of CFAR and CDR is utilized to select
the optimum threshold [23]. In this paper, considering all of the above aspects, the
optimal threshold is selected for ED which jointly achieve the targeted Pf and Pd at
low SNR.
where δ (0 ≤ δ ≤ 1) is PU’s spectrum usage ratio and (1-δ) is PU’s channel is vacant &
noise occupied, (1-Pd ) gives probability of missed detection.
The adaptive threshold to minimize the error decision probability is
150 B. Rose and B. A. Devi
⎧ ⎫
⎪ (1−δ) ⎪
⎪
⎨ 4(2 + γ )ln. (1 + γ ). δ ⎪ ⎬
(1 + γ )
λ = σn2 . . 1+ 1+ (13)
(2 + γ ) ⎪⎪ Nγ ⎪
⎪
⎩ ⎭
where σn2 = noise variance. Thus, optimum threshold setting should consider both
the probability of false alarm and the probability of mis-detection in order to benefit
both PU and SU.
4 Conclusion
The overall throughput at low SNR is improved by the combinational use of the
optimum threshold. Thus, the optimal threshold is selected for ED which jointly
achieve the targeted Pf and Pd at low SNR. While the SNR increases Pd is also
increases.
References
1. Special emergency session of the broadband commission pushes for action to extend internet
access and boost capacity to fight COVID19. https://www.itu.int/en/mediacentre/Pages/PR05-
2020 Broadband Commission-emergency-session-internet-COVID-19.aspx, online. Accessed
20 Apr 2020
2. Cisco visual networking index Global mobile data traffic forecast update, 2018–2023. https://
www.cisco.com/c/en/us/solutions/collateral/executive-perspectives/annual-internet-report/
white-paper-c11-741490.html
3. Gavrilovska L, Denkovski D, Rakovic V, Angjelichinoski M (2014) Medium access control
protocols in cognitive radio networks: overview and general classification. IEEE Commun
Surveys Tutor 16(4):2092–2124
4. FCC (2002) Spectrum policy task force. Technical report 02–135, Rep. ET Docket, Federal
Communications Commission
5. Shared Spectrum Company (2010) General survey of radio frequency bands: 30MHz to 3 GHz.
Technical report
6. Haykin S (2005) Cognitive radio: brain-empowered wireless communications. IEEE J Sel
Areas Commun 23(2):201–220
7. Liu Y, Jun LIANG, Xiao N, Hu Y, Hu M (2016) Dynamic double threshold energy detection
based on Markova model in cognitive radio. J Electron Inf Technol 38(10):2590–259
8. Tian Z, Giannakis GB (2006) A wavelet approach to wideband spectrum sensing for cognitive
radios. In: Proceedings of 1st international conference on cognitive radio oriented wireless
networks and communications (CROWNCOM), pp 1–5
9. Kumar A, Saha S (2019) Discrete wavelet transform based hybrid edge detection technique
for wideband spectrum sensing in CRNs. In: IEEE region 10 symposium (TENSYMP) pp
529–534
10. Verma P (2020) Adaptive threshold based energy detection over Rayleigh fading channel. Wirel
Person Commun 113(1):299–311
Analysis of Signal Sensing with Adaptive … 151
11. Vladeanu C, Nastase CV, Martian A (2016) Energy detection algorithm for spectrum sensing
using three consecutive sensing events. IEEE Wirel Commun Lett 5(3):284–287
12. Martian A, Al Sammarraie MJ, Vlădeanu C, Popescu DC (2020) Three-event energy detection
with adaptive threshold for spectrum sensing in cognitive radio systems. MDPI J Sens (Basel)
20(13):3614
13. Nasrallah A, Hamza A, Boukaba T, Baudoin G, Messani A (2018) Energy detection with
adaptive threshold for cognitive radio. In: IEEE international conference on communications
and electrical engineering, pp 1–5
14. Yu S, Liu J, Wang J, Ullah I (2020) Adaptive double-threshold cooperative spectrum sensing
algorithm based on history energy detection. Hindwai J Wirel Commun Mob Comput
15. Yang H, Fu Y, Lei J (2018) A new cooperative spectrum sensing algorithm based on double
threshold. In: IEEE 2nd advanced IMCEC conference, pp 588–592
16. Hu L, Cao N, Mao M, Liao H, Shi R, Wang S (2019) Dynamic adaptive double-threshold
cooperative spectrum sensing with multi-level quantization. In: IEEE 3rd advanced IMCEC
conference, pp 1381–1385
17. Sarala B, Devi SR, Sheela JJ (2020) Spectrum energy detection in cognitive radio networks
based on a novel adaptive threshold energy detection method. Elsevier J Comput Commun
152:1–7
18. Mahendru G, Shukla AK, Banerjee P, Patnaik LM (2019) Adaptive double threshold based
spectrum sensing to overcome sensing failure in presence of noise uncertainty. In: IEEE 6th
international conference on SPIN, pp 466–471
19. Kockaya K, Develi İ (2020) Spectrum sensing in cognitive radio networks: threshold
optimization and analysis: research square
20. Urkowitz H (1967) Energy detection of unknown deterministic signals. Proc IEEE 55(4):523–
531
21. Verma G, Sahu OP (2016) Intelligent selection of threshold in cognitive radio system.
Telecommun Syst 63(4):547–556
22. Koley S, Mirza V, Islam S, Mitra D (2015) Gradient based real-time spectrum sensing at low
SNR. IEEE Commun Lett 19(3):391–394
23. Verma G, Sahu OP (2016) Opportunistic selection of threshold in cognitive radio networks.
Wirel Person Commun 92(2):711–726
Collaborative Design in Concurrent
Engineering of Industry 4.0 Through
Virtual Reality Simulation in Achieving
Accelerated Time-To-Market During
COVID-19 (Coronavirus) Pandemic
Outbreak
Abstract In the era of industry 4.0, it is important for businesses to conduct design
and engineering activities via data exchange in cloud computing. Collaborative
design in concurrent engineering through virtual reality simulation using digital
mock up (DMU) presented in this paper is to construct a simulation in digital envi-
ronment in order to reduce development time, cost and improve quality of product
which will impact to accelerated time to market. However, due to current COVID-
19 (Coronavirus) pandemic outbreak, the activities could be interrupted. The virtual
reality design collaboration seems to be as a solution with the support of cloud storage
of product design and satisfy design issue in concurrent engineering. It is also over-
come the geographical limitation and social distancing constraint in “face to face”
collaboration. Thus, stakeholders’ activities are carried out as usual without inter-
ruption in the challenging period during the pandemic. Finally, virtual simulation of
collaborative design in concurrent engineering of previous related works conducted
by the author are briefly presented.
1 Introduction
Competitive market has forced industry to improve new product development and
improve the speed of product launch through digitalization. The basic formulation
and solution of accelerated time to market in new product launch is a vision for the
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 153
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_14
154 A. F. Fudzin et al.
concept of verification using digital product and verification using real objects [14].
The DMU process lead to significant reduction in the number of real object verifica-
tion and just limited for testing purpose only [15]. As the DMU’s goal are to reduce
development time, cost and improve quality, this will impact to accelerated time
to market of product. Furthermore, the number of design change due to assembly
problems and the amount of prototype validation needed for production during the
product validation stage will be reduced. The advantage of virtual simulation in the
design process is a key enabler for manufacturers to have a faster and more flexible
approach to solving complex issues in the collaborative design environment.
In the early stage of design, face to face communication during collaboration design
of stakeholders will support faster feedback between designer and process engi-
neer. Figure 2 demonstrates the process of collaboration design in concurrent engi-
neering. The designer will issue design to process engineer for review process, so
that, product change request can immediately be considered for the change. At this
point of time, the industry 4.0 through cloud computing which shared platforms that
serve to multiple users (Shared Data Format) is very essential to support engineer
to review the same data and same format [9, 16]. Faster feedback can be achieved
from several parties in respond to the design issue aroused. It can be visualized as
designer and process engineer synchronized both the design and visualization data
formats on the same computer screen, eventhough they are from different part of the
world. This will help to reduce design-cycle time and cuts cost.
On design stage of product development, virtual computing analysis of manu-
facturing processes can be applied using 3D CAD/CAM modeling using CATIA
V5 (Dassault Systemes) [17]. It is a key enabler in the technology of Industry 4.0
and integrated process from Design for Manufacturing (DFM) [18] to visual and
augmented reality simulation can be achieved using seamless data transfer using
cloud computing [19]. The presented case study in automotive simulation analysis is
used in connection to 3D modeling. Three (3) stages of analysis investigated using
virtual simulation in development phase which are as followings;
At this stage clash and interference check is using human manikin virtual simulation.
This investigation is to provide a visual simulation on how the inference of tools clash
with components during assembly and disassembly. It is very important for automo-
tive maintenance and service operation [15]. All possible clash and interference are
analyzed.
158 A. F. Fudzin et al.
The case study of tools accessibility analysis is to verify assembly tools’ accessi-
bility applied during the assembly process during assembly process. Particularly, tool
placement constraints during tool applications are verified into the assembly product.
Tools accessibility analysis is an efficient approach in assessing the tools require in
production assembly line. Prior to the purchase of tools by the procurement team,
this activity may eliminate the use of incorrect tools for production. The analysis
utilized CAD data which retrieved from cloud data storage. Further, DMU CAD
models of the assembly product are imported into a virtual reality visualization tools
accessibility analysis.
The designer provided the components design to the cloud data storage and process
engineer as well as tool engineer have to analyze the tool accessibility to ensure that
the assembly process does not pose accessibility related problems. Prompt response to
designer in the case of accessibility issue can be rectified by the designer immediately.
In this case study, a virtual reality simulation is being utilized for the analysis as shown
in Fig. 3. This approach would detect any problem earlier in design phase.
Fig. 4 Virtual simulation of human hand movements position 1 to positon 2 (Diassembly process)
At this stage clash and interference check is using human manikin virtual simulation.
The verification of the reach-ability of tools to perform a task is very crucial. This
investigation is to provide a visual simulation on how the reach-ability and inference
of tools clash with components. In the design phase all possible class and interference
are analyzed virtually. In Fig. 5, tool clash with the surrounding can be identified.
Feedback from process engineer to designer can immediately improve in the first
review session. The improvement of product design is validated in the second review
before proceed to the next phase. In order to support accelerated time to market, this
rapid and accurate assessment is needed.
Another example of reach-ability is the interference analysis for disassembly of oil
filter from the engine. As explained in previous section the tool used to disassembly
by turning the tool from position 1 to position 2. In continuation of that, human
hand is used to completely disassemble the oil filter. Figure 6 shows the virtual
simulation interference study during the disassembly of oil filter using hand. During
160 A. F. Fudzin et al.
the disassembly proceed we found the interference occurred between hand’s manikin
and sub frame. The red circle in the diagram clearly visualize the interference. With
help of simulation software this occurrence can immediately feedback to designer
in the form of product change request. It is proven that early detection during DMU
virtual simulation check will eliminate any cost during prototype stage. Figure 7
shows the proven real serviceability check where the interference was eliminated in
design phase.
to continue moving forward using virtual reality simulation supported the design
review activities through the data exchange in cloud computing. It is a good tool
in order to increase social distancing and to adhere to the restriction on face to face
collaboration beside making simulation in economical ways. Based on the case study
presented in the previous section, stakeholders have been shown to be able to work
in same space of virtual environment without face to face collaboration as new norm
in working culture.
Collaborative design in concurrent engineering using DMU approach is part of
successful factor in virtual reality simulation where product introduction is unin-
terrupted due to pandemic. In the area of automotive design, the such operation
presented in the case study, the operations were checked on a physical vehicle must
now be checked by means virtual verification tools. The development of the virtual
verification is opens new approach of the analyses performed on the DMU. The
application not only allows designers to simulate the tools accessibility, service-
ability check, tools clash and interference check, but also offers a set of tools useful
for analyzing the DMU to detect potential modelling errors which could jeopardize
product functionality.
As virtual reality simulation has been presented during product development cycle
of the automotive industry, which it is known as an important tool in virtual veri-
fication. The advantage of virtual verification in development design phase, cost
reduction is achievable due to reduce and eliminate most of physical prototype. The
availability of Industry 4.0 as a key for future business is leading to these two goals,
as previously addressed in achieving uninterrupted product launch and accelerated
time to market in the current pandemic situation.
6 Conclusions
Today, automotive industries are required to transfer most of the activities in virtual
environment to meet the product launch time line in competitive market. In the avail-
ability of industry 4.0 platform, it is an important for businesses to conduct design
and engineering activities of the data exchange in cloud computing. The deployment
162 A. F. Fudzin et al.
of Industry 4.0 pillars in the COVID-19 pandemic outbreak, helps the engineer to
continue moving forward using virtual reality simulation supported the design review
activities through the data exchange in cloud computing. The Internet is becoming an
important element for concurrent engineering applications. The presented approach
supported a collaborative design in concurrent engineering through the virtual reality
simulation using DMU to construct a simulation in the digital environment in order
to reduce development time, cost and improve quality, which will impact to accel-
erated time to market of product. The stakeholders’ activities are carried out as
usual without interruption in the challenging period during the COVID-19 pandemic
outbreak. In this way, the virtual simulation design including cloud storage supports
product design and satisfy design issue in concurrent engineering. It is also over-
come the geographical limitation and social distancing constraint in “face to face”
collaboration in order to avoid possible infections among stakeholders. Therefore,
web-based virtual communication method is taking place in rapid pace. Finally, it
should be mentioned that the virtual simulation of collaborative design in concurrent
engineering of the previous related works has been implemented into the develop-
ment process of automotive manufacturer and finds its application in design phase.
In conclusion, with the presence of the latest Industry 4.0 application, there is no
significant impact on the disruption due to the COVID-19 pandemic in the virtual
reality simulation phase.
References
1. Fragner A, Kreis A, Hirz M (2020) Virtual tools to support design and production engineering:
early detection of stone chips to optimize production processes. In 2020 IEEE 7th international
conference on industrial engineering and applications (ICIEA), 2020, pp 399–403
2. Avdiu B, Nayyar G (2020) When face-to-face interactions become an occupational hazard:
jobs in the time of COVID-19, ed: The World Bank
3. del Rio-Chanona RM, Mealy P, Pichler A, Lafond F, Farmer D (2020) Supply and demand
shocks in the COVID-19 pandemic: An industry and occupation perspective. arXiv preprint
arXiv:2004.06759, 2020
4. Katapally TR (202) A global digital citizen science policy to tackle pandemics like COVID-19.
J Med Int Res 22:e19357
5. Wu D, Rosen DW, Wang L, Schaefer D (2015) Cloud-based design and manufacturing: a new
paradigm in digital manufacturing and design innovation. Comput Aided Des 59:1–14
6. Zhou J, Li P, Zhou Y, Wang B, Zang J, Meng L (2018) Toward new-generation intelligent
manufacturing. Engineering 4:11–20
7. Xu LD, Xu EL, Li L (2018) Industry 4.0: state of the art and future trends. Int J Prod Res
56:2941–2962
8. Hermann M, Pentek T, Otto B (2016) Design principles for industrie 4.0 scenarios. In: 2016
49th Hawaii international conference on system sciences (HICSS), 2016, pp 3928–3937
9. Erboz G (2017) How to define industry 4. 0: the main pillars of industry 4. 0. In: 7th international
conference on management (ICoM 2017), At Nitra, Slovakia, 2017, pp 1–2
Collaborative Design in Concurrent Engineering … 163
10. Ružarovský R, Holubek R, Sobrino DRD, Janíček M (2018) The simulation of conveyor control
system using the virtual commissioning and virtual reality. Adv Sci Technol Res J 12
11. Vila C, Ugarte D, Ríos J, Abellán J (2017) Project-based collaborative engineering learning to
develop Industry 4.0 skills within a PLM framework. Procedia Manuf 13:1269–1276
12. Fuh JY, Li W (2005) Advances in collaborative CAD: the-state-of-the art. Comput Aided Des
37:571–581
13. Sułkowski Ł (2020) Covid-19 pandemic; recession, virtual revolution leading to de-
globalization? J Intercult Manage 12:1–11
14. Döllner G, Kellner P, Tegel O (2000) Digital mock-up and rapid prototyping in automotive
product development. J Integr Des Process Sci 4:55–66
15. Gynn M, Steele J (2015) Virtual automotive maintenance and service confirmation. In: SAE
technical paper 0148–7191
16. Zheng P, Sang Z, Zhong RY, Liu Y, Liu C, Mubarok K et al (2018) Smart manufacturing systems
for Industry 4.0: conceptual framework, scenarios, and future perspectives. Front Mech Eng
13:137–150
17. Ghionea IG, Devedžić G, Ćuković S (2015) Parametric modeling of surfaces using CATIA v5
environment. In: Applied mechanics and materials, 2015, pp 93–98
18. Myung S, Song K, Lee J (2002) Integration of DFM and virtual NC manufacturing process.
In: CIRP ISMS, pp 175–180
19. Posada J, Toro C, Barandiaran I, Oyarzun D, Stricker D, de Amicis R et al (2015) Visual
computing as a key enabling technology for industrie 4.0 and industrial internet. IEEE Comput
Graphics Appl 35:26–40
20. Kang X, Peng Q (2007) Analysis of tool accessibility in fixture setup planning. In: Interna-
tional design engineering technical conferences and computers and information in engineering
conference, 2007, pp 1007–1016
21. Bruno M, Milite A, Monacelli G, Pina M, Sessa F (2006) Serviceability analyses in virtual envi-
ronment for the automotive industry. In: DS 36: proceedings design 2006, the 9th international
design conference, Dubrovnik, Croatia, 2006
Stability Analysis of Semi-Markovian
Discrete-Time Neural Networks
with Time-Varying Leakage Delays
1 Introduction
Devoting much attention to the study of neural networks (NNs) in last few years
has tremendously impacted the application prospect in the real world scenario. NN’s
are extensively applied in varied areas of research in large data and artificial intelli-
gence which includes applications in pattern recognition, robotics, optimization etc.
Significant number of researches has also been carried out in dynamic analysis of
various NNs. Stability analysis via state estimation procedure is one of the hottest
research topic in the current trends of digital technology using discrete-time NNs.
In the areas of control and signal processing application problem procedures,
discrete-time NNs’ play a vital role along with time delays. There are a lot of
K. Maheswari (B)
Department of Mathematics, Kumaraguru College of Technology, Coimbatore, India
S. N. Shivapriya
Department of Electronics and Communications Engineering, Kumaraguru College of
Technology, Coimbatore, India
C. Ramkumar
Department of Biomedical Engineering, Dr. N. G. P. Institue of Technology, Coimbatore, India
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 165
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_15
166 K. Maheswari et al.
approaches for the state estimation procedure in discrete time NNs that holds its
consideration with time delays several times and leakage delays at fewer number
of times. These time delays in the state estimation problems has drawn increasing
research attention in today’s digital world. This state estimation problem along with
Markovian jump parameters in discrete-time NN’s has been discussed in various
dynamical problems and the references therein [1–3]. Due to the strong modeling
ability of Markovian jump neural networks (MJNNs), greater attention has been
drawn in the field of hybrid dynamical systems like in the case of biology, medicine,
industry production, in aerospace Engineering.
But as is well known, in MJNNs, the transition matrix is converted into an invariant
matrix caused due to exponential distribution and obeys the stochastic process and is
not relevant with the past modes. This is one of the disadvantages of MJNNs. There-
fore, in order to overcome this shortcoming, semi-Markovian Jump Neural Networks
(SMJNNs) has been used where the probability density function matrix takes the key
role. Some of the recent researches have emphasized the synchronization controller
using SMJNNs given in [4] and the robust stochastic stability conditions have also
been incorporated in [5]. Inspired by all these works, in this paper, SMJNNs has been
applied to the model that has time varying coupling delays with randomly occur-
ring uncertainties [6]. By employing Lyapunov Karsovskii functional (LKF) with
different decision variables, sufficient conditions to attain the stability is achieved.
On the basis of the these discussions, the aim of this paper is to investigate the
asymptoticity for SMJNNs with delays that are mixed and time varying in nature.
By constructing suitable LKF’s and proceeding with the LMI technique, we derive
the feasibility of the stability conditions by the addressed NNs.
Notations:
Throughout the paper T and −1 represent the transpose and the inverse of the matrix
respectively. The n-dimensional Euclidean space is given by R n for the set of m x
n matrices in R mxn and λmax(min) (A) is the corresponding eigenvalue of the spectral
norm matrix H.
ς (t + 1) = B(ς (t − ψ(t))) + x(t) + C f (ς (t)) + D g(ς (t − d(t))) + J
(2)
where ς (t) = [ς1 (t), ς2 (t), . . . , ςn (t)]T ∈ R n is the state vector and the activa-
tion functions are given as f (◦) = [ f (ς1 (t)), f (ς2 (t)), . . . , f (ςn (t))]T ∈ R n ,
g (ς (t)) = [g(ς1 (t)), g(ς2 (t)), . . . , g(ςn (t))]T ∈ R n and J = [J1 , J2 , . . . , Jn ]T
is the input external vector. Each of ψ(t) in the system represents the delay due to
leakage which satisfies 0 < ψm ≤ ψ(t) ≤ ψ M where ψm is the lower bound and
ψ M is the upper bound of ψ(t) and d(t) represents the delay in the transmission
satisfying 0 < dm ≤ d(t) ≤ d M with the positive integers dm representing the lower
bound and d M representing the upper bound of d(t).
Assumption 1
For any c1 , c2 ∈ R, c1 = c2 , the activation functions f i (◦) and g i (◦) satisfy,
f (c1 ) − f i (c2 )
Mi− ≤ i ≤ Mi+ ,
c1 − c2
g (c1 ) − g i (c2 )
Ni− ≤ i ≤ Ni+ , i = 1, 2, . . . , n
c1 − c2
yi (t) = ςi (t) − ςi∗ , f i yi (t) = f i (ςi (t))
− f i (ςi∗ , ), gi (ςi (t − d(t)))
= g i (ςi (t − d(t))) − g i (ςi∗ )
By the above assumption given, it can be verified that for every f i (◦), gi (◦), i =
M − ≤ ( f i (c1 ) − f i (c2 ))/(c1 − c2 ) ≤ Mi+ ,
1, 2, . . . , n satisfy i− for any c1 = c2 and
Ni ≤ (gi (c1 ) − gi (c2 ))/(c1 − c2 ) ≤ Ni+
f i (0) = gi (0) = 0.
The model is associated with the initial condition as
where δ = max{ψ M , d M }.
given by
f i (l) gi, j (l)
g j,i (l) ≡ gi, j (l), ≥ 0. (6)
gi, j (l) f i j l)
Lemma 2 For any matrices ℘1 , ℘2 and ℘3 where ℘1T = ℘1 and ℘2T = ℘2 > 0, then
℘1 ℘3T
℘1 + ℘3T ℘2−1 ℘3 < 0 with < 0.
∗ −℘2
3 Main Results
In this section, we apply the semi-Markovian and investigate the delay dependent
sufficient conditions of LKF.
Theorem 3.1 The NNs system (3) is globally asymptotically stable, under the
assumption (1), if there exists the symmetric matrices Pi > 0, i = 1, . . . , 6, Q r >
0, r = 1, . . . , 4, R1 > 0, R2 > 0, T j > 0 with the diagonal matrices 1 > 0, 2 >
0 and matrices U j , j = 1 to 4 with weights W0 , W1 of appropriate dimensions, such
that following LMI holds with.
Stability Analysis of Semi-Markovian … 169
T1 T2
>0 (7)
T3 T4
where P̃i = γi j (h)P j ,
j∈S
1 = ẽ8 P̃1 ẽ8 + 2ẽ8 P̃1 ẽ1 + 2ẽ8 P̃3 (ẽ5 − ẽ7 )
+ 2ẽ8 P̃2 (ẽ2 − ẽ4 ) + 2ẽ8 P̃3 (ẽ11 − ẽ12 )
+ 2ẽ1 P̃2 (ẽ2 − ẽ4 ) + 2ẽ1 P̃3 (ẽ5 − ẽ7 )
+ (ẽ2 − ẽ4 )T P̃4 (ẽ2 − ẽ4 ) + 2(ẽ2 − ẽ4 )T 2ẽ1 P̃2 (ẽ2 − ẽ4 )
2(ẽ2 − ẽ4 )T P̃5 (ẽ5 − ẽ7 ) + 2(ẽ2 − ẽ4 )T P̃5 (ẽ11 − ẽ12 )
+ 2(ẽ5 − ẽ7 )T P̃6 (ẽ5 − ẽ7 )2(ẽ5 − ẽ7 )T P̃6 (ẽ11 − ẽ12 ).
= 2(ẽ8 W0T + ẽ1 W1T )(B ẽ3 + C ẽ9 + D ẽ10 − ẽ8T − ẽ1 ).
−ψm
t−1
t−1
V2 (t) = y T (s)Q 1 y(s) + y T (s)Q 2 y(s) (9)
s=t−ψ(t) j=−ψ M +1 s=t+ j
170 K. Maheswari et al.
t−1 −dm
t−1
V3 (t) = y (s)Q 3 y(s) +
T
y T (s)Q 4 y(s) (10)
s=t−d(t) j=−d M +1 s=t+ j
−dm
t−1
T
y(s) R1 0 y(s)
V4 (t) = d1 (11)
η(s) 0 R2 η(s)
j=−d M +1 s=t+ j
T
R1 0
ŷ(t) ŷ(t)
V4 (t) = d12
η̂(t)
0 R2 η̂(t)
y(t) T R1 0 y(t)
t−dm −1
− d1 (15)
η(t) 0 R2 η(t)
s=t−d M
where ψ1 = ψ M − ψm and d1 = d M − dm .
Also, by zero initial conditions, it is seen that,
172 K. Maheswari et al.
d1 y T (t − d(t))U1 y T (t − d(t)) − y T (t − d M )U1 y T (t − d M )
t−d(t−1
− η T (s)U1 (η(s) + 2y(t)) = 0,
s=t−d M
d1 ŷ T (t − dm )U2 ŷ T (t − dm ) − ŷ T (t − d(t))U2 ŷ T (t − d(t))
t−dm −1
Also,
T (t)
⎡ T ⎤
y (t), y T (t − ψm ), y T (t − ψ(t)), y T (t − ψ M ), y T (t − dm ), y T (t − d(t)), y T (t − d M )
⎢ m −1 m −1
⎥
=⎢ ⎣ η(t) f T (y(t)) g T (x(t − d(t))
t−d t−d ⎥,
⎦
η(s) y(s)
s=t−d M s=t−d M
t−dm −1 T
ŷ(t) ŷ(t) R1 R2
− d1
η(t) η(t) R3 R4
s=t−d(t)
+ d1 ŷ T (t − d(k))U1 ŷ T (t − d(k))
− ŷ T (t − d M )U1 ŷ T (t − d M )
+ ŷ T (t − dm )U2 ŷ(t − dm )
− ŷ T (t − d(t))U2 ŷ T (t − d(t))
t−d(t)−1
− η T (s)U1 η(s) + 2 ŷ(s)
s=t−d M
t−dm −1
t−d(t)−1 T
d1 ŷ(t)
< (t)4 (t) −
T
d M − d(t) η(t)
t=t−d M
Stability Analysis of Semi-Markovian … 173
t−d(t)−1
ŷ(t)
R 1 U1
∗ R 2 + U1 η(t)
t=t−d M
⎛ ⎞
t−dm −1 TT
d1 ⎝ ŷ(t) ⎠
−
d(t) − dm t=t−d(t) η(t)
⎛ ⎞
t−d m −1
R 1 U2 ⎝ ŷ(t) ⎠
∗ R 2 + U2 η(t)
t=t−d(t)
⎢ η(t) ⎥ ⎡R ⎤
⎢ t=t−d M ⎥ 1 U1 T1 T2
⎢ ⎥ ⎢
⎢ ŷ(t − d(t) − ŷ(t − d )) ⎥ ⎢ ∗ R 1 + U1 T3 T4 ⎥ ⎥
−⎢ M
⎥ ⎣
⎢
t−dm −1 ⎥ ∗ ∗ R 1 U2 ⎦
⎢ y(t) ⎥
⎣ t=t−d M ⎦ ∗ ∗ ∗ R 2 + U2
ŷ(t − dm ) − ŷ(t − d(t)))
174 K. Maheswari et al.
⎡ ⎤
t−dm −1
⎢ η(t) ⎥
⎢ t=t−d M ⎥
⎢ ⎥
⎢ ŷ(t − d(t) − ŷ(t − d M )) ⎥
×⎢ ⎥
⎢
t−dm −1 ⎥
⎢ y(t) ⎥
⎣ t=t−d M ⎦
ŷ(t − dm ) − ŷ(t − d(t)))
≤ T (t)(4 + 5 )(t) (17)
where
8
If λ̃ = λmax () < 0, = s with
s=1
Stability Analysis of Semi-Markovian … 175
2
V (t) ≤ λ̃ ŷ(t) , ∀ ŷ(t) = 0. (21)
Remark These results can also be applied to the system of randomly occurring
parameter uncertainties, where the stability conditions does not depend upon the
transmission delays caused due to delay bounds but also depends upon the time-
varying leakage delays.
4 Numerical Example
Example 1 Let us take the uncertain NNs with the subsystem as:
0.8 0 −2.3 0 1.3 0.2
B= ,C = ,D = ,
0 0.7 1.7 0.4 0.1 −0.02
0.04 0.02 0.07 0 0.03 0.02
Na = , Nb = , Nc =
0 −0.03 0.04 0.03 0.01 −0.04
The time-varying
delay
and
the leakage delay are given
by.
d(t) = 21 7 + 5 sin tπ2 and ψ(t) = 21 5 + 3 sin tπ2 with corresponding
bounds given by dm = 3, d M = 8, ψm = 1, ψ M = 5. By solving the LMIs,
a solution is obtained.
Figure 1 showcases NNs converge smoothly in the absence of leakage delay and
there is some kind of dynamic behavioral change in the network in the presence of
leakage delay. Hence, the effectiveness of the given result is thus verified.
5 Conclusion
In this paper, the analysis of stability problem is discussed with leakage delays.
Sufficient conditions are encountered to ensure the delay-dependent conditions and
to ensure the robustness by using the LMI approach. Numerical example has also
been for the derived theoretical result.
176 K. Maheswari et al.
1.2
x (k)
1
1
x (k)
2
0.8
0.6
0.4
x(k)
0.2
−0.2
−0.4
−0.6
−0.8
0 1 2 3 4 5
Time k
References
1. Das S, Moura JMF (2015) Distributed Kalman filtering with dynamic observations consensus.
IEEE Trans Sign Process 63(17):4458—4473
2. Hu J, Wang Z, Gao H (2018) Joint state and fault estimation for time-varying nonlinear systems
with randomly occurring faults and sensor saturations. Automatica 97(2018):150–160
3. Maheswari K (2016) H∞ state estimation for discrete-time neural networks with multiple
missing measurements using second order reciprocal convex approach. Adv Intell Syst Comput
412:102–119
4. Rakkiyappan R, Maheswari K, Sivaranjani K (2018) Young Hoon Joo, Non-Fragile finite-time
l2 − l∞ state estimation for discrete-time neural networks with semi-Markovian switching and
random sensor delays based on Abel lemma approach. Nonlinear Anal Hybrid Syst 29:283–302
5. Rakkiyappan R, Maheswari K, Velmurugan G, Park JH (2018) Event-triggered H∞ state esti-
mation for semi-Markov jumping discrete-time neural networks with quantization. Neural Netw
105:236–248
6. Nagamani G, Ramasamy S (2016) Dissipativity and passivity analysis for discrete-time T-S
fuzzy stochastic neural networks with leakage time-varying delays based on the Abel lemma
approach. J Franklin Inst 353:3313–3342
Forecasting PM10 Concentration Based
on a Hybrid Fuzzy Time Series Model
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 177
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_16
178 Y. Alyousifi and M. Othman
1 Introduction
forecasting and to produce better model accuracy. The proposed model is validated
by comparison with several existing models.
2 Methodology
FTS is a forecasting model based on the fuzzy set theory and Fuzzy Logic proposed
by Zadeh in 1965 [14]. FTS models are considered as advanced statistical time series
analysis applying to fuzzy sets. due to their advantages. Song and Chissom [15, 16]
are firstly introduced the first-order FTS model. To reach the high accuracy of the
forecasting model, developing FTS models have been done by enhancing the three
main stages of the FTS algorithm, which are fuzzification, fuzzy inference, and
defuzzification [16]. For instance, Chen [17] has improved the FTS model proposed
by Song and Chissom [15, 16] through the use of fuzzy logical relation tables for
the purpose of reducing the computational complexity of the model. Also, Huarng
[18, 19] have developed Chen’s model [17] by handling the issue of the length of the
interval.
The C-Means clustering technique [20, 21] is one of the famous unsupervised
learning algorithms. It is a partitioning clustering algorithm, which partitions a given
data set into a set of C clusters. The result of the C-Means clustering Technique
depends on the number of clusters. Apart from that, the main matter in partitioning
clustering is determining the optimal number of clusters in a data set. For instance,
the C-Means clustering requires the user to specify the number of clusters k in order
to be generated. Accordingly, in this paper, an optimal number of clusters is deter-
mined based on three different methods, which are elbow, average silhouette, and gap
statistic methods [22], through using two functions in R, which are fviz_nbclust() and
NbClust() [23]. By using these functions, the optimal number of clusters determines
the best partition selected. For further details about these methods see [24].
The algorithm of this study is adopted the arithmetic processes proposed by Tsaur
[9] in order to conduct the proposed model. The steps of the model are as follows:
Step 1. Define U.
Step 2. Divide U into subintervals based on the C-Means Clustering technique.
Step 3. Define the fuzzy sets Ai for each time series observation on U.
180 Y. Alyousifi and M. Othman
Step 4. Fuzzify the actual values of the time series (observations) into fuzzy
numbers based on the maximum membership value.
f Ai (u 1 ) f A (u 2 ) f A (u n )
Ai = + i + ··· + i (1)
u1 u2 un
Ni j
Pi j = , i, j = 1, 2, . . . , n (2)
N i.
n
where Ni j is the frequencies and Ni. = Ni j is the total frequencies.
j=1
Step 6. Calculate the forecasted values by considering the following cases:
Case 1. If the FLRG of Ai is one-to-one, i.e., Ai → Ak , with Pik = 1 and Pi j =
0, j = k), then the forecasting of F(t) is m k , the midpoint of u k , k = 1, 2, ..n,
which is determined by using Eq. (3)
Three statistical criteria used in this study in order to validate the proposed model,
which are MAPE, RMSE, and Thiels’ U statistics, are given in Equations as follows:
N
1 Yi − Fi
MAPE = × 100, (6)
N i=1 Yi
N
i=1 (Yi − Fi )2
RMSE = and (7)
N
N
i=1 (Yi − Fi )2
Theil’s U = (8)
N N
i=1 Yi +
2 2
i=1 Fi
where Yi means the actual data, Fi the predicted values and N is the total number of
the actual data.
In this section, we will provide the results of the proposed model through the imple-
mentation and comparison of the model using two datasets. The algorithm of the
proposed model is implemented for predicting the concentration of PM10 . To vali-
date the proposed model, the datasets are used to assess the model performance and
compare it with other models.
For validation of the proposed model, the algorithm of the proposed model is also
applied for both training and testing datasets of PM10 concentrations are used to
evaluate the performance of the proposed model and compare it with some of the
existing FTS models. For implementing the Algorithm of the proposed model, we
start by defining U from data and then apply the C-Means clustering method.
As can be seen from Table 1 that the proposed model has been performed very well
with producing small errors. Figures 1 and 2 demonstrate the results of a compar-
ison of the proposed model and some existing models using the testing data of
PM10 concentrations. It observed that the model has predicted PM10 data well with
producing the smallest error. Besides, the proposed model outperformed the other
existing models. Also, the model shows its superiority as compared with the other
182 Y. Alyousifi and M. Othman
Table 1 A comparison of the hybrid model and some FTS models using PM10 data
Model Using training dataset Using testing dataset
RMSE MAPE U statistic RMSE MAPE U statistic
1 Chen’s model [17] 18.04 14.67 1.32 9.12 14.26 0.80
2 Sliva2 et al.’s model [25] 16.10 17.79 1.18 10.23 18.68 0.89
3 Yu’s model [26] 17.06 9.74 1.25 8.87 14.20 0.78
4 Cheng’s model [27] 16.70 10.88 1.22 8.67 13.69 0.76
5 Severiano et al.’s [28] 16.10 17.79 1.18 10.23 18.68 0.89
6 Sliva et al.’s model [29] 15.67 9.28 1.14 8.90 13.92 0.78
7 Sadaei et al.’s [30] 17.06 9.82 1.25 8.73 13.99 0.77
8 Tsuar’s model [9] 17.07 9.40 1.25 8.66 13.86 0.76
9 Hybrid Model 7.55 6.83 0.55 5.01 7.25 0.45
Fig. 1 A comparison of the hybrid model and some FTS models using the training PM10 Data
models. This implies the proposed model is an improved option for forecasting air
pollution events.
4 Conclusion
This study proposed a new hybrid fuzzy time series model, which is implemented
for predicting PM10 concentration. The model hybridization has been contributed in
Forecasting PM10 Concentration Based on a Hybrid … 183
Fig. 2 A comparison of the hybrid model and some FTS models using the testing PM10 Data
producing adequate partitioning and improving the model accuracy. The proposed
model was evaluated by comparison with eight FTS models, which are famous in
the literature. A comparison has demonstrated the ability of the model in avoiding
the arbitrary selection of intervals and dealing with recurrent observations, which
greatly improves model accuracy. The predicted values found by the model display
its flexibility in the FTS for predicting air pollution. Generally, the proposed model
has the flexibility to apply for many types of time series data.
References
1. Yan Y, Li Y, Sun M, Wu Z (2019) Primary pollutants and air quality analysis for urban air in
China: evidence from Shanghai. Sustainability 11(8):2319
2. Alyousifi Y, Othman M, Faye I, Sokkalingam R, Silva PC (2020) Markov weighted fuzzy
time-series model based on an optimum partition method for forecasting air pollution. Int J
Fuzzy Syst 22:1468–1486
3. Cheng CH, Huang SF, Teoh HJ (2011) Predicting daily ozone concentration maxima using
fuzzy time series based on a two-stage linguistic partition method. Comput Math Appl
62(4):2016–2028
4. Cagcag O, Yolcu U, Egrioglu E, Aladag CA (2013) Novel seasonal fuzzy time series method
to the forecasting of air pollution data in Ankara. American Journal of Intelligent Systems.
3(1):13–19
5. Koo JW, Wong SW, Selvachandran G, Long HV (2020) Prediction of Air Pollution Index in
Kuala Lumpur using fuzzy time series and statistical models. Air Q Atmosp Health 1–12
6. Dincer NG, Akkuş Ö (2018) A new fuzzy time series model based on robust clustering for
forecasting of air pollution. Eco Inform 43:157–164
184 Y. Alyousifi and M. Othman
7. Yang H, Zhu Z, Li C, Li R (2019) A novel combined forecasting system for air pollutants
concentration based on fuzzy theory and optimization of aggregation weight. Appl Soft Comput
105972
8. Wang J, Li H, Lu H (2018) Application of a novel early warning system based on fuzzy time
series in urban air quality forecasting in China. Appl Soft Comput 71:783–799
9. Tsaur RC (2012) A fuzzy time series-Markov chain model with an application to forecast the
exchange rate between the Taiwan and US dollar. Int J Innov Comput Inf Control 8(7):4931–
4942
10. Aripin A, Suryono S, Bayu S (2016) Web-based prediction of pollutant PM10 concentration
using Ruey Chyn Tsaur fuzzy time series model. AIP Conf Proc 17(46):20–46
11. Alyousifi Y, Othman M, Sokkalingam R, Faye I, Silva PC (2020) Predicting daily air pollution
index based on fuzzy time series Markov chain model. Symmetry 12(2):293
12. Van Tinh N, Vu VV, Linh TTN (2016) A new method for forecasting enrolments combining
time-variant fuzzy logical relationship groups and k-means clustering. Int Res J Eng Technol
3(3):1–32
13. Kai C, Fang-Ping F, Wen-Gang C v A novel forecasting model of fuzzy time series based on
k-means clustering. IWETCS, IEEE, 223–225
14. Zadeh LA (1965) Fuzzy sets. Inf Control 8(7):338–353
15. Song Q, Chissom BS (1993) Forecasting enrollments with fuzzy time series-Part I. Fuzzy Sets
Syst 54:1–10
16. Song Q, Chissom BS (1994) Forecasting enrollments with fuzzy time series-Part II. Fuzzy Sets
Syst 62(1):1–8
17. Chen SM (1996) Forecasting enrolments based on fuzzy time series. Fuzzy Sets Syst 81(3):311–
319
18. Huarng K (2001) Effective lengths of intervals to improve forecasting in fuzzy time series.
Fuzzy Sets Syst 123(3):387–394
19. Huarng K, Yu TH (2006) Ratio-based lengths of intervals to improve fuzzy time series
forecasting. IEEE Trans Syst Man Cybern Part B Cyber 36:328–340
20. Hartigan JA (1979) A K-means clustering algorithm: algorithm AS 136. Appl Stat 28(11):126–
130
21. Zhang Z, Zhu Q (2012) Fuzzy time series forecasting based on k-means clustering. Open J
Appl Sci 2:100–103
22. Rousseeuw PJ, Kaufman L (1990) Finding groups in data. Wiley, Hoboken
23. Charrad M, Ghazzali N, Boiteau V, Niknafs A (2014) NbClust: An R package for determining
the relevant number of clusters in a data set. J Stat Softw 61:1–36
24. Kaufman L, Rousseeuw PJ (2009) Finding groups in data: an introduction to cluster analysis.
John
25. Silva PC, Sadaei HJ, Guimarães FG (2017) Interval forecasting with fuzzy time series.
Conference: IEEE symposium series on computational intelligence
26. Yu HK (2005) Weighted fuzzy time series models for TAIEX forecasting. Physica A
349(34):609–624
27. Cheng CH, Chen TL, Chiang CH (2006) Trend-weighted fuzzy time-series model for TAIEX
forecasting neural information processing. Spring Berlin/Heidelberg Lect Notes Comput Sci
42(34):469–477
28. Severiano, C. A., Silva, P. C., Sadaei, H. J., Guimarães, F. G.: Very short-term solar forecasting
using fuzzy time series. In 2017 IEEE international conference on fuzzy systems (FUZZ-IEEE).
1–6 (2017).
29. Silva PC, Sadaei HJ, Guimarães FG (2019) Probabilistic forecasting with fuzzy time series.
IEEE Trans Fuzzy Syst 99
30. Sadaei HJ, Enayatifar R, Abdullah AH, Gani A (2014) Short-term load forecasting using a
hybrid model with a refined exponentially weighted fuzzy time series and an improved harmony
search. Int J Electr Power Energy Syst 62:118–129
Carbonated Water Injection for EOR
and CO2 Storage: A Simulation Study
Abstract CO2 Enhanced Oil Recovery (EOR) techniques have gained massive
attention by oil companies as they target the oil industry’s two main concerns of CO2
contribution to the climate change and the decline in oil production. Carbonated Water
Injection (CWI) is a promising EOR technique that promotes safe and economic CO2
storage, moreover, it mitigates the pitfalls of CO2 injection which include low sweep
efficiency, early CO2 breakthrough, and the risk of CO2 leakage in subsurface forma-
tions. Upon the contact of Carbonated Water (CW) with oil inside the reservoir, CO2
molecules cross the interface between the two fluids moving towards the oil phase
due to the concentration gradient and its higher solubility in hydrocarbons. Subse-
quently CO2 molecules diffuse inside the oil until thermodynamic equilibrium is
reached. CO2 dissolution in oil causes it to swell and consequently leads to changes
in its physical properties (viscosity and density). Such changes in oil properties lead
to improved sweep and displacement efficiency and thus higher recovery factor.
Several experimental studies have been reported in the literature, but little work has
been done on the simulation of CWI due to the complex multi-physics nature of
this process. In this paper, Schlumberger equilibrium-based compositional simulator
(ECLIPSE300) has been used to investigate the oil recovery and CO2 storage during
CWI. The carbonated water has been simulated using two injector wells placed at the
same location where they inject free CO2 and water at a certain volumetric ratio to
account for the mass fraction of dissolved CO2 inside the carbonated water. CO2 SOL
option has been used to account for CO2 solubility inside the water phase and a 2D
cartesian model (x, y) has been considered to avoid the effects of gravity which might
reduce the amount of CO2 dissolved inside the oil phase. A sensitivity analysis on
CW injection rate, and the effect of CO2 diffusion have been investigated. It was
found that low injection rate promotes longer contact time thus more CO2 molecules
A. Abdulrahman
Petroleum Engineering Department, Universiti Teknologi PETRONAS, Perak, Malaysia
J. Foroozesh (B)
School of Energy and Electronic Engineering, University of Portsmouth, Portsmouth, United
Kingdom
e-mail: [email protected]
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 185
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_17
186 A. Abdulrahman and J. Foroozesh
will get transferred to the oil leading to high oil recovery and CO2 storage. In addi-
tion, it was found that CO2 diffusion between grid cells has minimal impact on oil
recovery and CO2 storage in this case study.
1 Introduction
Carbonated water injection is a promising CO2 based EOR technique that injectes
CO2 in an effective and efficient way. In this method, the CW is prepared by dissolving
CO2 inside the water at surface conditions using gas infusion generator [1], then
the CW mixture is injected into the formation to increase oil recovery and store
the dissolved CO2 [2]. When CW contacts the oil phase inside the reservoir, CO2
molecules gradually cross the interface between the two fluids due to the concen-
tration gradient (cross phases diffusion) and higher solubility of carbon dioxide in
hydrocarbons, and subsequently, CO2 molecules diffuse over time within the resident
oil until thermodynamic equilibrium is reached. CO2 dissolution in oil causes it to
swell and consequently leads to a change in its physical properties (density–viscosity)
[2–6].
Several experimental studies have been carried out to evaluate the additional
recovery factor during CWI, these experiments have been conducted using different
setups such as coreflood rigs, sand-packs, and high-pressure transparent micro-
models. It has been found that CWI in secondary mode (after primary depletion)
yields a higher recovery factor compared to CWI in tertiary mode (after water injec-
tion). Sohrabi et al. [7] have performed coreflood experiment in a water-wet reservoir
core (sandstone), in the first part of the experiment, CWI was injected in a secondary
mode and the recorded increment in the recovery factor was 19%. In the second
part of the experiment, CWI was injected in a tertiary mode, the recorded incre-
ment the recovery factor was determined to be 9%. Mosavat and Torabi [2] have
conducted a flooding experiment in a sand-pack apparatus filled with sand made of
99.88% silicon, it was found that the injection of CO2 -saturated water resulted in
improving the conventional water flooding oil recovery factor by about 19.0% and
12.5% of OOIP for the secondary and tertiary scenario, respectively. Riazi et al. [8]
have injected CWI (tertiary mode) in a micro model made of two flat glass plates,
the recorded recovery factor by water injection was 27% and the increment in the
recovery factor due to CWI (post water flood) was found to be 6%.
Carbonated water injection EOR has the potential to store CO2 safely and econom-
ically in the geological formations. Kechut et al. [9] have performed an experimental
investigation of CO2 storage in sandstone core samples. During secondary CWI, 45%
of the dissolved CO2 was stored while in tertiary CWI it was found that about 51%
of injected CO2 was stored. Tavakolian et al. [10] have performed series of coreflood
experiments using sandstone core plugs and crude oil at real reservoir conditions. The
Carbonated Water Injection for EOR and CO2 Storage … 187
experiments were carried out to compare the performance of CWI and CO2 flooding.
CWI resulted in a less and more controllable CO2 production profile compared to
CO2 flooding. Also, it was found that duringCO2 flood experiments, CO2 production
exceeded that of CO2 storage very quickly and after that large volume of CO2 had
to be reinjected (recycled), whereas CO2 production during CWI was significantly
less than CO2 storage even after a large volume of CWI.
The additional oil recovery reported during CWI can be attributed to the effect
of several mechanisms simultaneously. These recovery mechanisms include oil
swelling, viscosity reduction, wettability alteration, IFT reduction, and CO2 exsolu-
tion [3–6, 11–13]. Sohrabi et al. [4–6] have investigated oil swelling during CWI by
conducting many flooding experiments in a micro model using a light oil (viscosity
= 0.83 cp) and also a viscous oil (viscosity = 16.5 cp). They have found out that
the swelling in the light oil was 105% while the swelling in the viscous oil was 23%
which led to the conclusion of swelling degree is a function of oil viscosity. A series
of experimental studies have been conducted by Miller and Jones [3] to investigate
the change in oil viscosity during the CWI process. They have run the experiments at
different values of pressures and temperatures, different oil samples with a viscosity
of 10 and 17 API were used. It was found out that in all the cases the viscosity of the
oil in the absence of CO2 increased as the pressure increased, while the viscosities
of the oil with CO2 in the solution decreased as the pressure increased. Mosavat
and Torabi [13] have performed several experiments to investigate the effect of CO2
diffusion on wettability alternation. It was also concluded that once CO2 molecules
reach the rock surface (in contact with the oil droplet), the molecules started replacing
the hydrocarbon molecules were adsorbed on the surface. This gradual replacement
caused the rock surface to shift its wettability towards the water-wet condition. Yang
et al. [11] have conducted lab experiment to measure dynamic interfacial tension of
oil-brine systems compared to oil-brine CO2 systems at different values of pressures
and temperatures. They realized that in the crude oil-brine system, the dynamic inter-
facial tension reduction was primarily caused by the natural surfactant which was
possibly generated due to chemical reactions between the crude oil and the reservoir
brine. While For the crude oil-brine CO2 system, the dissolution of CO2 into crude
oil was an additional factor to reduce the dynamic interfacial tension. Riazi [12] have
conducted a fluid flow experiment on CWI using a high-pressure micro-model. They
have noticed an interesting phenomenon of gas nucleation subsequent to depressur-
izing of the CW flooded. This pressure reduction appreciably affected the production
process, leading to an additional oil recovery (due to CO2 exsolution).
An accurate reservoir model with the ability to reliably predict the performance of
different recovery methods is vital for successful reservoir management. There are
black oil and compositional simulation approaches available depending on type of
reservoir fluids and recovery methods, Compositional simulation approach should
188 A. Abdulrahman and J. Foroozesh
Kechut et al. [9, 17, 18] used compositional simulators to simulate a set of carbonated
water coreflood experiments. The coreflood experiments had been performed using
two types of core (a reservoir water-wet core and Clashach water-wet sandstone core)
and two different oil samples: a light oil (n-decane) and a viscous stock-tank crude
oil with known composition. Secondary water injection, secondary CWI and tertiary
CWI were simulated using the ECLIPSE300 commercial reservoir simulators. First
Carbonated Water Injection for EOR and CO2 Storage … 189
water injection (WI) was simulated. The input water–oil relative permeability (Krw-
o) curve into the simulator was obtained based on Corey correlations when produc-
tion data was history matched. Next CWI was simulated. The ECLIPSE300 composi-
tional simulator was used with a tuned EOS, which could acceptably predict the PVT
properties of the oil-CO2 system. The Krw-o obtained for WI tests were input into
the ECLIPSE300 during CWI simulation. The simulation results by ECLIPSE300
showed an over prediction of oil recovery in comparison with experimental data
for all coreflood tests. They explained that compositional simulators were developed
based on the inherent assumptions of instantaneous equilibrium and complex mixing
while during carbonated water flooding, CO2 transferring process from water into the
oil phase happens gradually. The resistance against CO2 migration at the interface
between water and oil phases plays an important role, which prevents immediate
distribution of CO2 between the phases. Consequently, the assumption of instanta-
neous equilibrium may not be realistic during CWI process. To support their expla-
nation, they mentioned the micro-model observation reported by Sohrabi et al. [5],
who used image processing technique during the conducted visualisation tests. Riazi
et al. [8] have reported that in the micro-model system the swelling of oil blobs as
a result of CO2 diffusion happens gradually and it needs time to stabilise and reach
equilibrium state (170 h to reach its final shape corresponding to maximum swelling
value). Embid and Rivas [19] explained that this assumption of instantaneous equilib-
rium can lead to a large error in the cases where mass transfer resistances are large.
They mentioned that this may happen when there are short contact times for the
mass transfer process (laboratory displacement in cores) or large diffusion patterns
are available for components to diffuse through (field scale), moreover, if there are
slow diffusion velocities due to large viscosity. All these can lead to not having an
instantaneous thermodynamic equilibrium state.
The effect of CO2 diffusion within the oil phase is another challenge encountered
during the simulation of CWI. Molecular diffusion is defined as the movement of
molecules caused by Brownian motion or composition gradient in a fluid mixture
that can facilitate the dissolution of CO2 into the oil systems [20]. Injected CO2
during CO2 based injection projects could just pass through the reservoir, causing
early breakthrough, if there is no molecular diffusion. This would lead to lower
oil recovery prediction as some parts of the reservoir are not touched by the CO2
causing no benefit from the CO2 presence. This may lead to wrong prediction as
diffusion mechanism can play important roles during CO2 based injection projects
[21, 22]. Alfarge et al. [23] reported that most of the studies in the literature identified
molecular diffusion as an important mechanism for CO2 to enhance the oil recovery in
oil reservoirs. Grogan and Pinczewski [24] highlighted the importance of molecular
diffusion during the oilfiled life. As such disregarding the diffusion mechanism in the
model can lead to overestimation or underestimation of the recovered oil [22, 23].
190 A. Abdulrahman and J. Foroozesh
Foroozesh et al. [25, 26] have come up with a novel 1D compositional simula-
tion approach whereby the assumption of instantaneous equilibrium was relaxed by
adding a new term to the governing equation called mass transfer coefficient (MTC).
This term used to capture the slow CO2 mass transfer from CW to oil within the
grid cell. Their kinetic based model was used to simulate a secondary CWI experi-
ment reported by Sohrabi et al. [7] for validation purposes. MTC was tuned to match
differential pressure (DP) and recovery factor (RF) data of the experiments and a
good match was achieved. Additionally, Foroozesh and Jamiolahmady [25] inves-
tigated the role of rock wettability and wettability alteration in the performance of
CWI process. They simulated CWI coreflood experiments in mixed wet sandstone
cores. The relative permeability, capillary pressure and MTC values were tuned and
a good matching was achieved. The also used their kinetic based model to map the
result of CWI in core to reservoir (scale-up practice). Although, Foroozesh et al.’s
simulation approach was a novel way of simulating CWI but they did not capture the
effect of CO2 diffusion within the phases.
Fig. 1 a Oil water relative permeability and b gas oil relative permeability
A sensitivity analysis has been done on CW injection rate to investigate its impact on
oil recovery, CO2 storage and spatial spreading inside the reservoir. Three different
injection rates have been studies and the mass fraction of dissolved CO2 inside
injected CW for each scenario was 5%. In the first scenario, a total of 2000 STB/day
CW has been injected into the reservoir (1900 STB/day of water, 0.922 MSCF/day
192 A. Abdulrahman and J. Foroozesh
of free CO2 ) for 7500 days, while in the second scenario a total of 5000 STB/day
CW has been injected (4750 STB/day of water, 2.305 MSCF/day of free CO2 ) for
3000 days, and in the last scenario 8000 STB/day of CW has been injected (7600
STB/day of water, 3.688 MSCF/day of free CO2 ) for 1875 days.
It was found that the oil recovery factor for the injection rate of 2000 STB/day is
the highest (68.6%) and the wide spatial spreading of CO2 inside the reservoir led
to early CO2 breakthrough compared to the other scenarios thus less CO2 storage
(96%). This can be attributed to the long contact time between CW and oil phase
which allows more CO2 to be transfered based on the thermodynamic equilibrium
calculation. While for scenario three, less oil recovery was found (58%) and the
limited CO2 spatial spreading inside the reservoir caused 100% CO2 storage as CO2
did not reach the producer well. The flowing figures show the oil recovery, CO2
spatial spreading and CO2 storage for the three scenarios (Figs. 5, 6 and 7).
Carbonated Water Injection for EOR and CO2 Storage … 193
RF %
60
55
50
Fig. 6 Mass fraction of dissolved CO2 inside the oil phase at the end of simulation for a 2000
STB/day, b 5000 STB/day, c 8000 STB/day
Fig. 7 Cumulative injected free CO2 and cumulative produced CO2 for the different scenarios of
CW injection rate
194 A. Abdulrahman and J. Foroozesh
Fig. 8 Oil recovery vs time for the 3 scenarios of different CO2 -Oil diffusion coefficients
As CO2 gets transferred from CW to the oil phase, CO2 molecules diffuse inside the
oil due to the concentration gradient until the equilibrium state is reached. DIFFCOIL
keyword added in PROPS section provides CO2 -Oil diffusion coefficient between
adjacent grid cells. To study the effect of CO2 diffusion, three scenarios have been
simulated where in the first scenario the diffusion coefficient was set to zero while in
the second and third scenario the diffusion coefficient was set to 0.007758103 ft2/day
(1d) and 0.02327 ft2/day (3d) respectively. It was found that CO2 diffusion has a
minimal (negligible) impact on both oil recovery and CO2 storage. The following
figure shows the oil recovery for all the scenarios (68.4%) thus led to the same CO2
storage (100% as CO2 did not reach production well) (Fig. 8).
4 Conclusion
Carbonated water injection is a promising EOR technique that promotes high oil
recovery and safe CO2 storage. Upon the contact of Carbonated Water (CW) with
oil inside the reservoir, CO2 molecules cross the interface between the two fluids
moving towards the oil phase due to the concentration gradient, and subsequently,
CO2 molecules diffuse inside the oil until thermodynamic equilibrium is reached.
CWI was studied numerically using ECLIPSE300 where the CO2 SOL option was
used to account for CO2 dissolution inside the water phase. A sensitivity analysis
on injection rate and CO2 diffusion coefficient was done to understand their impact
on oil recovery and CO2 storage. It was found that low CW injection rates leads to
Carbonated Water Injection for EOR and CO2 Storage … 195
a higher oil recovery and a wider spatial spreading of CO2 inside the oil thus a higher
storage, while in the case of high injection rates, a less oil recovery was found and
the limited spreading of CO2 causes a lower CO2 storage. Additionally, the effect of
CO2 diffusion was found to be minimal (negligible) on both oil recovery and CO2
storage in this case study.
Acknowledgements The first author (AA) gratefully acknowledges the financial support of the
Ministry of Higher Education Malaysia and Universiti Teknologi PETRONAS under FRGS grant
(Grant No.: 015MA0-030).
References
1. Mosavat N (2014) Utilization of Carbonated Water Injection (CWI) as a Means of Improved Oil
Recovery in Light Oil Systems: Pore-Scale Mechanisms and Recovery Evaluation. University
of Regina, Faculty of Graduate Studies and Research
2. Mosavat N, Torabi F (2014) Application of CO2 -saturated water flooding as a prospective safe
CO2 storage strategy. Energy Procedia 63:5408–5419
3. Miller, J.S. and R.A. Jones. A laboratory study to determine physical characteristics of heavy
oil after CO2 saturation. in SPE/DOE Enhanced Oil Recovery Symposium. 1981. Society of
Petroleum Engineers.
4. Sohrabi M et al (2011) Safe storage of CO2 together with improved oil recovery by CO2 -
enriched water injection. Chem Eng Res Des 89(9):1865–1872
5. Sohrabi M et al. (2009) Mechanisms of oil recovery by carbonated water injection. in SCA
annual meeting
6. Sohrabi M et al (2011) Carbonated water injection (CWI)–A productive way of using CO2 for
oil recovery and CO2 storage. Energy Procedia 4:2192–2199
7. Sohrabi M et al (2012) Coreflooding studies to investigate the potential of carbonated water
injection as an injection strategy for improved oil recovery and CO 2 storage. Transp Porous
Media 91(1):101–121
8. Riazi M, Sohrabi M, Jamiolahmady M (2011) Experimental study of pore-scale mechanisms
of carbonated water injection. Transp Porous Media 86(1):73–86
9. Kechut, N.I., M. Sohrabi, and M. Jamiolahmady. Experimental and numerical evaluation
of carbonated water injection (CWI) for improved oil recovery and CO2 storage. in SPE
EUROPEC/EAGE Annual Conference and Exhibition. 2011. Society of Petroleum Engineers.
10. Tavakolian, M., et al. Significant improvement in oil recovery and CO2 storage by carbonated
water injection (CWI). in Third EAGE CO2 Geological Storage Workshop. 2012.
11. Yang D, Tontiwachwuthikul P, Gu Y (2005) Interfacial tensions of the crude oil+ reservoir
brine+ CO2 systems at pressures up to 31 MPa and temperatures of 27 C and 58 C. J Chem
Eng Data 50(4):1242–1249
12. Riazi, M., Pore scale mechanisms of carbonated water injection in oil reservoirs. 2011, Heriot-
Watt University.
13. Mosavat N, Torabi F (2016) Micro-optical analysis of carbonated water injection in irregular
and heterogeneous pore geometry. Fuel 175:191–201
14. Schlumberger, ECLIPSE Technical Description. 2014.
15. Schlumberger, ECLIPSE Reference Manual. 2014.
16. Whitson, C.H. and M.R. Brulé, Phase behavior. Vol. 20. 2000: Henry L. Doherty Memorial
Fund of AIME, Society of Petroleum Engineers ….
17. Kechut NI, Jamiolahmady M, Sohrabi M (2011) Numerical simulation of experimental carbon-
ated water injection (CWI) for improved oil recovery and CO2 storage. J Petrol Sci Eng
77(1):111–120
196 A. Abdulrahman and J. Foroozesh
18. Kechut, N.I., et al. Tertiary oil recovery and CO2 sequestration by carbonated water injection
(CWI). in SPE International Conference on CO2 Capture, Storage, and Utilization. 2010.
Society of Petroleum Engineers.
19. Embid S, Rivas O (1994) Simulation of miscible displacement with interphase mass transfer
resistance. SPE Advanced Technology Series 2(01):161–168
20. Mohebbinia, S. and T. Wong. Molecular diffusion calculations in simulation of gasfloods in
fractured reservoirs. in SPE Reservoir Simulation Conference. 2017. Society of Petroleum
Engineers.
21. Bear, J., Dynamics of fluids in porous media. 2013: Courier Corporation.
22. Peksa, A.E., K.-H.A. Wolf, and P.L. Zitha. Molecular Diffusion of CO From Carbonated Water
(CW) Into the Oil-Experimental Observations. in SPE Asia Pacific Oil and Gas Conference
and Exhibition. 2013. Society of Petroleum Engineers.
23. Alfarge D, Wei M, Bai B (2017) Factors affecting CO2-EOR in shale-oil reservoirs: numerical
simulation study and pilot tests. Energy & Fsuels 31(8):8462–8480
24. Grogan A, Pinczewski W (1987) The role of molecular diffusion processes in tertiary CO2
flooding. J Petrol Technol 39(05):591–602
25. Foroozesh J, Jamiolahmady M (2016) Simulation of carbonated water injection coreflood
experiments: an insight into the wettability effect. Fuel 184:581–589
26. Foroozesh J, Jamiolahmady M, Sohrabi M (2016) Mathematical modeling of carbonated water
injection for EOR and CO2 storage with a focus on mass transfer kinetics. Fuel 174:325–332
Study of Efficient FIR Filter Architecture
for Fixed and Reconfigurable
Applications
1 Introduction
High Complexity in Field Programmable Gate Array (FPGA) is the major demerit
for the program to execute in a running period which increases the delay duration,
which makes the parameter of power to be high. The complexity in the number of
slices used in Field Programmable Gate Array(FPGA) is increased in large numbers
and gradually increases the storage memory space. Earlier, the duration period of
high Look Up Table (LUT) were introduced, so that delay is increased in execution
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 197
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_18
198 B. Aruna Devi et al.
period of field arithmetic and one of the methods used is Distributed Arithmetic
(DA). Here the multiplications are rearranged and for the explicit multiplications
using Read Only Memory lookups, in analysis of FPGA based implementation, with
advantages of reduction in memory in FPGA. The timing of the DA based FIR filter
models is done using minimum clock periods depending on the delay of the final
adder stage in the PSAT (Pipeline Shift-Add Tree).However, which depends on the
critical path calculation of the adder. The throughput rate is the amount of output
produced by the critical path over a period of time.
In this paper, the study of various models of Finite Impulse Response Filter
with energy efficient architecture for reconfigurable applications is done. The paper
presents a comparative study of optimum low complexity, reconfigurable digital FIR
filter for low power with reduction in delay implementation.
2 Literature Survey
Many studies have been undergone towards FIR efficient implementation for various
applications [1]. The implementation of filter aids in proposing various models for
efficacy of power, delay, area and test bed analysis.
The key study deals with distributed FIR filter using the least mean block square
algorithm using Look Up tables, reducing the number of flip flops and adder. It
provides n/6 adders with reduction in area delay product, with block size tables less
than four block size tables. The implementation of the parallel structure with weighted
adaptation phrase, output filter computation and modified distributed arithmetic helps
to reduce the logic and power consumption of the look up tables with less words to
be updated per output. The algorithm includes decomposition into smaller vectors of
both input and weight vector, and filtering analysis of the inner product computation
[2].
Shah et al. [3] has implemented a competent FIR filter towards reconfigurable
implementation for an coherent FIR filter for fixed and reconfigurable applications
by implanting delay and area efficient addition based carry selector by optimizing
the superfluous operations, by performing the carry process before the ultimate sum
function analysis. The ability of the transpose form FIR filter implementation resulted
in 26.6% of reduction in delay and area of 15% in reconfigurable models.
An architecture based on distributed arithmetic FIR filter for increased speed
and less power consumption was proposed using block least mean square algo-
rithm. The implementation design provided an efficiency of 17.3 times, lower power
consumption of 17.3 times and 95% higher in terms of through output. The variable
coefficient adaptive proposed filter units includes the following operations: Register,
partial product computation, weight update and error generation, simultaneously
with shift accumulator look-up tables, emulate multiply and accumulate functions.
A systolic array architecture with parallel data processing units has been suggested
for the architecture. The parallelism was found to be successful with the number of
clock cycles reduced [4].
Study of Efficient FIR Filter Architecture … 199
The least-square linear phase FIR filter (LLFE) is frequently used in the analysis
of ECG signals. It aids in removal of low frequency mode using discrete wavelet
transform.
Distributed Arithmetic (DA) is a technique that is bit-serial in nature. It can there-
fore appear to be slow. It turns out that when the number of elements in a vector is
nearly the same as the word size, DA is faster in nature, reinstate the explicit multi-
plications overdone by read only memory look up table, by ROM look-ups and an
effective model to realize and to implement on Field Programmable Gate Arrays
(FPGAs). The coherent model is found effective for estimating of inner product or
multiply and accumulate (MAC). In Digital Signal Processing algorithms, the MAC
operation is popular. Multiplications are reordered and combined in DA so that the
arithmetic becomes “distributed” rather than “lumped” through the structure. When
DA is implemented in FPGAs, to execute the MAC operation, one can take advantage
of memory in FPGAs. When DA was devised (in the 1970s), the comparisons given
were in terms of number of TTL ICs required for mechanization of a certain type of
filter. In particular, for an eighth order digital filter operating at a word rate close to
1 MHz, 72 ICs with a total power consumption of about 30 W was stated with the
DA approach while 240 ICs with a power dissipation of 96 W was indicated for a
multiplier-based solution [5].
3 Proposed Work
FIR filter transpose type block was suggested with decreased register complexity. A
low-complexity design methodology that can be optimized for greater throughput
in reconfigurable applications for the block realization of fixed FIR filters using the
MCM scheme is proposed. It can be optimized for higher output in application that
are reconfigurable. A low-complexity design approach for the block implementation
of fixed FIR filters using the MCM scheme It can be configured for greater throughput
in reconfigurable applications as shown in Fig. 1. Here presented an optimum low
complexity design using the MCM scheme for the block implementation of fixed FIR
filters. MCM technique minimizes the number of adders and subtractors. As a result
the computation process is reduced. Since Power dissipation is directly related to the
hardware, some amount of power reduction is indirectly achieved by using MCM
4 Conclusion
Finite Impulse Response (FIR) digital filter is extensively used in various applica-
tions ranging from biomedical applications, to speech processing in various domains
of digital signal processing. Finite-Impulse Response (FIR) Reconfigurable Filter
whose filter coefficients alter dynamically during runtime has an important role. The
multiplier unit is the optical FIR filter’s building structure in its transposed form.
With the filter input realization, the multiplication filter coefficients can be intro-
duced, having a major effect on the complexity and design efficiency. Hence calcu-
lation of inner product or multiply and accumulate can be done efficiently. Hence we
have explored the possibility of realization of block FIR filter in transpose form
configuration for area-delay efficient realization of large order FIR filters for fixed
and reconfigurable applications.
References
1. Reddy KS, Suresh HN (2019) FPGA implementation of reconfigurable FIR filter using Vedic
design with CLA adder. Int J Adv Sci Technol 28(1):144–161. Retrieved from http://sersc.org/
journals/index.php/IJAST/article/view/228
2. Ahmed F (2019) The energy efficient fir filter for biomedical application using FPGA. Comput
Rev J 3:236–244
3. Shah S, Rajula S (2019) Design of FIR filter architecture for fixed and reconfigurable applications
using highly efficient carry select adder. Adv Intell Syst Comput 898:627–637 (2019)
4. Chowdari CP, Seventline JB (2020) Systolic architecture for adaptive block FIR filter for
throughput using distributed arithmetic. Int J Speech Technol 23:549–557 (2020). https://doi.
org/10.1007/s10772-020-09745-4
5. Mohanty BK, Kumar P (2013) A high-performance energy-efficient architecture for FIR adaptive
filter based on new distributed arithmetic formulation of block LMS algorithm. IEEE Trans
Signal Process 61(4):921–932
Improved Deep Learning Based
Prediction of Crop Yield Using
Bidirectional Long Short Term Memory
1 Introduction
Crop yield prediction (CYP), a challenging task, benefits farmers in order to take
managerial decisions and financial stands [1]. CYP is challenging because of factors
like environmental and field based features. Also, the food production of the world is
V. Saveetha (B)
Dr. N.G.P. Institute of Technology, Coimbatore, India
e-mail: [email protected]
R. Kingsy Grace · S. Birundha
Sri Ramakrishna Engineering College, Coimbatore, India
S. Sophia
Sri Krishna College of Engineering and Technology, Coimbatore, India
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 201
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_19
202 V. Saveetha et al.
dependent on CYP for precise forecast which leads to timely decisions [2]. Machine
learning (ML), that act has a stem of Artificial Intelligence, has been implemented
for agricultural prediction enormously in literature [3] due to its behavior of treating
outputs with respect to inherent function of the input features. Precision Agricul-
ture, the field of CYP manages to monitor measure and produce optimized outputs
with response to variability in crops [4]. Current research contributes vastly to esti-
mate the CYP justifiably but always a better performance is preferred [5]. Recently
ML techniques like Clustering, Classification, Regression and Neural Networks are
employed for CYP. In our study data of crop yield and other factors including statis-
tics and Deep Neural Networks (DNN) enables to produce effectual forecasting of
yield. DNN with hidden layers act to be more powerful to reveal the fundamental
nonlinear relationship between input and response variables [6]. In this paper Recur-
rent Neural Networks (RNN) [1] is used to predict and check crop yield. RNN works
with multiple stacked layers that convert the raw data into meaningful representa-
tions. RNN are well suited for modeling functions for which the data is composed of
vectors that involve a time dependency between the values like CYP [7]. It includes
a feedback loop that it uses to learn from sequences, including sequences of varying
lengths. The long short-term memory (LSTM) type of RNN has received enormous
attention in the realm of sequence learning. LSTM a popular variant of RNN has
internal memory to allow long-term dependencies to affect the output [8]. The LSTM
is good for CYP because it is designed to confine to the time dependencies of envi-
ronmental factors, it enables capability to predict crop yield without drop in accuracy
and combined with the bi directional method, the model predicts effectively.
This research aims to evaluate methods to detect rice cultivation in India, using
the LSTM and Bidirectional LSTM (Bi-LSTM) methods. The study compared the
results based on LSTM and Bi-LSTM with traditional machine learning techniques:
Support Vector Machines (SVM), Random Forest (RF) and Gradient Boost(XGB)
Regression. The rest of the paper is structured as Sect. 2 describes related work of
various models used for CYP. Section 3 presents the proposed methodology. Section 4
discusses the results and finally, concluding remarks is discussed in Sect. 5.
2 Review
Many methods exist in literature for CYP and every time an attempt is performed
by researchers to enhance the performance of prediction. In this paper an ensemble
based classification algorithm using Artificial Neural Networks (ANN), Decision
Tree and SVM is designed and implemented to predict crop yield. They try to increase
the accuracy of classification. The authors modeled an learning technique based on
AdaBoost ensemble algorithm and enhancement is performed using genetic opti-
mization technique [9]. Khaki and Wang suggested a DNN model for CYP across
from 2008 to 2016 which was analysed along with methods like Lasso, shallow
neural networks, and decsion tree [10]. You et al. designed a hybrid model with
Convolution Neural Network(CNN) and RNN to predict soybean yield [11]. Kim
Improved Deep Learning Based Prediction … 203
et al. suggested a DNN model for CYP using optimized input features from satellite
products and meteorological datasets from 2006 to 2015 [12]. Wang et al. [13] devel-
oped a Deep Learning(DL) framework to predict soybean crop yields in Argentina
and achieved good results with a transfer learning approach to predict Brazil soybean
harvests. Yang et al. (2019) analyzed the working of CNN to predict rice grain yield
on images and concluded that CNN model gives robust yield forecast through the
ripening stage [14]. This study shares a model to perform soybean yield forecasts in
Brazil using LSTM. They compare performance of linear regression, random forest
and LSTM for forecasting yield and evaluate accuracy [15]. The authors propose
a CNN-LSTM model for soybean yield prediction at the county-level. The model
was evaluated by crop growth variables and environment variables weather data.
They show the results of the experiment that performance of the proposed deep
CNN-LSTM model outperforms the CNN or LSTM model [16]. The above review
instigates that Deep learning can be used for CYP.
3 Methodology
Decision making in CYP require crop response models that make accurate estima-
tions to predict yields [17]. The purpose of the study is to find out whether RNN
can effectually predict crop yield for varying seasons, assess model performance
and compare the same with other models. The machine learning algorithms already
examined with respect to DNN are LSTM and CNN [8]. CNN is a specific class of
neural networks mostly used in image recognition. In this work Bi directional LSTM
is proposed for CYP in order to increase the performance accuracy. A comparison
of machine learning models and feature levels was performed based on the yield
estimation. Models were trained to predict yields from 1997 to 2015 for each state.
Model performance was evaluated using Root Mean Squared Error (RMSE), Mean
Absolute Percentage Error (MAPE), Mean Absolute Error (MAE) and Accuracy.
The input variables for the prediction model include the Crop yield year, Climatic
data and Area Cultivated. The yield prediction is based on the equation y = f(x),
where y is the yield predicted and x is the input data with respect to time.
Figure 1 shows the flow of the research performed. Initially Data Acquisition is
performed and the dataset obtained from [18] is explored for information following
which the data related to rice is extracted. Data preprocessing like encoding cate-
gorical variables, filling missing values and scaling are performed on the derived
dataset. The machine learning model and deep learning model are implemented and
the accuracy and error metrics for the same are analyzed.
LSTM works exclusively for prediction of sequence problems. The different archi-
tectures of LSTM are LSTM, CNN-LSTM, Encoder-Decoder LSTM, Generative
LSTM and Bi-LSTM. LSTM a RNN has an internal state, models parallel input
sequence, process variable-length input to generate variable-length output (Fig. 2).
The memory cell is the computational unit of the LSTM and consists of weights and
gates.
204 V. Saveetha et al.
Huang et al. [19] The objective of a Bidirectional LSTM is to make forward and
backward passes on the features of CYP in order to predict the yield. The past and
future features for a given time help to process the given sequence and thus provide
efficient implementation of the BDLSTM model. This model is trained using back
propagation through time (BPTT) [20] and the passes front and back are carried over
by unfolding the hidden states for all time steps. In fields like phoneme classification,
traffic prediction and speech recognition Bi-LSTM model are significantly better than
unidirectional ones and so an effort is made to use it in CYP [21]. The Bi-LSTM
Improved Deep Learning Based Prediction … 205
4 Result Discussion
The following indexes are selected to evaluate the forecast model performance: Accu-
racy, Mean absolute percentage error (MAPE), root-mean-square error (RMSE) and
mean absolute error (MAE) [9]. The MAPE is a measure of prediction accuracy in
statistics and is the most common measure used to forecast error, and works best if
206 V. Saveetha et al.
Table 1 Formula of
Performance metrics Formula
performance metrics
Accuracy Accuracy = P−A
P
N
Ak −Pk
MAPE M AP E = 1
N Ak
k=1
N
RMSE RMSE = 1
N (|Ak − Pk |)2
k=1
N
MAE M AE = 1
N |Ak − Pk |
k=1
there are no extremes to the data (and no zeros). RMSE is the standard deviation of
the residuals errors. Residuals tells how concentrated the data is around the line of
best fit. The RMSE is a measure of the differences between values predicted by a
model and the values actually observed. It is used to compare forecasting errors of
different models for a particular variable and not between variables, as it is scale-
dependent. The values for the above metrics are obtained from Actual Values and
Predicted Values of yield after training and testing the dataset. The sample Actual
versus Predicted trend is shown in Fig. 4. Table 1 shows the formula employed to
find the metrics.
4.3 Comparison
Table 2 shows the performance metrics values for the models which are compared
with Bi-LSTM. The accuracy graph shows good percentage for Bi-LSTM and LSTM.
The MAPE percentage error is minimum for Bi-LSTM. The RMSE error values
Improved Deep Learning Based Prediction … 207
are good for both XGB and Bi-LSTM. The MAE values are good for Bi-LSTM,
LSTM, XGB and RF. The results show that Bi-LSTM is better in predicting CYP
when compared with other methods with respect to accuracy and minimized error
metrics which is shown in Fig. 5.
5 Conclusion
This study analyzed the performance of crop yield prediction using different models
like SVM, RF, XGB, LSTM and Bi-LSTM of which four are machine learning models
and two are Deep Learning models. The yield prediction for test set is reasonable
high in accuracy. The Bi-LSTM model outperforms other models both in accuracy
and stability. The Bi-LSTM computationally scales better than SVM, RF and XGB.
These results contribute to a comprehensive understanding of the response of rice
208 V. Saveetha et al.
yield to variables and the comparative performance of crop yield forecasts across
different types of machine learning algorithms. Further improvement of the yield
prediction modeling could benefit from more explicit information on crop type. The
proposed method shows great potential in improving the accuracy of yield prediction
for other crops like corn, wheat, and potatoes at fine scales in the future.
References
1. Saeed K, Wang L (2019) Crop yield prediction using deep neural networks. Front Plant Sci 10
2. Horie T, Yajima M, Nakagawa H (1992) Yield forecasting. Agric Syst 40(1–3):211–236
3. Van Klompenburg T, Kassahun A, Catal C (2020) Crop yield prediction using machine learning:
a systematic literature review. Comput Electron Agricult 177:105709
4. Patrício DI, Rieder R (2018) Computer vision and artificial intelligence in precision agriculture
for grain crops: a systematic review. Comput Electron Agric 153:69–81
5. Filippi P, Jones EJ, Wimalathunge NS, Somarathna PD, Pozza LE, Ugbaje SU, Jephcott TG,
Paterson SE, Whelan BM, Bishop TF (2019) An approach to forecast grain crop yield using
multi-layered, multi-farm data sets and machine learning. Precision Agric 20(5):1015–1029
6. Yang X, Sun M (2019) A survey on deep learning in crop planting. In: IOP conference series:
materials science and engineering, vol 490, no 6. IOP Publishing, p 062053
7. Khaki S, Wang L, Archontoulis SV (2020) A cnn-rnn framework for crop yield prediction.
Front Plant Sci 10:1750
8. Kang Y, Ozdogan M, Zhu X, Ye Z, Hain C, Anderson M (2020) Comparative assessment of
environmental variables and machine learning algorithms for maize yield prediction in the US
Midwest. Environ Res Lett 15(6):064005
9. Koduri SB, Gunisetti L, Ramesh CR, Mutyalu KV, Ganesh D (2019) Prediction of crop
production using adaboost regression method. J Phys Conf Ser 1228(1):012005. IOP Publishing
10. Khaki S, Khalilzadeh Z, Wang L (2019) Classification of crop tolerance to heat and drought—a
deep convolutional neural networks approach. Agronomy 9(12):833
11. You J, Li X, Low M, Lobell D, Ermon S (2017) Deep gaussian process for crop yield prediction
based on remote sensing data. In: 31rst AAAI conference on artificial intelligence
12. Kim N, Ha KJ, Park NW, Cho J, Hong S, Lee YW (2019) A comparison between major
artificial intelligence models for crop yield prediction: case study of the midwestern united
states, 2006–2015. ISPRS Int J Geo-Inf 8(5):240
13. Wang AX, Tran C, Desai N, Lobell D, Ermon S (2018) Deep transfer learning for crop yield
prediction with remote sensing data. In: Proceedings of the 1st ACM SIGCAS conference on
computing and sustainable societies, pp 1–5
14. Yang Q, Shi L, Han J, Zha Y, Zhu P (2019) Deep convolutional neural networks for rice grain
yield estimation at the ripening stage using UAV-based remotely sensed images. Field Crop
Res 235:142–153
15. Schwalbert RA, Amado T, Corassa G, Pott LP, Prasad PV, Ciampitti IA (2020) Satellite-based
soybean yield forecast: Integrating machine learning and weather data for improving crop yield
prediction in southern Brazil. Agricult Forest Meteorol 284:107886
16. Sun J, Di L, Sun Z, Shen Y, Lai Z (2019) County-level soybean yield prediction using deep
CNN-LSTM model. Sensors 19(20):4363
17. Ji B, Sun Y, Yang S, Wan J (2007) Artificial neural networks for rice yield prediction in
mountainous regions. J Agricult Sci 145(3):249
18. Indian Agricultural Production. https://data.gov.in
19. Huang Z, Xu W, Yu K (2015) Bidirectional LSTM-CRF models for sequence tagging. arXiv
preprint arXiv:1508.01991
Improved Deep Learning Based Prediction … 209
20. Cui Z, Ke R, Pu Z, Wang Y (2020) Stacked bidirectional and unidirectional LSTM recurrent
neural network for forecasting network-wide traffic state with missing values. arXiv preprint
arXiv:2005.11627
21. Bolton DK, Friedl MA (2013) Forecasting crop yield using remotely sensed vegetation indices
and crop phenology metrics. Agric For Meteorol 173:74–84
Instrument Cluster IoT Enclosure Design
and Production Implementation in Self
Driven Vehicles
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 211
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_20
212 R. Thirumalai et al.
1 Introduction
The bounce vehicle principle is based on IoT, the major controls of the vehicle
depends on the IoT. From the powering ON to Powering OFF everything is controlled
through IoT. So the major essential electronic parts are being embedded in the instru-
ment cluster enclosure, that is designed and produced by Squraesri technologies.
IoT security: Review, blockchain solutions, and open challenges are discussed [1,
2]. Modeling of information processing in the internet of things at agricultural enter-
prises and solar powered buildings are more successful [3, 4]. Design and implemen-
tation of vehicle navigation system in urban environments, automatic garage doors
and weather stations using internet of things (IoT) are becoming familiar nowa-
days [5–7]. Managing IoT devices using blockchain platform is developed in many
countries [8, 9]. Secure integration of IoT and cloud computing is described with
several case studies [10]. IoT middleware survey on issues and enabling technolo-
gies is demonstrated [11]. Internet of Things (IoT) is the network of physical objects
embedded with sensors, and computers which will helpful in exchange of data and
information with several systems. There are too many applications with IoT avail-
able such as; Automobiles, Heart monitoring implants, Bio-chips etc., IoT consists
of web enabled devices to collect and send data across the environment using sensors
and processors. A machine to machine communication acts on the information they
get from each other. The lifecycle of IoT is to collect, communicate, analyze and
act on the data information. The sensors are useful in collecting the information in
the manufacturing plant and a cloud platform is created to send the data and events
through a network used in the industry. The analyzing of the data includes filtering
the data, building reports and visualizing the data. The final lifecycle of IoT—Act is
useful in communication to the other machine and send successful notifications to
the other systems effectively and efficiently. IoT works with RFID, Sensors, Smart
technology and Nano Technology. The applications of IoT is wider in the areas of
Medical and Healthcare systems, Manufacturing, Building and home automations,
Transportation, Energy management, Environmental monitoring and so on. Also
the IoT faces several challenges; technological standardization, complexity in the
software, wireless communication, power supply, data volumes and interpretation
and so on. IoT provides dynamic control of the industry and daily life, integrates
human society and physical systems, improves the resource utilization ratio, flexible
configuration, technological integrator.
2 Methodology
The electronic PCB board dimensions were taken into considerations with keeping
all the electronic components embedded into it. After that the area in vehicle which is
allocated for the installation of the enclosure is inspected. An 2D rough draft or sketch
were drawn of the enclosure in simple drawing sheet with all necessary dimensions
Instrument Cluster IoT Enclosure Design … 213
that were taken before. The 3D design is done with keeping 2D sketch as a reference,
in Solidworks software. When the deigning stage is over, the material for production
is selected. Preceding the screw for fastener as well as the manufacturing method
were selected as per the IPC standards. After all the above steps, the production
implementation of enclosure is done by the hot injection moulding method. Then
finally all the electronics are embedded into the enclosure and installed in the vehicle.
The process flow chart methodology is shown in Fig. 1.
A 2D rough draft/sketch
was drawn initially
Production implementation
Fig. 1 Methodology
214 R. Thirumalai et al.
While designing the 3D sketch in solidworks software, the 2D rough draft is kept
as the reference which was taken from various parameters while drawing. As earlier
mentioned, the enclosure was divided into two parts, and each part was designed sepa-
rately and then assembled together. So while designing in software several commands
were used to design the enclosure. One of the important step in the design process
is selection of the materials. This systematic selection of the materials will give the
minimization of the cost during the production process. The design engineer should
be aware of the properties of the material and behavior of the materials under several
working conditions. The thermal properties, machinability properties, strength, hard-
ness and other mechanical properties are to be well known the designer. The material
used for the production of the enclosure is PA66 30% GF. PA66 GF30 is a 30% glass
fibre reinforced PA 66(Poly Amide).Poly (hexamethyleneadipamide). Polyamide 66,
or Nylon 66 (PA66) is an engineering-grade thermoplastic.
PA66 provides lower impact resistance when compared to PA6. High strength,
creep strength, toughness are the mechanical properties which are the outstanding
properties of PA66. PA66 is a glass fibre reinforced about 30 percentage to form
PA66 GF30. Glass fibres exhibit high static load withstanding properties at elevated
temperatures.
For testing the corrosion resistance of the material, salt spray test is performed. The
materials tested for salt spray test are metallic in nature. These materials are also
finished with a surface coating for providing better test results. The material that is
to be tested provides a corrosion protection to the undercovering metal. Corrosive
attacks are produced in the Salt spray testing method. This salt spray test will evaluate
the correctness of the coated material for further usage. The viscosity is very sensitive
to temperature. Shrinkage is of the order of 0.01–0.02 mm/mm [1–2%]. The addition
of reinforcing glass fibers reduces the shrinkage to 0.2–1%. Differential shrinkage in
the flow and cross-flow directions is quite high. Mineral fillers yield more isotropic
moldings. PA66 is resistant to most solvents, but not to strong acids or oxidizing
agents. The material selected for the enclosure design provides sufficient success
rate and hence it is found satisfied.
5 Conclusion
The IoT consist of four layers namely; sensor layer, Gateway and Network layer,
Management service layer and application layer. Sensor layer collects and process the
real time information. Gateway and network layer allows the organizations to share
the information’s through the robust networks available in the industry. Management
service layer ensures security and privacy of data. Application layer provides a user
interface for the IoT. The enclosure for electronics to be embedded in it has been
successfully designed and produced by injection molding process. The enclosure has
been installed in vehicle without causing any interruption to other parts and system
in electric vehicle. Hence, the enclosure was designed initially from 2D sketch to
3D model by using Solidworks software, followed by selecting appropriate material
for production and also the fasteners as well and finally fabricated using injection
moulding process.
216 R. Thirumalai et al.
References
1. Khan MA, Salah K (2018) IoT security: review, blockchain solutions, and open challenges.
Fut Gener Comput Syst 82:395–411
2. Bagas AD, Budi WA (2018) IoT-based integrated home security and monitoring system. In:
IOP conferences series: journal of physics: conference series, vol 1140
3. Lvovich IY, Lvovich YE, Preobrazhenskiy AP, Preobrazhenskiy YP, Choporov ON (2009)
Modeling of information processing in the internet of things at agricultural enterprises. IOP
Conf Ser Earth Environ Sci 315(3):032029
4. Handayani TP, Hulukati SA, Jaya R, Tiandho Y, Abdullah R (2019) The prototype of solar-
powered building lighting IoT. IOP Conf Ser Mater Sci Eng 486(1):012079
5. Godavarthi B, Nalajala P, Ganapuram V (2017) Design and implementation of vehicle naviga-
tion system in urban environments using internet of things (IoT). IOP Conf Ser Mater Sci Eng
225:012262
6. Jianyun C, Yunfan S, Chunyan L (2017) Research on application of automatic weather station
based on internet of things. IOP Conf Ser Earth Environ Sci 104(1):012015
7. Uriawan W, Zulfikar WB, Sofa RM, Ramdhani MA (2018) Internet of things for automatic
garage doors using {ESP} 8266 module. IOP Conf Ser Mater Sci Eng 434:12057
8. Huh S, Cho S, Kim S (2017) Managing IoT devices using blockchain platform. In: International
conference on advanced communication technology (ICACT), IEEE, pp 464–467
9. Khan MA, Salah K (2018) IoT security: review, blockchain solutions, and open challenges.
Futur Gener Comput Syst 82:395–411
10. Stergiou C, Psannis KE, Kim BG, Gupta B (2018) Secure integration of IoT and cloud
computing. Futur Gener Comput Syst 78:964–975
11. Ngu AH, Gutierrez M, Metsis V, Nepal S, Sheng QZ (2016) IoT middleware: a survey on issues
and enabling technologies. IEEE Internet Things J 4(1):1–20
General Adversarial Networks: A Tool
to Detect the Novel Coronavirus from CT
Scans
Abstract Detection of the novel Corona virus in the early stages is crucial, since
no known vaccines exist. Artificial Intelligence- aided prognosis using CT scans
can be used as an effective method to identify symptoms of the virus and can thus
significantly reduce the workload on the radiologists, who have to perform this task
using their eyes. Among the most widely used deep learning convolutional neural
networks, research shows that the Xception, Inception and the ResNet50 provide the
best accuracy in detecting Covid-19. This paper proposes that using General Adver-
sarial Network (GAN) as a data augmentation technique, in combination with these
models will significantly improve the accuracy and thereby increase the chances of
detecting the same. The paper also compares and contrasts how each of the three
GANs namely DCGAN, LSGAN, CoGAN, perform in association with the afore-
mentioned models. The main aim of this paper is to determine the most credible GAN
network to carry out the task of data augmentation as well to prove that involving
GANs would improve the existing accuracy of our model, paving way for an effective
approach to train the model.
1 Introduction
The giant tide of SARS-CoV-2 virus, widely known as Covid-19 or novel corona
virus has swept off all of mankind in one fell swoop and with no known vaccine,
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 217
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_21
218 R. Shriram et al.
2 Related Work
A plethora of Deep Neural Networks have been designed to aid in the process of
detection since every new model may train a specific dataset in a better way. Some
of them include using a light CNN architecture based on SqueezeNet [4] for classi-
fication of CT scans that averages a computation time of just 7.81 s. Abdul Waheed
et al. [5] proposed an ACGAN (Auxiliary Classifier Generative Adversarial Network)
based GAN, called CovidGAN to generate synthetic Chest X-Ray images for data
augmentation the training sets along with a CNN model for better detecting COVID-
19. Another Deep Learning framework called the CovNet was proposed to identify
Covid-19 and also distinguish it from other Community-acquired pneumonia (CAP)
and non-pneumonic lung diseases that it can be confused with [6].
Various regression models including Lasso Regression model, Exponential
Smoothing model, etc. have been used to predict the number of COVID patients
to be anticipated in the future [7]. But these tools perform well only when they have
sufficient usable information, accumulating which is not easy. CNN models have
been used not only in classifying CT scans, but are also used in other domains such
General Adversarial Networks: A Tool to Detect … 219
as labelled Waste management [8] and detecting Breast cancer from mammographic
images by using the retrieved similar cases as reference [9].
Not only CNN’s but also VGGNet (which is a significantly more accurate CNN
[10]) has been used for classification purposes. Since 4D imaging is significantly
better in detecting diseases in motion-reliant organs like heart, lungs, etc. than 3D,
a variety of synchronization mechanisms including wavelet de-noising is used to
reconstruct the said 4D image[11].
3 GAN Architectures
3.1 GAN
GANs are basically neural networks that are trained in an adversarial manner to
generate a probability distribution of data that mimics the original probability distri-
bution. It consists of two models namely, the generator and the discriminator. Both
these models compete with each other with an objective of mimicking the original
distribution. The discriminator model performs the classification of whether a given
image is fake (labelled 0) or real (labelled 1). The generator generates images to fool
the discriminator by synthesizing artificial images that have a close resemblance to
the real ones, making it difficult for the discriminator to distinguish the fake (the one
generated from generator) from the real which is sampled from the true dataset. In
essence, the generative network is trained to maximize the final classification error
(between true and generated data), while the discriminative network is trained to
minimize it. Equilibrium is reached when the generator produces samples that follow
the original probability distribution and the discriminator predicts fake or not-fake
with equal probability. It is highly important that both networks learn equally during
training and converge together. So, during the training of the discriminator, we ignore
the generator loss and focus on the discriminator loss, which penalizes the discrimi-
nator for misclassifying real images as fake or vice-versa. Generator’s weights are not
updated. During the generator training, we use the generator loss, which penalizes
the generator for failing to fool the discriminator and generating an image that the
discriminator classifies as fake. The discriminator is frozen during generator training
and only generator’s weights are updated through back propagation.
Consider the training dataset as uniform distribution, p(X), a real image drawn
from it is taken as x and a random image in Rd to be z. Let G(z) and F(z) be the
generative and discriminative models, respectively. The generative model generates
an image as its output, G(z). Let p(G) be the probability distribution of G(z). The
discriminative model is tasked with computing the probability that an input image
220 R. Shriram et al.
E(x) is the Expected value (or First moment) of a function. f and g are the
respective functions of the generator and the discriminator.
However, in practice this is solved by alternating these two gradient update steps:
step 1 : θ t+1
f = θ tf − λt ∇θ f V f t , g t (2)
step 2 : θgt+1 = θgt − λt ∇θg V f t+1 , g t (3)
where λ is the learning rate and θ ƒ and θ g are the parameters of F and G. Goodfellow
et al. [12] show that the distribution, p(G), converges to p(X).
3.2 DCGAN
Deep Convolutional GANs are essentially tailor-made GANs that have certain archi-
tectural constraints that help in stabilizing the learning process of standard GANs
by learning a hierarchy of representations from object parts to scenes in both the
generator and discriminator. Standard GANs are known for being unstable to train.
DCGAN utilizes logical steps towards preventing this instability such as incor-
porating the all convolutional net which replaces spatial pooling functions (like
Max Pooling), the benefits of which are explained by Springenberg et al. [13] and
allowing the generator to determine the down-sampling on its own by providing
strided convolutional layers.
It also includes removing fully connected layers and applying Batch Normaliza-
tion (technique by which each input is transformed to have unit variance and zero
mean) to all but the input layer of the discriminator and output layer of the gener-
ator. Another unique feature is that generators employ the ReLu activation layer
whereas the discriminator uses the Leaky ReLu activation function. Generator and
Discriminator of a DCGAN architecture is shown in Figs. 1 and 2.
General Adversarial Networks: A Tool to Detect … 221
Fig. 2 Discriminator of a DCGAN with strided convolutional layers with the same dimensions
3.3 CoGAN
a tuple of GANs for each individual image domain, which under normal circum-
stances produces a product of marginal distributions instead of a joint distribution.
This inconvenience is overcome by using a weight-sharing constraint that favors a
joint distribution result over a product of marginal distributions result. This forces
the individual GANs, which decode the high- level semantics of the image to share
the weights with the other GANs. Then the layers of these GANs responsible for
decoding the low-level features translate the shared representation exclusively to fool
their corresponding discriminators. The structure of a CoGAN having two image
domains is shown in Fig. 3 of CoGAN proposal [14]. Although only two domains
are shown, it can be easily extrapolated to as many domains as required. As discussed,
the first few layers of the generators, which are responsible for extracting high-level
semantics, are forced to share the parameters from a common subset.
Also, the last few layers of the discriminators, which are responsible for encoding
the high-level semantics, are enforced to share the weights. Note that the flow of
information is in opposite directions. Thus, CoGANs effectively synthesizes pairs of
images without the need for correspondence in the input dataset and thereby aiding
in better training the models.
Here the forced weight sharing is represented using a dotted line.As mentioned
earlier, the weight sharing occurs in the first few layers of the generator and the last
few layers of the discriminator.
3.4 LSGAN
The Least Squares GAN is a slightly modified variant of the normal GAN. Standard
GAN employs the ‘Sigmoid Cross-entropy’ loss function in its discriminators while
classifying the synthesized images from the real ones. This loss function may lead to
this undesirable vanishing gradient problem. This problem arises in gradient based
learning methods that update the weights based on the partial derivative (gradient)
of the error function, where in some cases, the gradient becomes vanishingly small
and thereby effectively stops the training process. LSGAN provides a solution by
changing the loss function to the least squares loss function. Study [16] shows that
General Adversarial Networks: A Tool to Detect … 223
Fig. 4 a The decision boundaries of the two loss functions, b Fake images that may lead to vanishing
gradients, c Effect of LSGAN [15]
this approach is better than the standard GAN in two aspects-generating higher
quality images and learning in a more stable fashion.Least squares loss focuses
on the distance of the image from the discriminator boundary and thus penalizes
generated samples that, although appear on the right side of the boundary, are too
far away from the real data and subsequently the boundary and thus may lead to
vanishing gradients. Figure 4 shows how the LSGAN forces the generated samples
(in magenta) to move closer to the decision boundary.
4 Experiment
For this paper we have used an open source dataset of SARS-CoV-2 CT scan [17],
containing 2482 CT scans in total (1252 positive CT scans and 1230 negative CT
scans). This data has been collected from hospitals in the locality of Sao Paulo,
Brazil.
All the three different GANs were implemented with the help of pyTorch, an open
source ML library. Each scenario consists of a validation phase and testing phase.
In the validation phase, 15% of images were used whereas in the testing phase, 25%
of images were used. All images have a dimension of 75*75*3. A sample for the
fake images generated by the DCGAN is shown in Fig. 5 and their corresponding
generator and discriminator losses with respect to the iterations are shown in Fig. 6.
These images were generated after 5000 iterations. Better quality can be obtained
by increasing the number of iterations.
In addition to the deep transfer models, we added 3 dense layers with sigmoid
activation function in the last layer to classify outputs accordingly. Number of epochs
used to train is 100.
As a first step of approach, we trained all three deep transfer models using the
actual dataset (without GAN). Later, we trained them using the dataset obtained
by concatenating the actual dataset with the datasets obtained from various GANs
(artificial images) individually.
224 R. Shriram et al.
True Positives
Precision = (4)
True positives + False Negatives
True Positives
Recall = (5)
True positives + False Negatives
Precision*Recall
F1 Score = (6)
Precision + Recall
Here a True Positive refers to a Covid patient being correctly identified as one
whereas a True Negative refers to a healthy person identified correctly as one. A False
Positive is a healthy person incorrectly identified as a Covid patient whereas a False
Negative is a Covid patient was not identified as one. The plots of Validation Accuracy
Vs epoch and Validation Loss Versus epoch for all three of the deep transfer models
in each case (without GAN as well with each GAN) have been provided. Xception
has the most deviation among the models considered but it also has the best validation
accuracy among them. The series of tables drafted below shows the brief analysis of
the results obtained (Fig. 7).
Table 2 shows the results obtained by testing the deep transfer models trained
without GAN implementation. As we can say, Xception tops the list with its accuracy
being 97.26% which is a considerably good result.
Because of Xception’s complex network architecture, which simultaneously trains
3–4 small networks and concatenates all the results together for further processing,
Fig. 7 Validation accuracy and validation loss plot for deep transfer models without GAN
226 R. Shriram et al.
it can learn more details than ResNet50 consisting of 50 layers, whose architecture
is more of a repetitive one. Same analogy goes with InceptionV3 as its architecture
is less complex when compared to Xception’s but more than ResNet50’s.
Table 3 shows the results obtained by testing the deep transfer models trained with
the LS-GAN dataset along with the real dataset. This significant improvement in each
deep transfer model’s results attests the use of augmentation, especially using GAN.
Here, ResNet50 tops the list with an accuracy of 98.66%. It has improved its accuracy
to 4.14% which is to be noted. Due to the increase in dataset size, ResNet50 has
a greater exposure resulting in its improvement. Xception performed exceptionally
well with its accuracy being 98.57% which is much comparable with ResNet50, with
an improvement of 1.31%. InceptionV3 also showed improvement in its accuracy
with a difference of 1.47% which is more than Xception’s (Fig. 8).
Table 4 shows the results obtained by testing the deep transfer models trained with
CoGAN implementation. Here, Xception tops the list in this analysis with its accuracy
being 98.48%, followed by InceptionV3 and ResNet50. All the deep transfer models
Fig. 8 Validation accuracy and validation loss plot for deep transfer models with LSGAN
showed improvement in its scores with ResNet50 being the highest with an increase
of 2.8% (Fig. 9).
Fig. 9 Validation accuracy and validation loss plot for deep transfer models with CoGAN
Fig. 10 Validation accuracy and validation loss plot for deep transfer models with DC-GAN
228 R. Shriram et al.
Table 5 shows the results obtained by testing the deep transfer models trained with
DCGAN implementation. Here, Xception again tops the list in this analysis with its
accuracy being 98.31%, followed by InceptionV3 and ResNet50. Here, DCGAN
has the least incremented results among the GANs taken into consideration, yet
shows a significant improvement when compared to results obtained without GAN
implementation (Fig. 10).
From the above analysis, we can say that models trained with LSGAN augmented
images proved to be the best among the GANs considered in this proposal. This is
due to the ‘least squared’ loss function which solves the vanishing gradient problem
confronted by the other GANs here, resulting in a greater stability. LSGAN works
with the advantage of penalizing those samples that have no gradients though they
have been classified rightly (Fig. 11).
6 Conclusion
In this paper, the detection of the novel coronavirus proved to be more efficient by
augmenting the already available dataset [16]. The fabrication of artificial images
was primarily based on the adversarial network GAN. The main aim of this paper is
to determine the most credible GAN network to carry out the task of data augmen-
tation as well to prove that involving GANs would improve the existing accuracy of
our model, paving way for an effective approach to train the model. From our anal-
ysis, ResNet50 (98.66%) and Xception (98.57%) trained with LS-GAN augmented
General Adversarial Networks: A Tool to Detect … 229
images provides the best testing accuracy. The best improvement between archi-
tecture implemented without GAN and with GAN can be observed in ResNet50
implementation with LS- GAN (4.14%). The pre-processing phase involved here
solves the issue of limited datasets available in the biomedical areas of research and
also aids in training the Deep Learning models more efficiently. The work carried
out in this paper could be further experimented by venturing into other recent and
robust versions of GANs.
References
13. Springenberg JT, Dosovitskiy A, Brox T, Riedmiller M (2014) Striving for simplicity: the all
convolutional net. arXiv preprint arXiv:1412.6806
14. Radford A, Metz L, Chintala S (2016) Unsupervised representation learning with deep
convolutional generative adversarial networks
15. Liu M-Y, Tuzel O (2016) Coupled generative adversarial networks. Mitsubishi Electric
Research Labs (MERL)
16. Mao X, Li Q, Xie H, Lau RYK, Wang Z, Smolley SP (2017) Least squares generative adversarial
networks. In: 2017 IEEE international conference on computer vision (ICCV), Venice, 2017,
pp 2813–2821. https://doi.org/10.1109/ICCV.2017.304
17. Soares E, Angelov P, Biaso S, Froes MH, Abe D K (2020) SARS-CoV-2 CT-scan dataset: a
large dataset of real patients CT scans for SARS-CoV-2 identification. medRxiv. https://doi.
org/10.1101/2020.04.24.20078584
VLSI Implementation of Multipliers
for Artificial Intelligence Applications:
A Survey
1 Introduction
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 231
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_22
232 R. Karthick et al.
design time, and therefore, any changes in the requirements entail that further modi-
fication involves reconfiguration of the implementation. In cases where a tiny bit
span would be sufficient, the design runs at bottom efficiency since least mandatory
computation is taken on. To lighten this, mixed-accuracy algorithms intended to be
used in a lower bit span some partition of time, and a substantial bit space when
necessary. Two data paths that make the operation at different precisions which is
normally implemented in the circuit.
The target of the paper is to remove a portion of the computation completely and
significantly during the runtime which is a dynamic control structure. The goal is
accomplished using modified serial Booth multiplier, which is independent of loca-
tion and skips over encoded entire-zeroes or entire-ones computation. A parallel bit
arrangement of all operands both multiplicand and multiplier that is designed to give
a smart block which can be injected into existing AI bots, DSPs and ML. For certain
intake sets, the multiplier hits considerable striking improvements in computational
system performance. The main benefits of this method are as mentioned below.
(1) The data path of the booth multiplier has two sub-circuits. These circuits operate
at different critical paths.
(2) Implementing two critical paths in booth multiplier provides enhancement in
bit-pattern calculation and increases the throughput. Therefore its results in
reduced time and improved latency than conventional multipliers.
(3) The FPGA implementation of the proposed multiplier model for evaluating
the performance and comparison of throughput with conventional multipliers
(Fig. 1).
Fig. 1 Multiplier
architecture
VLSI Implementation of Multipliers for Artificial Intelligence … 233
Vasudeva et al. [4] proposed a “Study of 8 Bits Fast Multipliers for Low Power
Application”. In VLSI the design of multipliers is the most complex architecture
block. Designing the multiplier may face challenges in optimizing power, delay and
area. When the number of stages is increased the width of the multiplier is also
increased which leads to complexity in Areas. For accuracy and high sampling rate
wide bit width is required. So the requirement for designing and developing an IP
for high speed applications. The design is proposed utilizing the basic multiplier
building blocks. In this paper we analyze the different types of 8-bit multiplier and
compare their performance. We are comparing three multiplies namely modified
Booth, Wallace and Dadda.
Modified Booth’s Multiplier:
Booth’s algorithm is commonly used for multiplication of signed numbers because
it is the most powerful algorithm to perform multiplication of signed numbers. This
multiplier frequently performs addition of A or S with P and performs arithmetic
right shift in P register. The algorithm of booth multiplier as follows:
x—multiplicand and y—multiplier
Step 1: fill the A with x bits and y + 1 zeros, similarly S with −x and y + 1 zeros
also p with x + y + 1 zeros
Step 2: check the last two bit of p
If it is 01 perform P + A.
If it is 10 perform P + S.
If it is 00,11 do no operations.
Step 3: Shift the values obtained to one bit right.
Step 4: Repeat step 2 and three y times.
Step 5: Eliminate the MSB of P, and the resultant is obtained.
Wallace Multiplier:
In Wallace multiplier the delay is always equivalent to logarithm of operand size.
The operation of wallace multiplication follows:
Step 1: The bits are grouped for the product.
Step 2: Using carry save adders the product matrix is reduced to a two rows
Step 3: fast carry propagate adder will add the remaining two rows and the final
product is obtained.
The Wallace tree is a family of column compression multipliers (Fig. 2).
Dadda Multiplier:
The Dadda multiplier with carry look ahead reduces the number of rows in the
multiplication process. The counter numbers are less in this method. This reduction
in counter placement will provide a reduction in partial product. The figure shows
the flow of the Wallace multiplier (Fig. 3).
234 R. Karthick et al.
After analyzing the performance parameters like speed, area and power consump-
tion of Booth, Wallace, and Dadda multiplier it can be concluded that Dadda multi-
plier is best in considering area, Booth multiplier provides better power optimization
and Wallace tree multiplier provides better delay optimization. When the number of
interconnections between the building blocks increases there will be increase in the
intermediate stages of the multiplier. This paper shows that the Dadda multiplier has
a 76.5% increase in interconnections as it has more number of inter-stages.
VLSI Implementation of Multipliers for Artificial Intelligence … 235
Sri Lakshmi et al. [1] proposed “Design and Implementation of High-Speed Vedic
Multiplier Using Hybrid Full Adder”. A novel multiplier architecture is introduced
which includes the algorithms like Array, Modified Booth, Dadda multipliers etc. A
multiplier based on Vedic mathematics is discussed in this paper. Vedic mathematics
is an ancient method of sutras for mathematical calculations. The special purpose
of Vedic maths is it reduces the complicated calculations by calculating the partial
products in prior before the beginning of actual operations of multiplication so that
the optimization in time and delay can be achieved. A novel hybrid vedic multiplier
approach has been discussed. In this paper a Vedic Multiplier with CSA is used to
increase the speed. The half adders and full adders in existing Vedic multipliers are
replaced by hybrid adder and compressor adders and comparison has been performed.
Compressors are the logic circuits that can add 3 or more bits simultaneously. Hybrid
full adder is also a kind of full adder with a lesser number of gates so that the area
is reduced. Therefore, by observing all the above results, we can conclude that the
hybrid adder is more power and delay efficient. Though the compressor has more
power and area efficiency compared with normal conventional full adder design,
hybrid adder is giving even better results compared with compressors. So, hybrid
adders give better performance among these three adders circuits (Fig. 4).
Compared with conventional adder the hybrid adder has a delay of 7.04%, power
of 11.12% and area has 1.46% respectively. In comparison of hybrid adder with
compressors the delay, power and PDP got reduced by 2.02%, 1.43% and 3.74%
respectively. Finally conclude that hybrid adder with vedic is gives better efficiency
compared with conventional full adder and compressors.
Sundhar et al. [5] proposed the “performance analysis of wallace tree-multiplier with
Kogge stone Adder using 15–4 compressor” (Fig. 5).
Wallace multiplier using full adder and half adders to bring down the partial
product from three to two rows and then two partial products are added using the
final adder. In this paper an approximate 16 × 16 bit Wallace multiplier using 15–
4 compressor architecture is designed and if we compare the performance of this
multiplier with Kogge stone adder, it is using the same architecture of multiplier
using parallel adder. This architecture infers that kogge stone adder is faster when
we are compared to the multiplier with parallel adder since Kogge stone adder is
basically a parallel prefix adder. This adder has the fastest addition based on design
time. Hence this multiplier occupies about 15% of total coverage area and dissipated
power and delay are also highly reduced.
Devi et al. [6] proposed “low power optimized Array multiplier with reduced area”.
In this paper a 32-bit design of power and reduced area optimized unsigned array
multiplier is proposed. This multiplier is a 32bit array multiplier which makes use of
Carry Select Adders (CSAs) to execute partial product additions. This adder’s area
is expanded but delay is reduced. Ripple Carry Adders (RCAs) are the adders that
have a closely packed area but they are the slowest type of adders (Fig. 6).
On the other hand, carry look ahead adder (CLA) is the high-speed adder but they
are the worst from the area point of field of vision. Hence carry select Adder (CSA)
have been considered as the best comparing with RCA and CLA because they offer
a good counterbalance between compact areas of RCAs and short delay of CLAs.
The array multiplier with CSA uses modified booth algorithm with greatest power
compression technique to minimize the power dissipation upto 22% and reduction
in area is also achieved.
Aruna et al. [2] proposed “A low power and high speed Array multiplier using on-
the-fly conversion”. In this multiplier design with single precision truncation, the
product of 2 N-bits produces 2 N but partial products, excluding these 2 N bit partial
products, are going to be divided into 2 N-(n/2) bits and N/2 bits. As a result finally,
2 N bits are created by adding the above bits using ripple carry adder (Fig. 7).
The array multiplier outlined in this paper is designed and enforced with no
truncation or addition approach; instead it is executed using a typical array multiplier
scheme. The proposed array multiplier produces the high order bit (MSB) of the
final product. The multiplier design leverages on the fly conversion converter that is
implemented as the tail end of the multiplier. This is to achieve the expedited carry
propagation in the last leg of multiplication. The OTF conversion logic is achieved to
provide constant delay irrespective of input bits processed. Hence the array multiplier
with OTF conversion has a minimum array, smaller area and less energy.
See Table 1.
See Table 2.
6 Conclusion
The proposed multiplier is divided into two sub circuits which operate at different
critical paths. The design of CMOS Multiplier that will be highly adaptable to AI,
ML and DSP kind of applications. The Multiplier is designed for bit width of 32 and
64, expecting an improvement of 3 × and 3.56 × respectively. On comparing the
240 R. Karthick et al.
standard parallel multiplier, this design gives 3.64 × improvement in area and delay.
Upcoming works on application will take complete feature of two speed optimization.
References
1. Sri Lakshmi Y, Vigneswaran T (2020) Design and implementation of high-speed vedic multiplier
using hybrid full adder. Int J Adv Sci Technol 29(3):10663–10669
2. Aruna S, Venkatesh S (2019) A low power and high speed array multiplier using on fly
conversion. Int J Recent Technol Eng 7(5S4)
3. Moss DJM, Boland D, Leong PHW (2019) A two-speed, radix-4, serial-parallel multiplier. IEEE
Trans VLSI Syst
4. Vasudeva G, Cyril Prasanna Raj P (2015) Study of 8 bits fast multipliers for low power
application. Int J Soft Comput Eng 5(1). ISSN: 2231–2307
5. Sundhar A, Tharshini SD, Priyanka G, Ragul S Saranya C (2019) Performance analysis of
wallace tree multiplier with Kogge Stone Adder using 15–4 compressor. In: 2019 International
conference on communication and signal processing (ICCSP), Chennai, India, 2019, pp 0903–
0907
6. Devi P, Singh G, Singh B (2011) Low power optimized array multiplier with reduced area.
In: Mantri A, Nandi S, Kumar G., Kumar S. (eds) High performance architecture and grid
computing. HPAGC 2011. Communications in Computer and Information Science, vol 169
Automated Boneage Analysis Using
Machine Learning
Abstract Bone age assessment is done to analyse the skeletal maturity according
to their chronological age. This is done by radiograph method considering the left
hand or wrist. Bone age assessment is done by comparing the chronological age to
assess the endocrine disorders and pediatric syndromes. Earlier the manual method
was used, where the radiologists compare the radiograph image with the atlas and
estimate the age of the bone. In this study, the analysis and classification of the x-ray
image of the left hand is experimented to determine the bone age. Here the bone
age analysis method involves the segmentation of the image, feature extraction and
classification using support vector machine(SVM). The results obtained is future
used to assess the skeletal abnormalities.
1 Introduction
Bone is connective tissue. There are 300 soft bones which are formed during the birth,
then the cartilage grows and slowly replaced by hard bone. During the adolescence
period there are 206 bones formed. These bones have their own blood vessels and
made up of living cells. The longitudinal growth in long bones is due to process
J. K. Krithika (B)
Dr.N.G.P Institute of Technology, Kalapatti Road, Coimbatore, India
Y. Norashikin
UniversitiTeknologi Petronas, Seri Iskandar, Perak, Malaysia
e-mail: [email protected]
K. Porkumaran
Sri Sairam Engineering College, Chennai, India
S. Prabakar
Sona College of Technology, Salem, India
e-mail: [email protected]
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 241
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_23
242 J. K. Krithika et al.
2 Problem Statement
Bone age helps to evaluate growth but the children does not mature at the same time,
which makes it difficult to evaluate the growth. Usually the bone age assessment
depends on radiologist experience, it is carried out by visual evaluation of the skeletal
development using GP atlas. In manual approach, bone age assessment varies from
one scientist to another. It is also time consuming and may lead to errors.
In this research method, it will focus on analysis of the left hand x-ray image. This
is done by taking the ROI part and Extracting the image. This research will result
in development of automated bone age prediction algorithm using machine learning
technique.
3 Objectives
The assessments that will be carried out are about the bone age analysis from x-ray
image radiograph, to achieve few goals:
1. To analyze the bone from the x-ray image at various level of age.
2. To develop and evaluate bone age prediction algorithm by using machine
learning technique.
4 Literature Review
one is the receiver. Radiograph method where considered as the predominant method
for bone age analysis. A study was conducted by using the mammography x-ray,
which has the high image quality. The method used was dual energy x-ray absorp-
tiometry (DEXA) and sonography. The DEXA is more accurate but it is expensive
and radiation emitted is more. It is used in spine and hip. Sonography method is
radiation free and its cost is low. It is usually used to assess the bone mineral density
of peripheral bones (heel, knee cap) [5].
Studies were not only done in hand bone but few studies done the bone age analysis
by dental maturity, it was used for forensic purposes, but not used in diagnosis of
endocrine diseases. The other method is by iliac bone, it was determined based on the
degree of maturation of the iliac crest apophysis, but this was a drawback one as the
ossification is not uniform in iliac crest apophysis. The other method is by femoral
head, which is the alternative way by assessing the depth of the epiphysial cartilage
of femoral head which is continuously being ossified. Ossification is complete and
most of the cartilage is replaced by bone remaining cartilage, which is called as
hyaline articular cartilage. The bone age analysis was first done manually in tooth,
later in bones such as skull, foot bones, knees, spinal cord, pelvis, rib, femur, carpals
and epiphyses of the ulna and radius. The main characteristics of the bone age
assessment are fusion of bones. The problems in manual approach is time consuming,
it varies from scientist to scientist, it may also lead to error and during surgery
over estimation or under estimation of bone length may lead to deformity. But in
automated bone age analysis, according to Dinesh M.S the radiograph must have
preprocessing, segmentation, image enhancement, rotation of bone pixels to proper
angel, measurement analysis and decision analysis [6].
An attempt using the SVM classification method has been introduced by Markus
Harmsen et al. [4]. The semi-automatic BA analysis developed by combining the
support vector machine with cross correlation to a prototype image for each class.
The evaluation of BAA is presented by comparing the nominal and real- valued
SVM with k nearest neighbor classification. According to this automated BAA,
there are some step that have present: (1) Extract 14 epiphyseal regions from the
radiographs; (2) retain image features using the image retrieval in medical application
framework; (3) Use these features to build a classifier model (training phase); (4)
Evaluate performance on cross validation (testing phase); (5) Classify unknown hand
image (application phase). These all steps have been implement by using C++ and
SQLite database version 2.8.17. Therefore, based on the 1097 hand radiographs
of 30 diagnostic classes in range 0–19 years, the nominal and real-value SVM are
resulting 91.57% and 96.16% accuracy respectively. However, in all the study done,
can observe that the SVM classification method have the higher accuracy and state
that the simple classification for the BA analysis. However, some improvement of
feature extraction is required as to achieve the BA assessment more efficient and can
classify to the age level categories.
244 J. K. Krithika et al.
Detection of ROI
(midfinger)
Middle finger
Calculation of the im-
age properties of mid-
finger
Classes (5-
12yrs)
SVM
Classes (13-
18yrs)
5 Proposed Work
The detection and feature extraction of the hand bone x-ray of different image is
the significant work done in the bone age analysis. The input images are taken from
the website, and feature is extracted and the trained data is feeded into the classifier
(Fig. 1).
The image dataset of the left hand is taken from the website kaggle.com. The test
dataset and trained datasets are downloaded, about 10,000 and above datasets for
each dataset. The dataset of the images is obtained with the CSV, which contains the
information regarding the person’s age, sex that are used for training.
The first stage used for segmentation. The middle finger is labeled using the toolbox
in mat lab for all 120 images. Middle finger is considered because of prominent
Automated Boneage Analysis Using Machine Learning 245
growth of bones. After labeling of the image the data is exported into the workspace
in tabular form.
Feature extraction is the next stage. From the ROI image the mean, standard deviation,
maximum value, energy, variance is determined. Normalization was also considered
and the output was compared between before normalization and after normalization.
5.4 Classification
The feature extraction will undergo the classification method. All the classifier was
used to find the better classifier for classification. From the work it was found that,
linear and quadratic svm has better classification accuracy.
From the 120 images it is found that before normalization, the accuracy is 89.2%,
which is greater when compared to after normalization. Because during normal-
ization the raw data of the image may be missed. For less number of images, the
parameters used for feature extraction was enough and from this it is proven that
linear classifier gives the good result.
Automated Boneage Analysis Using Machine Learning 247
deviation, energy, minimum value, maximum value, variance for feature extraction
is not enough. The accuracy always dependent on the feature extracted.
6 Classifier Result
Once the extracted features are fed into classifier, it is trained and the result is gener-
ated as a code in the workspace. From the generated code, the Validation accuracy and
validation prediction of the images are obtained. For testing the obtained program,
more than 10 images are labeled again and feature is extracted from it and fed into
the code generated for the result Fig. 14.
Automated Boneage Analysis Using Machine Learning 249
From the figure the result is obtained. The accuracy for classification is 89.2% as
the code generated is by the linear classifier at accuracy 89.2%. 8 images for each
class was fed., Totally 16 images were fed into the generated code and the validation
accuracy is 0.8750 and for class 1, only 1 image was predicted wrong and for class
2, 1 images where predicted wrong.
250 J. K. Krithika et al.
7 Conclusions
In conclusion, the bone age reflects the skeletal maturity when differing significantly
from chronological age. There were a lot of method to automate the analysis of the
bone age that will be time consume for the radiologist in analysis the radiographs of
Automated Boneage Analysis Using Machine Learning 251
bone age but in this analysis, the reliable feature extraction is the most important key
point for analyze the development of the bone age. However, in the implementation
of the image processing for feature extraction as inspected ROI image have been
done as the proposed methodology. The method contains the detection of ROI using
252 J. K. Krithika et al.
parameters like mean, variance, energy and then the images are trained in the SVM
classifier.
References
1. Lee H, Tajmir S, Lee J, Zissen M, Yeshiwas BA, Alkasab TK, Choy G (2017) Fully automated
deep learning system for bone age assessment. Springer, Aug 2017
2. Vicente G, Osman R, Hand bone age. Springer
3. Mughal AM, Hassan N, Ahmed A (2014) Bone age assessment methods. PMC, Jan–Feb 2014
4. Harmsen M, Fischer B, Schramm H, Seidl T, Deserno TM (2013) Support vector machine
classification based on correlation prototypes applied to bone age assessment. IEEE J Biomed
Health Inform 17(1):190–197
5. Huo Z, Jyang CY, Purdum M, Bone Mineral density assessment mammography system. United
States patent patent no.: us 7.965,813 b2 hu0 et al
6. Bakthula R, Agarwal S, Automated human bone age assessment using image processing
methods. Int J Comput
Multi-Class SVM Prediction Model
for Lung Cancer Diagnosis
1 Introduction
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 253
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_24
254 D. Lakshmi et al.
(AUC) of 0.71, 0.75 and 0.76 for the Ghent, Toronto and Leuven datasets respec-
tively [4]. Selection of textural features is another major challenge in the character-
ization of lung tissues. Interobserver variability in texture analysis including lesion
morphology, image acquisition parameters, segmentation parameters, the number of
texture features and the category must be dealt that result in satisfactory outcome in a
clinical trial [5–9]. The role of non-linear classifiers in medical imaging diagnostics
is significant in the non-invasive characterization of lung abnormalities. Multiple
kernels based SVM has nowadays become effective in the characterization and
staging of lung diseases from CT images.
2 Literature Survey
Cancer is a multifaceted disease and many automated cancer detection systems are
proposed in literature. A predominant research work is based on classification of later
stages of lung tumors commonly termed as malignant. Early detection of cancer also
known as prognosis is not straight- forward as information about cancer biomarkers
are not known clearly. Detection of cancer from CT images involves feature extraction
and classification using machine learning algorithms such as ANN, Decision Tree,
Linear Discriminant Analysis, SVM, Nearest Neighbor classifier and Bayes Clas-
sifier. Recent literature reports that SVM classifier outperforms other conventional
classifier systems in lung cancer detection. SVM performs a two-class classification
by transforming the input data to higher dimensional feature space through a linear or
a non-linear mapping function. It identifies a decision boundary known as hyperplane
that maximizes the separation between the input variables in feature space. Keshani
et al. [10] proposed a SVM based lung nodule detection system from CT images. The
classifier was trained on 2D stochastic and 3D anatomical features to detect the nodule
areas. They reported an overall detection accuracy of 89% and false discovery rate of
7.3% per scan. Thomas et al. [11] proposed a data fusion technique to combine the
genomic and proteomic data for breast cancer detection. They proposed a weighted
average Least Squares SVM (LS-SVM) classifier for cancer classification. They
report classification performance of various conventional classifiers based on ROC
analysis. Shen et al. [12] proposed a bidirectional chain coding scheme and used SVM
classifier to segment the lung nodules from CT images. They reported a segmentation
accuracy of 92.6% with the proposed scheme. Huang et al. [13] compared the clas-
sification performance of standalone SVM and SVM ensemble classifier in breast
cancer analysis. They reported that SVM ensemble classifier based on RBF kernel
function outperformed other classifi-ers used in the analysis. Froz et al. [14] proposed
a combination of textural features based on artificial crawlers and used SVM classi-
fier for differentiating lung nodules from non-nodules of CT images. They reported
a mean accuracy of 94.3% and AUC value of ROC curve of 0.922 for RBF kernel
SVM classifier. A study done by Wang et al. [15] compared several classical machine
learning methods like Random Forest(RF), Support Vector Machine(SVM), Adap-
tive Boosting (Ada-Boost), Back Propagation Artificial Neural Network(BP-ANN)
Multi-Class SVM Prediction Model for Lung Cancer Diagnosis 255
3 Proposed System
lung tissues namely suspicious of TB, fibrosis and carcinoma. The work aims at the
implementation of highly accurate SVM prediction model with very less prediction
time.
3.2 Methodology
Fig. 1 Methodology
Multi-Class SVM Prediction Model for Lung Cancer Diagnosis 257
in Fig. 1. The strengths of our proposed work are 1. Processes the low-dose CT
image comprehensively. 2. Prevents human intervention during diagnosis. 3. Prior
knowledge of ROI is not required.
Prediction of SVM model is evaluated in terms of accuracy and prediction time with
respect to kernels and type of feature set. Many researchers use SVM for binary
classification problem and shown satisfactory performance. But, the use of binary
SVM is restricted for multi-class problems. Our three class SVM classifier model
performance with the three feature sets namely GLCM, GLRLM, both GLCM and
GLRLM are evaluated by the Confusion matrix as in Tables 1, 2 and 3 respectively. It
is well observed that the feature set type and the size contribute the classifier perfor-
mance to a greater extent. Classifier Model with 62 features reduces misclassification
of carcinoma and other cases significantly.
Our proposed system implements three class SVM prediction model trained with
GLCM based features, GLRLM based features and both GLCM and GLRLM based
features. Table 4 shows the classifier performance parameters namely true positive
rate, positive predictive value, false negative rate and false discovery rate respec-
tively. Table 5 presents the accuracy and prediction time of SVM classifier with
gray-level co-occurrence matrix parameters and gray level run length matrix with
respect to different kernels like linear, quadratic, cubic and Gaussian kernels respec-
tively. Classifier accuracy is influenced by the selection of the different feature set
numbering 18, 44 and 62 respectively. Moreover, the maximum accuracy is obtained
for quadratic and cubic kernels in both the glcm and glrlm based feature set. The
model categorizes the three cases with an accuracy of 95, 95.2 and 97.7% respectively.
Our experimental results clearly demonstrate that the proper selection of the infor-
mation data from the training dataset and exploiting that information in the predic-
tion function optimization reduces the prediction time. Figure 2 shows Prediction
time of SVM with different kernel function gives interesting results that model with
glcm features consumes maximum time of 7.65s compared to 1.05s for model with
glrlm features. Figure 3 gives the Quadratic, Cubic and Gaussian kernel functions
with GLRLM based features able to predict with a less timing of 1.11s, 1.069s and
1.246s compared with GLCM based features of 1.269s, 1.49s and 1.167s respec-
tively. Combining both GLCM and GLRLM features shown in Fig. 4 does not yield
significant change on prediction time. Finally, SVM prediction model with Quadratic
and Cubic Kernels proves that GLRLM features is efficient in the characterization
of the lung tissues namely fibrosis, suspicious of TB and carcinoma resulting in an
accuracy of 98.6% each as shown in Fig. 5.
The classifier models are usually evaluated by their prediction accuracy. Our three
class SVM classifier model is assessed by its prediction accuracy and is also compared
with all the binary class SVM state of the art as shown in Table 6. Many researchers
implemented Binary class SVM classifier model with GLCM and GLRLM features
yielding less accuracy. Our proposed three-class SVM Classifier model shows
significantly high value of prediction accuracy.
5 Conclusion
We developed three class SVM prediction model with 5 fold cross-validation for
the differentiation of lung cancer from other lung tissues like suspicious of TB and
Table 4 SVM classifier performance parameters with three feature set
True positive rate Positive predictive value False negative rate False discovery rate
Performance Healthy Other Carcinoma Healthy Other Carcinoma Healthy Other Carcinoma Healthy Other Carcinoma
Parameters Cases (in Cases Cases (in Cases Cases Cases (in Cases (in Cases Cases (in Cases (in Cases Cases (in
%) (in%) %) (in%) (in %) %) %) (in %) %) %) (in %) %)
Model with 100 89 96 100 95 92 – 11 4 – 5 8
18 features
Model with 100 95 92 100 91 96 – 5 8 – 9 4
44 features
Multi-Class SVM Prediction Model for Lung Cancer Diagnosis
Table 5 SVM prediction response for GLCM based features with respect to kernel function
Kernel GLCM GLRLM Both GLCM and GLRLM
function Accuracy Prediction Accuracy Prediction
(in %) time (insec) (in %) Time (in Accurac y Prediction
sec) (in %) time (in
sec)
Linear 95 7.614 95.9 1.0518 98.2 7.2094
Quadratic 96.4 1.2694 98.6 1.1142 97.7 1.363
Cubic 96.4 1.492 98.6 1.0694 97.7 1.1665
Gaussia n 94.8 1.167 93.9 1.2463 94.3 1.313
Fig. 3 Prediction time of SVM with glrlm features for different kernel functions
Table 6 (continued)
Author Features Images Classifier/class Accuracy (in %)
Our proposed GLCM based CT SVM/three class 95.0
work features – linear
– Quadratic 96.4
– Cubic 96.4
GLRLM based – Gaussian 94.8
features SVM/three class 95.9
– Linear
– Quadratic 98.6
– Cubic 98.6
– Gaussian 93.9
Both GLCM SVM/three class 98.2
features and – Linear
GLRLM – Quadratic 97.7
based features
– Cubic 97.7
– Gaussian 94.3
fibrosis. It is concluded that quadratic and cubic kernels for GLRLM based features
yield a maximum accuracy of 98.6% with prediction time of 1.23sec. Combining
these features and predicting with linear kernel increases the accuracy to 98.2%.
The main advantage of our Classifier model requires no prior knowledge of ROI
for segmentation, where the system process the low-dose CT comprehensively. The
work may be enhanced for the detection of metastasis of the lung abnormalities by
including dataset contributing images of different abnormalities.
References
1. http://www.who.int/nmh/countries/ind_en.pdf?ua=1
2. Veronesi G, Novellis P, Voulaz E, Alloisio M (2016) Early detection and early treatment of
lung cancer: risks and benefits. J Thorac Dis 8(9):E1060–E1062
3. Ito F, Kawasaki M, Ohno Y et al (2016) Noninvasive tissue characterization of lung tumors
using integrated backscatter intravascular ultrasound. An Ex Vivo Comp Study Pathol Diagnos
Chest 149(5):1276–1284
4. Hatt M, Majdoub M, Vallières M et al (2015) F-FDG PET uptake characterization through
texture analysis: investigating the complementary nature of heterogeneity and functional tumor,
volume in a multi-cancer site patient cohort. J Nucln Med 56:38–44
5. Jayasurya K, Fung G, Yu S, Dehing-Oberije C, Ruysscher D, Hope D (2010) A Comparison
of Bayesian network and support vector machine models for two-year survival prediction in
lung cancer patients treated with radiotherapy. Med Phys 37(4)
6. Fried DV, Tucker SL, Zhou S, Liao Z, Mawlawi O, Ibbot G, Court LE (2014) Prognostic value
and reproducibility of pretreatment CT texture features in stage III non–small cell lung cancer.
Int J Radiat Oncol Biol Phys 90:834–842
Multi-Class SVM Prediction Model for Lung Cancer Diagnosis 263
7. Doumou G, Siddique M, Tsoumpas C, Goh V, Cook GJ (2015) The precision of textural analysis
in 18 F-FDG-PET scans of oesophageal cancer. EurRadiol 25:2805–2812
8. Tixier F, Hatt M, Le Rest CC, Le Pogam A, Corcos L, Visvikis D (2012) Reproducibility of
tumor up take Heterogeneity characterization through textural feature analysis in 18 F- FDG
PET. J Nucl Med 53:693–700
9. Nyflot MJ, Yang F, Byrd D, Bowen SR, Sandison GA, Kinahan PE (2015) TU-AB- BRA-
04: quantitative radiomics:Sensitivity of PET textural features to image acquisition and
reconstruction parameters implies the need for standards”. Med Phys 42:3587
10. Keshani M, Azimifar Z, Tajeripour F, Boostani R (2013) Lung nodule segmentation and
recognition using SVM classifier and active contour modeling. A complete intelligent system,
Comput Biology Med 43:287–300
11. Thomas M, De Brabanter K, Suykens JAK, De Moor B (2014) Predicting breast cancer using
an expression values weighted clinical classifier. BMC Bioinformatics 15. https://doi.org/10.
1186/s12859 014- 0411–1
12. Shen S, Bui AAT, Cong J, Hsu W (2015) An automated lung segmentation approach using
bidirectional chain codes to improve nodule detection accuracy. Comput Biology Med 57:139–
4. https://doi.org/10.1016/j.compbiomed.2014.12.008
13. Huang MW, Chen CW, Lin WC, Ke SW, Tsai CF (2017) SVM and SVM ensembles in breast
cancer Prediction. PLoS ONE 2017;12:e0161501. https://doi.org/10.1371/journal.pone.016
1501
14. Froz BR, de Carvalho Filho AO, Silva AC, de Paiva AC, Nunes RA, Gattass M (2017) Lung
nodule classificatio using artificial crawlers, directional texture and support vector machine.
Expert Syst Appl. https://doi.org/10.1016/j.eswa.2016.10.039
15. Wang H, Zhou Z, Li Y, Chen Z, Lu P, Ang E, Wanyu Liu W (2017) Comparison of machine
Learning methods for classifying mediastinal lymph node metastasis of non-small cell lung
cancer from F-FDG PET/CT images. EJNMMI Res 7:11.https://doi.org/10.1186/s13550-017-
0260-9
16. Orozco HM, Villegas OOV, Sánchez VGC, Domínguezz HDJO. Alfaro (2015) Automated
system for lung nodules classification based on wavelet feature descriptor and support vector
machine, BioMedical Eng OnLine 14:9.https://doi.org/10.1186/s12938-015-0003-y
17. Makaju S, Prasad PWC, Alsadoon A, Singh AK, Elchouemi A (2018) Lung cancer detection
using CT scan images. Procedia Comput Sci 125:107–114
18. Hassen DB, Zakour SB, Taleb H (2016) A hybrid fuzzy-SVM classifier for automated lung
diseases diagnosis. Polish J Med Phys Eng 22(4):97–103.https://doi.org/10.1515/pjmpe-2016-
0017
19. Ajin M, Mredhula L (2017) Diagnosis of interstitial lung disease by pattern classification. In:
7th International conference on advanc in computing & communications, ICACC -2017, 22–24
August 2017, Cochin, India, Procedia Computer Science vol 115, pp 195–208
20. Zhou T, Lu H, Zhang J, Shi H (2016) Pulmonary nodule detection model based on SVM and CT
image feature-level fusion with rough sets. Hindawi Publishing Corporation BioMed Research
International, vol 2016, Articl ID 8052436, 13 p. https://doi.org/10.1155/2016/8052436
21. Dennie C, Thornhill R, Carolina V-V, Souza A, Gupta A, Maziak D, Bayanati H (2016) Role
of quantitative computed tomography texture analysis in the differentiation of primary lung
cancer and granulomatous nodules. Quant Imaging Med Surg 6(1):6–15
22. Mohana RM, Devi RDH, Bai A (2019) Lung cancer detection using nearest neighbour classifier.
Int J Recent Technol Eng (IJRTE) 8(2S11). ISSN: 2277-3878
23. Senthil S, Shubha BA (2019) Improving the performance of lung cancer detection at earlier
stage and prediction of reoccurrence using the neural networks and ant lion optimizer. Int J
Recent Technol Eng (IJRTE) 8(2). ISSN: 2277–3878
24. Lakshmi D, Santosham R, Ranganathan H (2014) Automated texture based characterization
of fibrosis and carcinoma using low-dose lung CT images. Int J Imaging Syst and Technol
24(1):39-44. Online ISSN: 1098-1098.https://doi.org/10.1002/ima.22077
Survey on Fire Safety Robot &
Implementation of Android Application
Abstract Detection and extinguishing of fire are dangerous job that puts the life of
fire fighters in danger. Using robots to do these things will reduce the loss of lives.
The roles of current systems are limited in certain ways so it can be improvised to a
quicker and effective method. This paper offers an overview of a fire fighting robot
with toxic gas calculation. Several gas sensors are used to determine the volume of
poisonous gases such a carbon monoxide, carbon dioxide and ammonia. The fire is
detected using a flame sensor and the information is processed by the processor and
the procedure to extinguish it is initiated. Data and informatics is the essential factors
for effective firefighting operation.
1 Introduction
Fire battle is a dangerous job, but firefighters are still putting their lives on line against
fire. The mortality rate of firefighters in service is high. The temperature steadily
increases when a fire threat arises. The harmful gases such as carbon dioxide and
carbon monoxide appear to impact humans. Fire fighters can also be replaced by
robots that can do the same thing safely. This article reviews and describes prob-
lems and criteria for robot-based firefighting and presents prospects of productive
fighting in the future. The programmed robot locates the fire and automatically extin-
guishes it. If the fire is not detected, the mobile application may monitor it manually.
With mobile device, the robot can be run easily. It tests the safe concentrations of
poisonous gases such as carbon dioxide, carbon monoxide and ammonia in the burn.
This helps us to consider the strength of the fire and the wellbeing of victims. A
mobile network for autonomous firefighting can track and manage a particular area
for the occurrence of fire, and locate the exact place of the fire, extinguish the fire and
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 265
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_25
266 K. Sakthisudhan et al.
return safely after completion of the extinction. The obstacle monitoring is carried
out using A / D conversion in the microcontroller to convert analogue output from
the ultrasonic sensor to digital, and it is compared with the pre-specified threshold
value for the AFFMP to decide whether there are any obstacles too close [1]. The
Snake Fighter Anna Konda can drive the fire away from outer obstacles apart from
a flat floor, and can use the nozzle that is fixed to the front with a hydraulic medium
in joint action to extinguish the fire. It is a combined use of water to power, cool and
extinguish hydraulic joints [2]. Robots will reach the area of the enemy and submit
information via wireless camera for night vision. The fire sensor senses the fire inside
the warfare and sends the fire extinguisher information, so the water pump can be
triggered whenever possible. An ultrasonic sensor is used to detect all obstacles and
to detect the distance between the robot and the obstacle. Without causing damage,
ultrasonic sensors calculate the distance. When the robot identifies obstacles and
responds. The ultrasonic sensor is incorporated such that the entire world is laid out
[3]. The conventional robot firefighting IOT based firefighting robot can be replaced.
The fire alarm from the cloud is issued to an android application. Then, the auto-
matic receiver in the robot provides a live video of the fire position. The recipient
also allows the consumer and fire victims to connect. This helps the people trapped
in the position where fire has occurred to get guidance on evacuation routes. By
calculating the amount of carbon monoxide within the fire area, the possible health
effects can be calculated roughly [4]. A water tank and a wireless communication
operated pump are used for pumping water. The robot body moves by the micro-
controller’s performance. The robot is operated through a Bluetooth app. Remote
position IR flame sensor is used. Facilities provide wireless networking. Exactness
and performance are maximized after fire detection with a minimum amount of time
[5]. The robot is wirelessly operated and remotely controlled. The fire is sensed by
a sensor. The robot is loaded with a water tank and all motors are operated by the
Microcontroller. The fire and gas sensors are sent to the microcontroller when the
fire threshold is over 450C or the gas threshold is over 300 parts per million. The
message is received via Bluetooth technology on a mobile device [6]. On its own
axis the robot rotates to scan for fire at various angles in the building. If the sensor
LM35 detects heat, the robot stops and moves the fire by means of a centrifugal
pump that casts water, if the temperature increases. In darker areas, the fire flame
is best sensed. It is a robot that prevents fire because it can immediately detect and
extinguish [7]. Nodes where fire incidents happen sometimes are located in an indoor
environment. A sensor and microcontroller node is included. The information shall
be transmitted to the coordinator unit if the node detects fire. The co-coordinating
unit is the main unit linking all nodes. The coordinators unit performs two actions,
one is to send fire alarm message to the department and the other to send details to
a mobile robot consisting of an indoor map. An algorithm for the planning of a star
path designs a route from the current position to the fire location, which is called the
global path. Mobile robots planner [8] transforms the route information into separate
movement instructions. Using its onboard camera the robot will conduct monitoring.
The operator of the robot will direct it remotely around the entire site. The extinction
of Water and CO2 is conducted. Nebulous sprays are being used to avoid burns in the
Survey on Fire Safety Robot & Implementation … 267
3 Conclusion
This proposed work is aimed at using robots to detect and extinguish fires that are
minimally harmful to humans. It also tests the amount of poisonous gases including
carbon dioxide, carbohydrate and ammonia using gas sensors, as well as controlling
fire. The gas sensor data helps to predict approximately the potential health effects
of the injured, to ensure that medical assistance is prepared in advance.
Survey on Fire Safety Robot & Implementation … 269
4 Future Work
Water is used to extinct the fire in the proposed work, but the extinct material may
be modified by fire type. The water capacity in the water tank can be increased to
meet requirements. To find their safe levels, other gas sensors can be installed.
References
1. Khoon TN, Sebastian P, Saman ABS (2012) Autonomous fire fighting mobile platform. In:
International symposium on robotics and intelligent sensors 2012 (IRIS 2012). Elseiver
2. Liljeback P, Stavdahl O, Beitnes A (2006) Snake fighter—development of a water hydraulic
fire fighting snake robot. In: 9th International conference on control, automation, robotics and
vision. ICARCV ’06, pp 1–6
3. Kolambe K, Pote R, Jadhav A, Chennur V (2018) Spy robot with fire detection and water
sprinkling. In: 2nd International conference on electronics, communication and aerospace
technology (ICECA 2018) IEEE conference record # 42487; IEEE Xplore. ISBN:978-1-5386-
0965-1
4. Kanwar M, Agilandeeswari L, IOT based fire fighting robot. IEEE
5. Zaman HU, Khan TA, Falgunee SR, Rashid GMS, Talukder FH (2018) Autonomous firefighting
robot with optional bluetooth control. In: 2018 4th International conference on computing
communication and automation (ICCCA). IEEE
6. Bose JSC, Mehrez M, Badawy AS, Ghribi W, Bangali H, Basha A, Development and designing
of fire fighter robotics using cyber security. IEEE
7. Rakib T, Rashid Sarkar MA, Design and fabrication of an autonomous firefighting robot with
multi sensor fire detection using PID controller. In: 5th International conference on informatics,
electronics and vision. (ICIEV) –IEEE
8. Anantha Raj P, Srivani M (2018) Internet of robotic things based autonomous fire fighting
mobile robot. In; 2018 IEEE international conference on computational intelligence and
computing research (ICCIC)
9. Mittal S, Rana MK, Bhardwaj M, Mataray M, The fire fighting robot. In: IEEE international
conference on advances in computing, communication control and networking
10. Fahim SR, Das SK (2018) High-Performance open-loop tracking control of a small scale fire-
fighting robot. In: IEEE international conference on advancement in electrical and electronic
engineering 22–24 Nov 2018
11. Kim JH, Keller B, Lattimer BY (2013) Sensor fusion based seek-and-find fire algorithm
for intelligent firefighting robot. In: 2013 IEEE/ASME international conference on advanced
intelligent mechatronics (AIM) Wollongong, Australia, 9–12 July 2013
12. Su KL (2006) Automatic fire detection system using adaptive fusion algorithm for firefighting
robot. In: 2006 IEEE international conference on systems, Man, and cybernetics Oct 8–11,
2006.Taipei, Taiwan
13. Amano H, Osuka K, Tarn TJ (2001) Development of vertically moving robot with gripping
handrails for firefighting. In: Proceedings of the 2001 IEEE/RSI international conference on
intelligent robots and systems Mad, Hawaii, USA, Oct. 29 -Nov 03, 2001
Detection of Emergency Vehicles Using
Radio Frequency Identification (RFID)
Abstract Road traffic is a major problem in major cities. With the increasing number
of vehicles, the resolution of the problem is very crucial. Problems related to traffic
not limited to road congestion, but also difficult for the passage of emergency vehi-
cles. In order to build an intelligent traffic signal system, a complete traffic manage-
ment system must be built that deals with all traffic-related problems. The aim of this
article is to propose an efficient system for detecting and prioritising emergency vehi-
cles on the junction. In this paper, emergency vehicles are identified and prioritized
using Radiofrequency Identification (RFID). RFID tags serve to identify vehicles
in a unique way. In this article, ultra-high-frequency RFID readers are used with a
frequency of 865–928 MHz and cover a maximum distance of 9 m.
1 Introduction
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 271
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_26
272 U. Mittal and P. Chawla
may be done as an emergency vehicle approaches with the nearest traffic light. During
an emergency, the RFID tag acts as a transducer and transmits signals to the lighting
system [2].
RFID is a technology that allows the automatic identification of an individual,
packet or element with radio signals [3]. It relies on RFID tags to do so. These
are small transponders that provide identity information when requested over a short
length. The majority of RFID labels have at least two parts. One is a built-in circuit that
stores certain information, modules and demodulates the signal for radio frequency
and other special functions. The second antenna serves the purpose of receiving and
transmitting the signal. There are primarily two types of RFID tags: active RFID tags
that contain batteries and passive RFID tags that do not contain batteries [4].
To transmit, RFID labels and readers must be adjusted to a similar frequency. There
are various frequencies which may be used by an RFID gadget. Generally speaking,
the more widely recognized are.
• Low recurrence, or LF, (125–134 kHz)
• High recurrence, or HF, (13.56 MHz)
• Ultra-high recurrence, or UHF, (433, and 860–960 MHz).
On the various frequencies, the radio waves act in a contrasting manner, so it is
essential to choose the adjustment recurrence for the application.
Low-frequency labels, for example, have a frequency and may penetrate delicate
metallic surfaces even more appropriately. LF RFID frameworks are likewise ideal
for perusing high water content articles, for example, natural product or beverages,
however the read go is just centimeters or centimeters. Common place LF RFID
applications incorporate creature labeling and get to control.
High-frequency labels work very well for metal items and can be used to handle
goods containing medium to high water. HF RFID frames usually operate in inches,
but can have a greater reading range of about three feet (1 m). Following library
books, following the patient stream, and travel tickets are the average HF RFID
applications.
Generally, UHF frequencies provide essentially a better understanding of the
extent and can move information quickly (e.g., read many more labels every second)
than lower or high recurrence frequencies (inches to 50 + ft. as per the RFID frame-
work arrange- ment). Since, in all cases, the frequency of UHF radio waves is shorter,
their signal would be lowered (or weakened) and would not be communicated via
metal or water. Other basic uses of UHF RFID integrate the electronic assortment
of costs and the control of discontinuation of access because of the more stretched
scope of understanding.
Detection of Emergency Vehicles Using Radio Frequency… 273
2 Literature Survey
Because of wasteful traffic frameworks, fuel time and cash are incredibly lost, it
some of the time prompts lost life. So a great deal of work is being done to deal
with those issues. In 2009, the IEEE issued a smart traffic light framework that uses
RFID to filter the vehicle through the vehicle’s RFID label and collect its EPC data
[1]. The traffic volume, vehicle need and street and time are determined by sort of
vehicle. A guideline calculation that suggests how the traffic signals are worked is
given in the Decision Making Section (DMS).
In 2009 [5], the developer proposed a technique for the use of the system’s sensors
as well as the innovation implemented. Accordingly the exchanging of traffic signal
builds traffic and street limit and may prevent traffic jams [6].
In 2014 [7], a web-based traffic management system that could be managed
remotely and monitored simultaneously at different crossings was built using new
web server technologies. The primary node is the Central Traffic Management Unit
(CTMU), which uses Internet technology to remotely track and control individual
nodes.
The author used the RFID label in 2015 [8] and consists of three parts: automatic
monitoring of signals, clearance of vehicles and detection of stolen vehicles.
In [9], the utilization of RFID exercises controls is talked about, specifically those
recognized in the image arrangements and in impedance techniques, to keep up a
vital good ways from issues which generally emerge from standard movement control
sys- tems. This multi-vehicle, multilateral and multi-track convergence framework
is over- seen by RFID. It provides a beneficial time organization plan, in which, a
component schedule proves to be consistently for the segment of each movement
segment. The current use of the framework emulates the fundamental choice of a
functioning cop. The amount and monitoring of vehicles in each segment is the
property where estimates and decisions are completed. The disadvantage of this
examination is that the techniques used for the emergency unit and the control frame
of the moving panels are not attempted.
In 2013, the author [4] proposed a framework for RFID and GPS-enabled crisis
vehicles. The focal errand of this work is to lessen the time the crisis vehicle arriving at
the recuperating office by normally clearing up the course through which the salvage
vehicle continues before the moving signs are reached. This should be possible when
a crisis vehicle is at a specific division of its action convergence, turning the action
signal in transit toward the green. The use of RFID perceives the hand brake and non-
emergency cases, in which unnecessary movement blockage is seen. Correspondence
is delivered through the phone and GPS between the crisis vehicle and moving signs.
When development converges, the instrument is completely mechanical and does
not require human mediation. The restriction of the structure is that it needs all
the information about the start, the end of the tour. This that not work if the salvage
vehicle needs to follow an alternate course or if the beginning stage is unquestionably
not comprehended ahead of time.
274 U. Mittal and P. Chawla
Ongoing advances have been utilized to recognize a crisis vehicle nearness and to
figure the real traffic dense[10], for example, infrared (IR) and worldwide situating
framework (GPS). [11]. So as to recognize the nearness and check vehicles, RFID
labels were utilized to distinguish the nearness of crisis vehicles [12]. The most
widely recognized techniques for traffic recognition incorporate video, microwave,
radar identification, ultrasonic location and so on [9, 13].
For crisis vehicle detection [2, 14] and vehicle tracking [5], RFID innovation
has been recommended. So as to control traffic signals [15], RFID and GSM were
additionally proposed. Such projects, be that as it may, depend on specifically taking
care of various issues.
3 Proposed Methodology
Every vehicle has passive RFID labels with explicit Electronic Product Code (EPC)
RFID label number. No outer force source should be connected. The intricacies,
such as the vehicle number, the vehicle type and the owner’s data are also stored on
each RFID tag. The Ultra-High Frequency (UHF) recurrence band recognizes labels
in the RFID radio wire. The peruser module peruses this information and it is then
sent for additional preparing to the worker. Table 1 shows the details of emergency
vehicles taken for experiment. The proposed system includes 2 main units that are
linked together, namely: emergency vehicle detection and vehicle priority based on
the type of vehicle.
Table 1 Details of
EPC vehicle Vehicle type Owner’s Details priority
emergency vehicles
number
1011 XXX Ambulance XXX 1
1012 XXX Firebrigade XXX 2
1013 XXX Ambulance XXX 1
1015 XXX Ambulance XXX 1
1016 XXX Police XXX 3
1017 XXX Firebrigade XXX 2
1018 XXX Ambulance XXX 1
1019 XXX Police XXX 3
1020 XXX Police XXX 3
Detection of Emergency Vehicles Using Radio Frequency… 275
Detection of Emergency
Vehicle
Multiple Yes
Emergency
Vehicles Prioritize
present at lanes
different
lanes?
No
1. Signals sent by the RFID reader are collected by the transceiver. After recog-
nizing the received signals, system identifies the type of emergency vehicle and
assigns priority to it. For example, if more than one different or same emergency
vehicle are present at different lanes, then according to their type, a numeric
value say 1 for ambulance, 2 fire truck and 3 for police van.
2. System then prioritize the lanes according.
This paper has a great impact on the traffic problems faced in urban areas. Normally
the traffic is controlled by a predetermined signal light controlled system. The
working of the modules is based on the RFID tag information. In this paper, ultra high
frequency RFID tags are used. The only requirement of this paper to be implemented
is that the vehicles should be provided with a unique RFID tag. Proposed method
is implemented and simulated in Python and the results are shows in the Table 2.
Traffic light switching starts from north direction. Emergency vehicle present column
represents whether any emergency vehicle is present or not. Third column represents
number of vehicles present. Fourth column shows type and direction of the emer-
gency vehicle with RFID information. Last columns shows the decision of traffic
light controller i.e. in which direction next signal to be provided.
Any RFID label that exists on a path for over 20 min will distinguish an anomaly.
This might be an accident of the vehicle or a mishap by street. The vehi- cle with
this EPC number will be tried for issues utilizing the phone number of the proprietor
and, where pertinent, will give prompt help.
5 Conclusion
This paper presents a methodology for the arranging of traffic crisis vehicles. The
pro- posed crisis location and organizing framework depends on radio recurrence
ID. With the crisis vehicle location module it will guarantee that the crisis vehicles
will arrive at the goal as quickly as time permits. Contrasted with recently executed
plans, the program is both financially savvy and simple to actualize. The cost level
of the gad- get is exceptionally valuable. Each RFID tag is gathered progressing
by the system. That vehicle may utilize the RFID labels to spare significant data,
for example, vehicle number, proprietor’s information, vehicle model and number
of encroachments by that vehicle. The accompanying can likewise be utilized to
screen crooks, checking vehi- cles taken and e-ringing vehicles, assessing the most
recent details for different sorts of street vehicles, identifying bottlenecks, analyzing
territories for customary blockages and deciding the causes.
References
1. Bansal CS, Chandra A (2009) An intelligent traffic control system using RFID. IEEE Potentials
28(3):40–43. https://doi.org/10.1109/mpot.2009.932094
2. Sharma S et al (2013) Traffic light priority control for emergency vehicle using RFID. Int J
Innov Eng Technol 2(2):363–366
3. Hua LC (2010) Automatic vehicle identification system based on RFID. Anti-Counterfeiting
security and identification in communication (ASID) (2010), pp 281–284
4. Chellani N, Tahilyani C (2013) Traffic congestion detection and control using RFID technology.
Int J Eng Res Technol
5. Chavan SS (walke) et al, Design of intelligent traffic light controller using embedded system.
In: Proceedings of the second international conference on emerging trends in engineering and
technology, ICETET-09
6. Bharadwaj R et al (2013) Efficient dynamic traffic control system using wireless sensor
networks”. In: Proceedings of the IEEE international conference on recent trends in information
technology (ICRTIT). pp. 668–673
7. Pallavi P, Hussain MA (2014) Intelligent traffic control system using embedded web technology.
Int J Comput Sci Inf Technol Secur 4(2)
8. Sundar R, Hebbar S, Golla V (2015) Implementing intelligent traffic control system for conges-
tion control, ambulance clearance, and stolen vehicle detection. IEEE Sens J 15(2):1109–1113.
https://doi.org/10.1109/jsen.2014.2360288
278 U. Mittal and P. Chawla
Abstract In the time of the pandemic like CORONA, Covid-19, everyone is ftghting
against this deadly virus. Besides, governments are looking for a barrier that stops
spread of virus until the vaccine is made. In modern era, technology plays an impor-
tant role. This paper brings the way by using a powerful technology called Big data.
Big data know for handling a large amount of data and provide powerful insights into
the data. Big data integrated with Artificial Intelligence is a powerful tool to ftght
against this pandemic. Many countries like Taiwan, China with the use of Big Data
stop this pandemic up to some extent. But the collection of data itself comes up with
the big challenge of PRIVACY AND SECURITY. In the recent times, the world has
seen the effect of data leaking whether by Facebook or by Google. Many European
countries due to this big challenge will not be able to use this technology.
1 Introduction
1.1 Bigdata
The term describes it all Big data refers to data which is in large volume and different
formats or variety. Generally, Big data is described using the following characteristics
also known as 3 V’s of Big Data. i.e., Volume, Variety, Velocity [1].
Volume—In Today’s world every day Over 2.5 quintillion bytes of data produced
in a single day i.e., approximate.1.7 MB of data in a single second. Big data helps to
manage this huge amount of data with the help of its various frameworks and tools.
Velocity—It refers to the rate at which data is generated and by that time only acts.
on that data. It is challenging but big data provides the various solution for this alsoby
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 279
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_27
280 D. Saikrishna et al.
using Various frameworks like Apache Spark, Flink, Apache Storm we can perform
the data analytics as soon as data enters the system.
Variety—It refers to different types of data that are produced. It is been
Categorized into 3 formats i.e., structured, unstructured, semi-structured.
Structured—It consists of a traditional relational table. Data is represented in a
tabular manner.
Unstructured—The data which is not in a particular format is referred to as.
unstructured data. Audio, video is the type of unstructured data.
Semi-structured—The data which has format, but it not structured is called.
structured data. JSON, CSV formats are the common example of this type of data.
Data has a crucial role in knowing the hidden facts. By collecting Data of individuals
big data helps to provide the insights. Data which helps fight against this pandemic
is:
Data provided by hospitals will play an eminent role. Through recent research, it is
found that a person with heart disease, lung disease or a diabetic person will have
more affected by coronavirus by the prior knowledge of this type of person [2].
Govt. can target this type of person and provide them more instructions during this
coronavirus. Also, By the research, it is found that coronavirus has more effect on a
certain age group of people, i.e., children and old age people.
Usage of location data is quite very useful [3]. Because as we have a medical history
of a patient, we can use the location of a person easily to provide sufficient aid to that
person. Also, by this Govt. also able to see people are following social distancing or
not.
Big Data-Enabled Solutions for Covid-19 281
By getting this data, it is easy to identify which person has travel history (most
favorably out of the country) [4]. So that quarantine period is assigned to that person.
And a person is tested. By this government’s priority to handle these people first.
Before stepping forward let us know what the difference between Privacy and security
is.
Privacy refers to control over how personal information is being collected or used.
It is the power of an individual to stop information about themselves disclose to the
public or using it for personal use.
Security refers to defend the information of an individual by using technology,
process and by using training data generation is categorized into two ways i.e., Active
Data Generation, Passive Data Generation. In Active data Generation data owner will
give its data to the third party while in passive data generation data is generated by
data owner while online activity and data owner do not know its data is being collected
by the third party. Both risks are minimized by access restriction or by falsifying the
data.
Access Restriction—In Access restriction, if data is collected by active data
generation user cannot give its data if it discloses its sensitive information. If it
is passive data generation data is restricted by using an anti- tracking extension or
adblocker and encryption tools.
282 D. Saikrishna et al.
Storing information is not hard in big data due to advancements in technology. But it is
high-risk work. Because once a big data database is being compro- mised the privacy
of an individual is disclosed. Moreover, Big data follows distributed approach data
is placed in several datasets are several locations so privacy is a priority for this. The
conventional security mechanism is divided into four categories Application- level
security, Database level Security, file- level security, and media level security [6].
3 V’s of big data depicts its storage mechanism to be scalable.
Privacy is being a major issue not only now but also, in the past. So, some techniques
were used earlier.
De-Identification—It is a traditional technique for privacy- preserving, for
persevering individual privacy. first, we need to generalize the data then suppres-
sion we must give the data for data mining, following up de- Identification comes
with the concepts of K-anonymity, l-diversity, t- closeness to increase the privacy
of information [7]. Migrating de-identification in bigdata is a great deal for privacy
preserving. Thus, big data also offer many methods of re-identification so using
de-identification is not a great idea for privacy protection. For a description of
K-anonymity, l- diversity, t-closeness let us first discuss some important terms.
Privacy-Preserving Aggregation—It is a homographic encryp- tion technique
used for data collection. This works as follows given a homo- graphic public key
source can convert their data into cipher texts. Different sources can use the same
public key then the whole result is aggregated and recovered by using a specific
private key. But this technique is purpose spe- cific. We can only use this technique
for storing and collecting data. For data exploitation and mining, this technique has
no use. So, it does not help in big data analytics.
Big Data-Enabled Solutions for Covid-19 283
2 1
Query
Database
3 4
Query Data
As Discussed earlier many algorithmic approaches of adding noise to the data but
the fact is chances of reidentification is still in a method. Meanwhile, this algorithm
also adds noise to the data, or we can say data to the noise, so it is quite hard. For the
hackers to find the original data as the same way of finding a needle in the haystack
284 D. Saikrishna et al.
concept or we can say finding salt in sugar. This will surely create a more cost of
computation for making a haystack (noisy data) to hide the needle. In Fig bellow, the
white circles are noisy data while the black circles are the original data. The service
provider adds a dummy as noisy data as original transactional data collected by the
data provider. Likewise, a unique code is assigned both dummies as well as original
data. This code is maintained by the service provider so that during the extraction of
data dummy data and original data got to separate [8].
The way in which medical records are being digitized gives a new face to the health-
care system. Thus, doingthis medical industry experiencing a heavy load of data. By
heavy lead of data then big data comes into the play. Due to the real-time availability
of records and more accurate test result helps doctors to take more constructive deci-
sions in treatment. The real-time monitoring of embedded sensors attached to the
patient gives the warning whenever there is an anomaly. Healthcare digitization with
integrated analytics is one of the next big waves in healthcare information technology
with electronic health records (EHR) being a crucial building block for this vision. By
this healthcare organization came to result that with the EHR incentive programs the
results are more accurate and thus eventually lead to improved healthcare of patients.
But with great power comes great responsibility. Like that the ever- changing risks
and vulnerabilities, security exploitations are expected to grow in the coming years.
Big Data-Enabled Solutions for Covid-19 285
Big data offers a comprehensive survey of different tools and techniques used in
pervasive healthcare in a disease-specific manner [9]. It provides a quick result by
finding major diseases and disorders to give a quick treatment. The adoption of big
data in healthcare significantly increases the security risks and patient privacy risks
[10]. As we have seen traditional techniques were not directly applied to the large
and inherently diverse datasets. In today’s world, we have seen there is a demand for
cloud-stored.
Health data is usually located from different sources, and with different databases,
this leads to anomalies in the data lot of missing values are there. These things should
be handled because dirty data leads to incomplete analysis [12].
7 Conclusion
In this paper, we discussed how big data is helpful in Fighting against COVID- 19 all
we need to take care of challenges that come with big data. We have seen different
techniques to secure the privacy of an individual. Big data leads to knowledge that
provide better analytics. Yet a small amount of data is being used for analytics
due to privacy issues. Privacy techniques with its advantages and disadvantages are
286 D. Saikrishna et al.
References
Abstract In the modern era, traffic congestion is one of the major issues in smart
cities. The existing traffic management is unable to handle the growing number
vehicles. The main aim of this work is to automate the traffic signal by monitoring
the entire system and allowing vehicles to pass through lanes based on the traffic
density. The existing method consists of a signal with fixed timing and also human
intervention for managing traffic. In this proposed work, a system is designed to
program the traffic signal timing and reassign the green signal glowing time based
on traffic density. The lane with higher traffic density will have maximum timing to
pass through and the lane with less traffic density will have minimum timing to pass
through. A system that gives highest priority to emergency vehicles is also developed.
1 Introduction
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 287
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_28
288 V. Radhika et al.
roads without any conflicts. But if a particular traffic signal fails it leads to increase
in accidents, waiting time, increase in greenhouse emissions also. If a traffic signal
in all the lanes have uniform timing it leads to more crowdie in a particular lane than
the other lanes. If the timing is allotted to lanes based on the density of vehicles and
also on any other specific criteria traffic jam can be avoided. So traffic signal timer
programmed based on the traffic density can reduce the traffic congestion problems.
Traffic light system consists of traffic lights and traffic light control system to
control the traffic signals. In the existing system the controllers used are based on
electro mechanical mechanism. These controllers are based on open loop mecha-
nism without any feedback mechanism and sensing network. So they lack in intel-
ligence to adapt to the change in traffic density and control the traffic signal effec-
tively. Traffic signal timer that works according to the traffic density by utilizing the
image processing technique is also proposed in earlier works. In those methods high
resolution cameras are used to sense the changing traffic density around the traffic
signal and programs the traffic signal timing accordingly. In these methods images
are captured on regular intervals, then the images are processed and given to the
controller. Controllers on receiving the input, compares it with predefined thresh-
olds and sends the required signal to traffic signal timer. But these methods need a
high resolution camera and images also to be captured continuously for predefined
time. So an alternative method to automate traffic signal that reduces the traffic jam
effectively and minimize the air pollution is needed for current busy scenario.
This paper proposes an intelligent traffic management utilizing IoT to handle
the different traffic situations. This proposed work reduces the flaws in the existing
system traffic system more effectively. This system utilizes the ultrasonic sensors to
detect the number of vehicles and the RFID tag placed in the emergency vehicles
like ambulance, VIP vehicles and fire brigades to detect and prioritize the emergency
vehicles. So this proposed system detect traffic congestion, provide predictable infor-
mation about the traffic condition can reduce the problems like traffic congestion,
accidents and parking issues.
In this paper Sect. 2 explains the related works, Sect. 3 elaborates the proposed
work, the results and discussion is explained in Sects. 4 and 5 gives the conclusion
and future scope.
2 Related works
This section presents an overview on the techniques and approaches used in the
related works considered for monitoring the traffic system.
For working with the traffic congestion issues, predictive type or real time knowl-
edge about the traffic is necessary. Traditional methods uses sensors like pneumatic
tube and automatic counter to obtain real time data on traffic signal data’s. These tech-
niques fail to detect the vehicle movement in multiple lanes. Video/image processing
techniques utilizing the number plate recognizing system to find the traffic density is
proposed. This technique fails due to non-uniformity in the number plates and visi-
bility problems in poor weather condition. Wireless Sensor Networks (WSN) and
Smart Traffic Monitoring System 289
Vehicle to Vehicle network is utilized to measure the traffic density is also proposed.
Also various related works using fuzzy logic and neural networks learning algorithms
are also utilized for controlling the traffic light is found in literature.
Combinational Interaction Testing (CIT) and Combinational interaction testing
with constraint approaches supports the prioritized based test generation for traffic
signal generation. Optimization algorithms were proposed for scheduling traffic light
conditions. Simulated annealing applied to CIT technique is proposed for computing
the test case generation for traffic signal conditions. Traffic analysis toolbox volume
IV: guidelines for applying corsim micro simulation modeling software and Direct
signal timing optimization: strategy development and results works uses Genetic
Algorithm (GA) for optimizing the programming cycle for traffic lights [2–4]. In
Intelligent Testing of Traffic Light Program approach a PFGM algorithm is proposed
to generate the prioritize test suite for a given a traffic feature model. In this work
an expert system with reduced set of scenarios with similar fault detection capacity
is developed with minimum cost. Developed expert system numerically quantify
and qualify the traffic cycle program for different traffic conditions. In the Isolation
niches particle swarm optimization applied to traffic lights controlling work, particle
swarm optimization algorithm is used for scheduling traffic lights. In this approach
one way road with two intersections was proposed. A multimodal traffic responsive
strategy using Particle Swarm Optimization (PSO) algorithm is proposed in this
approach multi-objective PSO was used. This work uses predictive control model
and simulations were carried out with private, and public vehicle model were carried
out on a virtual urban road network.
3 Proposed Work
Existing traffic signal management system consists of traffic signal that are
programmed such that the yellow, green and red signals glow for fixed interval
of time. The system proposed here identify the traffic density on different lanes and
adjust the traffic signal timing. But traffic jams happening in today scenario signs
that traffic signal must be programmed based on the density of road. So the proposed
system aims to reprogram the traffic signal based on the traffic density. This traffic
monitoring system has two modules such as hardware and software module. The
hardware module has ultrasonic sensor, RFID sensor, RF Module, LEDs, Arduino
UNO microcontroller and relay module [5, 6]. The software module uses Arduino
microcontroller programmed with Ardiuno IDE.
The primary element of this system is the ultrasonic sensor and RFID sensor.
The ultrasonic sensor is used to detect the density of vehicles in a lane and feed
the response to the centralized controller. Centralized controller make decisions to
regulate the traffic signal based on the data given by the ultrasonic sensor about the
traffic density. LEDs are used as traffic signal in this work. RFID Tag is fixed to
special vehicles like Ambulance, fire brigades and VIP vehicles. RFID is a technique
that automatically identifies the object through radio frequency waves. Various types
290 V. Radhika et al.
of RFID tags are available but majorly used tags are active and passive RFID tags in
this traffic management system. Passive tags don’t contain internal power source but
the life time of it is more than active tag. Active tags have internal battery as power
source and range of coverage is more than passive tags. Passive type tag is used in
this work. RFID tag has an integrated circuit for processing the data, modulating
and demodulating the RF signal and an antenna for transmitting and receiving the
data. The RFID reader module is fixed to the particular traffic signal and RFID tag
is placed to the emergency or special vehicles.
Arduino is an ATMega based microcontroller is used as a centralized traffic
controller and receives the ultrasonic sensor data and manages the timing of switching
of traffic light between red, green and yellow. The ultrasonic sensors fixed at four
sides of road in a four way lane that detects the vehicle density in a particular lane.
Ultrasonic sensor is generally used as a distance measuring sensor with a transmitter
and a receiver [7, 8]. Ultrasonic sensors are placed at regular interval of distance
across the lane. First ultrasonic sensor placed in the lane sends the ultrasonic wave
that hits the target surface and it reflects the wave back reaches the sensor and the
second sensor placed after 10 m of distance sends the ultrasonic wave get reflected
back to the sensor. The microcontroller counts the number of vehicle in a lane based
on the distance and timing for reflection of waves between sensors. RF module is
used for the communication between Arduino and the sensor. Figure 1 shows the
ultrasonic sensor placed in four lanes.
The flow diagram for the count of the vehicles with ultrasonic sensor and
microcontroller is shown in Fig. 2.
Data from ultrasonic sensor is processed by the Arduino microcontroller and
programs the LEDs in traffic signal according to the density of vehicles. The
maximum time for tracking 5 min and minimum time for tracking the density is
1 min. Based on the tracking data received from the ultrasonic sensor the micro-
controller gives the priority to the lane with large vehicle density else the priority is
programmed based on Round- Robin method.
This system can consider the density of traffic of road by utilizing the ultrasonic
sensor and can also consider the priority of different types of vehicles [9, 10]. For
this to happen a RFID tag is fixed to the special vehicles or emergency vehicles.
RFID reader reads the RFID tag fixed to the special vehicles. If a special vehicle
enters the road with heavy traffic RFID reader gets the signal from RFID tag and
sends the signal to the microcontroller. Green signal is enabled in the particular lane
till the emergency or special vehicle passes away and traffic get sequenced. The
use of RFID tag differentiates the emergency and special vehicles from the other
292 V. Radhika et al.
Traffic
vehicles and avoids traffic congestion. The block diagram for the proposed work for
identifying the special vehicles is shown in Fig. 3.
The Arduino UNO is used as the microcontroller communicates with RFID reader
and the traffic signal. It establishes the communication between different modules and
generates the control signal for controlling the traffic light and different peripherals.
So the Microcontroller using RFID tag can alter the time of glowing of traffic lights
upon the arrival of the emergency signals and reduce the waiting time of those
vehicles.
Proposed work insights for controlling and managing the traffic by dynamically
changing the timing of traffic light. This proposed work reduces the congestion,
accident, waiting time and pollution. Results presented below shows the effective-
ness of proposed smart traffic monitoring system. The ultrasonic sensors along with
RF modules are placed at a regular intervals of distance in the road. The system is
placed near road as a stand-alone device. Microcontroller detects the number of vehi-
cles whenever the vehicle passes between the ultrasonic sensors in a particular lane.
The microcontroller control the timing of glowing of traffic light based on the vehicle
count and reduce the congestion at road intersections. Delays of traffic light are clas-
sified as LOW, MEDIUM and HIGH range. These ranges can be predefined based
on the vehicle count. Emergency vehicles are equipped with RFID tags and these
RFID tags are read by RFID reader placed at the traffic junction. If the emergency
vehicles are detected by the RFID reader it is given as an interrupt to the microcon-
troller. Microcontroller on receiving the interrupt changes the glowing conditions of
LEDs in the traffic signal. It makes the green LED in the particular lane to glow till
the emergency crosses the junction and after that system returns to normal working
based on the density of vehicles in each lane. The proposed system works effectively
in managing traffic at junctions of road and also gives priority based on the type of
vehicle. A prototype has been made to evaluate and show the arrangement of our
entire system.
Smart Traffic Monitoring System 293
In this work the traffic is managed based on the priority and the density of road
in different lanes of junction. The vehicle density is calculated as a first step in
this work and control the signal timing accordingly. The time of glowing for green
signal is reassigned. The lane with maximum density of traffic will receive maximum
timing for the vehicles to pass through. The lane with less traffic density will receive
Smart Traffic Monitoring System 295
minimum timing for the vehicles to pass through. Ultrasonic sensors and microcon-
troller plays a key role in this work to automate the traffic management system. This
work also gives preference to the emergency vehicles such as ambulance, fire vehicle,
and police cars by allowing traffic signal in the particular lane to glow to green till
the vehicle crosses the signal. RFID tag fixed to emergency vehicle and the RFID
reader at the traffic junction allows the emergency or special vehicle to pass through
with the help of microcontroller.
This work can be combined to a single system for detecting emergency vehicle
and density of vehicle in a particular lane without using separate system.
Acknowledgements Authors thank the Management, Principal and Head of the Department of Sri
Ramakrishna Engineering College for providing facilities and support to carry out this work.
References
1. Sanchez J, Gal M, Rubio E (2008) Applying a traffic lights ´evolutionary optimization technique
to a real case: ‘Las Ramblas’ area in Santa Cruz de Tenerife. IEEE Trans Evol Comput 12(2)
2. Ferrer J, García-Nieto J, Alba E, Chicano F (2016) Intelligent testing of traffic light programs:
validation in smart mobility scenarios. Hindawi Publishing Corporation, vol 3871046, no 1
3. Harrison C, Eckman B and R. Hamilton (2010) Foundations for smarter cities. IBM J Res Dev
54(1)
4. Mendonca M, Cowan D (2010) Decision-making coordination and efficient reasoning
techniques for feature-based configuration. Sci Comput Program 75:311
5. Sivasankar BB (2016) IoT based traffic monitoring using raspberry Pi. Intl J Res Eng Sci
Technol 1:2454
6. Perumalla BK, Sunil BM (2015) An intelligent traffic and vehicle monitoring system using
internet of things architecture. Int J Sci Res 6:391
296 V. Radhika et al.
7. Parkhi1 AA, Peshattiwar AA, Pande KG (2016) Intelligent traffic light system using vehicle
density. Int J Electr Electron Eng 8(172)
8. Lanke N, Koul S (2013) Smart traffic management system. Int J Comput Appl 75(93)
9. Tyagi V, Kalyanaraman S, Krishnapuram R (2012) Vehicular traffic density state estimation
based on cumulative road acoustics. IEEE Trans Intell Transp Syst 23:253
10. Shruthi KR, Vinodha K (2012) Priority based traffic lights controller using wireless sensor
networks. Int J Electron Signals Syst 1:2231
Harvesting Electrical Energy from Body
Heat to Power Bio Medical Wearables
M. Preethika, Mohamad Radzi Bin Ahmad, K. Porkumaran, S. Prabakar,
and R. Sundar
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 297
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_29
298 M. Preethika et al.
1 Introduction
A fast-growing class of such devices is wearable, where sensor nodes are tightly
coupled with the human body, low power consumption is crucial in wearable systems
due to the light weight and size constraints for batteries, which severely limits
the energy that can be stored in the device. Wearables are; however, very tightly
constrained in terms of the size and weight and must also be compatable with the body.
Therefore, the possibilities for EH systems are more restricted than for other applica-
tions. Energy can be harvested from various environmental sources [1] including light
using photovoltaics [2], movement of the wearer [3], from radio frequency energy
(RF), or from temperature differences using thermoelectric generators (TEG). The
human body in contrast is a constant heat source. Typically, a temperature difference
exists between body core and the environment. Because of the voltages produced by
thermal harvesting are typically too low to power wearable electronics, a conversion
stage (DC-DC) with high conversion efficiency needs to be included into the wearable
system. A complete system analysis from the body heat to the application is required
to maximize both the output power and wearability. Thermal energy harvesting is
an another method of reducing the dependability on non-renewable sources. There
are various energy harvesting available, among them are electromagnetic, electro-
static, solar, and thermal methods. However, these mechanisms only produce a very
limited power. To meet the power necessity for most biomedical devices, we use
a boost converter to obtain the required output in order to satisfy the biomedical
wearable’s power requirement. The chosen target locations for power generation
here are the finger, palm, and wrist.
Thermal energy from human body is rarely used to generate energy. The existing
thermal energy harvester is minimally used for Biomedical applications. Hence, there
is a need to develop an alternative charging method using thermal energy harvesting
to prolong the life of the battery.
1.2 Objectives
iii. To boost the minimial output using a boost converter for required power to
charge the biomedical device.
iv. A study of thermal energy harvesting method along the Electrical and
Electronical Engineering to produce a successful Energy Harvesting device.
2 Literature Review
Research on energy scavenging from body motion has been investigated to eval-
uate the feasibility of powering a wearable or implantable systems. Energy from
walking has been previously extracted using generators placed on shoes, backpacks,
and knee braces while producing power levels ranging from milliwatts to watts.
Thermal energy harvesting is an another method of reducing the dependability on
non-renewable sources. The energy harvested from heat uses Thermoelectric Gener-
ators (TEG), which is based on the principle of Seebeck effect [4]. According to [5],
it is a certain type of thermoelectric effect in which the temperature difference is
converted to electrical energy which can be represented by this equation:
E emf = Se T
or
V =∝ (TH − TC ) (1)
• Se / ∝ : Seebeck coefficient
• T: Temperature difference
• TH : Hot side temperature
• TC : Cold side temperature
Fig. 1 Semi-conductor
thermocouple seebeck effect
∝2 T 2
PL max = (2)
4Rin
However, in [1], it is also known that the open-cicuit output voltage, Voc of a TEG
module is also directly proportional to the temperature difference between the two
surfaces. The equation for that can be expressed below:
VOC =∝ mT
the thermal resistance in a normally clothed person changes a little or not at all in the
trunk if the ambient temperature decreases as the average body’s thermal resistance
is 140 cm2 K\W.
Figures shows the miniaturized thermopile attached to the skin [2] (Fig. 3).
From Figure (a):
The miniaturized thermopile attached to the skin is shown in the figure provided
with 1. Ceramic plates which is a metal that acts as an outer covering for the ther-
mopile holding the 2. Thermocouple legs which are interconnected along the 3. Metal
interconnects altogether attached to the ceramic plate (Fig. 4).
From Figure (b):
4. The human skin is kept in contact with the 5. hot plate which is placed parallel
to the 6. cold plate from where the temperature difference T is noted between the
two plates. The 7. holding, encapsulating, and anti-shock elements are provided for
holding, covering, and providing resistance to the shock, respectively. Primarily the
302 M. Preethika et al.
3 Proposed Work
Primarily, after understanding the concept behind the thermal energy harvesting,
I have decided to use a method for heat generation from the human body and to
construct a mathematical model for the thermal energy harvesting.
Secondly, based on the constructed mathematical model, a simulation was done
using MATLAB Simulink and that was to test whether it is feasible or not for the
required output, in order to charge the battery of Biomedical wearables. In order to
obtain the output in volts, I used a boost converter to boost the output obtained from
the thermal energy harvester and is then fed to the battery of the biomedical wearables.
Voc = α × T (1)
where Voc is the output voltage, α is the Seebeck coefficient, and T is the difference
in temperature between the hot and cold surfaces.
P=η (2)
VOC 2
P= (3)
4Re
VOC 2
Q= (4)
4ηRel
α × (T )2
Q= (5)
4η R
The output voltage of thermal energy harvesting is shown in figure based on the
simulation done using MATLAB Simulink.
Harvesting Electrical Energy from Body Heat … 305
According to the theoritical calculatuons of Eq. (1) of the output voltage, VOC , it
is found that VOC = 0.98 V. However, the experimental value based on the simulation
done shows that the output voltage, VOC = 0.976 V, which is close to the theoritical
value. Figure representing the output power, P of the thermal energy harvester.
306 M. Preethika et al.
3.4 Thermocouple
Table 1 Output voltage results from multiple points in hand on two surfaces
Plastic lid output in V Counter top output in V
Part of hand Thermal Boosted Part of hand Thermal Boosted
output in V output in V output in V output in V
10.8 mV 4.95 V 17.8 mV 8.0 V
Finger Finger
17.88 mV 8.4 V 22.5 mV 9.94 V
Wrist Wrist
Palm Palm
Considering the below boost converter circuit as a sample for the generation of output.
The values from the TEG are fed to the boost converter circuit for output appraisal
and then, it is fed to the biomedical wearables (Fig. 7).
4 Problem Solution
Based on the results in Table 1, it is found that when the thermocouple is on a counter-
top, the voltage produced is higher than the plastic lid. However, it is difficult for the
thermocouple to maintain its stability and temperature. Thus, the voltage fluctuates
and does not give a stable reading.
308 M. Preethika et al.
Provided with the output of the finger, palm, and wrist readings by Multisim, the
given input is boosted to obtain the required output in order to charge the biomedical
wearables. The output will be generated by using the boost converter Multisim and
fed to biomedical wearable devices such as the blood pressure monitor, pulsoximeter,
heart rate sensor, temperature sensor, and pulse sensor.
5 Conclusions
References
1. Udalagama CJ (2010) Electrical energy generation from body heat. IEEE ICSET
2. Arabee SBM (2018) A study of wireless charging and thermal energy harvesting methods for
IOT applications
3. Thielen M, Sigrist L, Magno M, Hierold C, Benini L (2017) Human body heat for powering
wearable devices: from thermal energy to application. Energy Convers Manage 131:44–54
4. Alhawari M, Mohammad B, Saleh H, Ismail M (2013) A survey of thermal energy harvesting
techniques and interface circuitry. IEEE, pp 381–384
5. Ghosh A, Meenakshi Khalid S, Harigovindan VP (2015) Performance analysis of wireless body
area network with thermal energy harvesting. In: Proceedings of 2015 global conference on
communication technologies (GCCT 2015), pp 916–920
6. Ahiska R, Mamur H (2013) Design and Implementation of a new portable thermoelectric
generator for low geothermal temperatures. IET Renew Power Gener
7. Leonov V, Su J, Vullers RJM (2010) Calculated performance characteristics of micromachined
thermopiles in wearable devices. EDA Publishing/DTIP
The Significant Relationship Between
E-Governance System and Government
Operation Excellence
Abstract Each government is looking for to provide the best services to estab-
lish efficiency and quality of performance. This goal could be accomplished by
improving the service performance of entire sectors in society. The government of
Syria has realized the importance of moving in the direction of information tech-
nology. Therefore, E-governance initiatives were launched in Syria as a part of
overall country information technology in 20 s century. Each government sector has
since upgraded the performance by having its websites and e-services application.
However, there are gaps and loose connections exist among the sectors, which has
accordingly tarnished the image of Syrian E-governance. This has led to significant
questions about the requirement of modification and enhancement of such service.
Hence, the purpose of this research is to investigate and explore the factors that drive
the E-governance implementation and affect government performance as well as the
government-citizen relationship in Syria.
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 311
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_30
312 A. A. Saleh et al.
1 Introduction
In recent years, new technologies and concepts have therefore altered government
contact with relation aspect such as organizations and citizens [1–3] by launching
original service styles, similar to e-learning, e-banking, e-health, e-commerce, e-
business, e-voting and others, the electronic government has been popularized with
the term E-governance.
This refers to the services supplied by governmental to citizens via providing new
information and communication technology technologies that may allow creating
provision faster and easier. It had been first presented at the ‘90 s, it is regarded as an
innovation introduced by using high-tech development from information technology
increase [4].
Most scholars have divided E-governance into three main categories which are:
Government to Government (G2G), Government to Citizens (G2C) and Government
to Business (G2B). According to Fang (2002) and a few scholars have invented a
fourth category which is Government to Employee (G2E) [5, 6], however, none of
the scholars has explored the interaction between the opposite categories beyond
the government. Category as they consider them to be the e-commerce a part of
the E-governance. For instance, Business to Citizen (B2C), Business to Employee
(B2E) and Employee to Citizen (E2C), don’t seem to be yet identified within the
E-governance system.
Moving closer to E-governance would provide several benefits to a republic’s
kind of services, collectively with finances procedure and control. Certainly, it would
enhance the processing of the government in various components for the improve-
ment of Government Operation Excellence (GOE). A range of these benefits is
decreasing expenses and growing activity thanks to the discount of printed paper
and acquisition of documents and area for storage [7–9].
Reducing the workload on employees is some other advantage to the work envi-
ronment due to the fact employees are concerned with the minimal quantity of citizens
and this can create an increased and serene environment. E-governance will advan-
tage country areas together with large cities since citizens can practice for services
lengthy as they’re a part of the republic and in some instances and services, citizens
may additionally practice for and request a service whilst they’re overseas. Further
advantages to the environment, knowledge, etc., should even be counted for the
improvement of the country [10]. Additionally, E-governance has some challenges
and downsides, and these should not be forgotten or ignored. Just to name a few of
these difficulties are illustrated underneath cost, unreachability, hyper-surveillance,
loss of transparency and accountability [11, 12].
The Significant Relationship Between E-Governance System … 313
The present research will use a quantitative approach to research given that it will
examine the relationship between several independent variables as well as the accep-
tance of E-governance system by users in Syria. The dependent variables of the
study are the Government Operation Excellence (GOE), while the independent vari-
ables are Citizens-centricity, facilitating conditions, performance expectancy, effort
expectancy, flexibility system. And the mediation variable are behavioral intention.
Given the quantitative leanings of the research, the study will thereby focus on
existing statistical data, facts and figures. The accompanying data collection tools
have been identified as questionnaires. Questionnaires were particularly effective in
collecting quantitative data, particularly if researchers are aware of how the variables
are to be tested [13]. In addition to this, questionnaires are also effective when it comes
to collecting data from a wide sample of respondents, dispersed across a number of
locations. Using a questionnaire also allows researchers to ensure that respondents
are presented with a uniform set of questions, this will later result in uniform sets of
data which are not challenging as far as analysis is concerned.
Questionnaires, if designed effectively are unlikely to be time-consuming, there-
fore allowing data to be collected in a relatively short space of time [13–15]. Given the
fact that the research sample is geographically dispersed across Syria, the researcher
relied on a questionnaire approach to investigate.
3 Research Method
According to [16], the method is a set of principles that in any particular situation has
to be reducing to a method uniquely suited to a given situation. The research method-
ology for this research, it presents a systematic and organized procedure that enables
the investigation of the relationship between the different variables attached, namely
E-governance, behavioral intention to use and government operation excellence.
In achieving the objectives, the research methodology deals with procedures
involved in this research. Questionnaire targeting to survey the adoption information
technology for promoting E-governance system among citizens in Republic of Syria
are including in the instrument.
Series of tests such as Structural Equation Model (SEM), and reliability analysis
were conducted to confirm that the instruments were valid (Reliability and Conver-
gent Validity). Also, the research methodology that was followed and adopted by
this research was detailed out. Includes all the procedures, process, and guidelines
that according to research literature is suitable for this kind of the investigation.
The main aim of this research is to study and investigate the mediation effect of
behavioral acceptance on the relationship between E-governance and government
operation excellence in Syria.
314 A. A. Saleh et al.
The review of the information system relevant literature indicates that the acceptance
of information systems, especially E-governance systems are influenced by three
factors, namely: (1) individual factors (2) system factors (3) environmental factors
[17–22].
Concerning theoretical contributions, the principal theories that are used in this
study are the Implications of DeLone and McLean model, Technology Acceptance
Model (TAM), the Task-Technology Fit (TTF) and Unified Theory of Acceptance
and Usage Technology (UTAUT). In this research we combining thus theories into
a new framework in order to achieve the optimum GOE for the Syrian citizens.
Hence, the evaluation of GOE will be supported by the equal common criterion
as is stipulated within the SERVQUAL approach. It will be classified into assurance,
tangibles, reliability, responsiveness and empathy [23].
The focus of this study is to examine the effect of E-governance on GOE.
H1: There is a significant relationship between an E-governance system and
Government Operation Excellence.
Conceptualization of Factors.
Information and communication technology review of the literature has shown
that a number of studies have been conducted to investigate users’ acceptance of
E-governance systems [24–26].
The majority of previous studies have focused on employees and citizens envi-
ronments within economically developed nations. While developing countries have
much to gain from utilising information technology, they usually receive little to no
attention in regards to research [24–30]. To do so, they need to improve upon three
aspects: the implementation environment, individual factor, and system factor. In
that respect, the current study will aim to develop an understanding of the various
factors which influence the acceptance of E-governance system by Syrian citizens.
Behavioral intention.
Behavioral intention to use: Citizens get motivated to use a certain system or
application if it was used or marketed by social groups, friends or family or any
matter that touches the person’s behavior or feeling [31]. A number of authors [31–
34] have attested to the fact that BI has been employed and utilized as a dependent
variable in the examination of E-government acceptance.
Compeau and Higgins [34] Argues that intention is the determinant of the theoret-
ical foundation in information system research in the sense that individual behavior
dictates whether or not to use or not use the technology. Also, BI refers to specific
prompts that cause individuals to act in certain ways or perform certain behav-
iors [35]. This study adding three related determinants of behavioral intent to use
E-governance. They are as follows: system flexibility, citizens-centricity, and facil-
itating conditions. They are added to reflect the individual factor because they have
a significant influence on the intention to use of E-governance.
The Significant Relationship Between E-Governance System … 315
[49] Study concluded that flexibility greatly impacts behavioral intent. Moreover,
[50] explored the differing variables which influenced user satisfaction in regards
to online courses. [50] Concluded that, in most cases, perceived flexibility is vital
regarding user satisfaction. System Flexibility, which is the ability to implement
different processes and apply different facilities to achieve the same goals. It consists
of items such as: product volume flexibility, people flexibility, etc.; [51].
E-governance literature has also provided us with many studies that indicate the
predictive ability of flexibility of E- governance system in determining behavioral
intention [21]. In this research, Syrian citizens will adopt the E-governance system
if they perceive that they can use the system at any time from any place.
H5: System Flexibility will have a positive influence on behavioral intentions to
use E-governance services.
Facilitating Conditions (FC).
Facilitating conditions, as defined by Venkatesh (2003) are the factors which
lead users to believe that the technical and organizational infrastructure of Syria
governance can support E-governance application system. In that respect, facilitating
conditions can result in the rejection of information technology systems due to lack
of faith in the respective organizations and their systems themselves [37]. [39] Also
stated that facilitating conditions resources, such as money and time would motivate
a person to use a particular system. Several studies in the information technology
context have indicated that the facilitating conditions construct has a significant direct
influence on user behavior [52–54].
Adapting this construct to E-governance system suggests that Syrian citizens
perceive that technology infrastructure, and governance infrastructure will enable
them to use the E-governance system without any problem, and to interact with other
users [55].
H6: Facilitating conditions will have a positive influence on behavioral intentions
to use E-governance services.
SERVQUAL.
Jong and Wang [56] Refined their previous service quality evaluation model [57]
and created the SERVQUAL. In short, the authors suggest that a practical service
quality perception stems from the chasm between consumer expectations and there-
fore the performance of the service that’s actually delivered. Thus, finishing up new
studies, with new data collections and analysis, they were able to improve their
scale, reducing the initial ten dimensions to seven: (1) Tangibles; (2) Reliability; (3)
Responsiveness; (4) Communication, Credibility, Safety, Competence; 5) Courtesy;
6) Knowing the customer; and (7) Access.
During a second stage, which entailed another improvement of the study,
SERVQUAL was reduced to 5 dimensions (Table 1). According to [57], in spite
of the kind of service, consumers basically use an equivalent criterion to assess
quality. Service quality could be a general opinion the client forms regarding its
delivery, which is constituted by a series of successful or unsuccessful experiences.
It is difficult to live the standard of service operations because they need the
characteristic intangibility. Geared toward solving this problem, [57] developed a
The Significant Relationship Between E-Governance System … 317
Once the uni-dimensionality of the constructs was achieved, each of the constructs
was assessed for their reliability and validity. Reliability is assessed using Cronbach’s
alpha, construct reliability (CR) and average variance extracted (AVE), while for
validity using construct, including convergent and discriminant Table 2. Represents
the result of Cronbach alpha and convergent validity for the modified measurement
model with 42 reminder items.
As shown in Table 2, the remaining indicators have high factor loadings ranging
from 0.863 to 0.915 indicating that these indicators have preserved the meaning of
the factors.
Table 2 also shows that the AVE, which reflects the overall amount of variance in
the indicators accounted for by the latent construct, was above the cut-off 0.5 for all
constructs as suggested by [65], ranged from 0.531 to 0.684.
The composite reliability values, which depict the degree to which the construct
indicators indicate the latent construct, exceeded the recommended value of 0.6 for
all constructs as recommended by [66], ranging from 0.797 to 0.915.
The Cronbach’s Alpha values, which describes the degree to which a measure
is error-free, range from 0.795 to 0.914 which were above the threshold of 0.7 as
suggested by [58]. Therefore, the achieved Cronbach’s Alpha for all constructs was
considered as sufficiently error-free.
Based on the current GOF results, the chi-square is substantial at the 0.010 level. The
normed fit index of lowest difference chi-square can be removed if the sample size
for the research is more than 200 [82, 83]. After correcting for degrees of freedom
The Significant Relationship Between E-Governance System … 319
in based on the number of variables, the adjusted GFI (AGFI) was 0.881, which was
greater than the recommended cut-off value of according to [84, 85]. The model
predicted 86 percent of the variances and covariance in the survey results, according
to the report. The model showed a satisfactory fit of collected data on the CFI, TLI,
and IFI indices with values more than the cutoff value of 0.9 (i.e., 0.976, 0.973, and
0.975, respectively) [86, 87]. In addition, the root-mean square error of approximation
(RMSEA) was 0.020, and that was lower than the suggested threshold of 0.1 [88, 89,
90]. Furthermore, the fact that the Comparative CMIN/df (1.224) has been less than
5 revealed the good fit of the model [89, 90]. See Table 3.
5 Conclusion
The study has distributed 600 questionnaires to Syrian citizens. From the total, only
423 questionnaires were returned, with only 393 questionnaires were usable (65.5%).
16 respondents returned the questionnaire with missing data of more than 30% for
each questionnaire. According to [67–70] the questionnaires that have missing data
of more than 25% should be omitted. Additionally, 14 of the respondents returned
the empty questionnaire. Therefore, only 385 cases have been used for the study
analysis.
A total of six hypotheses has been tested in achieving the objectives of the research.
From the total, five direct relationships have been found to be statistically significant
to users’ intention to use framework of E-governance system, since the study indi-
cates that the first significant relationship is between Performance Expectancy and
Behavioral Intention (0.000). The second significant relationship is between Effort
320 A. A. Saleh et al.
References
1. Lee M-C (2010) Explaining and predicting users’ continuance intention toward e-learning: an
extension of the expectation-confirmation model. Comput Educ 54:506–516
2. Rokhman A (2011) E-Governance adoption in developing countries; the case of Indonesia. J
Emerg Trends Comput Inform Sci 2:228–236
3. Gil-Garcia JR, Zhang J, Puron-Cid G (2016) Conceptualizing smartness in government: an
integrative and multi-dimensional view. Gov Inf Q 33:524–534
4. Grönlund Å, Horan TA (2005) Introducing E-Gov: history, definitions, and issues. Commun
Assoc Inf Syst 15:39
5. Bertot JC, Jaeger PT, Mcclure CR (2008) Citizen-Centered e-governance services: benefits,
costs, and research needs. In: Proceedings of the 2008 international conference on digital
government research, 2008. Digital government society of North America, pp 137–142
6. Carter L, Belanger F (2004) The influence of perceived characteristics of innovating on e-
governance adoption. Electron J E-Governance 2:11–20
7. Tolbert CJ, Mossberger K (2006) The effects of e-government on trust and confidence in
government. Public Adm Rev 66:354–369
8. Helbig N, Gil-García JR, Ferro E (2009) Understanding the complexity of electronic
government: implications from the digital divide literature. Gov Inf Q 26:89–97
The Significant Relationship Between E-Governance System … 321
32. Hung S-Y, Chang C-M, Kuo S-R (2013) User acceptance of mobile e-governance services: an
empirical study. Gov Inf Q 30:33–44
33. Alkhuwaylidee A, Almahdy AS (2019) Syrian e-government framework toward government
excellence service. Texnologii Pazpabotki Infopmacionnyx Cictem Tpic-2019
34. Compeau DR, Higgins CA (1995) Computer self-efficacy: development of a measure and initial
test. MIS Q 189–211
35. Hsia J-W, Tseng A-H (2008) An enhanced technology acceptance model for e-learning
systems in high-tech companies in Taiwan: analyzed by structural equation modeling. In:
2008 International conference on cyberworlds. IEEE, pp 39–44
36. Hasan H, Ditsa G (1999) The impact of culture on the adoption of it: an interpretive study. J
Glob Inf Manage (JGIM) 7(1):5–15
37. Barua M (2012) E-Governance adoption in government organization of India. Int J Manag
Public Sector Inf Commun Technol 3(1):1
38. Venkatesh, V., Morris, M. G., Davis, G. B. & Davis, F. D. 2003. User Acceptance Of Information
Technology: Toward A Unified View. Mis Quarterly, 425–478
39. Davis F, Bagozzi R, Warshaw P (1992) Extrinsic and intrinsic motivation to use computer in
the workplace. J Appl Soc Psychol 22(14):1111–1132
40. Naor, O., & Geri, N. (2008). Easy as e-mail? Probing the slow adoption of an online assign-
ment submission system. In: Proceedings of the chais conference on instructional technologies
research 2008, vol 5, pp 94–101
41. Taylor S, Todd P (1995) Understanding information technology usage: a test of competing
models. Inf Syst Res 6(2):144–176
42. Venkatesh V, Davis FD (2000) A theoretical extension of the technology acceptance model:
four longitudinal field studies. Manage Sci 46:186–204
43. Goussal DM, Lezcano MSU (2003) Synchronous distance learning and virtual classrooms: a
case study on student expectations and conditioning factors. Australas J Educ Technol 19
44. Sahin I, Shelley M (2008) Considering students’ perceptions: the distance education student
satisfaction model. Educ Technol Soc 11(3):216–223
45. Marchewka J, Liu C, Kostiwa K (2007) An application of the utaut model for understanding
student perceptions using course management software. Commun IIMA 7(2):93–104
46. Karim MA (2003) Technology and improved service delivery: learning points from the
Malaysian experience. Int Rev Adm Sci 69:191–204
47. Themistocleous M, Irani Z, Love PE (2005) Developing e-governance integrated infrastruc-
tures: a case study. In: Proceedings of the 38th annual Hawaii international conference on
system sciences. IEEE, pp 228–228
48. King S, Cotterill S (2007) Transformational government? The role of information technology
in delivering citizen-centric local public services. Local Gov Stud 33:333–354
49. Hermans CM, Haytko DL, Mott-Stenerson B (2009) Student satisfaction in web-enhanced
learning environments. J Instr Pedagogies 1
50. Davis FD, Bagozzi RP, Warshaw PR (1989) User acceptance of computer technology: a
comparison of two theoretical models. Manage Sci 35:982–1003
51. Elsheikh Y, Azzeh M (2014) What facilitates the delivery of citizen-centric egovernment
services in developing countries: model development and validation through structural equation
modeling. Int J Comput Sci Inf Technol 6(1):77
52. Hsia JW, Tseng AH (2008) An enhanced technology acceptance model for e-learning systems
in high-tech companies in Taiwan: analyzed by structural equation modeling. In: International
conference on cyberworlds, pp 39-44
53. Almahamid SM (2013) E-Governance system acceptance and organizational agility: theoretical
framework and research agendas. Int J Inf Bus Manage 5:4
54. Selim HM (2005) Critical success factors for e-learning acceptance: confirmatory factor
models. Comput Educ 49(2):396–413
55. Folorunso O, Ogunseye OS, Sharma SK (2006) An exploratory study of the critical factors
affecting the acceptability of e-learning in nigerian universities. Inf Manage Comput Secur
14(5):496–505
The Significant Relationship Between E-Governance System … 323
56. Jong D, Wang TS (2009) Student acceptance of web based learning system. In: 2009
International symposium on web information system and application (Wisa’09), pp 533–536
57. Tarcan E, Varol ES, Kantarcı K, Fırlar T (2012) A study on kazakh academicians’ information
technology acceptance. Bilig 62:205–230
58. Li, H. And Suomi, R., 2009. A Proposed Scale For Measuring E-Service Quality. International
Journal Of U-And E-Service, Science And Technology, 2(1), Pp.1–10
59. Nunnally JC, Bernstein IH (1994) Psychometric Theory. Mcgraw-Hill, New York
60. Bagozzi, R.P. And Yi, Y. (1988) ‘On The Evaluation Of Structural Equation Model’, Journal
Of Academy Of Marketing Science, Vol. 16, No.1, Pp.74–94
61. Hoyle RH (1995) Structural equation modeling: concepts, issues, and applications. Sage
Publications
62. Chau PYK, Hu PJH (2001) Information technology acceptance by individual professional: a
model comparison approach. Dec Sci 32(4):699–719
63. Byrne BM (2013) Structural equation modeling with lisrel, prelis, and simplis: basic concepts,
applications, and programming: psychology press
64. Hair JF, Black WC, Babin BJ, Anderson RE, Tatham RL (2006) Multivariate data analysis, 6th
edn. Pearson Prentice Hall, United State Of Amreica
65. Parasuraman A, Zeithaml VA, Berry LL (1988) Servqual: a multiple-item scale for measuring
consumer perc. J Retail 64:12
66. Parasuraman A, Zeithaml VA, Berry LL (1985) A conceptual model of service quality and its
implications for future research. J Mark 49:41–50
67. Ho R (2006) Handbook of univariate and multivariate data analysis and interpretation with
Spss. United States of American: Chapman & Hall/Crc, Taylor & Francis Group
68. Schumacker RE, Lomax RG (2010) A beginner’s guide to. Structural equation modeling, 3rd
edn. Taylor & Francis Group, New York
69. Razzaq A, Mohammed AA (2020) Cloud Erp in Malaysia: benefits, challenges, and opportu-
nities. Int J 9.5
70. Mohan NR et al (2014) Using cloud–based web application for universityes: a case of university
of technology in Iraq. Adv Comput 4.1:15–17
71. Saleh AA, Alkhuwaylidee AR, Thangiah M (2021) Successful factors determining the user
behaviour in information system. J Phys Conf Ser 1963(1):012161
Trajectory Tracking Control
of Industrial Robot Manipulator Using
VS Control System with STHWS Method
Abstract In this paper, an experimental investigation has been made to evaluate the
effectiveness of proposed integration of single term haar wavelet series method with
a variable structure control system. The proposed method is applied to non linear
system to obtain solution for trajectory tracking problem by considering dynamics
of the robot as singular system time varying model. The aim is control the variation
or error occurs when the robot is working. In order to minimize the error, integration
of STHWS method with VS control has been designed and implemented in two dof
industrial robot. The experiment has been conducted with various time intervals and
results are analyzed. The obtained results reveals that the proposed integration of
STHWS method with VS control system works well and robust with singular system
time varying model for control of positions.
1 Introduction
In today’s competitive world, it is very clear that the product manufactured should
satisfy the needs of the customer. This scenario increases demand of quality in manu-
facturing and it tends to improve the infrastructure to automation and robotics. This
demand increases the considerable research in robot dynamics and control. The
correct robot manipulator dynamic model plays a vital in control mechanisms. The
formulation of the dynamic model involves various internal and external factors. As
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 325
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_31
326 S. Nandhakumar et al.
it operates, certain physical parameters like load and speed controls these forces [1–
4]. In order to achieve the proper control the exact dynamic model is very important.
In this proposed research, the dynamic model is formulated based on the Euler–
Lagrange principle [5]. To attain the required response and performance laws and
strategies are to be determined within the dynamic model.
Many literatures are available and the performance of robots has been studied
using differential equations. These differential equations are easy to solve and
provides a simple analytical solutions without non linearity’s [6–8]. For representing
any robot configurations where constraints causes rise in contact forces, differential
equations with singular systems are more suitable ([9] McClamroch 1986). Robots
that are applied for advanced manufacturing jobs, the complexity involved in building
a mathematical model is a direct indication. Hence, it is necessary to study the
behavior of proposed model robot arm and its performance for designing robustness
arm for industrial use. Moreover, it is essential to study various control techniques
to analysis problem and issues pertaining to control robot.
Due to various control issues and limitations of traditional controllers, the VS
control is highly desirable which has numerous interesting and significant properties
such as exact knowledge of the manipulator’s model is not needed and under certain
conditions, the system performance is insensitive to bounded external disturbances.
These properties are unique and cannot be easily obtained by other controllers which
makes VS control a perfect model for manipulator controls [10–14].
The objective of proposed work is to investigate the effectiveness of the proposed
integration of STHWS method with VS control by analyzing trajectories of each
joint of 2 DOF industrial pick and place robot. The analyses involves the tracing the
positions and velocities of each joint for find out the experimental results follows the
desired trajectories which is obtained by using cubic polynomial method with zero
speed at end point. During continuous operations, the trajectories has been traced and
plotted. Based on the task, the control system calculates the torque if a trajectory goes
beyond limits, the mathematical block programmed in the controller calculates the
new torque and controls the variations in position by varying input torque. The time
taken for completing one cycle is 48 s. The results are plotted and compared with the
reported values. From the results it has been observed that STHWS to VS control non
singular system time model and supports in attaining preferred trajectories within
the accuracy range expected.
The dynamics of a robot arm can be represented in terms of system of non linear
equations by Nandhakumar et al. [15]
T = P(Q) Q̈ + R Q, Q̇ + S(Q) (1)
Trajectory Tracking Control of Industrial Robot Manipulator … 327
where:
P(Q)
Isthe Coupled inertia matrix,
R Q, Q̇ Is the matrix of coriolis and centrifugal force
S(Q) Is the Gravity matrix,
T Is the Input torques applied at various joints.
For a manipulator with 2 DOF, under the assumption of lumped equivalent masses
and massless- links, the dynamic equations is indicated below
where
K11 : (W1 + W2 ) d1 2 + W2 d2 2 + 2W2 d1 d2 cos(q2 )
K12 : W2 d2 2 + W2 d1 d2 cos(q2 )
K21 : W12
K22 : W2 d2 2
K112 : −2W2 d1 d2 sin(q2 )
K122 : −W2 d1 d2 sin(q2 )
K211 : W122
K1 : [(W1 + W2 )d1 sin(q1 ) + W2 d2 sin(q1 + q2 )]g
K2 : [W2 d2 sin(q1 + q2 )]g
A linear time-varying singular system discussed by Nandhakumar et al. [15],
which is in the following form
The experiment is conducted using 2 dof industrial robot manipulator. The config-
uration and hardware details are discussed by Nandhakumar et al. [5] as shown in
the Fig. 1. In the experiment has carried out in four stages which include two rota-
tions and two linear movements. Initially the arm is in home position i.e. in the top
position. So it has to move to some degree in rotation (2 radians) followed by linear
328 S. Nandhakumar et al.
down movement (0–10 cm) to pick the component. After picking up the object, it has
to move upward from (10–5 cm) and reaching the third position, it rotate and place
the object. During operations, each joint trajectory is traced. Whenever new position
is out of control, the proposed model evaluates the desired torque and compares it
with the existing one. If it goes out of control, the proposed integration estimates the
torque based on the varying positions, and compares with existing torque. Finally,
the exact trajectories have been achieved and results of velocity errors and positions
errors have been plotted and shown in Figs. 2, 3, 4, 5, 6, 7, 8 and 9. From these
figures, it has been observed that the proposed integration controls the errors and
provides the projected range of accuracy.
Figures 2, 4, 6 and 8 shows the positions and velocity profiles of Joint 1 and 2
respectively arising out of the proposed algorithm. From these profiles, it has been
observed that trajectories very closely follow exact trajectories which are obtained by
cubic polynomial method. The position errors and velocity errors are as shown in the
Figs. 3, 5, 7 and 9. The errors obtained are large than the numerical simulations [15].
The obtained level of accuracy is comparatively good with steady errors. A detailed
comparison of the results with [5] shows that the position and velocity error for joint
1 and joint 2 is less. From Fig. 3, at joint 1—Position error—4.0 × 10–3 rad,Velocity
error—3 × 10–2 and at joint 2—Position error 5.0 × 10–2 rad and Joint velocity
error 4 × 10–2 rad/s which is approximately 20% less than the model reported by
Nandhakumar el al 2013.
Trajectory Tracking Control of Industrial Robot Manipulator … 329
5 Conclusions
References
1. Sekar S, Prabakaran K (2011) Numerical Solution for linear and nonlinear singular systems
using single term haar wavelet series, Int J Commun Comput Inf Sci 140:204–212
2. Sekar S, Prabakaran K, Paramanathen E (2012) Single-term Haar wavelet series technique for
time varying linear and non-linear singular systems. In: International conference on pattern
recognition, informatics and medical engineering (PRIME-2012), pp 364–369. https://doi.org/
10.1109/ICPRIME.2012.6208373
3. Nandhakumar S, Selladurai V, Sekar S (2009) Numerical investigation of an industrial robot
arm control problem using haar wavelet series. Am J Eng Appl Sci 2(4):584–589
4. Nandhakumar S, Selladurai V (2011) Haar wavelet approach to second order robot arm control
problem. Eur J Sci Res 56(4):573–583
5. Nandhakumar S, Muthukumaran V, Soorya Prakash K, Shunmughanaathan VK (2013) Position
control of industrial robotic manipulator using variable structure control system with single
term haar wavelet series method. J Vib Control 21(12):2465–2483
6. Ata AA, Johar H (2004) Dynamic simulation of task constrained of a rigid-flexible manipulator.
Int J Adv Robot Syst 1:61–67
7. Park BS, Yoo SJ (2018) A low-complexity tracker design for uncertain nonholonomic wheeled
mobile robots with time-varying input delay at nonlinear dynamic level. Nonlinear Dyn
89(3):1705–1717
8. Yamawaki T, Yashima M (2007) Effect of gravity on manipulation performance of a robotic
arm. In: IEEE international conference on robotics and automation, Rome, pp 4407–4413
9. Liu W, Li F, Wan Y et al (2020) Design of manipulator based on two-dimensional force
feedback. Intel Serv Robot. https://doi.org/10.1007/s11370-020-00321-y
10. Castri C, Messina A (2007) Exact modeling for control of flexible manipulators. J Vib Control
18(10):1526–1551
334 S. Nandhakumar et al.
11. Cheng-Yuan C, Hsu K-C, Chiang K-H, Huang G-E (2008) Modified fuzzy variable structure
control method to the crane system with control dead zone problem. J Vib Control 14(7):953–
969
12. Hannan MA, Hussin I, Ker PJ, Hoque MM, Hossain Lipu MS, Hussain A, Blaabjerg F (2018)
Advanced control strategies of VSC Based HVDC transmission system: issues and potential
recommendations. IEEE Access 6:78352–78369
13. Pai MC (2015) Sliding mode control of vibration in uncertain time-delay systems. J Vib Control
16(14):2131–2145
14. Quynh NX, Nan WY, Yen VT (2020) Design of a robust adaptive sliding mode control using
recurrent fuzzy wavelet functional link neural networks for industrial robot manipulator with
dead zone. Intel Serv Robot 13:219–233. https://doi.org/10.1007/s11370-019-00300-y
15. Nandhakumar S, Selladurai V, Muthukumaran V, Sekar S (2010) Haar wavelet approach to
time varing model robot arm control problem. Int J Acad Res 2(4):66–74
16. Bujurke NM, Salimath CS, Shiralashetti (2007) Numerical solution of stiff systems from
nonlinear dynamics using single-term Haar wavelet series. Int J Nonlinear Dyn 51(4):595–605
17. Hsiao CH (2004) Haar wavelet approach to linear stiff systems. Math Comput Simul 64(5):561–
567
18. Hsiao CH (2005) Numerical solution of stiff differential equations via Haar wavelets. Int J
Comput Math 82(9):1117–1123
Medical Fitness Device for Driving
Assessment
K. Chamundeswari, R. Saranya, M. Smitha, S. Prabakar,
and K. Porkumaran
Abstract This project is to check Medical Fitness for accidental Drivers and paral-
ysed patients health condition using pressure and heart beat sensor with RTC unit.
This can be helpful to analyze the accident person’s health improvement condition
or periodic monitoring of that person. In hospitals, accidental patients’ body condi-
tions (Leg and Hand) must be monitored constantly, which is typically done by
doctors or other paramedical staff. And when the conditions go abnormal then we
sense those values by Displaying in LCD and buzzer sound. The suitable parameters
which should be monitored are first determined and also the respective sensors are
connected to the patient‘s body and periodically checked by a PIC microcontroller.
And these parameters have fixed threshold levels which are monitored and also the
parameters are transmitted to a LCD for continues monitoring purpose.
1 Introduction
An accident is knock or miss event that at times has convenient or more undesir-
able consequences, other times being inconsequential. The term implies that such
an incident might not be preventable since, their leads to circumstances go unrecog-
nized and unaddressed. Accident may lead to injuries which can affect the nervous
system. This may also lead to paralysis, trauma, stroke etc. Paralysis may be a loss
of muscle function partly of our body which may be temporary or permanent. This
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 335
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_32
336 K. Chamundeswari et al.
could flow from to attack within the nerves of the peripheral systema nervosum by
the body’s system which may eventually result in temporary paralysis of arms and
legs [Guillain –Barre syndrome]. The patients with such disabilities don’t seem to
be ready to perform their everyday actions like feeding, toilette usage, movement
through space and driving. After recovery, to assess their driving ability this project
are going to be helpful by measuring the parameters like pressure with which they
apply the break and therefore the time with which they react and heart beat. Since,
till now only the software and scans like MRI, CT, X-RAY are present to assess
the anatomical functions and therefore the overall fitness. Occupational therapists,
as well as certified driver rehabilitation specialist are required to create important
recommendation concerning the fitness to the drive of their clients on daily. To help in
formulating these recommendations they’re going for scanning technologies. Hence
the goal of our project is to assess the fitness without scanning. This project will have
high accuracy than the prevailing ones. This device are going to be helpful for simple
observation of patients exact health condition and improvement. it’s easy to control
and handle. It may be of great use thanks to it’s compactness and price effectiveness.
2 Related Works
A wireless and reliable health monitoring and warning system was proposed by
Khemapech [1] in that the vital signals has been recorded using the wearable device
and further analyzed for providing real time warnings. The heart rate and blood pres-
sure and pulse rate are the measured as the vital signals for depicting the patient’s
cardiovascular symptoms by comparing ranges with the standard values given by the
doctors. One of the key drawback they faced is the, range between wearable device
and the software of the system is the not in the limited communication aspects. Omer
Al-Salihi [2] described about an android mobile application which was successfully
built and consequently the data transmission using internet (WiFi) to both the appli-
cation and to the remote server by was done perfectly,. In addition to the current, the
next objective of this was achieved by examining the vital parameters remotely and
predicts the consequences. Although the results were far considerable there still exist
a problem in both hardware and software interface. Appelboom et al. [3] proposed a
project by implementing smart wearable sensors into the practice patient care could
improve the relationship between doctor and patient which increased the indepen-
dence and involvement of patients regarding to the healthcare aspects and remote
monitoring provides the most effective way of monitoring which brings a drastic
change in the healthcare organization.
Kumara et al. [4] designed a device to assist these patients fingers of the hand
that play a major role in this system. The system ensures the patient to precise his
needs by converting the finger bending into auditory speech. If the identical action
is repeated thrice, then there is an efficiency of the system to send a text message
to the concerned caretaker with the assistance of GSM module. In future more flex
sensors can be accustomated to perform multiple tasks and more health Parameter
may be included. Aminian and Naji [5] proposed a system which implements all
Medical Fitness Device for Driving Assessment 337
sufficient aspects of monitoring to measure the vital parameter for the multiple users
or the patient. In this work a coordinator node is assigned which is attached to the
body of the patient which in turn it collects all the vital information’s by the wireless
sensors and it transmits to the base station for future analysis. The placed sensors
on the body of the patient act as a wireless body sensor network (WBSN) and are
capable of sensing the vital parameters. In case any abnormalities in the measured
parameters the system indicates an alarm to the patient and message also has been
passed to the doctor’s end by an Email or SMS.
While comparing with the pervious findings there is a large amount of reduction in
the energy consumption and also there is a speed in the network lifetime which in turn
improves the communication coverage to extent the comfortness and independence
for ensuring the quality of life, these are considered as a most advantageous of the
proposed system. Deepasri et al. [15] proposed a monitoring system, for indicating
the patients health aspects. In this work the most efficient biosensors are used to
observe or record the physiological parameters such as blood pressure, pulse rate,
heart rate and measured data are continuously sent to the nurse through by using
GSM. Future it been processed in Microcontroller (MSP430). Roshdy et al. [14]
proposed a design for the paralyzed hand people a easy portable device with light
weight and low cost. Based on all these consideration we proposed a device for
monitoring the parameters after recovery of the paralyzed patient in the aspects of
driving.
3 Methodology
The project uses Pressure and heart beat sensor for transferring data to a micro-
controller unit fitted in doctors monitoring area. The pressure sensor value and
reaction time take by patient is displayed through LCD screen. Respective sensors
are connected to the patient‘s body whenever medical test is carried out and peri-
odically values are checked by a PIC Microcontroller. All these parameters have
fixed threshold levels which are monitored. And the parameters are transmitted to a
LCD for continues monitoring purpose. Depending upon the value treatment maybe
carryout further.
The project uses Pressure sensor for transferring data to a microcontroller unit
fitted in doctors monitoring area. The pressure sensor value and latent period take
by patient is displayed through LCD screen. Depending upon the worth treatment
maybe carryout further.
4 System Implementation
The proposed system makes use of two sensors namely flexi force or pressure sensor
and heart beat sensor to measure the appropriate parameters (Fig. 1).
338 K. Chamundeswari et al.
In this system we use the PIC 16F877A microcontroller for the conversion of the
parameters. The pin of the this microcontroller is shown in the Fig. 2.
FSR are used to continuously monitoring the pressure of the patients especially the
pressure applied to the peddle of accelerator and brake at the time of driving. The
Fig. 3 shows the FSR used in the proposed work.
The heart sensor is placed on the patient’s body to continuously measure the heart
rate of the patient while driving. This is value is send to the physicians at the other
end. This is a digital value.
Medical Fitness Device for Driving Assessment 339
LCD (Liquid Crystal Display) is used to display the values of the pressure and the
heart rate. Here in this work we use a 16 × 2 LCD display is very basic module and
is very commonly used in various devices and circuits. The LCD display used in our
work is shown in the Fig. 4.
340 K. Chamundeswari et al.
The power supply we use in this work is ac voltage, typically 220V rms, is connected
to a transformer, which steps that ac voltage down to the level of the desired dc output.
• Embedded C.
• PIC Boot loader.
• CCS C Complier.
Medical Fitness Device for Driving Assessment 341
The working modules of the proposed system are shown in Fig. 5 and Fig. 6 which
are very helpful for the paralytic patients and accidental people.
The system continuously monitors heart beat rate and blood pressure of patient.
The system also has the feature to check the physical fitnesss of the patient for
driving with a high accuracy. Patients exact health condition can be easily observed.
Due to its compactness and low cost this product can be used easily in all healthcare
centers, even in home. Also it is easy to operate and handle.
6 Conclusion
This project presents the design and implementation of accidental health monitoring
system by using flexi force sensor module. The proposed system will be a helping
hand for accidental and paralytic patients. It is concluded that Programmable Inter-
face Controller (PIC) has been the low cost implementation used for recording and
transmitting the bio-medical signals and very useful to the patients. This system was
developed to minimize the device’s size and allow for daily life usage.
References
1. Khemapech I (2015) A real time health monitoring and warning system. TRON Forum
2. Omer RMD, Al-Salihi NK (2017) Healthmate: smart wearable system for health monitoring
(SWSHM). In: 2017 IEEE Students conference on electrical, electronics and computer science
3. Appelboom G, Camacho E, Abraham ME, Bruce SS (2014) Smart wearable body sensors for
patient self—assessment and monitoring. Archives of Public Health
4. Kumara KR, Kadam A, Rane N, Gouda A, Sensor based wearable system to assist paralytic
patient with continuous health monitoring. Int J Future Revolution Comput Sci Commun Eng
5. Aaminian M, Naji HR, A hospital healthcare monitoring system using wireless sensor networks.
J Health Med Inf
6. Nelson A, Shyam Kumar P (2013) Wearable multi sensor gesture recognition for paralysis
patients. In: 2013 IEEE, University of San Francisco
7. VasanthRao Y, Swamy Bhukya (2019) Gait Assistive exoskeleton device for semi paralyzed
stroke survivors. Int J Eng Adv Technol ( IJEAT) 8(5c)
8. Agrawal SK, Sangwan V (2007) Assessment of motion of a swing leg and gait rehabilitation
with a gravity balancing exoskeleton. IEEE Trans Neural Syst Rehabil Eng 15(3)
9. Yorozu A, Moriguchi T (2011) Improved leg tracking with data association considering gait
phase during turning motion in walk tests. In: 2015 IEEE International conference on advanced
intelligent mechatronics [AIM] July 7 2011
10. Milovanovic I (2008) Radial basis function (RBF) networks for improved gait analysis. In: 2008
IEEE 9 th Symposium on neural network applications in electrical engineering, NEUREL
11. Sayegh F, Fadhli F, Karam F, BoAbbas M, Mahmeed F, Korbane JA (2016) A wearable reha-
bilitation device for paralysis. In: IEEE 2016, American University of the Middle East Egaila,
Kuwait
12. Hasegawa Y, Mikami Y, Watanabe K, Firouzimehr Z (2008) Warble handling support system
for paralyzed patients. In: 2008 IEEE/RSJ International conference on intelligent robots and
systems acropolis convention center nice, France, Sept 22–26, 2008
13. Hasan HN (2019) A wearable rehabilitation system to assist partially hand paralysed patients
in repetitive exercises. Biomedical Engineering Department, College of Engineering Quality
Assurance , University of ThiQar, First International Scientific Conference Al-Ayen University
IOP. J Phys Conf Series 1279 (2019) 012040
14. Roshdy A, Al Kork S, Said S (2019) A wearable exoskeleton rehabilitation device for
paralysis—a comprehensive study. Adv Sci Technol Eng Syst J 4(1):17–26
15. Deepasri. T Gokulpriya. M Arun kumar. G (2017) Automated paralysis patient health care
monitoring system. Dep Electron Commun Eng South Asian J Eng Technol 3(2):85–92
Application of Artificial Intelligence
for Reservoir Storage Prediction: A Case
Study
Abstract There are many relevant and interesting contributions using Artificial
Intelligence (AI) based techniques, with different purposes. It has been used as an
effective way for estimating the forecasted data of reservoir daily storage value. The
efficiency of various AI methods was explored in this article and later the best method
is selected for reservoir storage level prediction. In estimating reservoir storage levels
several regression algorithms and artificial neural network (ANN) approaches have
been evaluated. There is better agreement between the ANN model compared to
regression algorithms. The findings were demonstrated by significant correlation
coefficient (R2 ) rate among the expected and calculated training outcome variables
up to 0.91 and the highest validity outcome of Root Mean Square Error (RMSE)
was 5.1989. Consequently, this method is therefore adequate for robustness and
generalizability abilities and is ideal for forecasts.
Nomenclature
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 343
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_33
344 A. S. Azad et al.
LM Levenberg marquardt
ML Machine learning
MSE Mean square error
R Correlation coefficient
RMSE Root mean square error
1 Introduction
A reservoir is a physical (artificial and natural) system used to preserve water for the
storage, management and regulation of the flow of water [1]. Reservoir level control-
ling parameters are inflows, water storage, water release, evaporation, and infiltration
from reservoir. These parameters are unpredictable and must be considered in the
operation of water supplies. The higher results can be achieved if the dam reservoir
level is calculated correctly [2]. Unfortunately, if a reservoir is full on the maximum
storage level due to excessive inflow, the water level rises and the discharge rate over
the drainage line rise. This event will lead to downstream flooding [3]. Therefore,
reservoir water level forecasting is important for flood caution, flood management,
river water potential assessment, hydroelectric power generation, domestic manage-
ment and for river water potential assessment [4]. Also, the significance of water level
forecasting would give the operator the advantage of optimum water level drawing
and a sustainable hydropower generation strategy [5].
There have been several methods used in the forecasting of hydrological activities
in recent years. Previously, the reservoir water level forecasting software used a tradi-
tional linear mathematical relationship approach focused on operator knowledge,
mathematical curves and instructions [6]. Overestimation of features in data and wide
missed variables, however, causes low performance and undermines computational
models because of difficulty and shortage of explosion to the data. With regard to these
problems, the use of Machine Learning (ML) techniques has been adopted, provided
that nonlinear process modelling and forecasting effects are better than conventional
models, such as moving average processes [7–9]. Machine Learning (ML) focuses on
allowing systems through a provided data to forecast, clustered, absorb classification
models, or render choices. The main importance of this modelling is the software’s
ability to visualize scenarios of data without the above-mentioned understanding
from variables influencing the projection variables [10, 11]. Studies have continued to
implement this ML to forecast a range of modeling methods and variables to increase
efficiency and model’s durability prediction description [12]. Regression Algorithms
is considered an alternative approach to ANN for implementing nonlinear predic-
tion. It has been increasingly used in hydrological prediction over the years [13]. In
graphical or mathematical equations, predictive methods are typically provided using
computed historical data and statistical analysis [14]. This research hence emphasizes
on predicting improvements to the amount of storage water through ML algorithms.
Application of Artificial Intelligence for Reservoir … 345
Key objective in this research to come up with reservoir operant by reliable storage
level predicting method and to investigate the simulation methodology of Regres-
sion Algorithms likely linear regression (LR) models, decision trees (DTs), Gaussian
process regression (GPR) models, support vector machines (SVMs), as well as Arti-
ficial Neural Network (ANN) and ensembles of tree models the models were trained.
Secondly, the efficacy of the best method will be used in studying the water level
parameter pattern. Lastly, the selected method will be used for industrial operations
and development and increases the reliability of forecasting for water level storage.
It will help to establish better river operating strategies, leading to better condi-
tions for water users, including agricultural operations, manufacturing practices,
and hydropower production. As a consequence, the exact prediction for the level
of stored water would have a positive effect on improved planning for all relevant
industrial and commercial operations. In a summary, the purpose over this analysis
is to predict volume of stored reservoir water thus it is an essential consideration for
decision-making or the reservoir operator to manage the planning of water supplies.
Mangla Dam is located on the Jhelum River in the Mirpur District in Pakistan.
Coordinates are latitude: 33.142083 °N and longitude: 73.645015 °E with Height
147 m (482 ft) and Length 3140 m (10,302 ft). The total capacity 9.12 km3 (7,390,000
acre ft) with catchment area 33,334 km2 . Turbine numbers are 10 MW × 100 MW,
installed capacity 1150 MW (15% overload) 1500 MW (max. planned). Figure 1
illustrate the Dam location.
Data for this study were 10 years Meteorological Historical data points, collected
for average monthly temperature, rainfall, wind speed, humidity, sun hours and UV
index from January 2009 to October 2020. Hydrological average monthly data of
inflow, irrigation demand and reservoir storage for 2014–2020 was collected. The
average monthly data was used for the study purpose (Fig. 2a, b).
2.2 Methodology
In this study, the dataset was split into training and test dataset, to avoid over-fitting
of the method. Using the dataset of the training section for various statistical models
and Neural Network Regression (NNR).
346 A. S. Azad et al.
n 2
i=1 y pr ed,i − yex p,i
R =1−
2
n 2 (1)
i=1 y pr ed,i − ym
1 2
n
MSE = y pr ed,i − yex p,i (2)
n i=1
1
n
M AE = y pr ed,i − yex p,i (3)
n i=1
n
y pr ed,i − yex p,i 2
RMSE = (4)
i=1
n
where, respectively, ypred and yexp are the predicted and experimental values. In
the data, N denotes the total number of values. The mean answer value is denoted
by ym.
The RLA enables system to instantly train several standard regression techniques
simultaneously time analyze their findings and thus choose the most appropriate one.
The RMSE, quadratic regression (R-square) parameter, mean absolute error (MAE),
mean squared error (MSE), predicted speed, and time of training of Regression
models and Artificial Neural Network are shown in Table 1. The table shows that ANN
had better accuracy in terms of errors evaluation than regression linear algorithms.
Therefore, ANN is selected for further study purpose.
The efficacy of the ANN model at various training protocols is listed in Table 2
here. The most efficient learning strategies for ANN training was back propagation
(BP) algorithm. This algorithm can be quickly applied and incorporated in functional
forecasting systems as given in early research works. It is possible to split back
propagation into two phases: A process of feed-forward and a backward step. In
current study feed forward BP algorithm had been used based on its minimum value
of MSE. The input information was delivered to the input layer, the hidden layer and
the output layer in order to the feed-forward stage, to receive the output information.
ANN was trained and tested further to reach the near-optimal relationship between
the input data set and the output data set.
It must be remembered, though, that the participation of several neurons in the
hidden layer may induce the over-adaptation, resulting in the lack of network gener-
alizability, and that a limited number of hidden neurons may also fall short of the
Application of Artificial Intelligence for Reservoir … 349
data and that the network will not then remember. Therefore, if the structure fails to
meet the experimental outcomes, the water treatment plant production value becomes
disorganized and the plant management challenging. The trial and error approach
were used to identify the most effective network model, the ideal number of cached
layers and the maximum possible number of neurons in the cached layer. There
were two maximum number of hidden layers. The addition of more hidden layers
increases the system’s complexity and thus simulator time and expense. Authors have
shown that a framework with a hidden layer will estimate any continuous function
if adequate freedom is given.
The required design generated both training and testing data with the minimum
of errors. Table 3 tests the output of each network model by evaluating the MSE for
each test performed to find the required architecture. Backpropagation network and
14–20 neurons have been shown to achieve minimum MSE levels, so that all models
are limited to 20 neurons and evaluated.
Simulated data was analyzed for the testing and validating of model networks with
only one low, one medium, one in high-value storage datasets. The findings for the
predictive process modelling in Table 3 show the number of neurons in hidden layer.
Several experiments for each group have been performed to achieve the required
Application of Artificial Intelligence for Reservoir … 351
network layout, before the appropriate learning rate, the number of hidden layers
and the number of neurons per hidden layer have been achieved.
If a model ANNs can precisely define the highest standards in the Correlation
coefficient (R2) and minimum standards in the Root Mean Square of Error (RMSE)
it can simply be the accurate and superior ANN model. The number is noticeable in
Fig. 4 that the result directions for training (R-value = 0.9221), validation (R value
= 0.82371) and testing (R value = 0.65729) are aimed very strongly. These values
may equate to a total R value of = 0.86872. In this scenario, the network reaction
has so far been optimistic, and simulation can be used to reach new inputs.
The Regression (R) plot highlights the assessment between the target that is
believed to be the optimal output and the ANN output that is known to be the actual
output, which indicates that the ANN output is exactly the same as the target output.
Notably, the Regression (R) value indicates the significance of outputs and priorities.
In fact, since R can exceed 1, this means an accurate linear relevance between goals
and outputs. The validation and the performance checked display values of (R) are
both greater than 0.90. It shows high performance of the model.
352 A. S. Azad et al.
The trained model was used for the prediction of storage data for water level of
January–November 2020. Figure 5 shows that the data was well predicted using the
trained ANN model. The prediction accuracy was 92% and RMSE value 5.1989. The
lower value of error shows that ANN is a reliable ML tool for prediction of reservoir
storage data provided with suitable meteorological and hydrological data.
4 Conclusion
The networks used to train and forecast water storage levels using the Matlab software
in this study. The learning algorithm, such as Levenberg–Marquardt, was qualified in
training to train the model. The results were shown by a high value of the correlation
coefficient (R2 ) between the expected and observed training variables of up to 0.91
and the highest validity efficiency value of the RMSE was 5.1989. Consequently, it
Application of Artificial Intelligence for Reservoir … 353
is concluded that for accuracy and generalization capability, this developed model
has been satisfactory and is suitable for prediction purpose.
Acknowledgements The research work has been supported by the Graduate Assistantship (GA),
Fundamental and Applied Sciences Department (FASD), Universiti Teknologi PETRONAS and
SGRG research group.
References
10. Najah A, El-Shafie A, Karim OA, Jaafar O (2011) Integrated versus isolated scenario for
prediction dissolved oxygen at progression of water quality monitoring stations. Hydrol Earth
Syst Sci 15:2693–2708
11. Hipni A, El-shafie A, Najah A, Karim OA, Hussain A, Mukhlisin M (2013) Daily forecasting
of dam water levels: comparing a support vector machine (SVM) model with adaptive neuro
fuzzy inference system (ANFIS). Water Resour Manag 27:3803–3823
12. Zantalis F, Koulouras G, Karabetsos S, Kandris D (2019) A review of machine learning and
IoT in smart transportation. Future Internet 11:1–23
13. Wang Q, Wang S (2020) Machine learning-based water level prediction in Lake erie. Water
(Switzerland) 12:1–14
14. Zhao T, Minsker B, Salas F, Maidment D, Diev V, Spoelstra J et al (2018) Statistical and hybrid
methods implemented in a web application for predicting reservoir inflows during flood events.
J Am Water Resour Assoc 54:69–89
Fabrication of Parallel Mechanism
Actuated by Pneumatic Artificial Muscle
for Rehabilitation Therapy
Abstract Ankle–Foot injuries are amongst the most common injuries of the lower
limb and almost 25,000 people experience ankle injuries each day due to vigorous
activities. Traditionally, ankle injuries are rehabilitated via physiotherapy using
simple equipment like elastic bands and rollers, requiring intensive efforts of ther-
apists and patients. Currently, Stewart platform rehabilitation devices are actuated
by various methods including double acting pneumatic cylinder, hydraulic, electric
motor and shape memory alloy. The limitation of using these actuation methods is
that it provides lower range of motions and requires higher maintenances. A less
known type is Pneumatic Artificial Muscles (PAMs). PAMs are operated by pneu-
matic pressure and are contractible naturally upon inflation. The objective of this
study is to fabricate Stewart platform using Pneumatic Artificial Muscle, to analyze
the range of motion produce by various diameter of the PAMs. The diameters of PAM
used were 8, 10 and 12 mm. It was observed that Stewart Platform actuated by PAM
of diameter of 12 mm produced maximum platform angle of 31.73°, whereas PAM
with the diameter of 10 and 8 mm produced maximum platform angle of 28.62° and
25.31°, respectively.
S. Krishnan (B)
Manipal International University (MIU), Nilai, Malaysia
e-mail: [email protected]
A. M. A. Rani
Universiti Teknologi PETRONAS (UTP), Tronoh, Malaysia
L. G. Kurappa
University Nilai, Nilai, Malaysia
S. Paramasivam
UOW Malaysia KDU University College, Glenmarie Campus, Shah Alam, Malaysia
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 355
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_34
356 S. Krishnan et al.
1 Introduction
Ankle joint is one of the significant joints in human body which helps maintain body
balance during ambulation. In a human skeleton, human ankle joint is a very complex
bony structure. Ankle sprain is a common injury and the number of injuries is rising
day by day due to vigorous activities carried out by people and this can happen
to athletes and non-athletes, children and adults. Beside vigorous activities, it can
also happen when you simply step on an uneven surface, or step down at an angle.
Ligament play important role in human ankle where it holds the ankle bones and joint
in position. For all the abnormal movements-especially twisting, turning, and rolling
of the foot, the ligament provides protection to the ankle joint. Since ligament is an
elastic structure it has the ability go back to their normal positions after it stretches
within its limits. Sprain occurs when ligament is forced to stretch beyond its limit.
Actual tearing of the elastic fibers caused by the severe sprain. Apart from ankle
sprain, stroke is one of the most typical impairment which causes the patient difficult
to lift their foot due to weakness in the dorsiflexion muscles. Figure 1 shows anatomy
of human ankle.
Physical therapies method which rehabilitates patient one-to-one is known as
traditional rehabilitation. Obtaining desired results from this traditional rehabilita-
tion method solely depends on the therapist skills. This traditional method affects the
therapists in a negative way since the rehabilitation process contains repetitive and
intensive exercises and it takes longer period of time to cure. Hence, there will be
difficulties in delivering the same rehabilitation process by the therapist. Therefore,
from the viewpoint of patients, therapists and rehabilitation process by studies, the
2 Literature Review
The braided pneumatics actuators were invented by McKibben to help the movement
of polio patients 1950. They are also called a McKibben Muscle or Pneumatic Muscle
which is actuated by compressed air. Pneumatic muscle was used as the original
applications such as, door opening arrangement and on industrial robot [3]. The
displacement movement performed by pneumatic muscle is shown in Fig. 3.
The main element of PAM is thin membrane which makes it lightweight and it
is an advantage during replacement of a defective muscle. There are different types
of braided sleeves discovered in the past literature such as latex and silk pneumatic
Muscle [5–7]. Another advantage of PAM in their inherent complaint behavior is
when force is exerted on the PAM, which gives in without increasing the force in the
actuation.
358 S. Krishnan et al.
Wickramatunge and Leephakreeda [8] reported that the low assemble weight and
high power to ratio of PAMs is most desirable to be considered for use in mobile
robotics. In certain situation, the impotency of the PAMs is higher than the conven-
tional pneumatic cylinder. The development of a model gives effective description
to understand the mechanical behavior of the PAMs. The stiffness parameter of the
muscle denotes as K and it is considered a function of the actuated air pressure, P
and total length, LS . The elastic force is denoted by Felastic and the expression given
in Eq. 1.
K = ao P 2 + a1 P L s + a2 L 2s + a3 (2)
where ao , a1 , a2 are constant cerfficient, which can obtained from experiment data.
Typical values of coefficient in K for operating on construction mode under low
pressure:
Upper platform
Double acting
pneumatics cylinder
Lower platform
Recently there has been great interest in some research center around the world to
propose and develop automated system for ankle rehabilitation. Much of the work
focuses on the limitation of existing commercial rehabilitation devices, i.e. obtaining
combinations of basic ankle movement. Figure 4 shows the Rutgers ankle platform.
Rutgers ankle platform is a Stewart platform that generates displacement to the
ankle–foot of the patients during rehabilitation process by actuating six pneumatic
cylinders. The Stewart platform was designed to control the displacement of the
cylinder in six degrees of freedom movement of the foot–ankle.
The electro pneumatic controller is used to actuate the Stewart platform. The
electro pneumatic controller modulates the air pressure in the platform. The Rutgers
actuation is based on pneumatic system, and this has disadvantages in terms of
portability due to the heaviness of the device and sounds that is unsuitable for medical
rehabilitation environment.
3 Methodology
The Pneumatic Artificial Muscle used in this experiment is enclosed with nylon
braided sleeves. The PAM is pressurized from 1 bar up to 5 bars and diameter of PAM
used is 12 mm. The gyroscope is fixed in between the upper platform to measure the
top angle and force sensitive resistor (FSR-402) is fixed on top of the upper platform
to measure the force applied on the platform. The developed Stewart platform is
depicted in Fig. 6.
Force Sensitive
Upper platform
Gyroscope
PAM
Delrin panel
Lower platform
Force (N)
2
0
1 2 3 4 5
Pressure Supplied (Bar)
40
produced by various 12mm 10mm
diameter PAMs 30
20
10
0
1 1.5 2 2.5 3 3.5 4 4.5 5
Pressure Supplied (Bar)
362 S. Krishnan et al.
of motion of 2.21° and the maximum range of motion of 25.31°. The difference
between minimum and maximum range of motion of the platform is 23.1°. PAM with
the diameter of 10 mm obtained minimum range of motion of 2.43° and maximum
range of motion of 25.31°. The difference between maximum and minimum range
of motion is 22.88°. There is a slight increase in range of motion when the diameter
increases. The maximum and minimum range of motion obtained for 12 mm PAM
is 31.73° and it was 24.8° ROM actuated by SMA [11, 12].
5 Conclusion
In this study, it is established that the use of Stewart platform actuated by Pneumatic
Artificial Muscle improves the range of motion (ROM) for patient’s ankle–foot reha-
bilitation in biomedical applications. As the PAM is pressurized, the muscle gets
expand and shorten in length pulling the upper platform downwards. Hence, dorsi-
flexion and plantar flexion movement is achieved. Pneumatic Artificial Muscle with
the diameter of 12 mm was fabricated in the Stewart platform. The Stewart platform
actuated by Ø 12 mm diameter of PAM produced maximum platform angle and force
of 31.73° and 3.5 N, respectively. The developed Stewart platform actuated by Pneu-
matic Artificial Muscle produced range of motion up to 31.73° which is acceptable
to promote dorsi/plantar flexion.
References
10. Krishnan S, Krishna A, Nordin A, Amirtham C, Rani A, Rao T (2020) Design and fabrication
of a parallel mechanism for foot/ankle rehabilitation therapy. In: Advancement in emerging
technologies and engineering applications. Springer, pp 133–141
11. Nagarajan T, Krishnan S, Rani A, Amirtham V, Rao T, Ambaraj W et al. (2015) A novel
approach of using SMA wires actuated stewart platform for continuous passive motion (CPM)
of ankle rehabilitation therapy. In: Advanced materials research, pp 494–498
12. Abdul-Rani AM, Krishnan S, Nagarajan T, Rao T, Ramiah R, Ambaraj W (2017) Early mobi-
lization using heat treated shape memory alloy (SMA) for rehabilitation phases. In: Key
engineering materials, pp 112–116
A Review of Artificial Intelligence-Based
Techniques to Estimate Atmospheric
Parameters Influencing the Performance
of Concentrating Photovoltaic/Thermal
Systems
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 365
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_35
366 F. Masood et al.
1 Introduction
Owing to the rapid population growth and the fantastic advancement of the indus-
trial sector, energy consumption is rising globally at a much faster pace. Renewable
energy resources have emerged as a promising alternative to conventional energy
generation systems due to their abundance and free availability. They have gained
significant importance in overcoming the energy crisis [1]. Among all renewable
energy resources, solar energy has particularly drawn enormous consideration in
various engineering and industrial applications. Solar energy can be harnessed either
by employing solar photovoltaic panels, which directly convert sunlight into elec-
tricity, or by utilizing solar thermal collectors, which extract heat energy from the
solar radiation and transfer it to a working fluid, usually air or water, for further
use in thermal applications [2, 3]. The combination of photovoltaic panels, solar
thermal collectors, and reflective or refractive optical concentrators, also collectively
called concentrating photovoltaic thermal collectors, has attracted the attention of
researchers and developers in the field of solar energy due to their hybrid nature and
tremendous performance as compared to their constituent elements operating alone
[4]. The optics attached to the PV module also called the CPV system, is used to
concentrate sunlight on PV cells to generate the same amount of power with the rela-
tively lesser solar cell material, thus replacing costly solar cells with cheap optical
concentrators. In contrast, the purpose of solar thermal collector appended to the
back of PV cells, called photovoltaic/thermal system, is to remove heat from solar
cells for enhancing their efficiency on the one hand and producing thermal energy on
the other hand [5]. A CPV/T collector generates an output of hybrid nature consisting
of both electrical and thermal forms.
The CPV/T systems are usually classified on the basis of the concentration ratio of
the incident sun rays over the receiver [6]. The concentration ratio demonstrates the
extent to which sunlight is concentrated on the target. The concentration is sometimes
expressed in the form of ‘suns,’ which indicates the level up to which the solar power
is aggregated. Based on concentration ratio (CR), the CPV/T systems are grouped
into low (1 < CR < 10), medium (10 < CR < 100) and high concentration (CR > 100)
photovoltaic/thermal systems [7].
In a CPV/T system, the prediction of electrical and thermal attributes is mandatory
for the detailed assessment of operational constraints and the economic viability of
these systems to promote their market proliferation [8]. However, the modeling of
CPV/T systems is entirely dissimilar and more complicated than CPV and PVT
systems operating alone in real operating conditions due to diverse factors like their
dependence on spectral distribution, difficulty in measuring cell temperatures, and
solar flux distribution on the receiver surface [9].
A Review of Artificial Intelligence-Based Techniques … 367
The HCPV/T systems usually employ refractive optical elements to concentrate the
sunlight over solar cells. Such systems respond only to direct normal irradiance
abbreviated as DNI. Hence, this segment of sunlight is compulsory for the perfor-
mance prediction and characterization of HCPV/T structures. On the other hand, the
performance of LCPV/T systems depends on both direct and diffuse parts of solar
radiation due to their wide acceptance angles. The prediction of direct and diffuse
components of solar radiation is difficult when using analytical methods. Therefore,
many authors have employed ANNs for their prediction.
Lopez et al. [16] proposed an artificial neural network based on Bayesian function
for modeling of direct normal irradiance. The authors demonstrated that clearance
index and air mass are major parameters pertinent to the operation of the system.
368 F. Masood et al.
Alam et al. [17] used an ANN model to estimate beam solar radiation using a param-
eter called reference clearance index. The root-mean-square error for the model
fluctuates in the range of 1.66–2.78% for different states in India. In reference [18],
the authors established two ANN models for the prediction of DNI indirectly from
the clearance index. A close agreement between predicted and measured values was
observed for the Indian region. An adaptive alpha model for forecasting the hourly
global, diffuse, and direct irradiances was developed by Millet et al. [19]. However,
the results were found to be inferior compared to a feed-forward neural network.
Premalatha and Naveen [20] developed six models based on ANN using the training
algorithm, TRAINLM, and a transfer function, log-sigmoidal, to predict global solar
radiation. The authors used 32 combinations of different input parameters for predic-
tion purposes. Renno et al. [21] developed a multi-layer perceptron model for the
prediction of global radiation and DNI. Two ANN models were developed using
different parameters like climatic, astronomic, and radiometric variables. For hourly
DNI estimation, the root-mean-square error value was found to be 3.15%. Hameed
et al. [22] proposed a low-cost methodology for predicting solar insolation, a so-
called substitute to a pyranometer, based on ANN. The authors used the open-circuit
voltage and short circuit current of a solar PV module and then used these values
as input parameters to ANN to estimate the irradiance. The authors concluded that
the proposed method could produce solar irradiance values with fair approxima-
tion. Table 1 summarizes the AI base techniques used for the prediction of both
components of insolation, i.e., direct, and diffuse.
From the literature survey, it was observed that most authors employed a partic-
ular type of ANNs called feed-forward back-propagation artificial neural networks
to estimate the constituents of solar radiation like direct and diffuse components
and global irradiance. Most techniques were found to be dependent on atmospheric
parameters like temperature and humidity. Astronomic and radiometric parameters
like longitude, latitude, and zenith angle were also used as input variables. The value
of root-mean-square error for the selected papers ranges from 0.20 to 24.5%. Few
authors used a combination of ANN and GA for the prediction of solar insolation.
Rodriguez et al. [23] used the ANN model and GA to predict global solar radiation
for Spain. The predicted values had an accuracy of 99% and a root-mean-square error
of 6.74%. In another study, Rodriguez et al. [24] used a combination of ANN and
GA models to predict DNI. The improvement in estimation performance was found
to be 35%, whereas the root-mean-square error was found to be 24.29%.
by Tasadduq et al. [25]. The percent difference between forecasted and measured
temperatures was 3.17, 4.18, and 2.84 for three years. For hourly temperature predic-
tion, Abdel-Al [26] developed an alternative machine learning model. The authors
used measured hourly temperature values for a period of seven years as input to the
model. The model was used to estimate hourly temperature for the following day
and the next hour temperature having a mean absolute error of 2.01 °F and 1.05 °F
respectively. In another study, Smith et al. [27] developed an ANN model for the
prediction of ambient air temperature for the entire year by using the actual weather
data set measured by an organization, namely the Georgia Automated Environmental
Monitoring Network. The present magnitudes of ambient air temperature together
with 24 h of advance values, solar insolation, relative humidity, wind speed, and rain-
fall were used as model inputs. The model was able to predict the values showing a
mean absolute error of only 0.516 °C for a one-hour horizon. To predict the average
370 F. Masood et al.
ambient air temperature in Denizli, Turkey, Dombayci, and Golcu [28] established an
ANN model. The authors used three years of temperature data noted by the Turkish
state metrological service to train the network. The average previous day tempera-
ture, the day, and the month were used as input to the model. The root-mean-square
value was found to be 1.96 for the testing data set. To estimate temporal and spatial
air temperature, Deligiorgi et al. [29] developed various ANN models. For one hour,
two hours, and three hours advance temporal air temperature prediction, the authors
trained different ANN models using Levenberg-Marquadrt back propagation algo-
rithm. The proposed ANN model performance was found to decrease with rising
forecasting lag. The mean absolute error was less than 1.4 °C in all cases. The radial
basis function network and the multi-layer perceptron non-linear feed forward neural
networks were compared for spatial forecasting. It was concluded that both models
produced accurate results for the air temperature.
An ANN model was developed by Almonacid et al. [30] for the prediction of
air temperature on an hourly basis. The daily maximum, minimum, and average air
temperature, latitude, and longitude of a site in Spain were used as inputs to the model.
The results were compared with those produced by conventional techniques. The
root-mean-square error value lies from 0.53 to 1.98 °C. Cobaner et al. [31] suggested
applying ANN together with Multiple linear regression models and adaptive neuro-
fuzzy inference system (ANFIS) for the prediction of the largest, smallest, and mean
temperatures for a whole month, taking into account the geographical coordinates and
the month number for a given site in Turkey. The input variables include longitude,
latitude and altitude, and the month number. When compared, the results obtained
by the ANFIS model were found to be the best, with a mean absolute error of
1.25% for the mean temperature. The value of R2 was found to be 0.966 for the
average temperature. The comparison of ANN models with ANFIS for predicting
air temperatures on a monthly basis was conducted by Kisi and Shiri [32]. For training
the network, data obtained from twenty different metrological stations located in Iran
were utilized. The longitude, latitude, altitude of the site, and the time span were used
as input parameters. The performance of the ANN model was found to be superior
to the ANFIS model in the test duration.
Venkadesh et al. [33] used a GA-based approach to adjudge the duration and
resolution of prior input data for an ANN model to predict air temperature. The
GA classified contributive roles of different input variables for the estimation of air
temperature for one to twelve hours prediction horizons. The authors concluded that
ANN models established based on the GA approach produced more accurate results
as compared to the usual ANN. The mean absolute error showed an improvement
with the proposed technique.
4 Conclusions
The atmospheric parameters like solar irradiance, ambient air temperature, and solar
spectrum strongly influence the performance of CPV/T systems. A comprehensive
A Review of Artificial Intelligence-Based Techniques … 371
Acknowledgements The authors highly acknowledge the support and resources provided by
Universiti Teknologi PETRONAS, for conducting this research under YUTP scheme (Cost Centre:
015LCO-024).
References
1. Baig H, Sarmah N, Heasman KC, Mallick TK (2013) Numerical modelling and experimental
validation of a low concentrating photovoltaic system. Sol Energ Mater Sol Cells 201–219
2. Chandan DS, Kumar PS, Reddy KS, Pesala B (2020) Optical and electrical performance
investigation of truncated 3X non-imaging LCPV/T systems. En Con Mgt 113056
3. Zhang G, Wei J, Wang Z, Xie H, Xi Y, Khalid M (2019) Investigation into effects of non-
uniform irradiance and photovoltaic temperature on performances of photovoltaic/thermal
systems coupled with truncated compound parabolic concentrators. Appl Enrg 245–56
4. Al-Nimr MA, Mugdadi B (2020) A hybrid absorption/thermo-electric cooling system driven
by a concentrated photovoltaic/thermal unit. Sus En Tech Assmnts 100769
5. Kim SM, Kim JH, Kim JT (2019) Experimental study on the thermal and electrical
characteristics of an air-based photovoltaic thermal collector. Energies 2661
6. Sharaf OZ, Orhan MF (2015) CPVT solar collector systems: part I—fundamentals, design
considerations and current technologies. Ren Sus En Rev 1500–65
7. Sharaf OZ, Orhan MF (2015) CPVT solar collector systems: part II—implemented systems,
performance assessment, and future directions. Ren Sus En Rev 1566–1633
8. Diwania S, Agrawal S, Siddiqui AS, Singh S (2019) Photovoltaic–thermal (PV/T) technology:
a comprehensive review on applications and its advancement. Int J En Env Eng 33–54
9. Almonacid F, Fernandez EF, Mellit A, Kalogirou S (2017) Review of techniques based on
ANNs for the electrical characterization of cpv technology. Ren Sus En Rev 938–953
10. Shahsavar A, Moayedi H, Al-Waeli A, Sopian K, Chelvanathan P (2020) Machine learning
predictive models for optimal design of BIPV/T collectors. Int J En Rsh 5675–95
11. Qazi A, Fayaz H, Wadi A, Raj RG, Rahim NA, Khan WA (2015) The artificial neural network
for solar radiation prediction and designing solar systems: a systematic literature review. J Cnr
Prdn 1–12
12. Motahar S, Bagheri-Esfeh H (2020) Artificial neural network based assessment of grid-
connected pv/t systems in heating dominated regions of Iran. Sus En Tch Assesments
100694
372 F. Masood et al.
13. Bermejo JF, Fernandez JF, Polo FO, Marquez AC (2019) A review of the use of artificial neural
network models for energy and reliability prediction. A study of the solar PV, hydraulic and
wind energy sources. Appl Sci 1844
14. Almaktar M, Rahman HA, Hassan MY, Saeh I (2015) Artificial neural network-based photo-
voltaic module temperature estimation for tropical climate of Malaysia and its impact on
photovoltaic system energy yield. Prog Phvltaics Res Appl 302–18
15. Almonacid F, Fernandez EF, Rodrigo P, Perez-Higueras PJ, Rus-Casas C (2013) Estimating
the maximum power of a HCPV module using an ANN. Energies 165–172
16. López G, Batlles FJ, Tovar-Pescador J (2005) Selection of input parameters to model direct
solar irradiance by using artificial neural networks. Energy 1675–1684
17. Alam S, Kaushik S, Garg SN (2006) Computation of beam solar radiation at normal incidence
using artificial neural network. Ren Enrgy 1483–1491
18. Mishra A, Kaushika ND, Zhang G, Zhou J (2008) Artificial neural network model for the
estimation of direct solar radiation in the Indian zone. Int J Sus En 95–103
19. Mellit A, Eleuch H, Benghanem M, Elaoun C, Pavan AM (2010) An adaptive model for
predicting of global, direct and diffuse hourly solar irradiance. En Cnvrsion Mgt 771–782
20. Premalatha M, Naveen C (2018) Analysis of different combinations of meteorological param-
eters in predicting the horizontal global solar radiation with ANN approach: a case study. Ren
Sus En Rev 248–258
21. Renno C, Petito F, Gatto A (2015) Artificial neural network models for predicting the solar
radiation as input of a concentrating photovoltaic system. En Con Mgt 999–1012
22. Hameed WI, Sawadi BA, Al-Kamil SJ, Al-Radhi MS, Al-Yasir YIA, Saleh AL, Abd-Al-
Hameed RA (2019) Prediction of solar irradiance based on ANN. Inventions 45
23. Linares-Rodriguez A, Ruiz-Arias JA, Pozo-Vazquez D, Tovar-Pescador J (2013) An artificial
neural network ensemble model for estimating global solar radiation from Meteosat satellite
images. Energy 636–45
24. Linares-Rodriguez A, Quesada-Ruiz S, Pozo-Vazquez D, Tovar-Pescador J (2015) An evolu-
tionary artificial neural network ensemble model for estimating hourly direct normal irradiances
from Meteosat imagery. Energy 264–73
25. Tasadduq I, Rehman S, Bubshait K (2002) Application of neural networks for the prediction
of hourly mean surface temperatures in Saudi Arabia. Ren Energ 545–554
26. Abdel-Aal, Radwan E (2004) Hourly temperature forecasting using abductive networks. Eng
App Artficial Intell 543–556
27. Smith BA, Hoogenboom G, McClendon RW (2009) Artificial neural networks for automated
year-round temp prediction. Comp Electro Agric 52–61
28. Dombaycı ÖA, Gölcü M (2009) Daily means ambient temp prediction using artificial neural
network method: a case study of Turkey. Ren Energ 1158–61
29. Deligiorgi D, Philippopoulos K, Kouroupetroglou G (2013) ANN based methodologies for the
spatial and temporal estimation of air temp. In: International conference on pattern recognition
applications and methods
30. Almonacid F, Pérez-Higueras P, Rodrigo P, Hontoria L (2013) Generation of ambient
temperature hourly time series for some Spanish locations by ANNs. Ren Energ 285–91
31. Cobaner M, Citakoglu H, Kisi O, Haktanir T (2014) Estimation of mean monthly air
temperatures in Turkey. Comp Electro Agric 71–79
32. Kisi O, Shiri J (2014) Prediction of long-term monthly air temperature using geographical
inputs. Int J Climato 179–86
33. Venkadesh S, Hoogenboom G, Potter W, McClendon R (2013) A genetic algorithm to refine
input data selection for air temperature prediction using ANNs. App Soft Comp 2253–60
On the Performance of MIMO-UVLC
System over Turbulence-induced Fading
Channels
1 Introduction
Optical signals are currently in target stream of researchers due to their bandwidth
and power efficiency, least latency, and reliability compared to traditional RF and
acoustic signals. However, the transmission of optical signals underwater faces many
challenges, such as absorption, scattering, turbulence, noise and pointing errors, that
need to be resolved. Absorption and scattering of optical signals have been widely
studied and modelled. Moreover, the blue and green window of optical spectrum
offers minimal absorption and scattering [1], making it to be an optimized option
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 373
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_36
374 S. Mumtaz et al.
Moreover, for increased spectrum efficiencies, MIMO space shift keying (SSK)
technique has been widely applied to OWC systems; where instead of signal
constellations, the data bits are transmitted via antenna indices providing reduced
system complexities. Abaza [24] assessed the performance of optical SSK-MIMO
under atmospheric turbulence and observed an increase in spectral efficiency of 3
bits/s/Hz. Normally the existence of correlated signals in MIMO always reduced the
system’s performance. But Priyadarshani and Jaiswal [25] reported that the usage
of SSK-MIMO over correlated FSO channel provided a significant increased BER
performance up to 13 dB.
This paper evaluates the ergodic capacity for MIMO UVLC systems under
Lognormal, Gamma and Weibull distributed channels based upon their achieved
SNR. These models characterize the optical scintillations for a particular range of
underwater turbulence conditions. Sections 2 and 3 cast light over the employed
system and the channel models, respectively. Section 4 explains about the ergodic
capacity with MIMO technology with some light over the correlated cannels. For
enhanced MIMO spectrum efficiency, an overview about Space Shift Keying (SSK)
is provided in Sect. 5. Section 6 presents result and discussion, and Sect. 7 gives the
conclusion and the future work.
2 System Model
where s ∈ {0, 1} denotes the transmitted signal, and Pt is the transmitted power. η
represents the optical-to-electrical conversion coefficient, R is the receiver’s respon-
sivity, n(t) ∈ R M×N is the AWGN with zero mean, and N0 /2 noise power spectral
density (PSD). Moreover, H is the MIMO channel matrix comprising of row and
column vectors, and s(t) is the transmitted intensity OOK modulated optical beam.
376 S. Mumtaz et al.
3 Channel Model
For the analysis, we considered Lognormal, Gamma and Weibull distribution models
that can characterize the optical scintillations due to weak, temperature-as well as
salinity-induced turbulence, respectively.
where μx = ln(I0 )/2 and σ 2 = E I 2 /E2 {I } − 1 is the logscale parameter (first
moment) and the scintillation index (second moment) of normal distributed log-
amplitude factor respectively, with I0 as the received irradiance in absence of the
considered channel.
with a and b as shape and scale parameters. (.) is the Gamma function which
is related to the S.I by the relation b = 1/a = σ I2 . For UVLC links with very
low temperature-gradients, i.e., σ I2 ≤ 0.001, this simple distribution can be used to
characterize the fluctuations of received optical beam.
Fig. 1 Lognormal, Gamma and Weibull distributions with zero mean and unit variance for intensity
fluctuations
a
a I (a−1) I
f I (I ; a, b) = exp − (4)
b b b
with a and b as shape and scale parameters, respectively. I , a and b are all
positive valued. The parameter b is related to the S.I by the relation σ I2 =
b2 {[1 + (2/λ)] − [1 + (1/λ)]}. It is noted that for a = 1, this distribution yields
f (I ; 1, b) = (1/b) exp(−(I /b)) which is negative exponential distribution, and
for a = 2, this distribution is converted to the Rayleigh distribution. The Weibull
distribution fits well for salinity-induced turbulence fading in underwater channels.
The graph between the probability of irradiance because of the turbulence and the
optical power per unit solid angle for the considered distribution models is shown in
Fig. 1.
4 Ergodic Capacity
The utmost mutual information that can be attained by changing the PDF of transmit
signal vector is called the capacity, which is an important parameter in characterizing
the performance of a wireless communication system. The instantaneous channel
capacity is defined as
Cinst = E log2 (1 + cI ) (5)
378 S. Mumtaz et al.
where I M denotes M × M identity matrix, and H is an M × N channel matrix with
elements being Imn . Since the instantaneous SNR is a random variable (RV) with a
given PDF, so we assume the channel capacity as a RV. The average of this RV is the
ergodic capacity, and can be written as
∞
C= B log2 (1 + I ) f I (I )d I (7)
0
∞
1 (a−1) I
C Gam = B log2 (1 + I ) I exp − dI (9)
(a)ba b
0
Similarly, the ergodic capacity expression for MIMO-UVLC for Weibull distri-
bution channel can be expressed as
∞ a
a I (a−1) I
CW b = B log2 (1 + I ) exp − dI (10)
b b b
0
On the Performance of MIMO-UVLC System over … 379
Fig. 2 a Channel capacity with varying number of receivers when number of transmitters is fixed,
b Channel capacity with varying number of transmitters when number of receivers is fixed
380 S. Mumtaz et al.
Fig. 3 a Channel capacity versus received SNR based on water filling algorithm, b compar-
ison between iid and correlated MIMO UVLC channel under lognormal, Gamma, and Weibull
distributions
channel capacity. For lower SNR values less than a specific threshold, an increment
in the number of receivers does not help in providing the higher capacity and the
data rates.
Figure 3a shows the channel capacity based on water-filling algorithm, where the
optical receivers adapt themselves according to the channel conditions to attain the
desired power points. When the channel state information (CSI) is not known at the
transmitter, then the transmitter throws the power isotopically towards the receivers.
Due to path losses, some of the power goes wasted, and the channel capacity is
affected. However, in presence of CSI known to the transmitter, the transmitted
power is then optimized, and it acts as a beamforming MIMO. For known CSI,
with increase of transmitters and receivers, the channel capacity increases and thus
offers high data rate transmission. Compared to SISO systems, MIMO provides an
increased capacity of up to 11 dB, that can further be increased employing a greater
number of transceivers, but at the cost of increased system complexities.
Much of the study is done for analyzing the capacity of MIMO channels when
all the optical signals are independent and identically distributed (iid). In a scat-
tering environment, the channel impulse response (CIR) has independent fingerprints
helping the receivers to identify and differentiate among the signals. On the receiver
side, this condition can be achieved by making separation between the receivers
more than the channel correlation length. However, in some practical scenarios, the
realization of this condition is not possible, and this makes the analysis of correlated
channels more crucial and important. In less scattering environments, there exists
the correlation among the signals, and hence the performance of the link is degraded.
Figure 3b shows the ergodic capacity of Lognormal, Gamma and Weibull distribu-
tion channels with varying SNR values. The results show that Gamma channel yields
higher capacity and data rates compared to Lognormal and Weibull channels. The
results also show that in presence of correlation among signals, the capacity as well
as the data rate under all channels reduce significantly.
On the Performance of MIMO-UVLC System over … 381
Fig. 4 BER versus SNR for optical SSK-MIMO UVLC system under Lognormal, Gamma and
Weibull distributions
The BER performance of optical SSK MIMO system under Lognormal, Gamma,
and Weibull distribution channels is manifested in Fig. 4. As SSK uses only the
antenna indices for data transfer, so the system and detection complexity are reduced,
and spectrum efficiency is achieved. BER versus SNR plot shows that OSSK-MIMO
under Gamma distribution channel outperforms Lognormal and Weibull channels.
However, its performance under Lognormal channel is much better than that of
Weibull distribution.
This paper evaluates the ergodic capacity and BER performance of MIMO
underwater visible light communication (UVLC) under Lognormal, Gamma, and
Weibull distributions which can characterize the optical scintillations due to weak,
temperature-induced, and salinity-induced turbulence fading, respectively. More
transceivers can transmit and receive more power, and that can be combined using
MRC and EGC techniques to produce high SNR value, and hence that of the channel
capacity. To optimize the transmit power, CSI is required at the transmitter, because an
unknown CSI yields a reduced amount of channel capacity. Moreover, it is observed
that in presence of correlation among optical signals at the receivers, the ergodic
capacity for all channels reduces. The results also show that Gamma channel provides
higher capacity and data rates; however, it can model only low temperature gradient
382 S. Mumtaz et al.
turbulence induced underwater fading. The results also show that the performance
of SSK-MIMO under Gamma distribution channels is much better than those of
Lognormal and Weibull channels.
The mixture distribution models, named Exponential and Lognormal, Exponen-
tial and Gamma, Exponential and Weibull, Exponential and Generalized Gamma
(EGG), are also found to fit well with the experimental data for different under-
water turbulence conditions. The performance and capacity analysis of MIMO as
well as MIMO-OFDM UVLC systems under such channels with suitable coding
and modulation techniques are still open research issues in such regimes.
References
1 Introduction
Paralysis is a common term used to define the defective sensation due to following
damage to the nervous system. In medical terms accurate level of injuries are used
which is supportive in determining the parts of body that affected by paralysis and also
loss of function. Generally, this disease can be within a temporary period or even worst
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 385
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_37
386 J. Duwaraka et al.
to be permanent. In common cases, paralysis affects half side of human body, called
as unilateral. Under classification of paralysis, there is a particular medical condition
defined as Paraplegia [4]. Paraplegia is defect affecting the lower extremities of
human body. This term describe the paralysis that affects both legs and partially the
trunk, but not the arms. Contracting the study of Paraplegia, there is another form
of paralysis that affects just one limb known as Monoplegia. Crucial Monoplegia is
scientific term for paralysis affecting the single lower limb [5]. Patients with these
disabilities have trouble in walking as well as reflecting their muscle. Conversely,
rapid development of technology has formed a solution with a design mechanism of
exoskeleton in conjunction of instrumented cane for the control.
Development in this field creates an enthusiasm towards patients to experience the
transaction from undergoing rehabilitation therapy to a “real world” walking prac-
tice. Human motion intention depends on certain bioelectrical signal of neurology
system. Because of variation on their bioelectrical difficulties in mobility. Research
community developed several wearable gait, exoskeleton, instrumented cane that
will compute and estimate the motion intention by disable. Nevertheless, there are
several constrains on the design mechanism and followings are:
1. Modelling with variable electronic hardware modules are composite.
2. In data transmission power supply for wireless is ineffectual.
3. Immense design and additional weight restrain, uncomfortable cane.
Movement transparency is critical when wearing a robot for gait rehabilitation.
In other words: When wearing the exoskeleton, its movement should be synchro-
nized and consistent with a patient’s natural movement. “If not, it exerts extra forces
on the human joint “And this extra force causes patient discomfort and unnatural
movements.”
1.2 Objectives
The expected result of this project is to develop an instrumented cane that able
to communicate and control a lower limb knee exoskeleton together with control
algorithm that able to notice human motion trajectory. This work is to contribute in
the use of rehabilitation for remote assessments as well as for the disable with their
day-to-day activities. Throughout this project, research will be performed on the
required balance in motion, ability, upper body strength, and the use of arm strength
with the instrumented cane. Besides, the scope of this project is also to concentrate
on the wireless communication system to assist in real-time control of exoskeleton
via a standard I2C serial communication protocol.
IMU sensor was attached to the instrumented cane to detect the movement trajectory
that is discharged. The output signal obtained from the sensor will be illuminated
using a microcontroller that is compatible with the sensor and wireless communi-
cation module. The data will then be mediate to the microcontroller used at the
exoskeleton in order to control the trajectory based on, the information is carried
via Bluetooth communication sensor. Instrumented cane and exoskeleton enhance
the compatibility using Bluetooth communication protocol as the best option. As the
information is conveyed to the exoskeleton at real-time, it begins to initiate the motor
to act as actuator in order to make motion. Fundamentally, humans target in motion
triggers first, in sequence to command the robotic structure make motion. The push
button will be assigned as a safety circumstance if there is any sudden detach or drop
of cane from the arm, it will not send any trajectory control data to the Exoskeleton.
This prevents the exoskeleton to make any false motion trajectory which could cause
weakness. In related parts for Exoskeleton section, the Microcontroller, Controller,
and Motor subsystem will be continuation from previous project.
388 J. Duwaraka et al.
3 Methodology
After a thorough study on several research papers, there are several different types
of instrumented cane with different design mechanism in order to control the trajec-
tory of exoskeleton. Two major factors is considered in this project, first the sensor
for angle reflective properties and second on the communication medium. Real-
time assessments on the motion intension can be establish using rotational attributes
computing sensing technology integrated with radio-communication technology in
order to control the trajectory of exoskeleton. Wireless communication medium plays
an important role as exoskeleton response based on acquired real-time data signal
that receive from the cane (Fig. 1).
4 Hardware Implementation
The chip uses I2C (inter integrated circuit) protocol for serial communication to
function according to various body condition. The weight of sensor is 5 g which is
helpful in reducing the weight of the prototype. Arduino nano,interpret information
from MPU6050 and sends through I2C protocol to Bluetooth HC05.The board is
user friendly and smaller in size. The power requirement is 5 V. The Bluetooth
module HC05 act as the transmitter and has the capacity to act as both master and
slave according to the preference. In this work it performs as master and can be
paired with any slave device automatically. The consumption of power is low and
the average current fluctuation range of 25 mA.
The communication range of distance required is 1–3 m and it is applicable for
HC05 module. The HC06 bluetooth module in exoskeleton executes as the receiver
and performs as slave. It weighs 0.9 g and suitable for wireless communication. The
exoskeleton consist of controller boards Arduino UNO and MEGA. UNO act as the
intermediate and act as middle device and information from Bluetooth receiver and
convert into digital signals and relay it to Arduino MEGA. The coded program in the
mega rotates the servo/cylindrical motor. The controller circuit controls the torque
and provides the flexion and extension of the exoskeleton.
The design and development of electromechanical system for instrumented cane
was done by modifying the mechanical design of a regular cane and instrumented
with respecting sensor and microcontroller boards. The main objective is to able to
modify an off-the-shelf cane such that it can house position, orientation and force
sensor together with wireless microcontrollers to interpret and send data in order
to control the exoskeleton trajectory. Based on the design of instrumented cane as
illustrates in Fig. 2, additional weight factor on the cane was took into consideration
in order to achieve one of the targeted objective. Execution on this project believed
to deliverable an instrumented cane that able to sense the motion trajectory either
forward or backward by the user and sends the real-time data in order to initiate
trajectory of exoskeleton. This instrumented cane will be developing strong spirit to
experience “real world walking” especially for those who suffering in Cerebrovas-
cular Accidents (CVA) and Spinal Cord Injury (SPI). Critical consideration on the
design specification whereby the instrument cane should be wireless, portable, light
in weight as well as meets clinical requirements so that this cane is use comfortably.
In conclusion, this instrument cane technology target to assist disability patients as
all of the objectives is achievable.
5 Conclusion
IMU sensor was attached to the instrumented cane to detect the movement trajectory
that is discharged. The output signal obtained from the sensor will be illuminated
using a microcontroller that is compatible with the sensor and wireless communi-
cation module. The data will then be mediate to the microcontroller used at the
exoskeleton in order to control the trajectory based on, the information is carried via
390 J. Duwaraka et al.
References
18. Zhang J (2016) Towards systematic controller design in rehabilitation robot, August 2016.
Proquest number 10308368
19. Bing C (2017) Design and implementation of a lower extremity exoskeleton for motion
assistance to paraplegic patients
20. Jhawar V (2018) Design of knee exoskeletongait assistance, August 2018. Proquest number
10846724
Semi-Automatic Detection
and Measurement of Fetal Parameters
from Ultrasound Images and the Scope
Automatic System Using LabVIEW
S. Prabakar (B)
Sona College of Technology, Salem, India
K. Porkumaran
Sri Sairam Engineering College, Chennai, India
e-mail: [email protected]
J. Samson Isaac
Biomedical Engineering, Karunya University, Coimbatore, India
R. Karthikeyan
Dr NGP Institute of Technology, Coimbatore, India
G. Gopu
ECE, Sri Ramakrishna Engineering College, Coimbatore, India
R. Kannan
EEE, Universiti Teknologi Petronas, Seri Iskandar, Malaysia
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 393
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_38
394 S. Prabakar et al.
consumes 20–30 min for a single subject scanning whereas in case of these proposed
techniques the time taken for automatic detection is less than 5 min based on the
system used and the manual intervention also be reduced in diagnostic procedures.
1 Introduction
Accurate fetal ultrasound measurements are one of the most important factors for
high quality obstetrics health care. Common fetal ultrasound measurements include
Biparietal diameter (BPD), Head circumference (HC), Abdominal circumference
(AC), Crown rump length (CRL), Femur length (FL), Humerus length (HL). These
measures are used to estimate the getational age(GA) of the fetus and are considered
as vital parameters in an important diagnostic tool to identify the viability of the
fetus.
In clinical settings, the manual measurement by specialists of BPD, HC, AC,
CRL, FL, HL present the following issues such as the quality of measurements are
user dependent, the exam can take more than 30 min and specialists can suffer from
Repetitive Stress Injury (RSI) due to these lengthy exams. Therefore, the automatic
system of these ultrasound measures has the potential of improving productivity
and patient throughput, enhancing accuracy and consistency of measurements, and
reducing the risk of RSI to specialists.
Table 1 Details of
S. No Trimester scans Fetal parameters
parameters for all trimesters
1 First trimester Crown Rump Length
2 Second trimester BPD, HC, AC, HL and FL
3 Third trimester BPD, HC, AC, FL an HL
The detailed study of literature in the area of fetal parameters measurement and
medical image processing techniques and algorithms, especially the problem asso-
ciated with the automatic segmentation and feature extraction for clear visualization
issues.
The points inferred from results of literature review are:
• Loss of traits and application in outcomes of the measurement systems
• Fetal anatomy measurement requires manual intervention
• Long computing time for various parameter measurements.
This procedure is followed for all the six parameters and the parameters measured
during each trimester are enlisted below (Table 1).
The proposed system for the measurement of fetal parameters such as BPD, AC, HL,
FL, HC and CRL is based on two processes namely semi-automatic and automatic
process which is performed with the help of the Image processing criterion in the
LabVIEW software known as VISION ASSISTANT.
The semi-automatic process is wherein the parameters are segregated into six
SUB VI’s and are programmed into one single MAIN VI. The trimester fetal scan
image is loaded and is converted into 1000 × 1000 resolution. This is then followed
by manually dragging the region of interest ROI which could be either circle or line
annotation or the values of the result is obtained.
The automatic process is being dealt only for four parameters namely AC, HC,
HL and FL. The trimester fetal scan image is loaded and by applying certain
threshold levels and performing convolution as well as morphological operations
using different icons such as contour extract, the search direction and the contour
selection are used to locate the images from left to right, top to bottom, maximum
length, etc. for automatic process. Here, it selects the appropriate ROI and the values
of the result are obtained.
396 S. Prabakar et al.
converted into a mask. A mask is nothing but a group of pixels with the programmer
specified value in it. Here the value is 255. The image is converted into an array with
the array elements corresponding to the pixel’s colour value/intensity. The output
will be a 2-D array of rows and columns pixel colour values of the input image.
Here, it is in the mask or other words, number of pixels in the ROI drawn by the user.
A function called unbundle by name is used. We use the element of global rect-
angle. The output from the global rectangle will be an array of 4 numeric elements
namely; ROI’s left, right, top and bottom coordinates respectively. Certain expres-
sions like negation, addition, subtraction, division, etc. are used. The DBL converts
a double-precision number to floating point number. The linear fit VI is used This
VI uses the iterative general Least Square method and the Levenberg–Marquardt
method to fit experimental data to a straight line of the general form described by
the following equation:
f = ax + b (1)
where x is the input sequence X, a is slope, and b is intercept. The expression node is
used to calculate the expressions that contain a single variable such as acosh, etc. The
calibration factor of 1000 × 1000 is done and multiplied, the sine of the expression
node is then divided to get the result. This is then linked with the main VI.
The trimester fetal scan image is loaded and by applying certain threshold levels and
performing convolution as well as morphological operations are used to locate the
images from left to right, top to bottom, maximum length, etc., for automatic system.
Here, it selects the appropriate ROI and the values of the result are obtained.
The image is created using IMAQ CREATE and IMAQ READ and a specific memory
is given to the image. This is then converted into an array to work with each pixel.
This is enclosed in a nested for loop, in which the outer loop runs according to the
number of rows in the pixels of the array and the inner loop runs for the number
of columns. A threshold is set inside the loop to convert grayscale image to binary
image.
Then the particles are removed from the image for proper segmentation of the
appropriate region from the surrounding particles. If there are large numbers of
particles it would be difficult to draw the region of interest ROI. The number of
398 S. Prabakar et al.
particles or pixels which has been removed from the image is identified with the help
of IMAQ PARTICLE ANALYSIS report. A filter of the nth order is used in which
the X axis coordinate is −1 and Y axis coordinate is 5. For every value in the X axis,
it is multiplied with 5 to define the organ of measurement as they are stronger in
densities. This filtering operation is done only for giving intensity for the appropriate
region.
Morphological operations such as dilation, erosion, opening and closing are
performed in the vision assistant tool of LabVIEW. Dilation is a transformation
that produces an image that is as the same shape as the original, but of different
size. Dilation stretches or shrinks the original figure. It increases the valleys and
enlarges the width of the maximum regions. Erosion is used to reduce the objects
in the image and known that erosion reduces the peaks and enlarges the widths of
minimum regions. Opening of an image is erosion followed by dilation with the same
structuring element and closing operation is the reverse of opening operation.
The forthcoming operation is performed to select the ROI automatically. Hence,
we use the IMAQ Extract Contour icon which extracts a single, best contour from an
image. If the input image is calibrated, the contour will be learned with calibration.
The main VI is the composition of all the six SUB-VI’s. The user or the sonographer
can select the type of trimester and perform the semi-automatic measurement. On
selecting the appropriate trimester, the parameters that are measured only remains
enabled while the other parameters are disabled. The main VI is displayed in Fig. 2.
The ultrasound scan image is loaded in the file path and the window shows two
sets of images. The first image shows the original image and the second image shows
the conversion of the original image to 1000 × 1000 resolution as shown in Fig. 3.
This is set as the standard resolution for all parameters for greater throughput and
accuracy.
In this semi-automatic process, the region of interest is drawn manually either
by using circular annotation or line annotation. The region of interest ROI of BPD,
HC and AC are drawn using circular annotation while the ROI for CRL, HL and
FL are drawn using line annotation. This ROI depends on the calibration factor. The
calibration factor for 1 pixel in the 1000 × 1000 resolution image is calculated as
0.016129 cm. The image is then converted back to array using Build array icon. This
is followed by the important mode where the values are then converted from degrees
to radians which are followed by other mathematical operations and the required
value for the result is obtained. The output of the ROI drawn manually is shown in
Fig. 4.
Semi-Automatic Detection and Measurement of Fetal Parameters … 399
6 Conclusion
Thus, the fetal parameters are measured manually and are similar to the process
carried out by the sonographers and also LabVIEW being highly effective software
with very minimal time duration of 5 min per parameter. The manual intervention
drawbacks in this work is analyzed and hence to overcome these we proceed with
400 S. Prabakar et al.
References
Abstract Internet of Things (IoT) is one of the current technology that gains the
highest acceptance from various industries and academia nowadays. The growth
of this technology has streamlined the day-to-day activities of individuals. It can
be stated that day-to-day activities in the industries without the aid of the IoT is
becoming harder and harder due to the manual process. This study investigates the
issues of energy consumption in the IoT environment due to denial of service (DoS)
attack in the IoT environment. Besides, a solution that remedies the issues is proposed
in this study. The contributions of this study are categorized into two aspects. Firstly,
the study described the components of the IoT, which includes the sensor nodes.
Secondly, the study emphasizes in improving the energy of the nodes to enhance
the network lifetime in the IoT environment. To achieve the milestone the study was
implemented using MATLAB, and thereby considering the enabling technology such
as wireless sensor network (WSN) and the RFID in the perception layer of IoT. The
results reveal that the network lifetime has improved significantly with about 6%
improvements as compared with security-based modified LEACH (MS-LEACH).
1 Introduction
Internet of things (IoT) is a collection of various objects (devices) that are intercon-
nected to perform a specific task without human endeavor. The connected devices
share data and information in order to achieve the intended goals in IoT application
areas [1]. Besides, IoT has been implemented in many application areas such as oil
and gas, transportation, health care, homes, education, and sports with the aim to
transform the industries and individual’s activities [2].
The widespread of using various technologies such as wireless sensor networks
(WSN) and smart devices that are equipped with several sensors, such as radio
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 401
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_39
402 A. A. Ghali et al.
The following sections will highlight the application area of IoT and its challenges.
These applications areas include transportation, health care, education, homes, sports,
and automobile. Table 1 summarizes the security challenges in the IoT environment
[6, 7].
Based on Table 1 the security challenges of the IoT application areas are summa-
rized with their security challenges and proposed recommendations. Thus, the secu-
rity challenges may caused the degradation of the network performance and commu-
nication denial between the sensor nodes and CHs. Hence, the study provides
A Framework for Enhancing Network Lifetime … 403
Table 1 Summary of the security challenges in the IoT application area of IoT and their solutions
Application area of IoT Security challenges Proposed solutions
Transportation DoS attack Authentication mechanism
Health care DoS and Zingbox Confidentiality and
authentication mechanism
Education DoS attack Authentication and
availability mechanism
Homes DoS and Replay attack Integrity and confidentiality
mechanism
Sports DoS and Snipping attack Communication security
Automobile Man in the middle and DoS attack Mutual authentication and
authorization technique
a framework that will improve the efficiency of the network lifetime in the IoT
environment.
3 Methodology
In this section the study presents the method on how clustering formation is formed.
Therefore, Fig. 1 displays the clustering formation method.
Initially the function name and number of the nodes are required to compute
the clustering formation arrangement. Next, the nodes are distributed based on the
number of scale design. After the nodes distribution, the CH is assigned. The calcu-
lation of the clustering rate is obtained to provide the adequate distribution between
the cluster and the CH (For further details see Sect. 5). The output of the formation
along with the CHs are displayed.
Based on the method presented, the result of the proposed method for enhancing the
network lifetime in the IoT environment is shown in Table 2. The illustration of the
figures and tables are displayed. Table 2 and Fig. 2 shows the nodes distributions.
404 A. A. Ghali et al.
As shown in Fig. 2, the initial field distributions of the nodes are distributed based
on 50 m ✕ 50 m. The nodes are randomly distributed in the clusters, before the CHs
are selected, the nodes are randomly installed into the scales waiting for the CHs
selection. The CHs can be placed at the center or the edge of the field scale but it
depends on the nodes selected to be the CH. The insertion of the CH in the center or
the edge is convenient to assist the node to finds its transmission range to CH. Table
3 and Fig. 3 shows the advertisement phase.
In Fig. 3 the nodes with the red colors are the two CHs. The CH sent an adver-
tising message to the BS for communication after a few rounds. The nodes usually
transfer their energy to CH in the initial data transmission rounds. Thus, the data are
aggregated and sent to the BS. Most importantly, the CHs uses their energy faster
than non-CH for data transmission. Table 4 and Fig. 4 display clustering formation.
Table 3 Advertisement
No Items Description parameter
phase
1 Simulation area 350 ✕ 500
2 No. of points 350
3 Radius meters 50 by 50
4 Field area X by Y in meters
5 Estimated no. of clusters 2
A Framework for Enhancing Network Lifetime … 405
Table 4 Clustering
No Items Description parameter
formation
1 Simulation area 350 ✕ 500
2 No. of points 350
3 Radius meters 50 by 50
4 Field area X by Y in meters
5 Estimated no. of clusters 2
Cluster Head 9
Cluster Head 3
Figure 4 indicates the clustering formation for the nodes in the cluster. In this
figure, the formation are divided into two which are brown and blue. The brown color
indicates the clustering formation with a circle node representing its cluster group
as the CH. Therefore, having the formation in the cluster will reduce the number
of energy consumption from the nodes significantly. The more energy consump-
tion is reduce the more network lifetime increased. Figure 5 indicates the energy
consumption per nodes in the cluster.
406 A. A. Ghali et al.
From Fig. 5, it is evident that the energy dissipation by the nodes in the cluster
is minimal. Among all the nodes, node nine (9), which served as the CH consumed
the highest amount of energy since all the remaining nodes in the cluster transmit
their data to the cluster head. The next highest node is (3), which is also the CH. This
signifies that the other nodes transmit their data to the CHs and this allow them to
have a rest in order to save their energy and by reserving their energy the network
performance increase significantly.
This study investigated the issues of energy consumption in the IoT environment
and further provides solutions that remedy the challenges. Besides, the IoT security
challenges specifically the DoS attacks can be addressed from the early stage of
the IoT implementation. The contributions of this study are categorized into two
aspects. Firstly, the study described the initial components of the IoT, which includes
the sensor nodes. Secondly, the study provides a solution that improve the network
lifetime in the IoT environment using a method clustering formation. The simulation
results revealed that network lifetime is increase with 6% improvement as compared
MS-LEACH.
Acknowledgements The research was fully funded by the Center of Graduate Studies (CGS).
Author completely acknowledges the financial support given by Universiti Teknologi PETRONAS
(UTP) for this study.
A Framework for Enhancing Network Lifetime … 407
References
1. Mahmoud R, Yousuf T, Aloul F, Zualkernan I (2015) Internet of things (IoT) security: current
status, challenges and prospective measures. In: 2015 10th international conference for internet
technology and secured transactions (ICITST). IEEE, pp 336–341
2. Asghari P, Rahmani AM, Javadi HHS (2019) Internet of Things applications: a systematic review.
Comput Netw 148:241–261
3. Dehkordi SA, Farajzadeh K, Rezazadeh J, Farahbakhsh R, Sandrasegaran K, Dehkordi MA
(2020) A survey on data aggregation techniques in IoT sensor networks. Wirel Netw 26(2):1243–
1263
4. Ghali AA, Ahmad R, Alhussian H (2021) A framework for mitigating DDoS and DOS attacks
in IoT environment using hybrid approach. Electronics 10(11): 1282
5. Xia Y, Li W, Clark W, Hart D, Zhuang Q, Zhang Z (2019) Entangled radiofrequency-photonic
sensor network. arXiv preprint arXiv:1910.08825
6. Ghali AA, Ahmad R, Alhussian HSA (2020) Comparative analysis of DoS and DDoS attacks in
Internet of Things environment. In: 9th computer science on-line conference. Springer, Cham,
pp 183–194
7. Kouicem DE, Bouabdallah A, Lakhlef H (2018) Internet of things security: a top-down survey.
Comput Netw 141:199–221
An Empirical Analysis on Big Analytics
for e-Healthcare and Agriculture
Abstract There is a lot being said and done in the field of data analytics. Using large
amounts of data for analytics has become one of the rising trends in the business
world but, implementing this business intelligence into different sectors of govern-
ment hasn’t still progressed well. We have discussed two major applications of data
analytics in government sectors where the government and eventually the citizens
could benefit from all the available big data. The applications include (i) Agriculture,
where the big data analytics could result into better crop planning, yield analysis,
improved soil health and irrigation as well as reduce the support cost incurred. (ii) The
section on data analytics in healthcare mainly points out the importance of predic-
tive analytics in improving personalized healthcare and healthcare infrastructure as a
whole. It also talks about how the government can unlock value through big data and
machine learning to provide better health insurance than the existing ones and how
data analytics is helping with fraud detection while providing the health insurances.
1 Introduction
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 409
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_40
410 N. Purandhar and S. Ayyasamy
of data, analyzing and interpretation of the data and reporting. Monetizing this data,
i.e. creation of data to wealth is a challenge as well as an opportunity for the future.
For different objectives different types of analytics is to be carried out. For which we
need to classify data analytics.
The IBM organization in 2013 categorized analytics into three important ways
like descriptive, predictive and prescriptive analytics. To which in 2015 diagnostic
analytics was added.
A. Descriptive Analytics: Data mining techniques are used along with business
intelligence to inquire “What has happened?” That means it analyses the histor-
ical data for insights and uses that data to determine the future [2]. Though it
provides insights into historical data and trends, this technique is most time-
intensive and produces the least value of all the types. Therefore, descriptive
analytics alone is not used very often. But, combining it with any other form
of analytics can be practical and productive.
B. Diagnostic Analytics: It is a form of sophisticated analytics which examines
data or information to give solutions for the question “Why did it happen?”. It
uses historical data with other data to identify patterns and find dependencies
[3]. Diagnostic analytics is used for root cause analysis of a problem.
C. Predictive Analytics: Statistical models are used to forecast “What could
happen?” [2]. It uses the results of descriptive and diagnostic analytics to find
out patterns, clusters and outliers to estimate future trends. Predictive analytics
is widely used because of its numerous advantages but, it still depends largely
on the data quality and therefore it requires continuous optimization.
D. Prescriptive Analytics: Methods for optimization and simulations are used
to ponder “What should we do?” [2]. This is the action-oriented type and
focuses on taking any endeavors possible to deal with the problem at hand.
This advanced type of data analytics needs historical data coupled with infor-
mation from external sources for applying statistical algorithms. Other than that,
prescriptive analytics uses machine learning as well as business intelligence,
which makes it expensive to implement and manage [4].
On Feb 1st 2018, while delivering his speech on the Union Budget, the Finance
Minister told the parliament that the government’s think tank—Niti Aayog, would
initiate a national program helping the government to focus on the development of
cutting edge technologies in digital space namely Big Data, Artificial Intelligence
(AI), Machine Learning, Internet of Things (IoT), 3D printing and others. Further
accentuating to invest in research, training and skill development in robotics, AI,
manufacturing, big data analysis and IoT, he announced to double the budget for
Digital India Programme to 3073 crores in 2018–19 [5]. This shows that the govern-
ment of India is keen on using the technological advancements in different fields for
good.
An Empirical Analysis on Big Analytics for e-Healthcare … 411
Farming is the cornerstone of India’s wealth and this agriculture is responsible for
18% of India’s gross domestic product (GDP) and facilitates employment to more
than 50% of the countries’ manpower but, because of rapid development of the
industrial and service sectors the agricultural industry has declined to a great extent.
For improving the yield of small farmers, the government of India had an agreement
with IBM to develop an Artificial Intelligence based model for crop yield prediction
[6]. When IBM will employ Artificial Intelligence for developing the technological
model for improvement of agricultural output and productivity of various types of
crops and soil species for the specified districts, NITI will use this data insights
generated through these AI models to help farmers and other stakeholders [7]. Also,
by providing soil health cards the government has empowered the farmers to better
analyze the requirements of the water for irrigation and the crop to be sown.
To add value chain and increase profitability of the agricultural industry the
government is keen on employing technology. Also, for creating opportunities in
the field, the Ministry of Agriculture and Farmers ”Welfare had a deal signed with
the Indian Space Research organization (ISRO) to use satellites for geotagging the
agricultural assets [8]. The big data acquired through such agricultural infrastructure
will help develop vigorous responses for the issues pertaining to climate change, pest
and disease outbreaks, and land degradation. Applying these “big data” approaches
by the government for agriculture guarantees to search out new ways to build a
food-secure society.
Big data is seen as a prospective technology for selection of appropriate agri-
cultural inputs but, acquiring data is the biggest challenge. For any agribusiness
increasing the profitability means increased yield and decreased support costs. Big
data aids in enhanced yield analysis. Increased yield and decreased support costs
play principal role to gaining profitability and enhanced customer experience for
any small or large agribusiness. Big data analytics can enable improved analysis of
crop yield and other crucial methods for a profitable and in-depth root-cause analysis
resulting in actions for enhanced crop quality and reduced overall cost [9]. The data
associated with turnover, utilization of available resources and the effectiveness of
equipment, can be combined with further analysis for improved quality.
Predictive analytics may be used to foresee the demand for seeds, requirement of
fertilizers and animal feed needs, for empowering the agribusiness supplier to take
appropriate steps to match production to demand. Table 1 shows different fields of
agriculture and how big data is used for their management.
412 N. Purandhar and S. Ayyasamy
Table 1 (continued)
Focus area Ref Proposed Possible improvements
techniques/solutions
Reduced support cost Many small and marginal To reduce cost of entire
farmers are utilizing agriculture value-chain the
M-trading platforms to sell farmers should be enabled
their farm produces right with data centric methods for
away to businesses across improving the efficiency of
the country through smart supply-chain and the
phones marketplace of agrarian items
Yield analysis [15] Different data mining With the soil health card, the
[16] techniques like linear and farmers can know the nutrient
multiple regression, contents of the soil better and
[17] clustering, decision tree and hence, get an idea of which
k-nearest neighbors have crops to invest in and what
been used with the big kind of fertilizers they need
datasets for finding the yield ultimately, increasing the
of the crop crop yield [11]. IBM”s AI
based model for crop yield
prediction is also in use by
small farmers
Crop protection and weed [18] Big data as well as machine There are no crop protection
control learning approaches have initiatives by the government
been suggested for available as yet
modeling and prediction of
herbicide/pesticide
resistance, detection of
non-native species plants
and flowers, decision
support system for crop
protection and using AI
based systems for plant and
flower control
A real-time alerting system for instant care through wearable devices is fast growing
trend in the first world countries [19]. The world is rapidly moving from tradi-
tional medical practices to wearable medical devices but, India is still struggling for
providing primary healthcare to majority of population. Despite of the government
recommendations for standards of Electronic Medical Records (EMR), there are
no concrete steps taken for centralizing EMRs. The Ministry of Health and Family
Welfare suggests a standard for structured record fields to be inserted into the database
viz., Name of the patient, DoB, Residential addresses, Doctor’s name, Hospital name
and addresses, and medical treatment reimbursement codes for easier management
through a database system but, the lack of acceptance to field-code the data and
use of EMRs by the medical practitioners and caregivers is a major barrier [20].
414 N. Purandhar and S. Ayyasamy
Moreover, the traditional databases are inadequate to handle the large chunks of data
being produced every day. There are three dimensions to the role of government in
Healthcare in India.
The major shortcoming of the government is its lack of insight for providing devel-
opment in personalized healthcare sector. The main focus needs to be on research and
development models for epidemiology research clubbed with big data. Big data can
help to identify problem areas where there is a need for improvement in personalized
healthcare. It can also be useful for statistically analyzing the course of treatment
employing the historical data. Whether it be biomedical sciences or pharmaceutical
researches data is everywhere. One-way Predictive analytics can help hospitals is
anticipate the number of incoming patients at different times during the day, week,
month and year. This can solve the difficulty caused by over-staffing or under-staffing
in hospitals because in the first scenario the cost of labor increases and in the latter the
patient services might be hampered. The other area where the predictive analytics can
help is by using data for identifying the cause of different diseases through available
data.
Providing healthcare at affordable prices is one of the main objectives of the govern-
ment in India. Under the initiative of Ayushman Bharat, the government of India
is firm on contributing to better healthcare for all. The Pradhan Mantri Jan Arogya
Yojana under Ayushman Bharat aims to cater to more than 50 crore beneficiaries.
It covers not less than 10.74 crore underprivileged families from rural areas and
workers’ families from urban and suburban parts of the country as per 2011 Socio-
Economic Caste Census (SECC) data [23]. It has a provision of health cover of
rupees 5 lakhs for secondary and tertiary healthcare services. This scheme has a lot
to offer to private hospitals, insurance companies as well as the main stakeholders—
the beneficiaries. According to media reports of 13th Feb, 2019, there have been
discrepancies in some of the procedures, and a few hospitals receiving a greater
number of patients than the other. Therefore, the government has employed analytics
methods to ensure Ayushman Bharat remains without any scams. The analytics is
also necessary for detecting fraudulent behavior as the Aadhar card—the 12-digit
unique identity number were not mandatory for availing the services and any identity
cards were accepted in the initial stages. Utilizing the research evidence Recommen-
dation engines can be built based on collaborative filtering for creating healthcare
plans for patients with similar symptoms. These recommendation engines use best-fit
classification algorithms to classify patients based on different parameters like, age,
disease type, symptoms and seasonal changes among many. This would help to build
a better prediction model for precision medicine. The most commonly used strategies
for predictive analytics include Bayesian networks, Decision Tree learning, Markov
and Monte Carlo Simulations [24]. In the paper [25], the authors suggested a combi-
nation of Principal Component Analysis and Neural Network for sentiment analysis.
They had promising results when checked with different performance parameters.
A feedback mechanism can be developed, where the patient or the caregivers can
provide feedback on different parameters relating to the overall experience at the
hospital. A mechanism can be implemented by applying sentiment analysis on the
feedbacks, where (1) this analysis can be provided to the recommender systems for
personalized healthcare plans, (2) the hospital can utilize the data to improve the
infrastructure and (3) it can help to improve the efficiency of the nursing team by
providing useful insights.
416 N. Purandhar and S. Ayyasamy
3 Conclusion
The applications discussed in this paper specify the potential of data analytics in their
respective fields and the possible solutions using different methods of data analytics.
It presents the sources from which the data is acquired for various applications, and
how this data can help to improve the citizen services provided by the government.
In Agriculture domain the government does the soil analysis of each farm and gives
suggestions to the farmers for crop planning, soil health and yield analysis. But this is
not enough for entire growth of the agrarian market and therefore, the paper discusses
different big data analytics methods and machine learning algorithms for an inclusive
agri-business setup. Healthcare is the sector where data analytics could do wonders
with personalized medical plans, for infrastructure development and patient service
improvement. Recommendation engines could be built for personalized medicine
based on EMR’s. But, first there is a need to develop and deploy centralized electronic
health records systems. Once these systems are in place the data analysis could be
done to increase efficiency and save lives.
References
1. Sondergaard P (2019) Information is the oil of the 21st century, and analytics is the combustion
engine. Gartner Research, Datasciencecentral.com
2. Delen D, Demirkan H (2013) Data, information and analytics as services. Decis Supp Syst
55(1):359–363. Available https://doi.org/10.1016/j.dss.2012.05.044
3. Fleckenstein M, Fellows L (2018) Modern data strategy, 1st ed. Springer, Cham, Switzerland,
p 133. ISBN 9783319689937
4. 4 types of data analytics to improve decision-making. Scnsoft.com, 2019. [Online]. Available
https://www.scnsoft.com/blog/4-types-of-data-analytics
5. Ministry of Finance (2018) Budget speech 2018–19
6. Misal D (2019) How big data is the game changer for indian government in E-governance.
Analytics India Magazine
7. D’Monte L (2019) To improve weather forecasting for farmers in India, IBM is relying on AI.
https://www.livemint.com
8. Deoras S (2019) Top use cases where modi government used big data, AI for reform. Analytics
India Magazine
9. (2016) Improving manufacturing performance with big data. Oracle Enterprise Architecture
White Paper
10. Lesueur D, Burra D, Bui D, Nguyen M, Zhong D (2019) Geocoded soil data and AI to provide
an estimate and indicator for soil health. CGIAR Platform for Big Data in Agriculture
11. Soil Health Card|National Portal of India. India.gov.in, 2015
12. Rajeswari S, Suthendran K (2019) C5.0: advanced decision tree (ADT) classification model
for agricultural data analysis on cloud. Comput Electron Agric 156:530–539
13. Rajeswari S, Suthendran K (2018) Chi-square mapreduce model for agricultural data. J Cyber
Sec Mob 7(1):13–24
14. Sekhar C, Kumar J, Kumar B, Sekhar C (2018) Effective use of big data analytics in crop
planning to increase agriculture production in India. Int J Adv Sci Technol 113:31–40
15. Sellam V, Poovammal E (2016) Prediction of crop yield using regression analysis. Indian J Sci
Technol 9(38)
An Empirical Analysis on Big Analytics for e-Healthcare … 417
16. Fan W, Chong C, Xiaoling G, Hua Y (2015) Prediction of crop yield using Big Data. In: 8th
international symposium on computational intelligence and design
17. Rani S (2017) The impact of data analytics in crop management based on weather conditions.
Int J Eng Technol Sci Res 4(5):299–308
18. Ip R, Ang L, Seng K, Broster J, Pratley J (2018) Big data and machine learning for crop
protection. Comput Electron Agric 151:376–383
19. Firouzi F et al (2018) Internet-of-Things and big data for smarter healthcare: from device to
architecture, applications and analytics. Futur Gener Comput Syst 78:583–586
20. Raghupathi W, Raghupathi V (2014) Big data analytics in healthcare: promise and potential.
Health Inform Sci Syst 2(1)
21. Central Bureau of Health Intelligence (2018) National health profile—2018. Ministry of Health
and Family Welfare, Government of India, New Delhi
22. CRISIL Opinion, Bharat A (2018) Improvement in quality of government infrastructure and
leveraging private sector at right price to be crucial tasks for the scheme. CRISIL
23. Ministry of Family and Welfare, PIB Delhi (2019) Beneficiaries of Ayushman Bharat Yojana
24. Janke A, Overbeek D, Kocher K, Levy P (2016) Exploring the potential of predictive analytics
and big data in emergency care. Ann Emerg Med 67(2):227–236
25. Vinodhini G, Chandrasekaran RM (2014) Sentiment classification using principal compo-
nent analysis based neural network model. In: International conference on information
communication and embedded systems (ICICES2014). https://doi.org/10.1109/icices.2014.
7033961
Enhanced Approach in VANETs
for Avoidance of Collision
with Reinforcement Learning Strategy
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 419
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_41
420 A. Ganesh and S. Ayyasamy
1 Introduction
Numerous vehicle applications include efficient traffic control, active road safety,
etc. The main goal is to eliminate car accidents and to provide a collision-free,
driver-safe environment by sharing information on the danger of collisions and obsta-
cles. To extend the driver’s perception by allowing him/her to respond much more
quickly is the fundamental idea. This is possible through the wireless communica-
tion of alert/warning messages. And also, the traffic jam situations are avoided by
maximizing the movement of vehicles. Discovering the nearby vehicles and sharing
the obstacle/hazard information at low latency is the best technique to maintain a
collision-risk free environment.
An Ultrasonic sensor measures the distance between sensor and object. It has trans-
mitter and receiver, in which transmitter sends an ultrasonic wave, if the wave comes
in to contact with any object, the wave gets reflected back to the sensor receiver. It’s
measuring range is 2–400 cm with the accuracy of 3 mm. Distance is measured using
the formula:
Distance (cm) = Speed * Time Speed = ultrasonic wave speed (0.034 cm/µs)
Time = time traveled by the wave to and from the object (consider half of the
time, time (μs)/2)
Figure 1 is the IR (Infrared) based LM393 speed measuring sensor. This is used to
measure the wheel rotation speed of a vehicle, the two vertical columns consist of IR
LED on one side and phototransistor on the other side. The speed is determined by
wheel encoder which passes in between the two vertical columns. RPM (Revolutions
per Minute) defines the completion of all the 20 slots passage between the two vertical
column slots. This gives the RPM.
From RPM, the speed (cm/s) is measured by: (2/r * rpm)/60, where r (cm) =
Radius of the wheel.
3 ESP-NOW Protocol
4 Reinforcement Learning
RL consists of an agent, environment, action, reward, state [9, 10]. The agent performs
the action in an environment which results in the state transition from s to s´. s, s´
is the current state and next state respectively. Interaction of the agent with the
environment with certain actions leads to maximum reward. The combination of
artificial neural networks and RL architecture enables defined agents to determine
the best possible actions in a simulated environment to achieve their goals defines the
DRL. In other words, it combines function approximation and target optimization,
mapping state-action pairs to the predicted rewards.
Figure 2 shows the overview of the RL architecture for this application. The main
task of the agent here is to perform an action, so that vehicle is prevented from
the collision. The sensor readings are observed by the agent to get an idea about the
environment state. After observations from the environment, the agent has to generate
correct actions like moving forward/backward, left/right directions, or stopping the
vehicle, this prevents the vehicle involving in a collision/accident. The agent learns
a policy, which helps in the generation of correct actions from the observations. The
repeated attempts of the agent to have the collision-free environment maximize the
reward. Depending on the observations, actions, and rewards received, the training
algorithm adjusts the policy of the agent. After training, the vehicle will be able to
travel without involving in any collisions using tuned policy and sensor readings.
Enhanced Approach in VANETs for Avoidance … 423
4.1 Q-Learning
Q in Q-learning indicates quality. Quality here defines how effectively an action helps
in gaining some reward in the future. Mapping of state-action pairs to rewards is done
by Q. For each state-action pair, a Q-value is present in Q-table. Q-learning employs
Temporal Differences (TD) to determine the value of Qπ (s, a). π is the policy, it
maps the states to actions, where the actions guarantee the maximized reward. s and
a represents the state and action respectively.
Figure 3 depicts the Q-learning algorithm. Q-table has state-action pairs in which
each action is chosen corresponding to the current state. The reward is calculated for
each action and Q-table is updated until the maximum reward is attained.
PG methods are based on the principle “Observe and act” [11, 12]. A policy is trained
here, training makes actions with high rewards more probable or vice versa. PG holds
what is working and dumps what is not. It focuses on optimizing the parameterized
policies with respect to the anticipated cumulative reward by gradient descent. In
this, the agent analyzes the scenario, and instinctively understands what to do with
424 A. Ganesh and S. Ayyasamy
it. The more time of training enables the policy to maximize the rewards. PG aims
to effectively optimize policy space. The neural network is used in PG to model the
probability of action directly. When the agent interacts with the environment, we
fine-tune the neural network parameters so that “good” actions are more likely to be
sampled in the future. This process is repeated until the policy network converges
towards an optimum policy.
The collision avoidance with and without DRL is conveyed here. The implementation
starts with the collection of sensor readings from the chosen ultrasonic sensor and
LM393 speed sensor. The sensors are mounted on the two-wheeler chassis which
has 2 DC motors got fixed for the movement of chassis. L298N motor driver is used
which enables to control speed and direction of 2DC motors. All these components
are connected to the ESP32 microcontroller and this setup is taken as a vehicle node
which is shown in Fig. 4, it is controlled by the code written into microcontroller
using the Arduino IDE platform.
Enhanced Approach in VANETs for Avoidance … 425
First, let’s see how the simple collision avoidance code performs without any deep
learning technique. For this method, we use a servo motor on which the ultrasonic
sensor is mounted to rotate the ultrasonic sensor in the left and right directions from
the center. A threshold distance value is set for the ultrasonic sensor, if ultrasonic
measured distance > threshold value (i.e. obstacle is at safe distance from the sensor)
vehicle node moves in forward direction. Otherwise (i.e., measured value ≤ threshold
value) vehicle moves backward and checks the distance between sensor and obstacle
in the left and right directions. The sensor compares the distance in both left and
right directions and then the vehicle moves in the left or right direction which has
more space to move.
Collision avoidance with DRL starts with designing of Q-learning framework by
having state-action pairs, discrete states represent the distance range varying from 0
to 400 cms, actions are forward, backward, turn right and left, halting the vehicle.
In this DRL method, servo motor is not used, the ultrasonic sensor is set to a fixed
direction. Training is done in episodes. The PG network consists of an input layer of
dimension 1 × 1. Distance to the nearest obstacle is taken as input, the hidden layer
of 3 × 1 dimension with RELU (Rectified Linear Unit) activation, and the output
layer of dimension 5 × 1, with soft max activation, corresponding to 5 possible
directions at which the vehicle node can travel. PG network samples the 5 actions,
performs them, stores, and updates the gained reward. This sampling process is
encapsulated as the policy π, which is responsible for the implementation of action
dependent on past environmental states {s}. PG agent here specifically determines
an optimal strategy that enhances long-term rewards. By using both Q-learning and
policy gradient, Q-values are determined from the policy preferred actions to which
we apply Q-learning updates.
426 A. Ganesh and S. Ayyasamy
One can see the episode number, the number of actions sampled in the serial
monitor of a particular vehicle node and this Deep reinforcement algorithm helps
in maintaining the safe distance from the nearest obstacle. The vehicle node takes
appropriate action based on the training when it finds an obstacle. In both with and
without DRL algorithm methods, V2V communication is the same. It includes the
transmission of vehicle’s distance from the obstacle, speed of the vehicle which
is sensed by the LM393 speed sensor and, also MAC address of the transmitted
vehicle. These parameters are displayed on the serial monitor of all peer devices
which are present in the ESP-NOW protocol range. ESP-NOW is responsible for
V2V communication.
Fig. 5 Vehicle V1 displaying it’s parameters and Receivingv2 parameters through ESP-NOW
protocol
Enhanced Approach in VANETs for Avoidance … 427
Fig. 6 Vehicle V2 displaying it’s parameters and receiving V1 parameters through ESP-NOW
protocol
VANETs are considered to be a significant research area and drawing the attention of
a noticeable number of researches. It is a vast area where there are many challenges
that are to be addressed. The training of ESP32 microcontroller with a deep learning
428 A. Ganesh and S. Ayyasamy
model is accomplished here using DRL. The proposed work is carried out with two
sensors and two vehicle nodes, but V2V communication can be extended by adding
GPS (Global Positioning System) module to the vehicle node setup which gives the
location information of a vehicle and number of vehicle nodes can be increased to
show the broadcasting of information effectively.
References
1. Kumar KV, Jayasankar T, Prabhakaran M, Srinivasan V (2017) Fuzzy logic based efficient
multipath routing for mobile adhoc networks. Appl Math Inf Sci 11(2):449–455. https://doi.
org/10.18576/amis/110213
2. Gopinath S, VinothKumar K, Jayasankar T (2019) Secure location aware routing protocol with
authentication for data integrity. Springer Cluster Comput 22:13609–13618. https://doi.org/10.
1007/s10586-018-2020-7
3. Ejmaa AME, Subramaniam S, Zukarnain ZA, Hanapi ZM (2016) Neighbor-based dynamic
connectivity factor routing protocol for mobile ad hoc network. IEEE Access 4:8053–8064
4. Wang M, Liu D, Zhu L, Xu Y, Wang F (2016) LESPP: lightweight and efficient strong privacy
preserving authentication scheme for secure VANET communication. Computing 98(7):685–
708
5. Chim TW, Yiu SM, Hui LC, Li VO (2011) SPECS: secure and privacy enhancing communi-
cations schemes for VANETs. Ad Hoc Netw 9(2):189–203
6. Bayat M, Barmshoory M, Rahimi M, Aref MR (2015) A secure authentication scheme for
VANETs with batch verification. Wirel Netw 21(5):1733–1743
7. Biswas SB, Iqbal MT (2018) Solar water pumping system control using a low cost ESP32
microcontroller. In: 2018 IEEE Canadian conference on electrical and computer engineering
(CCECE). IEEE, pp 1–5
8. Kushnir V, Koman B, Yuzevych V (2019, September) IoT image recognition system imple-
mentation for blind peoples using ESP32, mobile phone and convolutional neural network.
In: 2019 XIth international scientific and practical conference on electronics and information
technologies (ELIT). IEEE, pp 183–187
9. Wang W, Min M, Xiao L, Chen Y, Dai H (2019, May) Protecting semantic trajectory privacy
for VANET with reinforcement learning. In: ICC 2019–2019 IEEE international conference
on communications (ICC). IEEE, pp 1–5
10. Dai C, Xiao X, Xiao L, Cheng P (2018, December) Reinforcement learning based power control
for VANET broadcast against jamming. In: 2018 IEEE global communications conference
(GLOBECOM). IEEE, pp 1–6
11. Luo Y, Chiu CC, Jaitly N, Sutskever I (2017, March) Learning online alignments with contin-
uous rewards policy gradient. In: 2017 IEEE international conference on acoustics, speech and
signal processing (ICASSP). IEEE, pp 2801–2805
12. Liu S, Zhu Z, Ye N, Guadarrama S, Murphy K (2017) Improved image captioning via policy
gradient optimization of spider. In: Proceedings of the IEEE international conference on
computer vision, pp 873–881
13. Vishnupriya E, Jayasankar T, Venkatesh PM (2015) SDAOR: secure data transmission of
optimum routing protocol in wireless sensor networks for surveillance applications. ARPN J
Eng Appl Sci 10(16):6917–6931
VBlock—Blockchain-Based Traceability
in Medical Products Supply Chain
Management: A Case Study in VietNam
Tuan Khoi Nguyen Huynh, Tuan Anh Dao, Trong Nhan Van Pham,
Khang Hy Vuong Nguyen, Nghi Cong Tran, and Hoang Huong Luong
1 Introduction
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 429
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_42
430 T. K. N. Huynh et al.
WHO estimated that the world needs up to 89 million medical face masks each
month. Having acknowledged the scarcity of medical facilities, many rogue traders,
and sellers distributed counterfeit products to the community and even hospitals
to gain money. These facilities, as their quality is not guaranteed [1, 2], can pose
significant threats to patients. Also, some stores advertise genuine products at unrea-
sonably unreasonable prices. Our team believes that these illegal activities impede
the progress in dealing with COVID. Therefore, they need to be eliminated.
Our team believes that the solution to rogue business is to use a proper supply-
chain management platform. We define a proper supply-chain management platform
as a platform that is: fast, immutable, and decentralized. As for speed, if the whole
distributing process of a particular product is retrieved quickly, we can save much
time. If the government detects a facility’s problem, all entities related to that facility
can be traced within seconds. About immutability, all sensitive data needs to be
tamper-proof, meaning that rogue traders cannot alter what they have uploaded, and
hackers cannot break havoc. We define sensitive data as business related information
such as supply chain information, or companies’ credentials. Such data is critical for
the government to check business legality and products’ origin. Decentralization is
the last characteristic to mention. A decentralized platform does not store data in a
central place, but copies and distributes it across the network. This ensures that all
entities have a copy of the whole database, limiting the impact of a data loss at a
node [3].
Many researchers have suggested the use of Blockchain to develop such a plat-
form. Blockchain is a database system, commonly known as a distributed ledger. It
includes traceable records or blocks that are linked via cryptography. Blockchain is
intrinsically immutable and transparent [4]. Our team agrees with the researchers’
perspective, so we want to develop a medical supply chain management system using
Blockchain, which we name “V-Block”.
In this paper, we will introduce V-Block and explain how it can stop fake and
overpriced products. We will begin by describing some previous work and documents
on implementing Blockchain for the medical supply chain. We will discuss their
strengths as well as their shortcomings. From that analysis, we can suggest our ideas
on improving the implementations. After pointing out our ideas, we will present
our system. There are three main identities in our system: customers, members,
and agencies. Each identity will have distinct functions that suit their needs and
help enforce data safety. Our paper is expected to have three main benefits. First of
all, current Blockchain implementations are analyzed, and their shortcomings are
identified. Additionally, we propose our own system, which includes improvements.
Finally, our system can be used to assist the medical industry in the battle against
COVID-19.
The rest of the paper is organized as follows. Section 2 contains previous
Blockchain applications in various areas, especially in the supply chain. Then, in
Sect. 3, we briefly explain our Blockchain system. We will also get insight into
the system’s functions in Sect. 4 before describing experiments in Sect. 5. We will
summarize our working Sect. 6.
VBlock—Blockchain-Based Traceability in Medical … 431
2 Related Research
Today Blockchain is applied in many areas such as mobile payments [5] introduce a
device that was made up of an Arduino ATMEGA256 board, A TFT-LCD touch
screen, a smartcard socket, and a Bluetooth Low Energy (BLE) module named
Bluetooth Blockchain payment terminal (PBT). PBT is a new generation of open
payment terminals dedicated to Blockchain transactions. A removable secure element
is attached to the BPT to keep the card safe from the internet-connected device. They
inserted a smartcard with a level of security up to EAL5+ to store private keys and
compute signatures to the PBT to make transactions via BLE. Medical [6] this study
mentions some of the benefits of Blockchain in the medical field such as: Firstly,
smart healthcare management. It used a continuously updated decentralized database
to give all medical professionals and doctors access to a patient’s treatment history
with previous doctors as well as disease health information to get more informa-
tion and to have the right treatment for that patient. In addition, the decentralized
database helps to solve problems such as information that is not being updated,
which causes time to verify when needed. Secondly, collect medical information for
research purposes: Build a Blockchain system to help users update their health infor-
mation (heart rate, blood pressure, weight, sleep, used medicine, …). Researchers
can buy that information from users to research and find ways to cure new diseases.
Finally, Blockchain prevents counterfeit medicine: The drug is not in the correct dose
as provided, or the drug is produced in an impure way that can be dangerous to the
patient. Blockchain can be used to provide accurate information about the medicine.
Blockchain can be used to store transaction records in dental care [7] so that partici-
pants can access and control authority. It has the transaction id number generated by
the hash number mechanism provided by the Hyperledger editor. This application
creates a relationship of trust to establish the set of participants in an ecosystem
to share information and interact between clinics so that it can save time and a
variety of costs for the patients. Another application of Blockchain in security is the
recommended Blockchain-based framework (B2MDF) [8] For detecting malicious
apps in the App Store before being downloaded by a user. They proposed using two
private Blockchains inside and outside to form a double private Blockchain to make
the correct decision. Internal Blockchains store feature blocks that are analyzed by
static and dynamic analysis programs to increase malware detection rates. External
Blockchains store detected results as blocks for current versions of the application.
B2MDF shares information with third parties to help security companies deliver anti-
malware. B2MDF does not limit implementation to any particular machine learning
algorithm.
Blockchain plays an important role in the supply chain [9]. We have referenced
research on Blockchain applications in the supply chain “A Blockchain-based Supply
Chain Quality Management Framework” [10]. In this paper, we study the Blockchain-
based SCQI framework for improving supply chain quality management. It consists
of 4 layers as follows: The first layer is the IoT Layer. It uses IoT sensors to collect
Logistics Process Selling Process Manufacture and Quality Inspection Process data.
432 T. K. N. Huynh et al.
The information collected is captured using RFID technology or a bar code. The
second layer is the Data Layer. Blockchains and secure distributed ledgers are in
this layer. The distributed ledger contains the information Quality Data, Logistics
Data, Transaction, Data, Assets Data. The third layer is the Contract Layer. This
layer contains the data to be protected. So, this layer’s data needs to be protected
by means of digital identity access control. The last layer is the Business Layer.
This layer contains the business operations of the enterprise. With Blockchain,
businesses manage product quality in their supply chains more efficiently. “Supply
Chain Management using Blockchain” [11] from this paper, we study how to apply
Blockchain to product tracking from the raw material until it reaches the consumer,
and how to solve problems such as inaccuracies in profile, fraud problem, payment
problem. This study also shows that using Blockchain in product tracking can reduce
the cost, time, and manpower in monitoring and controlling raw materials and prod-
ucts. Besides, this study also suggests the use of smart contracts to handle product
quality problems such as handling compensation when factors are affecting product
quality such as temperature, humidity, not in accordance with smart contract require-
ments. We also looked at the case of Blockchain use in the agricultural supply chain,
“Blockchain-based Traceability in Agri-Food Supply Chain Management: A Prac-
tical Implementation” [12]. This article is about AgriBlockIoT, a fully decentralized
application used to monitor agricultural and food products from seeding to consump-
tion with the desire to monitor to improve food safety and maintain trust and reliability
throughout the supply chain. After reviewing the feasibility study for application in
Vietnam, we realized that AgriBlockIoT has some disadvantages when it comes to an
application as follows: farmers must be trained to store information into the system,
causing feeling constrained for farmers when being monitored by IoT sensors; Only
some common values, such as weight, were measured, but special indicators such as
the number of pesticides that were applied to agricultural products were measured.
To measure the plant protection drug index, specialized sensors are required with a
high cost and not suitable for practical implementation; IoT sensors cannot be widely
equipped because of the high cost of setting and maintaining. That causes false data
when farmers can affect agricultural products in areas without sensors. Assuming
the case of sensors being placed throughout the garden, the farmer can still cheat by
masking the sensors when using a pesticide so that the system recognizes “clean”
and high yielding produce.; The last problem, the sensor fails to measure correctly,
but the data is still within the allowable range and will not be detected and cause an
incorrect data supply. We have looked at the challenges of applying Blockchain to the
supply chain [13]. This article outlines the opinion that supply chain management
(SCM) is the key factor in achieving financial, environmental, and social benefits
in the supply chain industry. But the traditional SCM mechanisms are outdated and
have many weaknesses in information sharing, data retrieval, and reliability in finding
products. These problems can be solved by Blockchain due to Blockchain features
such as immutability, transparency, and decentralization. What have we learned from
this article? These are the requirements of supply chain management when adopting
Blockchain, the technical challenges of applying Blockchain to the supply chain,
VBlock—Blockchain-Based Traceability in Medical … 433
3 System Overview
V-Block is a system used to store and retrieve information about medical products’
supply chain. It differs from traditional systems in terms of data storage, as it uses
Blockchain technology. This storage implementation is expected to improve data
security, thereby assisting violation detection, and increasing business transparency.
In general, there are three main entities, including customers, manufac-
turers/distributors, and agencies. Manufacturers or distributors will upload data about
the origin and prices of products. Agencies and customers, when in need, can query
data and analyze it for anything questionable. Agencies have the right to view data
without restrictions.
The variety of users, along with different levels of rights and permissions, means
that V-Block requires a layer of security over data access and performed activi-
ties. For this reason, our team develops V-Block to be a permissioned Blockchain.
Permissioned Blockchain is a well-known model, acknowledged by enterprises and
researchers worldwide [14, 15]. The system has three main parts: users, access
control, and data storage.
3.1 Users
There are three kinds of users, namely: customers, members, and agencies. Customers
are the least-privileged identities in the system. They are hospitals or any other
institutions that purchase a medical product and want to check for its authenticity.
Customers are only able to query information about a particular product. Members
are manufacturers and distributors of a product. They are allowed to submit data to
the system and also view their uploaded data. Agencies refer to medical institutions
from the government. They are granted the most rights and permissions. They are
permitted to view any information as well as new changes made to the system. This
enables them to check the validity of a particular item or a manufacturer.
We acknowledge the importance of keeping uploaded data safe and private in supply
chain management [16]. That is why we implement different controls to ensure that
only allowed individuals can access private data. Normal customers do not have to
login or sign up to use our system, as they only want to query information about a
434 T. K. N. Huynh et al.
specific product. However, members must prove their identity before proceeding with
any operations. Upon registration, they first need to fill a form, in which they include
some personal details and business licenses. After this step, they will have to wait to
be checked and approved by agencies. What users can query is also controlled. For
example, only agencies can view all information about the system, while members
can only ask for information about their own products. This is done by implementing
smart contracts for searching and querying data [17]. Additionally, agencies can
create full nodes, but members only have light nodes. Having full nodes means that
a user can save everything on the system into their machines, as opposed to light
nodes, which only stores limited information.
Any data created will be stored on Blockchain. New data, after verified by peers,
will be kept in a block, which then is broadcasted via multiple network protocols [4].
Data input and output will be handled by different consensus algorithms [18] and
smart contracts. This serves the system’s functions and ensures the validity of data
being stored and submitted.
4 System Functions
customers are able to know who the manufacturers are and where the facilities are
produced within a limited amount of time. If they find something suspicious about
the supply chain of any facility, they can report it to the agency. The report includes:
the name of the business, the name and the code of the product. Reports will be kept
away from companies, which means companies cannot know the reporters.
436 T. K. N. Huynh et al.
If the agencies are reported about a potentially fake medical facility, they can
use our system as a great tool. The agencies can quickly get all information about
individuals and groups involved in the supply chain thanks to the origin-tracing
feature of V-Block. All they need to do is to submit the reported product’s code and
wait for the result. This will greatly speed up the investigation. Additionally, as all
data stored on our system is tamper-proof, agencies can use it as evidence against
any illegal trades (Fig. 3).
Overpriced facilities
V-Block allows users to analyze fluctuations in products’ prices. Manufacturers and
distributors will upload the product’s new cost after each stage in the supply chain. As
a result, price changes can be recorded and retrieved whenever needed. Agencies and
consumers can query the data and check if a distributor or a seller has unreasonably
increased the cost.
Furthermore, our system attempts to block suspicious cost-related data upload.
If a facility’s cost is raised 20% after only one stage in the supply chain, it will be
flagged, and the upload will be stopped (Fig. 4).
VBlock—Blockchain-Based Traceability in Medical … 437
5 Experiment
5.1 Introduction
one account for a customer, another for an agency, and the remaining one for a seller
(member).
Fig. 5 Example of QR
VBlock—Blockchain-Based Traceability in Medical … 439
6 Conclusion
In this paper, we explained the issue of counterfeit and overpriced medical facilities.
We also described our solution to that issue, which is to develop a supply chain
management system using Blockchain, which we call ‘V-Block.’
440 T. K. N. Huynh et al.
References
1. Cheng M (2001) A guide for the development of medical device regulation. Washington DC
2. Corciovă C, Andritoi D, Ciorap R (2013) Elements of risk assessment in medical equipment.
In: 2013 8th international symposium on advanced topics in electrical engineering (ATEE)
3. Xiao T, Choi TM, Edwin Cheng TC (2018) Pricing and benefit of decentralization for competing
supply chains with fixed costs. IEEE Trans Eng Manag 65(1)
4. Antonopoulos AM (2017) Mastering bitcoin, 2nd ed., June 2017
5. Urien P (2020) High security bare metal Bluetooth Blockchain payment terminal for trusted
ethereum transaction. In: 2020 IEEE 17th annual consumer communications and networking
conference (CCNC)
6. Mettler M (2016) Blockchain technology in healthcare: the revolution starts here. In: 2016 IEEE
18th international conference on e-health networking, applications and services (Healthcom)
7. Wutthikarn R, Hui YG (2018) Prototype of Blockchain in dental care service application
based on Hyperledger composer in Hyperledger fabric framework. In: 2018 22nd international
computer science and engineering conference (ICSEC)
8. Homayoun S, Dehghantanha A, Parizi RM, Choo KKR (2019) A Blockchain-based framework
for detecting malicious mobile applications in app stores. In: 2019 IEEE Canadian conference
of electrical and computer engineering (CCECE)
9. Caro MP, Ali MS, Vecchio M, Giaffreda R (2019) Blockchain-based traceability in agri-food
supply chain management: a practical implementation. In: 2018 IoT vertical and topical summit
on agriculture—Tuscany (IOT Tuscany) Wu H, Cao J, Yang Y, Tung CL, Jiang S, Tang B, Liu Y,
VBlock—Blockchain-Based Traceability in Medical … 441
Wang X, Deng Y (2019) Data management in supply chain using Blockchain: challenges and a
case study. In: 2019 28th international conference on computer communication and networks
(ICCCN)
10. Chen S, Shi R, Ren Z, Yan J, Shi Y, Zhang J (2017) A Blockchain-based supply chain quality
management framework. In 2017 IEEE 14th international conference on e-business engineering
(ICEBE)
11. Bhalerao S, Agarwal S, Borkar S, Anekar S, Kulkarni N, Bhagwat S (2019) Supply chain
management using Blockchain. In: 2019 international conference on intelligent sustainable
systems (ICISS)
12. Caro MP, Ali MS, Vecchio M, Giaffreda R (2018) Blockchain-based traceability in agri-food
supply chain management: a practical implementation. In: 2018 IoT vertical and topical summit
on agriculture—Tuscany (IOT Tuscany)
13. Wu H, Cao J, Yang Y, Tung CL, Jiang S, Tang B, Liu Y, Wang X, Deng Y (2019) Data
management in supply chain using Blockchain: challenges and a case study. In: 2019 28th
international conference on computer communication and networks (ICCCN)
14. Mitani T, Otsuka A (2019) Traceability in permissioned Blockchain. In: 2019 IEEE interna-
tional conference on Blockchain (Blockchain)
15. Islam MdA, Madria S (2019) A permissioned Blockchain based access control system for IOT.
In: 2019 IEEE international conference on Blockchain (Blockchain)
16. Roy A, Gupta AD, Deshmukh SG (2012) Information security in supply chains—a process
framework. In: 2012 IEEE international conference on industrial engineering and engineering
management
17. “SmartContracts”. https://solidity.readthedocs.io/en/v0.7.1/introduction-to-smart-contracts.
html
18. Chaudhry N, Yousaf MM (2018) Consensus algorithms in Blockchain: comparative analysis,
challenges and opportunities. In: 2018 12th international conference on open source systems
and technologies (ICOSST)
Development of Real-Time Internet
of Things (IoT) Based Water Quality
Monitoring System
1 Introduction
The water quality status of rivers and lakes in Malaysia has always been a reason
of concern for various local authorities, government agencies and the public at large
scale. Water quality monitoring system is important for various applications that
requires a real time monitoring to determine the contents and substances present
inside the water. According to Departmental of Environmental (DOE) Malaysia, it is
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 443
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_43
444 H. F. Hawari et al.
stated that two main parameters are most likely to affect the river water quality that
are urbanization and development [1]. WQI is a method or a tool used to measure
the pollution level of water using SI calculation whereas National Water Quality
Standard (NWQS) is the national standards used to indicate the level of pollution
after measuring the water quality using WQI [2, 3].
IoT is defined as internetworking physical objects embedded with electronics,
sensors, network connectivity and actuators which can be used to collect and
exchange the data [4–6]. IoT can connect billions of devices over the network which
will eventually create much more need for data storage and much larger traffic to
control the connectivity of those devices using TCP/IP protocol stack to communicate
between network hosts for today’s Internet [7].
In this proposed IoT system, users can monitor the quality of water parameters
in real time using mobile application developed using suitable software and alert the
user for water quality alarm system. This alert system is important for the user to
alert them when the value of certain parameters exceed the threshold set. According
to UN Water, it is stated that around 780 million of people did not manage to get
safe and clean water while around 2.5 billion of people do not have suitable adequate
clean drinking water provided by UN-Water [8].
The problem with existing technology is water quality monitoring involves the
manual intervention of collecting water sample in different locations. This is followed
by further laboratory analytical techniques to characterizes the water quality based
on measurement parameters of WQI. This manual approach takes longer time and
is considered as less efficient for water quality monitoring. It is also not considered
as real-time as the person responsible must take samples from particular locations
from time to time.
Therefore, this project is implemented to develop and implement a real-time appli-
cation based on IoT for water quality monitoring system, to integrate water quality
sensors together with IoT communication platform for an efficient real-time water
quality monitoring system and to develop mobile application suitable in analysing
the water quality conveniently.
Table 1 Current consumption and response time for sensors and Arduino Nano
Components/Sensors Current consumption (mA) Response time (ms)
Arduino Nano 19 0.9
Temperature 10 0.8
Turbidity 14 0.3
pH 30 0.4
446 H. F. Hawari et al.
The data from sensors are transferred over the network using Wi-Fi connection. A
Wi-Fi module (EP8266) is integrated with Arduino Nano as controller to provide
the connectivity of sensor device with cloud computing by connection to a Wi-Fi
gateway to connect to Internet. It supports 802.11 b/g/n/e/i of network protocol with
Wi-Fi Direct (P2P). It is advantageous of using this module as it supports standard
network protocol, has antenna diversity and selection (to extend wireless range) and
relatively low power consumption which is about 170 mA during Transmit Mode
(Tx Mode). Figure 2 shows wireless communication architecture for this project.
In the cloud, all data collected in the microcontroller are stored in the database through
cloud computing system. The cloud used for this project is Ubidots which provide
simple IoT implementation in cloud computing system. This cloud can provide suffi-
cient features needed to implement this project successfully. To enhance the porta-
bility and mobility of data monitoring system, mobile application for tablets and
smartphones is designed to capture those data stored in the cloud and visualized in
charts and gauges representation for water quality monitoring purposes. Real time
data will be measured and saved in the cloud to produce historical data for certain
period.
This prototype uses two lithium-polymer batteries as the power sources to all elec-
tronic components involved. To manage the power consumption, 1st battery will
provide power to ESP8266 to set up the module and network. Once finished, 1st
battery is turned-off while 2nd battery is switched on providing power to sensors to
get measured sensor data from water. After all sensors data have been achieved, 2nd
battery is switched off while turning on 1st battery to send sensors data to cloud via
Development of Real-Time Internet of Things (IoT) … 447
Figure 4 shows the system architecture of the system starting with sensors getting all
the data from lakes and transmit to cloud via wireless connectivity and monitored in
mobile application.
448 H. F. Hawari et al.
The prototype is placed on top of water supported by three pillars anchored in the
water attached with stretchable strings. This is to prevent the device to flow away
from desired location. Figures 5 and 6 shows the location of prototype development
in UTP lakes and deployment technique of the device in UTP lake respectively:
Fig. 5 Prototype
deployment in UTP lake
Development of Real-Time Internet of Things (IoT) … 449
Data for water quality parameters include turbidity, pH value and temperature were
taken in two different locations in UTP. The duration for each data measurement is
1 h with 14 h of total data measurement per day for 7 days in a week.
The sensor data are inserted into the formula to get water pollution index value.
Table 2 shows the pollution index value for each pollution level.
The value of water pollution index in UTP lake is calculated based on formula
constructed with the value of temperature, pH and turbidity measured. Data for WPI
is calculated for overall 4 weeks of data collection to indicate the level of pollution
of UTP lake. From Fig. 7, we can observe that the trend of WPI is stable during 1st
week until 3rd week, but there is slight drop in WPI value in 4th week. This is due
to lower value of turbidity which contributes to lower WPI value in the calculation.
WPI increases during 5th week to 6th week because it has been measured in location
C where the water quality is found to be slightly polluted.
3.2 Temperature
Figure 8 shows the data for water temperature in UTP lakes for 6 weeks of data
collection:
The average of temperature value for location A is 29.50 °C, for location B is
29.10 °C and for location C is 29.45 °C. During 1st week until 6th week of data
collection, water temperature of location A, B and C shows nearly the same value
except for 4th week. During 1st week to 3rd week, the mean average water tempera-
ture is higher than during 4th week due to warmer ambient condition. This increases
the solubility and toxicity of certain compounds in the water. These compounds
include heavy metals elements such as zinc, lead, cadmium and ammonia. With
higher water temperature (more than 30 °C), oxygen consumption, metabolic rate
and tissue of aquatic organism permeability increases. This will reduce the survival
rate of aquatic organisms in the water as the lethal concentration 50 (LC50) drops for
a long period. The lower the value of LC50, the lower the concentration of chemical
inside the water, causing lower life time for aquatic organisms.
3.3 PH
Figure 9 shows the data for water pH in UTP lakes for 6 weeks of data collection:
The average pH value for location A is 6.84, for location B is 6.83 and for location
C is 8.65. Based on WQI, the permissible range of value for normal lake water pH
is 6.01–8. As the average of pH value for each week and total average of pH value
for 1st week until 4th week falls in between the permissible range of value, the water
pH measured contributes normal value of water pollution index of UTP lake.
Unlike water pH in location C where the average pH value is 8.65, it is included
in slightly polluted range. This value has possibility to contribute slightly polluted
water condition for water pollution index. When the pH value is below or higher
than the optimal value or permissible range of value, aquatic life become susceptible
to fungal infections and other physical damages thus it is not suitable for aquatic
organisms to stay in water that has non-optimal pH value.
This can harm those organisms and reduce water quality. Related to spectrum, high
pH levels can damage the gills and skin of aquatic organisms and can cause death at
pH level higher than 10.0. While the mean pH value of location C is still below that
pH level, the aquatic organisms are safe, but it also can harm those organisms inside
the water.
3.4 Turbidity
Figure 10 shows the data for water turbidity in UTP lakes for 6 weeks of data
collection. The average of turbidity value for location A and B contributes to normal
water pollution index while the average of turbidity value for location C contributes
to slightly polluted water pollution index. The lower the value of turbidity, the higher
the water quality based on water pollution index. This can be achieved by reducing
the Total Suspended Solids (TSS) present in the water which higher total suspended
solids causes the water to be turbid. TSS are particles that are larger than 2 µ, therefore
any particles in the water that are smaller than 2 µ are considered as Total Dissolved
Solids (TDS). TSS is contributed by algae and bacteria inside the water that can
cause water to be more turbid.
Usually, clear water is considered as clean water. Excessive value of turbidity
can impair the water quality for aquatic and human life. Other than that, higher
water turbidity will increase water temperature and reduce Dissolved Oxygen (DO).
Warmer water cannot hold DO more than cold water. Based on the data from those
locations, location C is showing least DO quantity compared to the other 2 locations
as it has highest water turbidity and included in slightly polluted range based on
WQI. This lead to higher heat absorption from solar radiation and transferred to
surrounding water by conduction. Therefore, location C is considered as slightly
polluted and not suitable for aquatic life and human usage.
On the other hand, pollutants are most likely to be found in high water turbidity.
The contaminants include bacteria, mercury, lead and other material that encourage
the development of harmful substances inside the water. Since those 3 locations are
used for recreational use, as turbidity reduces visibility of underwater structure, it
negatively affects the body’s recreational use for users in UTP.
Figure 11 shows the overall current consumption for device used in water quality
monitoring:
Analysis is made to calculate the battery life for both batteries (battery A and B)
using the formula shown below:
From the calculation made, the battery life for the prototype is estimated up to
30 days.
4 Conclusion
In conclusion, Internet of Things (IoT) assist and help people in deciding the best
solution for any problems faced in daily life especially water quality because water
is the main element required by human beings for their life every day. Without good
quality of water, people tend to have problems with their healthcare and welfare
because everyone need water to survive in any situation. Development of innovative
technology help them to monitor the water quality especially in the river as it is the
main source of water supply for users in entire world. Once water quality can be
454 H. F. Hawari et al.
monitored in real time, it is easy for the department responsible in maintaining the
water quality to get all required data of water quality, visualize them in most effective
representation and decide the best solution for water treatment. The performance of
this system is comprehensive with good techniques and methods used for every
scope included. The system has good response time sending all data successfully to
cloud with minimum communication error. This is a good practice as it increases
reliability and practicality of this project. Without this technology, people are forced
use manual intervention continuously by taking samples from site locations regularly
and analyse them in the laboratory using effective laboratory analysis methods. This
is time consuming and not considered as efficient water quality monitoring.
Acknowledgements I would like to thank Dr Radzi Ahmad from Electrical and Electronics
Engineering department, Universiti Teknologi PETRONAS who despite of being busy with their
schedule, managed to take time out to provide me with untiring assistance, direction, encourage-
ment, continuous guidance, and support indeed provides a huge contribution for me to complete
this project successfully.
References
1. Suratman S, Sailan MI, Hee YY, Bedurus EA, Latif MT (2015) A preliminary study of water
quality index in Terengganu River basin, Malaysia. Sains Malaysiana 19(44):67–73
2. Hossain MA, Sujaul IM, Nasly MA (2013) Water quality index: an indicator of surface water
pollution in eastern part of Peninsular Malaysia. Res J Recent Sci 10(2):10–17
3. Gazzaz NM, Yusoff MK, Ramli MF, Aris AZ, Juahir H (2012) Characterization of spatial
patterns in river water quality using chemometric pattern recognition techniques. Department
of Environmental Science, Faculty of Environment Studies
4. Khan R et al (2012) Future internet: the internet of things architecture, possible applications
and key challenges. DITEN Department University of Genova (UNIGE), 16145 Genova, Italy
5. Hawari HF, Zainal AA, Ahmad MR (2019) Development of real time internet of things (IoT)
based air quality monitoring system. Indonesian J Electr Eng CS 13(3)
6. Chung Hua BD, Fahmi H, Yuhao L, Kiong CC, Harun A (2018) Internet of things (IOT)
monitoring system for elderly. In: International conference on intelligent and advanced system
(ICIAS), p 201
7. Minerva R, Biru A, Rotondi D (2015) Towards a definition of the Internet of things (IoT). IEEE
Internet Initiative. Telecom Italia, Torino
8. UN-Water (2013) An increasing demand, facts and figures, coordinated by UNESCO in collabo-
ration with UNECE and UNDESA. Available: http://www.unwater.org/water-cooperation-201
3/en/
9. Vinegre C, Madeira D, Narciso L, Cabral NH, Diniz M (2012) Effect of temperature on oxida-
tive stress in fish: lipid peroxidation and catalase activity in the muscle of juvenile seabass,
Dicentrarchus labrax
10. Schveitzer R et al (2013) Effect of different biofloc levels on microbial activity, water quality
and performance of Litopenaeus vannamei in a tank system operated with no water exchange.
Aquacult Eng 56:59–70
11. Jonsson et al M (2013) Foraging efficiency and prey selectivity in a visual predator: differential
effects of turbid and humic water. Can J Fish Aquat Sci
12. Norshidah B et al (2014) Characterization of spatial patterns in river water quality using
chemometric techniques. Sains Malaysiana 9(43):1355–1362
Edge Computing Technology:
State-of-the-Art and Challenges
in an Internet of Things
Abstract The Internet of Things (IoT) is a network of expedients that can intuiting,
collecting, managing and communicating data over the Internet without human inter-
vention. It is a concept that connects to nearby objects without user intervention, using
a wired or wireless network. In IoT, objects interact and share information to provide
the user or users with advanced, intelligent services. In the field of edge computing,
data engendered by IoT devices can be processed closer to where it was created,
and as an alternative to sending data to data centers or data centers in more ways.
New research and customer support in selecting seemly edge computing structures
for particular solicitations. This paper gives a complete overview of the methods of
introduced power adding structures, comparing Open Source devices through their
survival method. Finally, this work sheds light on smart home application tools using
Edge Computing Systems and also explores the challenges posed by Edge Computing
Systems.
1 Introduction
Kevin Ashton first coined the term Internet of Things in the presentation tile, which
introduced sensor tagging for large consumer company products in 1999 to strengthen
supply chain management. Things to the Internet describe the ecosystem, where
sensors and actuators are connected to the Internet. All objects in the IoT system speak
the same language and understand and operate without human intervention. The
Internet of Things has been labeled too late, the concept of connected networks is not
D. Raghunathan (B)
Research Scholar, Department of Computer Science and Engineering, Dr.N.G.P Institute of
Technology, Coimbatore, India
M. Krishnamoorthi
Professor, Department of Computer Science and Engineering, Dr.N.G.P Institute of Technology,
Coimbatore, India
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 455
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_44
456 D. Raghunathan and M. Krishnamoorthi
new and there are many developments like this in progress [1]. In the 1980s, the Coca-
Cola machine was connected to the Internet and monitored for availability, which
reduced unnecessary visits to vending machines. Sethi et al. The following statistics
review basic IoT systems, architectures, available protocols, and applications of IoT.
The Edge Computing model has gained a significant reputation in the academic
and industrial sectors over the past few years. It serves as the key to many future
technologies, such as vehicle-to-vehicle communication using 5G, IoT and cloud
computing services for future customers. It gives low-slung invisibility versatility
and arrangement mindfulness bolster for delay-sensitive solicitations. In the context
of the latest developments in cloud computing, mist computing, and versatile edge
computing, basic investigation within the space of edge computing, pioneered by
providing additional solutions to researchers in existing solutions and future appli-
cations [2]. Due to the characteristics and features of IoT services, smart home
applications have also become a major problem in the IoT application areas (Fig. 1).
Edge Computing speaks to an unused concept within the computing extract that
fetches the administrations and practicalities of cloud calculating closer to the conclu-
sion patron. It is characterized by two variables, such as quicker handling and speedier
application reaction time. Right now created Internet-facilitated applications such
as observing, cybernetic authenticity and instantaneous activity observing involve
quicker administration, speedier reaction times [3]. Conclusion clients ordinarily run
these applications on their portable gadgets whereas performing center administra-
tions and preparing on cloud servers. Cloud administrations are influenced by versa-
tile gadgets, coming about in tall idleness and versatility issues [4]. Gives an edge
computing application as per the prerequisites by moving the handling to the arrange
edge. Mist computing concerns can be illuminated by three distinctive authority
figuring simulations, such as cloudlets [5], mist computing [6] and versatile edge
computing [7].
This paper describes Edge Computing methods and models. Section 2 describes
the literature review of an IoT edge computing, fog computing, and mist computing.
Section 3 gives a brief computing model. Sections 4 and 5 provides conclusions
Edge Computing Technology: State-of-the-Art … 457
about the availability of equipment and, finally, Sect. 6 grants the supposition on this
work.
2 Literature Review
Assuncao et al. [15] describes emerging technologies for distributed data stream
processing, as well as cloud resource management techniques for elasticity manage-
ment. But it is difficult to develop and implement stream processing applications
under different distribution structures.
Bilal et al. [16] offered an outline of edge computing motivation and edge tech-
nology. Concerns in this paper include resource management and allocation, resource
scalability, methods, data capture, fault tolerance, and privacy in service quality,
conservation and edge computing. Table 1 shows some of the detailed reviews about
edge computing in different application domains, also with future directions.
This review commends of an Edge Computing worldview is silent at a primary
organize, and the logical communal requirements additional period to get it in what
way to utilize its preferences to provide the administrations accessible to end-user
clients. In expansion, the study centers as it were on the focal points of a superiority
computing deprived of tending to great contests, such as the determination of the
finest procedures that IoT operators got to take after in arrange to attain great benefit.
Edge registering brings information handling power at the edge of the system, closer
to the wellspring of the information.
Portable edge computing may be an arrange design that empowers computing and
capacity assets to convey substance over the radio get to organize, and to conclu-
sion clients to make strides the viability of the network [25]. To do this, the MEC
adjusts the load on the radio connection to increase network performance and reduce
the need for long-range backhauling. It is hoped that network requirements will
further spread to IoT and 5G compliant technologies and devices. Mobile Edge
Computing enables operators to more efficiently meet traffic and resource demands
while laying the foundation for the future smart, next-generation networks. Mobile
Edge Computing supports services such as augmented reality, augmented reality,
and the Internet of Things, to give startups and startups the opportunity to adapt to
emerging technologies.
Edge Computing Technology: State-of-the-Art … 459
Table 1 (continued)
Author Methodology Result and discussion Future scope
Zamora et al. [22] Precision agriculture Developed an open Experience the lack of
(PA) with three tier source three tier modularity and
platform platform like local delocalization. For
plane (cyber physical this reason future
system) to gather data, work aims to use
edge plane is for modular turnkey
monitoring and system to falling
managing PA tasks, industrial time
finally cloud plane in situation active and
will collect the current conservation
and past data records.
The system was
implemented for
tomato plants to
monitoring growth of
the crop
Sha et al. [23] Review based on IoT Checked on almost The virtual machine
security at edge layer edge centric IoT, edge and hypervisors are
based security plans basic for the security
and inquire about in entirety edge layer.
issues So these virtual
machine to be
lightweight conjointly
require secure
working framework
like SeL4
Khan et al. [24] Review edge Categorized the Tended restriction of
computing techniques up-to-date in edge edge computing
computing based on worldview. The
the application domain restrictions are
like cloudlet, mist and energetic charging
portable edge component, genuine
computing time application
bolster, security, joint
trade show, repetition,
adaptable design,
come up short over
capabilities and asset
administration
reduce the transmission, analysis and storage of data transmitted to the cloud. Data
collected from IoT sensors and other devices is conveyed to the mist for review and
dispensation. However, these sensors and devices are far from responsive, and cloud
calculations allow short-term analytics and processes at the arrange edge to shrinkage
the sum of information shown to the cloud [27].
3.3 Cloudlets
Cloudlets represent small-scale mist data hubs on the authority of the system and
the second tier of the three-tiered hierarchy: cloud-to-cloud phone or smart device.
Cloudlets aims to development resource-intensive and intelligently versatile appli-
cations by giving versatile gadgets with less idleness and more proficient computing
assets in geographical proximity [28]. This reduces the latency delay usually asso-
ciated with cloud computing in the WAN. While the 5G network is moving up on
the horizon, cloudlets have evolved to support resource-intensive applications such
as machine learning, speech recognition, and language processing applications. Due
to the increased demand, they provide the necessary support for the 5G network that
comes with them. Cloudlets need to be decentralized and spread widely to provide
the greatest amount of network coverage possible. It helps to get power from nearby
mobile computers.
Microdata centers are small, access-level systems that provide all the necessary
components of a traditional data center. Microdata centers are more suitable for some
edge computing applications than traditional data centers because they are usually
smaller in size and can be deployed indoors and outdoors in harsh conditions. This
makes them an ideal solution for edge computing because they can be used locally
in a data source to suit the needs of those who choose to implement them [29].
Microdata centers may be more attractive to small and medium-sized ones—they
may not have data centers because large companies have more resources and do not
need such a clarification. While the growth of the Internet of Effects is driving the
expansion of new shrewd policies and IoT sensors, the souk for the micro data center
is estimated to be worth $32 billion over the next two-and-a-half years.
Some of the edge computing techniques discussed in the next generation network
infrastructure are discussed above. The Internet of Things and the continued advance-
ment of the 5G wireless networking network are viable when changes in other areas,
such as versatile edge and haze computing, are ready and provide a secure platform
for innovation. MEC, Fog Computing, Cloudlets and Micro Data Center play impor-
tant roles when it comes to the future of network computing, but their first test is
when 5G is ready for roll-out.
462 D. Raghunathan and M. Krishnamoorthi
In general, the application landscape consists of ten different edge computing systems
with different end devices such as IoT devices, mobile devices, firewall nodes and
edge, as well as cloudlets, mobile devices, local servers, PCs, home gateway, router,
server, local cluster, FOG, MEC and seaport.
4.1 Cloudlet
Cloudlet could be a trusted, clever computer or computer cluster that is well connected
to the Web and is accessible for versatile gadgets within the encompassing region
[31]. It upgrades “Cloud to Mobile Device” from “Cloud to Mobile Device”, the basic
two-stage architecture of Mobile Cloud Computing. Cloudlet can support users as
a stand-alone cloud, which is currently becoming a “minor server” or “statistics
focus in the case”. However, cloudlet has not been developed under the name Edge
Computing, and its design and concept of edge computing are due to three key
features, such as softness, great resources, and proximity to customers. As an edge
computing system.
4.2 CloudPath
4.3 PCloud
PCloud [33] coordinating edge computing and capacity assets to bolster consistent
portable administrations. Basically, these assets are virtualized through the primary
virtualization layer called STRATUS [34]. Making a bunch of distributed resources
that can hunt for modern assets to memorize almost changes in assets. The runtime
method is capable for asset determination and allotment through the asset choice
interface; the runtime instrument chooses and interfaces fitting resources based on
diverse choice prerequisites. Beneath the Asset Get to Control Arrangement, the
reserves will be combined to supply comparative administrations for outside applica-
tions and make a modern worldview. In expansion, the PCLoud framework moreover
gives basic administration administrations such as authorization administration and
client information accumulation to control access to assets for other clients.
4.4 ParaDrop
4.5 SpanEdge
arranged for a consolidated information center. In any case, this strategy does not
switch enormous information created by numerous gadgets on the edge of orga-
nize impacts, and when the application requires less idleness and forecast, the
circumstance gets to be more awful.
Another scenario that plays an imperative character in edge reckoning is the manage-
ment of IoT strategies in insolent home environments. One common factor is the
privacy issue of a variety of home appliances. There are two different edge computing
systems, Vigilia and HomePad.
Vigilia
In [39], Vigilia proposed tightening the smart home system by limiting network
access to system devices. An authentication framework has been introduced for
applications that reject default access rules and implement authentication at the
network level for API-granularity. Only check the time spent on the router for those
communications, thereby helping customers secure their home appliances.
HomePad
The HomePad system in [40] also proposed setting IoT applications and presenting
privacy-aware centers to diminish security issues. Home Pad permits clients to char-
acterize protection approaches for how apps get to and handle their information. By
requiring applications to utilize clear data streams, they can utilize Preamble rules to
confirm that they have the potential to damage the security approach indicated amid
establishment.
Edge Computing Technology: State-of-the-Art … 465
IoT and Edge devices are used by lightweight Operating Systems and limited storage
capabilities projected for narrow, special tasks. For example AWS Lambda at the edge
platform the content delivery network (CDN) feature within Amazon Cloud Front
with benefits in the form of speed and the cost by pushing code closer the users
through limited edge compute executions.
Cost and Connectivity
IoT sensors and devices services are managed locally compared to the edge devices.
The edge devices are handled large services inside the infrastructure because of more
number of transactions are processed locally so that less number of data traverses to
and from of the cloud. The connectivity is a main problem while reaching out to an
on-premises data to the cloud.
Security
In distributed data processing, edge computing is the major challenge. The security
in edge computing usable for one edge computing application it’s not applicable
directly migratable to another situation because of variety of reasons for example
heterogeneity of edge devices and communication protocols. In this section, illustrate
that the Edge computing Challenges and Security issues in Table 2 [2].
466 D. Raghunathan and M. Krishnamoorthi
Edge Computing is a new model that gives users the ability to network and store
from remote cloud to IoT and 5G context. The Superiority Computing principle is
auspicious in providing a better user experience for better services and applications.
Newly developed edge computing systems and technologies normally decrease over-
head records dispensation and storage and increase the reliability and effectiveness of
transportable data analytics. This paper also discusses the basic concepts and systems
and tools related to edge computing and classifies them as advanced computing such
as cloudlets, mist, and portable edge computing and micro-data center application
domains. This work gives future researchers an excellent idea to understand the
superiority computing pattern and advance investigation to solve unexpected prob-
lems. Future research should explore investigation tendencies in edge-fog computing
systems.
References
1. Hassan N, Gillani S, Ahmed E (2018) The role of edge computing in internet of things. IEEE
Commun Mag 99:1–6
2. Yaqoob I, Ahmed E, Gani A (2016) Mobile ad hoc cloud: a survey. Wirel Commun Mob
Comput 16(16):2572–2589
3. Liu M, Yu FR, Teng Y (2019) Distributed resource allocation in blockchain-based video
streaming systems with mobile edge computing. IEEE Trans Wireless Commun 18(1):695–708
4. Pace P, Aloi G, Gravina R (2019) An edge-based architecture to support ancient applications
for healthcare industry 4.0. IEEE Trans Ind Inf 15(1):481–489
5. Shaukat U, Ahmed E, Anwar Z (2016) Cloudlet deployment in local wireless networks:
motivation, architectures, applications, and open challenges. J Netw Comput Appl 62:18–40
6. Bao W, Yuan D, Yang Z (2017) Follow me fog: toward seamless handover timing schemes in
a fog computing environment. IEEE Commun Mag 55(11):72–78
7. Ahmed E, Akhunzada A, Whaiduzzaman M, Gani A (2015) Network centric performance
analysis of runtime application migration in mobile cloud computing. Simul Model Pract
Theor 50:42–56
Edge Computing Technology: State-of-the-Art … 467
8. Granjal J, Monteiro E, Silva JS (2015) Security for the internet of things: a survey of existing
protocols and open research issues. IEEE Commun Surv Tutorials 17(3):1294–1312
9. Batalla JM, Vasilakos A, Gajewski M (2017) Secure smart homes: opportunities and challenges.
ACM Comput Surv 50(5):75:1–75:32
10. Jose AC, Malekian R (2017) Improving smart home security: Integrating logical sensing into
smart home. IEEE Sens J 17(13):4269–4286
11. Tiwary A (2018) Internet of Things (IoT) research, architectures and applications. Int J Future
Revolution Comput Sci Commun Eng 4(3):23–27
12. Sharma V, Tiwari R (2016) A review paper on IOT and it’s smart applications. Int J Sci Eng
Technol Res (IJSETR) 5(2):472–476
13. Sarhan QI (2018) Internet of things: a survey of challenges and issue. Int J Internet Things
Cyber-Assur1(1):40–75
14. Shi W, Cao J, Zhang Q, Li Y, Xu L (2016) Edge computing: vision and challenges. IEEE
Internet Things J 3:637–646
15. de Assuncao MD, da Silva Veith A, Buyya R (2018) Distributed data stream processing and
edge computing: a survey on resource elasticity and future directions. J Netw Comput Appl
103:1–17
16. Bilal K, Khalid O, Erbad A, Khan SU (2018) Potentials, trends, and prospects in edge
technologies: fog, cloudlet, mobile edge, and micro data centers. Comput Netw 130:94–120
17. Alonso RS, Corchado JM (2019) A review of edge computing reference architectures and a
new global edge proposal. Futur Gener Comput Syst 99:278–294
18. Luo Y, Peng K (2019) A computation offloading method over big data for IoT-enabled cloud-
edge computing. Futur Gener Comput Syst 95:522–533
19. Alamgir Hossain SK, Anisur Rahman MD (2018) Edge computing framework for enabling
situation awareness in IoT based smart city. J Parallel Distrib Comput 122:226–237
20. Ray PP, Dash D (2019) Edge computing for Internet of Things: A survey, e-healthcare case
study and future direction. J Netw Comput Appl 140:1–22
21. Dhanvijay MM, Patil SC (2019) Internet of Things: a survey of enabling technologies in
healthcare and its applications. Comput Netw 153:113–131
22. Zamora Izquierdo MA, Santa J (2019) Smart farming IoT platform based on edge and cloud
computing. Biosyst Eng 177:4–17
23. Sha K, Yang TA (2019) A survey of edge computing based designs for IoT security. Dig
Commun Netw (DCN) 1–12
24. Khan WZ, Ahmed E (2019) Edge computing: a survey. Futur Gener Comput Syst 97:219–235
25. Ahmed A, Ahmed E (2016) A survey on mobile edge computing. In: 10th International confer-
ence on intelligent systems and control (ISCO). pp 1–8. https://doi.org/10.1109/ISCO.2016.
7727082
26. Yi S, Li C, Li Q (2015) A survey of fog computing: concepts, applications and issues. In:
Proceedings of the workshop on mobile big data ACM. pp 37–42
27. Bao W, Yuan D, Yang Z, Wang S, Li W, Zhou BB, Zomaya AY (2017) Follow me fog: toward
seamless handover timing schemes in a fog computing environment. IEEE Commun Mag
55(11):72–78. https://doi.org/10.1109/MCOM.2017.1700363
28. Shaukat U, Ahmed E, Anwar Z, Xia F (2016) Cloudlet deployment in local wireless networks:
motivation, architectures, applications, and open challenges. J Netw Comput Appl 62:18–40
29. Yaqoob I, Ahmed E, Gani A, Mokhtar S, Imran M, Guizani S (2016) Mobile ad hoc cloud: a
survey. Wirel Commun Mob Comput 16(16):2572–2589
30. Liu F, Tang G (2019) A survey on edge computing systems and tools. Proc IEEE 107(8):1–24
31. Satyanarayanan M, Bahl V, Caceres R, Davies N The case for vm-based cloudlets in mobile
computing. IEEE Pervasive Comput
32. Mortazavi SH, Salehe M, Gomes CS, Phillips C, de Lara E (2017) Cloudpath: a multitier
cloud computing framework. In: Proceedings of the second ACM/IEEE symposium on edge
computing. ACM, p 20
33. Jang M, Schwan K, Bhardwaj K, Gavrilovska A, Avasthi A (2014) Personal clouds: sharing
and integrating networked resources to enhance end user experiences. In: Proceedings of IEEE
INFOCOM. IEEE, pp 2220–2228
468 D. Raghunathan and M. Krishnamoorthi
34. Jang M, Schwan K (2011) Stratus assembling virtual platforms from device clouds. In:
Proceedings of the IEEE international conference on cloud computing. IEEE, pp 476–483
35. Liu P, Willis D, Banerjee S (2016) Paradrop: enabling lightweight multi-tenancy at the networks
extreme edge. In: IEEE/ACM symposium on edge computing (SEC). pp 1–13
36. Sajjad HP, Danniswara K, Shishtawy A, Vlassov A (2016) Spanedge: towards unifying stream
processing over central and near the-edge data centers. In: IEEE/ACM symposium on edge
computing (SEC). IEEE, pp 168–178
37. Xu Z-W (2014) Cloud-sea computing systems: towards thousand-fold improvement in
performance per watt for the coming zettabyte era. J Comput Sci Technol 29(2):177–181
38. Fielding RT, Taylor RN (2000) Architectural styles and the design of network-based software
architectures, vol 7. University of California Irvine Irvine, USA
39. Trimananda R, Younis A, Wang B, Xu B, Demsky B, Xu G (2018) Vigilia: securing smart
home edge computing. In: Proceedings of the IEEE/ACM symposium on edge computing
(SEC). IEEE, pp 74–89
40. Zavalyshyn I, Duarte NO, Santos N (2018) Homepad: a privacy-aware smart hub for home
environments. In: Proceedings of the IEEE/ACM symposium on edge computing (SEC). IEEE,
pp 58–73
41. Xiao Y, Jia Y, Liu C (2019) Edge computing security: state of the art and challenges. Proc IEEE
107(8)
42. Liu S, Liu L, Tang J (2019) Edge computing for autonomous driving: opportunities and
challenges. Proc IEEE 107(8)
Comparison of Rectangular Tunnel
with Shield Jacking Support and Pipe
Roof Support in Finite Element Method
Abstract Methods of the Gaussian distribution curve and Finite Element Method
(FEM) were applied in predicting ground settlement however, those equations are
still incomplete with a trough width parameter, K of soil condition that has evolved
from the developing of Kenny Hills tunnel in Malaysia. Therefore, this paper of
research analyses pipe roofing and shield support of rectangular tunnel using FEM
by PLAXIS 2D software to estimate in occurrences of ground deformations. Charts
of maximum ground surface settlement and trough width parameter are developed to
show the difference between both lining supports through parametric study comprises
variation of soil properties. Results show that pipe roofing support is suitable to be
adopted in weak soil condition while shield support is suitable in soil condition
with higher stiffness value. K can be concluded with value equal to 0.6 for the
rectangular pipe roofing support and 0.7 for rectangular shield support in Kenny
Hills soil formation.
1 Introduction
Rectangular tunnel is becoming popular and has been used in constructions especially
by developed country. Hence, for the past few decades improvements have been
done on its support systems in order to prevent deformation of ground especially on
a weak ground condition. Generally, shield and pipe roofing support were used for
constructing rectangular shaped tunnel.
Fang et al. [1] have go through on shield support and Koyama [2] has discussed
further on the benefits of the shield support. Koyama [2] found shield support can
be applied in different type of soil condition [2] due to rigid structure of shield while
pipe roofing support can prevent the settlement of ground above the tunnel using a
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 469
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_45
470 M. Y. M. Nasir et al.
series of steel pipe by consolidate the ground stress and disperse the ground stress
to reduce the excavation stress during the tunnel excavation work [3]. The design
of the pipe roof support may vary between the project because it solely based on
experience [4].
Ground deformation is a major problem that crucial to be solved prior to develop
the underground systems which is not able to be seen directly through underground
layout. Furthermore, different properties and different characteristics of underground
soil at different coordinate makes prediction on ground deformation hardly to be
calculated manually. Therefore, simulation will be done to predict amount of risk
and to identify the affected surrounding and structure constructed in underground
space.
In order to determine the settlement above the tunnel, Peck [5] proposed an equa-
tion to develop transverse settlement above the tunnel known as Gaussian distribution
curve as shows in Fig. 1.
Equation 1 give a settlement at various point of the trough,
x2
Sx = Smax e− 2i x 2 (1)
where:
Sx the settlement profile at the surface
S max the maximum vertical settlement
ix the trough width parameter which, physically, is the distance from the tunnel
axis to the point of inflection of the curve.
Volume of the surface settlement profile is given by,
√
Vs = 2πi x Smax (2)
Comparison of Rectangular Tunnel with Shield Jacking Support … 471
i x = K z0 (3)
where
K constant value
z0 depth of the tunnel axis.
The width through parameter, K, varies with the type of soil. Kimura and Mair
[6] suggested that K values must be equal to 0.5 for clay soil. Meanwhile, O’Reilly
and New [7] recommend that K values must be between 0.2 and 0.3 for granular
soil with condition of tunnel depth less than 10 m, 0.4–0.5 for stiff fissured clay,
0.5–0.6 for Glacial deposits and 0.6–0.7 for silty clay deposits. O’Reilly and New
[8] also had suggested K values must be 0.4 for stiff clays, 0.7 for soft silty clays. In
contrast, Mair and Taylor [9] suggested differently in K values where K values that
they emphasized were 0.5 for all clay soil and 0.35 for granular soils. Khoo et al. [10]
stated that an appropriate K to be used was 0.5 for soil type encountered in Klang
Valley of Malaysia. On the other hand, Yeates [11] suggested that K value should
be from 0.2 to 0.3 for granular material above the water table. Moreover, Rankin
[12] had proposed K value which were 0.4–0.5 for stiff fissured clay, 0.5–0.6 for
glacial deposits, 0.6–0.7 for silty clay. Overall K is depending on soil properties that
significant to consider even though K value does not include in the Gaussian curve
distribution that mostly used by researchers, due to none of the research had found
K value that specifically use for rectangular or box tunnel. In addition, all trough
width parameter is produced by analyze the circular tunnel cross-section. Hence,
this research study aims at producing charts for the tunnel designer in estimating the
maximum settlement and searching tough width parameter value for the rectangular
tunnel with either shield support or pipe roof support located at Kenny Hill soil
condition.
Mostly, Kuala Lumpur, Malaysia covers Kenny Hill soil formation. This type of
soil consisting of interbedded shales, mudstone, siltstone, and sandstones. Kenny
Hills soil has undergone some metamorphic events resulting in changes of sand-
stone/siltstone to quartzite and schist/phyllite respectively. As stated by Ooi [13], the
depth of Kenny Hills soil layer will be more than 10 m below the existing ground
level and the soil formation becomes very hard with SPT greater than N = 50.
The investigation of the Kenny Hills engineering properties has been done by
Refs. [14, 15]. From their study, the measured bulk unit weight mostly ranged from
15.8 to 21.9 kN/m3 for residual soil and for highly weathered rock (Grade IV) is
24.0 kN/m3 .
For the effective shear strength parameters, [15] has stated that for the residual
soil with SPT ≤ 100, the range for cohesion, c is from 5 to 10 kN/m2 and for angle
472 M. Y. M. Nasir et al.
of friction, ϕ is 28° while for the soil which has SPT greater than 100, the cohesion,
c is 15 kN/m2 and angle of friction, ϕ is 29°. For highly weathered rock (Grade
IV), the equivalent Mohr–Coulomb strength parameters are 30 kN/m2 for cohesion,
c and 34° for an angle of friction, ϕ [16].
3 Method of Analyses
Each simulation in PLAXIS 2D software will have one variable parameter and
constants for other parameters by Hardening Soil method in as in Table 2.
c = 5 kPa
E = 25 MPa
v = 0.3
Stiffness E (MPa) ysat = 20 kN/m3 E = 20, 30, 50, 75, 100, 150, 200, 250
yunsat = 19 kN/m3
c = 5 kPa
ϕ = 28
v = 0.3
v ysat = 20 kN/m3 v = 0.1, 0.15, 0.2, 0.25, 0.3
yunsat = 19 kN/m3
c = 5 kPa
ϕ = 28
E = 25 MPa
4 Result
The amount of soil settlements affects the degree damage of the ground surface.
Therefore, factors of settlement need to be identified at first in order to ensure least
ground deformation. Water content within the soil is one of the factors that determine
the level of settlement. Figure 2 proven a higher level of water in soil properties that
cause greater settlement from applying for rectangular shield supports. However,
rectangular pipe roof supports show a striking effect of constant settlement although
water content level is increasing. This condition occurs because pipe roof support
provide advance protection for soil deformation caused by tunnel excavation work
474 M. Y. M. Nasir et al.
while shield support provide soil protection just before the excavation. Next, cohesion
and angle of friction as illustrated in Fig. 3 also affecting soil interaction when the
pattern of settlement is decreasing inversely to cohesion and angle of friction of
the soil with rectangular shield support. On the other side, rectangular pipe roofing
support will be maintaining a constant settlement with increasing value of cohesion
and angle of friction. Hence, constructing a rectangular pipe roofing support can
minimize the soil interaction thus reduce soil deformation at any degree of cohesion
and angle of friction. Figure 4 illustrated a comparison between Young’s modulus
and Poisson’s ratio towards soil settlement and results show reducing settlement with
increasing Young’s Modulus value on rectangular pipe roofing support which similar
to the result of rectangular shield support. However, pipe roof supports produce less
settlement reduction compared to shield support because of steel pipe installation
for pipe roof support has disturbed the soil stiffness around the support’s perimeter.
Poison’s ratio shows no effect toward the ground surface settlement for both supports.
By fitting the Gaussian distribution curve graph into the FEM graph, the value of the
trough width parameter can be produced as a reference to estimate ground surface
settlement in constructing any rectangular tunnel project which has approximate
similar soil condition. Hence, investigations to determine a trough width parameter,
K were done and mostly K value for rectangular pipe roofing support is 0.6 and K
value for rectangular shield support is 0.7. In this study, the range of the trough width
parameters are between 0.4 and 0.7 effected by saturated unit weight, angle of friction,
cohesion and Young’s modulus values used. Figure 5 illustrated the increment of
water level within soil due to lesser in K value for rectangular shield support, while
Fig. 6 resulted in inconsistent K value which between 0.6 and 0.7 for both angle of
friction and cohesion properties. Hence, it is significant to further study in angle of
friction and cohesion properties of soil that affected the K value inn understanding
476 M. Y. M. Nasir et al.
Fig. 6 Relationship between trough width parameter and shear strength properties
the stress of soil. Meanwhile, Fig. 7 illustrated reduction on K value from 0.7 to
0.4 when the value of Young’s modulus increases for both types of supports with
approximately constant K value on Possion’s ratio properties. It can be concluded
that unsaturated unit weight and Poisson’s ratio properties give no effect to K value
for rectangular tunnel with shield of pipe roof support.
Comparison of Rectangular Tunnel with Shield Jacking Support … 477
5 Conclusion
Previous literatures are yet discovering the trough width parameter use in constructing
a rectangular tunnel cross-section. Hence, this simulation and analyses were done to
investigate on the maximum settlement and trough width parameter for rectangular
pipe roofing and shield support using Kenny Hill soil condition. The outcome of
applying rectangular pipe roofing resulted in the suitability of constructing this type
of support within the soil with properties of high-water content can still ensure the
reduction of ground surface settlement. Furthermore, the result obtains also shows
that by constructing a rectangular tunnel with pipe roof support, it is only suitable
to be applied for low soil stiffness because the affected area by the pipe excavation
in higher soil stiffness will increase soil plasticity area, hence reduction of ground
surface settlement is lower. Meanwhile, a rectangular tunnel with shield support gives
better respond in reducing the ground surface settlement in soil with high stiffness
as shield support only involves with jacking process. Besides that, stiffness of soil
is important to be considered prior to all construction works because it will affect
the soil deformation during the excavation either rectangular pipe roofing or shield
support. Finally, the trough width parameter value can be concluded with K equal to
0.6 for the rectangular pipe roofing support and 0.7 for rectangular shield support in
Kenny Hills soil formation.
478 M. Y. M. Nasir et al.
References
1. Fang Y, Chen Z, Tao L, Cui J, Yan Q (2019) Model tests on longitudinal surface settlement
caused by shield tunnelling in sandy soil. Sustain Cities Soc 47:101504. https://doi.org/10.
1016/j.scs.2019.101504
2. Koyama Y (2003) Present status and technology of shield tunneling method in Japan. Tunn
Undergr Space Technol 18(2–3):145–159. https://doi.org/10.1016/S0886-7798(03)00040-3
3. Wu K, Shao Z (2018) Effects of pipe roof support and grouting pre-reinforcement on the track
settlement. Adv Civil Eng.https://doi.org/10.1155/2018/6041305
4. Volkmann GM, Schubert W (2007) Geotechnical model for pipe roof supports in tunneling.
In: Proceedings of the 33rd ITA-AITES world tunneling congress. pp 755–760
5. Peck RB (1969) Deep excavations and tunneling in soft ground. In: Proceedings of the 7th
international conference soil mechanics and foundation engineering, Mexico City, State of the
art volume, pp 225–290
6. Kimura T, Mair RJ (1981) Centrifugal testing of model tunnels in soft clay. In: Proceedings of
the 10th international conference on soil mechanics and foundation engineering, Stockholm.
pp 319–322
7. O’Reilly MP, New BM (1982) Settlement above tunnels in the United Kingdom—their
magnitude and prediction. Tunnelling 82:173–181
8. New B, O’Reilly M (1991) Tunnelling induced ground movements: predicting their magnitudes
and effects. In: 4th International conference on ground movements and structures. Pentech
Press, Cardiff, pp 671–697
9. Mair R, Taylor R (1997) Theme lecture: board tunneling in the urban environment. In: Proceed-
ings of the 14th international conference on soil mechanics and foundation engineering.
Balkema, Hamburg, pp 2353–2385
10. Khoo CM, Idris NISI, Mohamad H, Rashid ASA (2018) Numerical evaluation of settlement
trough width parameter. MATEC Web Conf 203:1–9. https://doi.org/10.1051/matecconf/201
820304010
11. Yeates J (1985) The response of buried pipelines to ground movements caused by tunnelling
in soil. In: GEDDES JDF (ed) Ground movements and structures. Pentech Press, Plymouth,
pp 145–160
12. Rankin WJ (1988) Ground movements resulting from urban tunnelling: predictions and effects.
Geol Soc Eng Geol Spec Publ 5(5):79–92. https://doi.org/10.1144/GSL.ENG.1988.005.01.06
13. Ooi TA (1986) Design and construction problems of foundation for the high rise structure in
the Kuala Lumpur Areas. In: Proceedings of IEM-ISSMFE joint symposium on geotechnical
problem, Kuala Lumpur. pp 112–122
14. Toh CT, Ooi TA, Chiu HK, Chee SK, Ting WH (1989) Design parameters for bored piles in a
weathered sedimentary formation. In: Proceedings of the 12th international conference on soil
mechanics and foundation engineering, Rio de Janeiro, vol 2. pp 1073–1078
15. Wong J, Sing M (1996) Some engineering properties of Kenny Hill formation in Kuala Lumpur.
In: Proceedings of the 12th South-East Asian geotechnical conference, Kuala Lumpur. pp
179–187
16. Hoek E, Brown ET (1997) Practical estimates of rock mass strength. Int J Rock Mech Min Sci
34:1165–1186
The Effect of Skill Type on Skill-Gap
in the Nigerian Construction Industry
Abstract Skills gap and its effects has become a serious problem requiring urgent
attention in the Nigeria construction industry. Basic literatures on construction prac-
tice left no doubts that the successful realization of construction projects requires
the committed action of professionals and vocational trade operatives. The objective
was to identify effects of skills-type on skills-gap related to Building projects in
Abuja, Nigeria. Field survey of some building projects sites in Abuja compliment
the secondary information on the existence of skill gap. The sample comprises both
public and private projects, using both criterion-based and stratified random sampling
techniques. 254 questionnaire was administered, 215 were returned, resulting to a
response rate of 85%. The result shows that more than 80% of the Nigeria construction
professionals are all male with years of experience ranges between 16 and 20 years,
with academic qualification ranges from bachelor degree to Doctorate Degree respec-
tively. Number of employees ranges within 10–50 people in most of the construction
sites. From the result it shows that stake holders participation on training and devel-
opment was not encouraging resulting to Poor workmanship, Construction waste,
Rework as some of the major effects of skills-gap. The study therefore recommend
that policy makers should develop skill acquisition programmed which include public
private partnership, developmental training, on the job training. Government should
come up with special scheme for the improvement of skills through the vocational
and certification program to assist and encourage younger generation to various
apprenticeships to reduce occupational segregation, provide the necessary require-
ment and standard training for both young men and women. The national direc-
torate of employment (NDE), the national skills qualification framework (NSQF),
M. S. Aminu
Department of Building Technology, Jigawa State Polytechnic Dutse, Dutse, Nigeria
e-mail: [email protected]
S. U. Kunya (B) · I. Y. Mohammed (B)
Department of Building Technology, Abubakar Tafawa Balewa University, Bauchi, Nigeria
e-mail: [email protected]
S. A. Bustani (B)
Department of Quantity Surveying, Bayero University Kano, Kano, Nigeria
e-mail: [email protected]
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 479
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_46
480 M. S. Aminu et al.
the industrial training fund (ITF) and national board for technical education (NBTE)
should ready to face the challenges of the recent skills-gap economy as we enter the
third decade of the millennium for sustainability and productivity.
1 Introduction
The poor quality of project delivery can be attributed to the gradually diminishing in
size of competent skilled construction workers and the influx of unskilled, inefficient
and dissatisfied workers who see the sector as a last resort. There are indeed a few
that see construction crafts as career worth coming to. This scenario will inevitably
result in skill-gap, an issue of major concern in view of its invariable impact to the
construction industry, citizens and national economy [12]. There are no exact data
statistics of trade skill-gap in the construction industry, however, Medugu et al. [13]
established that, there is skilled craftsmen shortage that needed urgent attention in
the Nigerian construction industry.
Odediran and Babalola [15] observed that, the most needed large quantity
manpower for building construction in Nigeria and globally are artisans and
labourers. Okuntade [16] also observed that, despite the importance of craftsmen
to the construction industry, many of them still remain untrained, even though the
construction industry, all over the world have been implementing skills acquisition
programmed to meet the demand of occasional change in technologies. Bilau et al.
[3] suggested that, construction craftsmen are operatives who contribute skillfully
with their hands in the practical realization of projects in a construction industry.
Siboe [18] stated that, training in the construction industry has a specific purpose,
it should provide experience which develop the labor of employees in the area of
skills, knowledge and attitude. Overtime, training has been a precursor to human
resource planning which is a process for determining and ensuring that an organi-
zation will have enough number of skilled and experience persons available at the
right time, quantity/quality and place, carrying out jobs that can meet the desires of
the organization, which provide satisfaction for workers involved [18].
Skill shortage is the major problem due to the rapid urbanisation in developing
nations as a result to substantial increase in construction activities. This has tremen-
dous consequences resulting to poor workmanship [10]. Lack of adequate training
centers and training programmed in Nigeria, dominantly utilizing unskilled labour to
maximize profit, resulting to constant rework leading to time and cost overruns, and
The Effect of Skill Type on Skill-Gap … 481
construction waste. Ameh and Itodo [2] assert that, most managers of the construc-
tion industry put little emphasis on the effects of construction waste contributed by
unskilled labour to construction projects.
The shortage of skilled craftsmen in Nigeria has also been attributed to boom in
the agricultural sector in recent times. Most craftsmen engage themselves on farming
being a lucrative business, living closer to their families, making them happier than
working for a long period of hours with a meager amount of pay. The quality shortage
was as a result of lack of passion for the job, seeing the sector as a last resort that
doesn’t get promoted, eager of finishing work on time irrespective of its quality.
This has drastic consequences on productivity in Nigeria. Construction waste sent
to landfills has a drastic effects of polluting the environment, reducing the quality of
construction, delay in construction activities leading to time and cost overruns in the
study area.
In spite of the contributions of skilled craftsmen to the development of the
construction industry, little efforts is being put on the issues that relate to the types
of skill-gap and skill-gap effects in the industry. As a result, this is rarely been
addressed and is little understood. There is need to address this problem by providing
a clear theoretical and conceptual framework to understands of the basics and related
concepts of skills-gap and its effects in the construction industry in Abuja Nigeria.
The aim of this research was to investigate the effects of skills-gap on types of
skills-gap in the construction industry in Abuja Nigeria.
To achieve the above aim, the objective was:
(i) To study the effects of skill-gap on types of skills-gap in the Nigerian
construction industry.
According to Kurushi [12], skills is what an individual possesses which may be learnt
informally and/or on-the-job. Skill-Gap is the difference in the skills required to do
the job and the actual skills possessed by the employees. A “Skill-Gap” arises when
a worker does not have all the skills necessary to do the job effectively. Therefore,
skill-gap refer to the situation where a company has employees but they are not
skilled enough to meet the organization’s objectives.
482 M. S. Aminu et al.
Most construction projects has frequent changes in design often result in construction
waste, time delays, cost overruns, quality defects and other negative impacts on
construction. In recent years, many researchers investigated the effects of design
change and more importantly the common causes behind the changes from different
perspectives. Reviewing the existing literature on design changes, causes and effects,
developing two taxonomies for change causes and change effects; and illustrating
how the taxonomies can be used during the design change management process to
reduce construction waste [20].
4.2.2 Rework
One of the most disturbing issues facing organizations in the construction industry
is their inability to become quality focused. As a result sub-standard products and
services often emanate, which inadvertently result in rework. Typically, rework is
caused by errors made during the design and construction process. These errors
appear downstream in the design process and therefore have a negative impact on
a project’s performance [17]. The lack of attention to quality, especially during the
design and construction process, has meant that rework has become an inevitable
feature of the construction work, and the costs have been found to be as high as
12.4% of total project costs. Such costs could be even higher because they do not
represent schedule delays, and other intangible costs of poor quality. To reduce the
cost and effect of rework, an understanding of its causal structure is needed so that
effective prevention strategies can be identified and the effects of rework reduced or
eliminated [17].
The Effect of Skill Type on Skill-Gap … 483
The rate of client dissatisfaction is adversely affecting the timely delivery of construc-
tion projects. This assesses construction stakeholders’ perception of the causes of
client dissatisfaction and its effects on project delivery in a bid to proffer solution in
minimizing its occurrences. This suggest that client’s dissatisfaction related problems
are the main causes of delays while time and cost overruns are the major identifiable
effects of dissatisfaction in construction projects [1].
Change of design or nature of the work is one of the most significant causes of vari-
ation in construction projects, the client financial status can affect project progress,
and this status often leads to change in work schedules, causing delay, affecting
the quality of the project. Hindrance to prompt decision-making process leading to
decrease in productivity and profitability [22].
5 Sample Size
z2 × P 2
S=
E2
where
Z is the z-score having a constant value of 1.96 at 95% confidence level [21].
P is the standard deviation ranges from 20 to 100 distributions (my P = 56).
E is the margin of error having a value of 4.
Causes of
Skills-gap
Current
Effect of Types of
measures of
skills-gap Skills-gap
bridging the gap
Challenges in
bridging the gap
the professionals that participated in the survey. Where Builders has the highest
percentage (39.5%), Architect (18.6%), Civil engineers (37.2%), Quantity surveyors
(9.3%) and other Professionals (4.6%).This indicates that the respondents that partic-
ipated in the survey are all professionals. Table 1 identify the working experience of
respondents are within the range of 11–15 years has the highest percentage (32.5%),
and those within (16–20) year’s also. and above 20 years (9.3%). Despite the highest
percentage goes to the group of 11–15 and 16–20 years of experience, therefore,
the perception of the experienced Expertise is respected in this research. Table 5
further represent the organizational position of respondents in their various compa-
nies, projects manager has (41.8%), (37.2%) for site engineer, and other positions has
13.9%. Table 5 further shows the number of employees in their various organization,
486 M. S. Aminu et al.
with (65.1%) between 10 and 50 workers, (23.3%) between 50 and 200 workers and
(11.6%) above 200 workers.
The objective was achieved by evaluating the measurement and structural models
using the Partial Least Squares Structural Equation Modeling (PLS-SEM) analysis.
The PLS-SEM approach involves a two-staged process of framework evaluation that
involved test of convergent validity, discriminant validity, collinearity diagnostic,
determining R2 and path coefficients [9]. The output of these criteria were gauged
against the recommended benchmarks provided in the extant literature. Table 2 shows
the recommended bench- marks for decision on the parameters.
Figure 2 shows the measurement model for effects of skills-gap on types of skills-gap.
The model consist of five constructs measured in reflective mode [8]. The constructs
are linked with one-headed arrows which indicate causal effect between them. Being
a reflective model, the assessment of the measurement model involves testing the
validity of the model, the indicator loadings, Average Variance Extracted (AVE),
composite reliability and discriminant validity needs to be computed [8]. The main
purpose of assessing measurement model is to test for construct validity. According
to Hair et al. [9] construct validity is the degree to which the indicators of a latent
construct represent the underlying latent construct.
Construct validity is measured by testing the convergent and discriminant validity.
The convergent validity indicates the extent of correlation among the indicators of a
latent construct. On the other hand, discriminant validity test the degree of uniqueness
of a construct relative to other constructs in the research model [19].
According to Hair et al. [9] convergent validity is the extent of correlation among
the indicators of a latent construct. Hair et al. [8] recommended the use of indicator
loading and AVE as criteria for judging convergent validity.
The Effect of Skill Type on Skill-Gap … 487
Table 2 (continued)
Criteria Description Acceptable/Recommended
Measurement model assessment threshold
criteria
R2 (Coefficient of Determination) Amount of variance R2 = 0.20 is considered
explained in endogenous high in consumer
latent variable by all behaviour studies [9]
exogenous latent variable in R2 = 0.25, 0.50, 0.75 are
structural model considered weak,
moderate and substantial
respectively in [8]
R2 = 0.19, 0.33, 0.67 are
considered weak,
moderate and substantial
[5]
β coefficient Standardized regression Higher magnitude
coefficient that allows for a indicates more explanatory
direct comparison between power [7, 9]
coefficients as to their
relative explanatory power
of independent variable
Source Sinkovic et al. [19]
The second condition for ensuring convergent validity is the AVE. It is defined as
the overall mean value of the squared loadings of the indicators of a latent construct
[9]. The AVE measures the amount of variance a latent construct explained the
measurement items with respect to the amount explained by measurement error [6].
Hair et al. [9] recommended 0.50 and above as suitable to ensure that at least two-third
(2/3) of the variance in the indicator is explained by the construct.
Composite reliability (CR) measures the internal consistency of the measurement
model. Hair et al. [9] described it as the degree of inter-correlations among indicators
of a latent construct. The recommended value is 0.70. In an exploratory research,
however, Hair et al. [8] pointed that 0.60 is equally acceptable.
Table 3 shows the outer loading of the indicators, AVE and composite reliability
of the measurement model. As shown in the Table 3, the outer loadings, the AVE
and CR for the two constructs all fall within the recommended thresholds of the
respective parameter desired for convergent validity. The AVEs reported 0.884 while
the CR reported 0.991 for causes of Skill Gap. The AVEs reported 0.881 while
the CR reported 0.991 for challenges in bridging the gap. The AVEs reported 0.887
while the CR reported 0.992 for effects of skills-gap. The AVEs reported 0.895 while
the CR reported 0.989 for measures of bridging the gap. The AVEs reported 0.886
while the CR reported 0.987 for types of skills-gap. The values were well above the
recommended minimum of 0.50 and 0.70 for AVE and CR respectively. Similarly,
all outer loadings of the indicators measuring the constructs reported values above
0.70 which further indicate the achievement of convergent validity.
Table 3 (continued)
Construct Measurement Outer loading T statistics P Values CR AVE
items
ESGG 0.959
ESGH 0.935
ESGI 0.911
ESGJ 0.948
ESGK 0.962
ESGL 0.938
ESGM 0.969
ESGN 0.958
ESGO 0.930
ESGP 0.956
Measures of MBGA 0.949 0.989 0.895
_Bridging MBGB 0.918
the_Gap
MBGC 0.934
MBGE 0.971
MBGF 0.956
MBGG 0.926
MBGH 0.962
MBGJ 0.934
MBGK 0.935
MBGL 0.959
MBGM 0.959
Type TSGI 0.945 0.987 0.886
of_Skill_Gap TSGII 0.916
TSGIII 0.947
TSGIV 0.922
TSGIX 0.952
TSGV 0.930
TSGVI 0.959
TSGVII 0.958
TSGVIII 0.921
TSGX 0.961
shows the result of the HTMT test as effects of skills-gap. The correlation matrix
indicates that the HTMT value reported is below 0.90 that is (0.889 and 0.892) which
implies that discriminant validity was achieved.
492
After assessing the measurement model, next stage is to evaluate the effects of skills-
gap on types of skills-gap. The structural model is evaluated using collinearity assess-
ment, test of significance of path coefficients determination of R2 and effect sizes
[9]. Figure 3 shows the structural model extracted from the PLS software .
The path coefficient measured the relative significance of the independent variables
in a model. According to Hair et al. [8], path coefficient ranges from +1 to −1 where
values tending towards 1 indicate stronger relationship. Apart from magnitude of the
path coefficient, it is equally important for the estimates to be statistically significant.
On the other hand, R2 is considered the overall measure of model performance. It is
an indicator of the combined effects of all the independent variables in explaining the
variation in the dependent variable in a model. Hair et al. [9] provided a threshold
for interpreting the R2 value where R2 of 0.75, 0.50 and 0.25 are considered to
be of substantial, moderate and weak effect respectively. However, in behavioural
research, consumer satisfaction studies and other related areas in management and
social science fields, R2 value of 0.20 are considered of high effect [9]. Table 5
presents the path coefficients for the relationships among the two constructs in the
model. The reported path coefficients showed a very strong positive and statistically
significant effect of the independent variable (Effects of skills-gap) on the types of
skills-gap (β = 0.765, t = 13.781, p = 0.000). As shown in the Table 5 explained
about 97% variation in Measures of bridging the skills-gap.
494
The findings from the research indicated that the skills shortage has directs effects
on all the types of skills-gap. Major effects of these skill-shortage were poor work-
manship, construction waste, rework which has definitely affects the productivity
and profitability of projects in the study area. Some current measures of bridging the
skills-gap include Establishing industry base skills acquisition centers, good wages
and remuneration to skilled artisans. The introduction of apprenticeship training
under the new national skills qualification framework (NSQF) across the country for
the informal skills sector should be upgraded, polytechnic should lead this National
Skills Qualification Framework intervention as trainers.
Poor workmanship apparent in buildings across the country make up skilling
existing Artisans and craftsmen imperative. On the job training, developmental
training will help to improve the situation. However, appropriate framework for
setting up vocational training, building trade schools/centers devoted to skill training
development no doubt will enhance workforce development.
Performance Enhancement: Employers should develop standard for improving
productivity of workforce with appropriate incentives. In addition, they need to have
in place clear and definite strategies, goals and performance metrics.
The Nigerian construction industry should diversify its awareness by establishing
a special link with their foreign counterparts in creating the best ways of handling
sustainable skills development internationally.
The federal government should formulate a regulation mandating all construction
project stake-holders to attend a compulsory workshop on the issues leading to skills
development, as well as their management principles.
The national directorate of employment (NDE), the national skills qualification
framework (NSQF), the industrial training fund (ITF) and national board for tech-
nical education (NBTE) should ready to face the challenges of the recent skills-
gap economy as we enter the third decade of the millennium for sustainability and
productivity.
496
References
1. Akinsiku O, Akinsulire A (2012) Stakeholder’ perception of the causes and effects of construc-
tion delays on project delivery. J Constr Eng Proj Manag 2. https://doi.org/10.6106/JCEPM.
2012.2.4.025
2. Ameh JO, Itodo ED (2013) Professionals’ views of material wastage on construction sites.
Organ Technol Manage Constr Int J 5(1):747–757
3. Bilau AA, Ajagbe AM, Kigbu H, Sholanke AB (2015) Review of shortage of skilled craftsmen
in small and medium construction firms in Nigeria. J Environ Earth Sci 5(1):35–42
4. Becker J, Klein K, Wetzels M (2012) Hierarchical latent variable models in PLS-SEM: Guide-
lines for using reflective-formative type Models. Long Range Plan 45:359–394. https://doi.org/
10.1016/j.lrp.2012.10.001
5. Chin WW (1998) The partial least squares approach to structural equation modeling. Mod
Methods Bus Res 295(2):295–336
6. Fornell C, Larcker DF (1981) Structural equation models with unobservable variables and
measurement error. Algebra Stat
7. Field A (2009) Discovering statistics using SPSS. 3rd Edition, Sage Publications Ltd., London
8. Hair J, Sarstedt M, Ringle C, Mena J (2012) An Assessment of the use of partial least squares
structural equation modeling in marketing research. J Acad Mark Sci 40:414–433. https://doi.
org/10.1007/s11747-011-0261-6
9. Hair JF, Hult GTM, Ringle CM, Sarstedt M (2014) A primer on partial least squares structural
equation modeling (PLS-SEM). SAGE Publication, Inc., Thousand Oaks, California
10. Ibrahim S (2016) Management of material waste and cost overrun in the Nigerian construction
industry: Unpublished PhD thesis submitted to Nelson Mandela Metropolitan University
11. Krejcie RV, Morgan DW (1970) Determining sample size for research activities. Educ Psychol
Measur 30(3):607–610
12. Kurushi P (2015) Construction practices and skills gap in the construction industry. In:
Proceedings of the 45 builder’s conference of the Nigerian institute of building
13. Medugu NI, Majid MR, Bustani SA, Bala K, Abdullahi U, Mbamali I (2011) Craft skills
availability in the Nigerian construction industry: perception of contractors and consultants.
IUP J Infrastruct 9(3):63–69
14. Nunnally JC, Bernstein IH (1994) The assessment of reliability. Psychom Theory 3:248–292
15. Odediran SJ, Babalola MO (2013) Assessment of business development strategies in the
Nigerian construction industry. J Bus Manag 2(1):34–45. https://doi.org/10.12735/JBM.V2I
1P34
16. Okuntade TF (2014) Building construction technician training: it’s relevance to modern
construction industry in Nigeria. Int J Technol Enhancements Emerg Eng Res 2(3):234–4289
17. Proverbs DG, Holt GD, Cheok HY (2000) construction industry problems: the views of UK
construction directors. In: 16th annual ARCOM conference, pp 6–8
18. Siboe IW (2016) Investigating the adequacy of construction planning in Kenya. Doctoral
dissertation, Jomo Kenyatta University of Agriculture and Technology
19. Sinkovics RR, Henseler J, Ringle CM, Sarstedt M (2016) Testing measurement invariance of
composites using partial least squares. Int Mark Rev
20. Sun M, Meng X (2009) Taxonomy for change causes and effects in construction projects. Int
J Project Manage 27(6):560–572
21. Vogt WP, Johnson RB (2011) Dictionary of statistics and methodology: a nontechnical guide
for the social sciences. Sage
22. Zaghloul SM (2006) Effect of poor workmanship and lack of smoothness testing on pavement
life-cycle costs. Transp Res Rec 1539(1):102–109
Propagation Model of Molecular
Communication Based Targeted Drug
Delivery for Atherosclerosis Disease
Therapy
P. Murugesan (B)
Department of BME, Dr.N.G.P Institute of Technology, Coimbatore, India
S. Prabakar
Sona College of Technology, Salem, India
K. Porkumaran
Sri Sairam Engineering College, Chennai, India
e-mail: [email protected]
R. Karthikeyan
Department of ECE, Dr.N.G.P Institute of Technology, Coimbatore, India
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 499
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_47
500 P. Murugesan et al.
1 Introduction
The research and development of Body Area Networks (BAN) has gained maturity
in recent years. Hence the researchers are now working for the In body networks
using Nano machines [1]. The first useful applications of Nano machines are in
medical technology, where they could be used to identify pathogens and toxins from
the samples of body fluid. The reduced size of Nano machines translates into higher
operational speed. They are designed according to their applications. Specialized
Nano machines called nanobots might be designed not only to diagnose, but to treat
the disease conditions, perhaps by seeking out invading bacteria and viruses and
destroying them. In the field of biomedicine, drug delivery and health monitoring
systems are achieved by biological Nano machines. Targeted Drug Delivery System
(TDDS) is an emerging technology that plays a major role in specialized thera-
peutic purposes. It promises to tackle the conventional hurdles in order to achieve a
controlled rate of drug release with precise location [2]. A new network Paradigm
called IoNT is developed for providing solutions in the drug delivery system. There
are four different types of communication in IoNT. The first type is Electromagnetic
communication. It enables Nano machines to communicate over electromagnetic
waves. The second type is Acoustic, where Acoustic energy is used for communi-
cation. The last type is called as Molecular communication. Here, the molecules are
used for communication between each other. This MC plays a major role in Nano
networks. In this article, focus is put forward on analysis of Diffusion based molec-
ular communication, its propagation model and finally numerical analysis is carried
out for MC-TDDS for Atherosclerotic disease therapy.
2 Methodology
Mathematical models are developed in order to compare the quality of the molec-
ular communication. Propagation period is calculated on the probability basis using
Gaussian distribution [5].
For Brownian motion in the blood medium, we apply the Gaussian distribution
as follows,
√ −d 2
f (t) =
1/2
4π Dt exp (1)
4Dt
This is known as the PDF of first hitting time (i.e., time when the molecule first
hit the Nano-receiver or absorption time).
Where D is the diffusion coefficient of the propagating molecules during the prop-
agation process and d is the distance between the Nano-transmitter and Nanoreceiver
in a one dimensional interval (−∞, d]. Diffusion coefficient represents the inclina-
tion of the propagating molecules during the propagation process through the fluid
medium and it can be obtained by the following formula [6]
Kb ∗ T
D= (2)
b
where Kb is a fixed value called the Boltzmann constant, T is the temperature of
the blood medium, and b is also a fixed value representing the drag constant of the
molecule in the fluid medium. The drag constant b is derived for two conditions
based on the size of the drug carrier molecule Scm and the size of the propagating
molecule Sfluid in the fluid (blood) medium [6],
When Scm = Sfluid
b = 4π ηrnm (3)
b = 6π ηrnm (4)
η represents the viscosity of the fluid medium and rnm is the radius of the propa-
gating molecule in the fluid medium. According to author [7], liposomes play a major
502 P. Murugesan et al.
role in drug delivery systems. At the systemic level, drug bioavailability is increased
due to the high relative surface area of nano particles and it has been shown that
liposomes around 150–200 nm in diameter remains in the blood stream larger than
those with diameters less than 70 nm. Liposomes exhibit Brownian motion which is
directly proportional to temperature and inversely to the liposome size.
Since the diseased vessel has a blood flow with increased velocity, i.e., fluid
medium has a medium drift velocity, we go with Inverse Gaussian model [8] as
follows,
√
νt − d 2
f (t) = 4π Dt exp −
d/2
3 (5)
4Dt
This part deals with the numerical analysis for the following conditions,
PDF of latency for different distances in blood,
• PDF of Latency for different shear rates with different Drift velocity
• Comparison of different shear rates.
Fig. 1 PDF of Latency for different distances {2, 6, 10}µm, constant drift velocity 1 µm/s and rnm = 75 nm
503
504
Fig. 2 PDF of Latency for different distances {2, 6, 10}µm, constant drift velocity 1 µm/s and rnm = 100 nm
P. Murugesan et al.
Propagation Model of Molecular Communication Based … 505
decreasing temperature and minimum shear rate has a greater chance for propagating
the molecule with very less time delay.
In this part, we compare different shear rates in a blood medium for the PDF of the
latency. Here, we have considered the latency in the blood medium with different
shear rates {1sˆ−1, 10sˆ−1, 50sˆ−1, 100sˆ−1, 1000sˆ−1}.
506
Fig. 3 PDF of latency for different drift velocity {1, 2, 2.5, 3}µm/s, minimum distance −2 µm and rnm = 75 nm
P. Murugesan et al.
Propagation Model of Molecular Communication Based …
Fig. 4 PDF of latency for different drift velocity {1, 2, 2.5, 3}µm/s, minimum distance −2 µm and rnm = 100 nm
507
508
Fig. 5 PDF of latency for different drift velocity {1, 2, 2.5, 3}µm/s, maximum distance −10 µm and rnm = 75 nm
P. Murugesan et al.
Propagation Model of Molecular Communication Based …
Fig. 6 PDF of latency for different drift velocity {1, 2, 2.5, 3}µm/s, maximum distance −10 µm and rnm = 100 nm
509
510 P. Murugesan et al.
In this section, the higher PDF rate obtained from different conditions using Inverse
Gaussian distribution methodology is compared for two different Liposome radii.
Propagation Model of Molecular Communication Based …
Fig. 7 PDF of latency for different shear rates {1s−1 , 10s−1 , 50s−1 , 100s−1 , 1000s−1 } Minimum distance 2 µm and Minimum drift velocity 1 µm
511
512
Fig. 8 PDF of latency for different shear rates {1s−1 , 10s−1 , 50s−1 , 100s−1 , 1000s−1 } for minimum distance 2 µm and Maximum drift velocity 3 µm
P. Murugesan et al.
Propagation Model of Molecular Communication Based …
This comparison predicts the role of molecular radius inside the propagation
medium. Let consider two different radius of the Liposome molecule 75 nm and
100 nm respectively.
From the Table 1, it is very clear that Liposome molecule with radius of 100 nm
has a greater efficiency to propagate inside the blood medium with less time delay. By
comparing the higher probability values for both cases, the higher most probability
takes place in the case 2 for all the conditions. Hence, molecule with increasing
radius will play a major role in the propagation process in targeted drug delivery
system.
4 Conclusion
References
1. Tiwari G, Tiwari R (2012) Drug delivery system—an updated review. Pharma Investig
2. Nasrdlahzadeh M, Sajadi SM, Sajjadi M, Issaabadi Z (2019) An introduction to nanotechnology.
Interface Sci Technol
3. Akyildiz IF, Jornet JM, Pierobon M (2010) Propagation models for nano communication
networks. In: Proceedings of the fourth European conference on antennas and propagation.
pp 15–19
4. Kadloor S, Adve RS, Eckford AW (2012) Molecular communication using Brownian motion
with drift. IEEE Trans Nanobiosci 11(2):89–99
5. Berg HC (1993) Random walks in biology. Press Princeton University Press, Princeton
6. Kuran MŞ, Yılmaz HB, Tugcu T, Özerman B (2010) Energy model for communication via
diffusion in nanonetworks. Nano Commun Netw 1(2):86–95
7. https://www.pharmatutor.org/articles/liposomes-novel-drugdelivery-carrier
8. https://holisticprimarycare.net/topics/topicsn/healthy-aging/1299bloodviscosity-the-unifying-
parameter-in-cardiovascular-diseaserisk.html
Adaptation of Machine Learning
and Blockchain Technology
in Cyber-Physical System Applications:
A Concept Paper
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 517
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_48
518 M. Abdullahi et al.
1 Introduction
From past few years, cybersecurity has become an increasing demand in computa-
tions devices, digital transformation in Cyber-Physical System (CPS) such as smart
manufacturing, smart homes, smart transportation and smart factory, where trust
has to be crucial aspect [1]. Those devices use wireless network as a means of
communication to broadcast their data, hence it is easy to be attacked during normal
network communication [2, 3]. Cybersecurity [4] was major concerned because the
technology moving is moving to industry 4.0 [5] where a large amount of data are
generated by CPS and others smart devices such as IoT [6] are usually used for
real-time monitoring and predictions failure [7, 8]. Many cyber attacked has been
accoutered in past few days which include Ukraine power plant in 2016 and Stuxnet
worm which target a nuclear power plant, also there is cyber-attacked on Australia’s
Maroochy water services in 2002 [9]. Also, those CPS applications are generating
a huge amount of data, storage this communications process is vulnerable against
attack. To overcome this problem Machine Learning and Blockchain can be used,
BT has a distributed database network with time series stamped of tamper-proof
record, while ML will analysis the data traffic through effective, efficient designing
of algorithms [10, 11].
Machine Learning (ML) in the field of Artificial Intelligence focus on building
applications that learn through experience and they make self-decision without
human interference such as Logistic Regression, Linear Regression, and Support
Vector Machine etc. ML was applied to various field of discipline such as statistics,
computational complexity [12, 13] smart devices [14] internet of things [15, 16] and
artificial intelligence to machines are getting smarter [17, 18]. While Blockchain
Technology (BT) are cryptography immutable set of records or distributed ledger
which is link together for auditing, by linking its valid blocks together the current
block contains the hash of the previous block. This process makes the Blockchain
traceable and resistant to change [10, 19].
Cyber-Physical System (CPS) was introduced by Helen Gill in 2006 at National
Science Foundation, they are integrated of both physical and cyber to enhance
the control and operation which include computational, communication, sensing
and actuation, in general, they are dynamics joint of computers, software, phys-
ical process, event processing [20] and networks. The CPS includes operating in
real-time, transactional services and secure communication between components
[21].
The ML processes can be applied to BT for CPS applications to improve their security,
privacy, monitoring [22] and making them smarter. It also gives an opportunity with
improvement in developing good models by using the decentralized architecture of
Adaptation of Machine Learning and Blockchain Technology … 519
There are several research works which address the importance of combining ML
and BT, in this paper, we investigate how ML can be adopted in BT based on CPS
applications. The research contribution for this paper include the followings:
• We find out that ML and BT can be adopted together and a brief discussion on
ML with BT can be used together in CPS applications.
• We also proposed architecture on how ML and BT can be adopted and
implemented together.
The core concept and rationale of ML algorithms with BT can be adopted and imple-
mented in CPS applications to improve their performance in terms of security, moni-
toring and detection activities such as improved security and privacy were ML and
BT covers CPS applications such as data protection, data encryption, instruction
detection system (IDS), intrusion protection system (IPS). They provide an impor-
tant role in the monitoring and detection of malware. Furthermore, BT increased
the privacy between users for both private and public Blockchain through secure
authentication of data accessibility.
Improved efficiency and scalability increasing ability in multiple of CPS appli-
cations from multiple devices that consist of huge datasets due to their multiple
transactions through providing decentralize self-authorization, detection from ML.
Scalability form BT also provides capable of increment between users.
Secure of data sharing in CPS Applications was the BT contains chains blocks of
all data transaction for CPS applications by providing accountability, transparency
within the data access. ML components were included with its algorithms to provide
decentralised platform and removal of fraud fewer datasets.
In this section research question and methodology has been discussed, the research
question was (Can Machine Learning and Blockchain Technology be adopted
520 M. Abdullahi et al.
PRISMA guideline was used to structure the systematic literature review by creating a
question to guide the review process (Can Machine Learning and Blockchain Tech-
nology be adopted together in Cyber-Physical Systems Application?). The search
strategy used for this systematic review, we developed a search strategy to identify
relevant literature. The search strategy was tailored by one database: Scope and search
terms used were the followings: “Machine Learning for Blockchain Applications”
and “Blockchain for Cyber-Physical Systems Application”. All search spanned from
the database inception were until 2020, and included journals articles with few review
papers published in English only.
The selection criteria were mainly focused on the mapping on existing literature on
Machine Learning for Blockchain Applications and Blockchain for Cyber-Physical
Systems Application in the field of computer science, Decision Science and math-
ematics. The search span was from 2020 to 2018. All articles before 2018 below
were excluded from the search. The also was mainly focused on the global search,
not a specific country or region. A total of 95 research articles were excluded at this
stage. There were 42 research article included and extracted at this stage. The quality
assessment is based on the original research and a few review papers. For maintain
the quality of the review, all duplicated were checked thoroughly.
3.2 Methodology
The methodology investigates process, steps to overcome this problem with rational
for CPS applications, by using a combination of two technologies which are ML and
BT with a proposed framework to support this research study (see Fig. 1).
4 Findings
Firstly, providing detection with prevention in CPS applications to handle data secu-
rity and cybersecurity issues within its application. Those methods are intrusion
detection system (IDS), intrusion prevention system (IPS) with ML classifiers. BT
also add additional support through linked blocks using cryptography for prevention
of datasets against attack.
Adaptation of Machine Learning and Blockchain Technology … 521
Internet of
Smart Devices
Things
New
Block 1 Block 2 Block 3
Block
User
Blockchain
We conclude, that recent development of CPS application with their huge increasing
data generated this leads to the increase of cyber-attack. With the development of
ML and BT are helping to overcome these issues by adopting them together in
522 M. Abdullahi et al.
various CPS applications such as power systems, medical systems, smart buildings
and industrials automation. The research paper also presents brief details on ML
and BT with how they can be adopted together in CPS applications using a proposed
architecture. A study comparison has been made which shows methods, contributions
and limitation. We also have some suggestion for further scope.
The future scope on this topic is therefore recommended to provide in details on
how ML and BT can be adopted into specific CPS applications such as Industrial
Automation, Power Systems and Medical Systems or Internet of Things devices.
References
17. Bello A, Mahadevan V (2019) A cloud based conceptual identity management model for
secured Internet of Things operation. J Cyber Secur Mobil 8(1):53–74
18. Al-Saud M, Eltamaly AM, Mohamed MA, Kavousi-Fard A (2020) An intelligent data-driven
model to secure intravehicle communications based on machine learning. IEEE Trans Ind
Electron 67(6):5112–5119
19. Islam A, Young Shin S (2020) A blockchain-based secure healthcare scheme with the assistance
of unmanned aerial vehicle in Internet of Things. Comput Electr Eng 84:1–12
20. Roldán J, Boubeta-Puig J, Luis Martínez J, Ortiz G (2020) Integrating complex event processing
and machine learning: an intelligent architecture for detecting IoT security attacks. Expert Syst
Appl 149:1–22
21. Al-Hadhrami Y, Hussain FK (2020) Real time dataset generation framework for intrusion
detection systems in IoT. Futur Gener Comput Syst 108:1–10
22. Zhong CL, le Li Y (2020) Internet of things sensors assisted physical activity recognition and
health monitoring of college students. Meas J Int Meas Confed 159:1–9
23. Lu Y, Huang X, Dai Y, Maharjan S, Zhang Y (2020) Blockchain and federated learning for
privacy-preserved data sharing in industrial IoT. IEEE Trans Ind Inform 16(6):4177–4186
24. Lima E, Vieira T, de Barros Costa E (2020) Evaluating deep models for absenteeism prediction
of public security agents. Appl Soft Comput J 91:106236
25. Lin X, Li J, Wu J, Liang H, Yang W (2019) Making knowledge tradable in edge-AI enabled
IoT: a consortium blockchain-based efficient and incentive approach. IEEE Trans Ind Inform
15(12):6367–6378
Application of Artificial Neural Network
(ANN) and Adaptive Neuro Fuzzy
(ANFIS) Techniques for the Modelling
and Optimization of COD Adsorption
Process
Abstract Artificial neural network (ANN) and adaptive neuro fuzzy (ANFIS)
modelling techniques have been applied in this study to model and optimize the chem-
ical oxygen demand (COD) adsorptive removal in produced water. The models were
well trained and showed minimum error values for predicted data when compared
to experimental data. The error values were 0.4035 and 0.2886 for sum of squared
error (SSE), 0.1628 and 0.0832 for mean square error (MSE) and 0.13 and 0.23%
for average relative error (ARE) using ANN and ANFIS, respectively. Error anal-
ysis and coefficient of determination (R2 ) of the models determined that ANFIS was
better than ANN for the prediction of COD adsorption on the biochar. Also, ANFIS
required minimum run time as compared to ANN. Both artificial intelligence (AI)
based techniques well predicted the optimized values of adsorption process, when
compared with the experimental values. It is concluded that the use of AI tech-
niques can inevitably pave the way in the water treatment sector using adsorption for
improved efficiency and process automation.
1 Introduction
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 525
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_49
526 H. Khurshid et al.
complex properties are very difficult to handle when it comes to the adsorption treat-
ment process [2]. Due to the interaction of several factors in adsorption process,
the resulting correlations are strongly non-linear [3]. This problem leads to inade-
quate understanding and false quantification of interactions between process inputs
and outputs, and it is thus difficult to explain by linear mathematical models. The
standard regression methods are also not sufficient for mimicking the experimental
adsorption data as the methodology does not grasp the physics [1]. For interrelating
the input and output parameters of adsorption process, advanced computer-simulated
models are therefore required [4].
Artificial intelligence (AI) methods e.g. fuzzy logic (FIS) [5], particle swarm opti-
mization (PSO) [6], genetic algorithm (GA) [7] and artificial neural networks (ANNs)
[8], have been applied recently in water treatment and adsorption optimization and
automation systems. ANNs are particularly well known for their high capacity to
research and arrange data [9]. Researchers have reported different studies using
ANN for modelling and optimization of adsorptive removal of various pollutants
e.g. chromium removal using cyanobacterial biomass [10], copper removal using
light expended clay aggregate [11], removal of various dyes using nanoparticles and
activated carbons [12, 13], removal of As (III) and various other metals [5]. Fuzzy
systems also have some advantages over traditional approaches, particularly where
ambiguous data or prior experience is involved [14]. Few studies have reported appli-
cation of ANFIS in adsorption studies for modeling and optimization of adsorption
process e.g. indium (III) [15], dyes and metals [5]. Meta-heuristic optimization algo-
rithms e.g. GA and PSO, are the products of social actions or natural phenomena
and used to optimize the various operations. The PSO is a well-known optimization
algorithm motivated by the action of a flock of birds. It is now widely used in system’s
optimization because it is not conveniently stuck in a local minimum.
Most of the studies in literature are performed on synthetic waters or stock solu-
tions containing single pollutant as an objective. These solutions do not completely
represent the industrial effluents, which contain multiple types of pollutants together.
Hence, practical application of the adsorptive materials and automation of treat-
ment plants becomes limited considering single pollutant studies. Therefore, the
purpose of this research work was to present the design, implementation, compar-
ison and evaluation of the ANN and ANFIS approaches to the adsorption of chemical
oxygen demand (COD) in produced water (PW). The wastewater obtained during
oil and gas exploration is known as PW. The COD in the water is caused by various
carcinogenic organic pollutants. Organic pollutants have been confirmed to be one
of the top priority pollutants to date including PAHs, dyes, and pharmaceuticals etc.
COD is an important parameter for interpreting the organic pollutants in wastewater
and has been used widely for effluents quality measurement [16]. Various studies
have reported efficient adsorption of COD in wastewater which also represents the
elimination of organic pollutants in the water [17, 18].
To the best of our knowledge, no study has been found in literature containing the
application of ANN and ANFIS for modelling and optimization of COD adsorption
process for predicting organic pollutants removal in PW. Tea waste (TW) had been
transformed into biochar for this purpose and was applied under the control of various
Application of Artificial Neural Network (ANN) … 527
parameters (pH, dosage and contact time). To improve the automated adsorptive
removal of COD from the wastewater sample, adsorption process was optimized
using the ANN-PSO and ANFIS techniques. The detailed objectives of this study
are: to establish ANN and ANFIS models capable of predicting adsorption efficiency
for the removal of COD from PW; Optimizing the COD adsorption process using
ANN-PSO and ANFIS; and to incorporate ANN and ANFIS models such that they
can be used effectively for PW treatment system automation. This study will help
the decision makers and wastewater treatment plant operators in decision making,
strategic designing of system, automation, and optimization of adsorption process.
2.1 Materials
Tea waste was obtained from the nearby restaurants in Malaysia. High range (HR)
COD vials were obtained from the laboratory stock of Avanti’s Malaysia. About 98%,
H3 PO4 was purchased in Malaysia from local supplier Merck. The water sample was
obtained from an oil and gas company operating in South East Asia.
The biochar was produced using the chemical activation technique. For 12 h, tea
waste was immersed in 10% phosphoric acid solution and then left to dry at room
temperature. Biochar was developed by pyrolysis at 700 °C using the dried waste
in the tube furnace. Afterward, the prepared biochar was cleaned with filtered water
and kept in the oven to dry. In this study, the biochar was only used for exploring the
application of AI based techniques in automation of adsorption process and mecha-
nism evaluation was not the objective. Therefor the material was not characterized
and used only for the experimental purpose.
In this analysis, the commonly documented adsorption process was used to observe
the elimination of COD [19]. A significant segment of AI configuration is the selec-
tion of input variables from a variety of parameters that influence the mechanism
[1]. To select different important factors affecting the adsorptive removal of COD
from PW, batch adsorption tests were performed. From the experiments and litera-
ture review, it was found that initial pH of water, dosage of adsorbent and contact
528 H. Khurshid et al.
time were most significant input parameters effecting the adsorption process. There-
fore, the biochar was applied in varying dosage (25–300 mg/L) for 100 mL of PW at
varying initial pH (3–12) and shook for various contact times (10–60 min) at temper-
ature 20 ± 5 °C. The percentage of COD removal was detected by the formula given
[20]:
(Ci − Co ) × 100
Percentage removal (% ) = (1)
Ci
where, Ci and Co represent the initial and final concentrations of COD in mg/L,
respectively.
A three-layer feed forward neural network was formed using the neural network
toolbox using programming codes written in MATLAB R2019b mathematical soft-
ware. The input layer was given three inputs i.e. pH, contact time and dosage of
biochar and one output e.g. removal efficiency of COD as shown in Fig. 1. Number
of neurons in hidden layer were selected based on hit and trial method using 1–10
neurons. Higher number of neurons were not tested to avoid the over fitting of model.
Cascade forward backpropagation (CFBP), Levenberg–Marquardt backpropagation
(LMBP), Bayesian Regularization (BR) and Scaled Conjugate Gradient (SCG) algo-
rithms were tested to find the best suitable algorithm. The most suitable algorithm
was selected based on minimum value of mean square error (MSE). For all data sets
in ANN, the symmetric sigmoid transfer function (tansig) was used in the first layer
and the linear transfer function (purelin) was used at the output node for the simula-
tion and prediction of COD elimination. The backpropagation (BP-ANN) network
was selected for the tuning of parameters. It is a first-order gradient descent technique
to model experimental data [21]. Total 36 no. of data sets were taken through batch
experiments. 70% data was used for training of model and 30% data was used for
testing purpose. For validation of model and prediction of COD removal efficiency
13 more data values were taken through lab experiments using Box–Behnken Design
(BBD) of experiments.
ANN and PSO hybrid was used for the optimization of input variables using ANN
predicted data with 1000 no of population, C1 = 1.5, C2 = 2.5 and 1000 iteration
values. The schematic diagram of hybrid model is shown in Fig. 2. MSE was taken
as function to be minimized. Tolerance value was taken as 1. Boundary violations
were taken as 1.5 and −1.5. The algorithm was run for optimization further. The
purpose was to obtain the input variables causing the maximum removal of COD in
PW.
The fuzzy inference method (FIS) applied to ANFIS was developed in the MATLAB
using Neuro-Fuzzy designer. The developed ANFIS was of the Sugeno type, which
consist of four hidden layers [15]: (i) the fuzzification layer with two membership
functions; (ii) the inference layer with weight rules; (iii) the de-fuzzification layer; and
(iv) the output layer (Fig. 3). Data was randomly divided into training (70%) and
testing data (30%). Three variables (pH, Dosage and contact time) were selected
as inputs and removal efficiency of COD was taken as target. Range of influence
was selected based on minimum MSE. To obtain the optimized results for COD
adsorption using ANFIS, surface plots were generated.
The performance of the ANN and ANFIS techniques for predicted data was evaluated
using statistical equations i.e. (i) the coefficient of determination (R2 ), (ii) the sum
of squared errors (SSE), (iii) mean–squared error (MSE) and (iv) average relative
error (ARE) [22]. Following equations were used to measure the errors [23]:
Application of Artificial Neural Network (ANN) … 531
n 2
i=1 y pr ed,i − yexp,i
R =1−
2
n 2 (2)
i=1 y pr ed,i − ym
n
1 2
SS E = y pr ed,i − yexp,i (3)
n i=1
1 2
n
MSE = y pr ed,i − yexp,i (4)
n i=1
where, ypred and yexp denote the predicted and experimental values, respectively. n
denotes the total number of values in data. Mean value of the response is denoted by
ym .
The ANN structure was opted after analyzing the various parameters i.e. no. of
neurons and training algorithm. The algorithm and no. of neurons were selected
based on the best performance as shown in Fig. 4a, b. MSE was taken as perfor-
mance criteria. Total 10 no. of neurons and LMBP algorithm were selected based on
minimum MSE and simulation time. After selecting the structure of ANN, selected
model was trained by using trainlm training function. It was observed that, after
training, the output data had a correlation coefficient (R2 ) of 0.99 with the exper-
imental data (Fig. 4c). The MSE value was 0.0001 and training time was 400 s.
The lower value of error indicated efficient performance of selected ANN model for
the given dataset. For the validation and testing of model, correlation coefficients of
0.998 and 0.98 were obtained respectively (Fig. 4d).
Higher values of R2 showed that the model was well trained and could be used
effectively for the prediction of COD adsorption on biochar. The model was validated
further and a correlation coefficient of 0.99 was obtained for the predicted data when
compared with experimental data using 13 dataset points, as given in Table 1.
532 H. Khurshid et al.
Fig. 4 Modelling of ANN; a relationship between MSE and no. of neurons in hidden layer, b
relationship between CFBP, LMBP, BR and SCG algorithms and their performance, c ANN training,
d ANN testing
The PSO technique was hybridized with the generated ANN model for optimizing
process parameters with the goal of maximizing the COD removal efficiency. The
performance of ANN-PSO model was analyzed on basis of MSE i.e., 5.27. Minimum
value was obtained after 500 runs and it took time of 4000 s. The optimum conditions
for the COD removal process were: 11.79 pH, 300 mg/L concentration of biochar
and 60 min contact time. The COD removal efficiency obtained under optimum
conditions was 89.80%. The value was confirmed in the lab and obtained as 89.3%.
It indicated the good performance of ANN-PSO in predicting the COD adsorption
optimized process parameters.
Application of Artificial Neural Network (ANN) … 533
Table 1 Experimental and predicted results of COD removal efficiency using ANN and ANFIS
No pH Dosage Contact time COD removal efficiency (%)
mg/L min Actual Predicted ANN Predicted ANFIS
1 10 162.5 60 53.123 52.31 52.123
2 3 162.5 60 36.611 36.12 36.59
3 6.5 25 60 69.269 69.01 69.16
4 6.5 300 60 78.044 78.77 78.12
5 6.5 162.5 35 63.918 64.66 63.91
6 10 25 35 54.296 54.21 54.3
7 6.5 25 10 65.574 65.57 65.57
8 3 162.5 10 36.892 36.72 36.89
9 3 25 35 39.169 39.23 39.18
10 10 162.5 10 44.941 45.01 44.9
11 3 300 35 49.431 49.43 49.39
12 10 300 35 60.831 60.85 61.02
13 6.5 300 10 72.958 72.8 72.8
R2 0.9991 0.9995
SSE 0.4035 0.2886
MSE 0.1628 0.0832
ARE 0.13% 0.23%
Sugeno type sub-clustering ANFIS was generated for three inputs and one output.
For all input variables membership functions were taken as 35 in number and guassmf
type. Range of influence was selected as 0.00001 based on minimum MSE value,
when tested from 0–1 as shown in Fig. 5a. Whereas, squash factor, accept ratio and
rejection ratio were 1.25, 0.5 and 0.15, respectively. For training of generated FIS
structure hybrid optimization method was taken as method type due to lower MSE
(0.15721) as compared to backpropagation optimization method (2.32). Total 100 no
of epochs were taken for training of ANFIS model using 70% dataset. The model was
well trained and minimum MSE of 0.15721 was obtained at epoch 2 (Fig. 5b). The
run time of ANFIS was only 10 s for 100 epochs which was much lesser than ANN.
The model was further tested and checked using the 30% data. The MSE of 0.00017
and 0.0008 was obtained for the test and checked data, respectively (Fig. 5c, d).
After training, the ANFIS model was used for prediction of COD removal efficiency
using 13 input datasets. The predicted outputs were compared with the experimental
data as given in Table 1. A higher correlation of 0.9995 and lower MSE of 0.0832
was obtained for the predicted data. It showed that generated ANFIS model could
be used effectively for the optimization of COD adsorption process and selecting
suitable input parameters.
534 H. Khurshid et al.
Fig. 5 Modelling and contour plots for COD adsorption on biochar using ANFIS; a error analysis,
b ANFIS training data, c ANFIS testing data, d ANFIS checked data, e impact of pH and dosage
on COD removal efficiency, f impact of pH and contact time on COD removal efficiency, g impact
of dosage and contact time on COD removal efficiency
For the prediction of optimized values of adsorption data through ANFIS surface
plots were generated for pH range 3–12, dosage 25–300 mg/L and contact time 10–
60 min. Figure 5e–g shows that COD removal efficiency was highly affected by all the
three variables. It also shows the correlation between each variable. Maximum value
of COD removal was obtained at pH value of 12, dosage of 300 mg/L and contact
time of 60 min. The COD removal efficiency achieved at the optimized parameters
was 90.8%. The experimental value for the optimized input parameters was 89.3%.
It showed that the model was able to predict the optimized values with significant
accuracy.
Developed ANN and ANFIS models were well trained for predicting the COD
removal efficiency using pH, dosage of biochar and contact time as input variables.
Both models well fitted the experimental data and predicted values were near to
Application of Artificial Neural Network (ANN) … 535
experimental data (Table 1). Whereas ANN-PSO and ANFIS models also performed
very well for the optimization of adsorption process (Table 2). Further performance
of the models was analyzed using Eqs. 2–5 for error analysis of predicted data. The
analysis is given in Table 1. The value of R2 should lie between 0–1. Value near
to 1 shows good correlation between the data sets. SSE has values range from 0–1,
while best value is closer to 0 [4]. MSE minimum value is taken as better value
[24], whereas ARE is %age error and minimum value is recommended [25]. Table
1 shows that ANN and ANFIS had R2 values of 0.9991 and 0.9995, error values
of 0.4035 and 0.2886 for SSE, 0.1628 and 0.0832 for MSE and 0.13% and 0.23%
for ARE, respectively. Error analysis and coefficient of determination of the models
determined that ANFIS > ANN in performance for the prediction of COD adsorption
on the biochar. Also, for both methods no specific design of experiments was needed.
It helped the models to analyze the correlation between inputs and outputs in broader
range.
4 Conclusion
Tea waste biochar was synthesized and applied for treatment of produced water. Batch
tests were performed using pH, dosage, and contact time as input variables. The exper-
imental results were modelled and optimized using ANN and ANFIS models. ANN
model was modelled using 3 layered, feed forward Levenberg–Marquardt backprop-
agation algorithm and 10 no. of neurons. For optimization it was hybridized with
PSO. ANFIS was generated sugeno type using sub-clustering FIS type. The exper-
imental values were successfully optimized using surface plots correlations. Opti-
mization values were well matched with the experimental results. ANFIS required
minimum run time (10 s for 100 epoch) as compared to ANN model (400 s for
1000 epoch). Error analysis and coefficient of determination of three models deter-
mined that ANFIS was slightly better than ANN in performance for the prediction of
COD adsorption on the biochar. However, it can be stated that the ANN and ANFIS
methods are promising techniques and can be used effectively for the automation of
COD adsorption process for produced water treatment. Both methods have high capa-
bility to predict the adsorption efficiency of pollutants considering multiple factors
and can be implemented for sustainable and automated water treatment systems.
536 H. Khurshid et al.
Acknowledgements The authors would like to acknowledge the financial support provided for
this study under YUTP grant with cost center 015LC0-190.
References
14. Sargolzaei J, Haghighi Asl M, Hedayati Moghaddam A (2012) Membrane permeate flux and
rejection factor prediction using intelligent systems. Desalination 284:92–99. https://doi.org/
10.1016/j.desal.2011.08.041
15. Franco DSP, Duarte FA, Salau NPG, Dotto GL (2020) Analysis of indium (III) adsorption
from leachates of LCD screens using artificial neural networks (ANN) and adaptive neuro-
fuzzy inference systems (ANIFS). J Hazard Mater 384:121137. https://doi.org/10.1016/j.jha
zmat.2019.121137
16. Malakahmad A, Law MX, Ng KW, Manan TSA (2016) The fate and toxicity assessment of
polycyclic aromatic hydrocarbons (PAHs) in water streams of Malaysia. Procedia Eng 148:806–
811. https://doi.org/10.1016/j.proeng.2016.06.572
17. Mohammad-pajooh E, Turcios AE, Cuff G, Weichgrebe D, Rosenwinkel KH, Vedenyapina
MD et al (2018) Removal of inert COD and trace metals from stabilized landfill leachate by
granular activated carbon (GAC) adsorption. J Environ Manage 228:189–196. https://doi.org/
10.1016/j.jenvman.2018.09.020
18. Bansode RR, Losso JN, Marshall WE, Rao RM, Portier RJ (2004) Pecan shell-based granular
activated carbon for treatment of chemical oxygen demand (COD) in municipal wastewater.
Biores Technol 94(2):129–135. https://doi.org/10.1016/j.biortech.2003.12.009
19. Ademiluyi F, Amadi S, Amakama N (2010) Adsorption and treatment of organic contaminants
using activated carbon from waste Nigerian Bamboo. J Appl Sci Environ Manage 13(3). https://
doi.org/10.4314/jasem.v13i3.55351
20. Azari A, Mahmoudian MH, Niari MH, Eş I, Dehganifard E, Kiani A et al (2019) Rapid and
efficient ultrasonic assisted adsorption of diethyl phthalate onto FeIIFe2IIIO4@GO: ANN-GA
and RSM-DF modeling, isotherm, kinetic and mechanism study. Microchem J 150:104144.
https://doi.org/10.1016/j.microc.2019.104144
21. Olawoyin R (2016) Application of backpropagation artificial neural network prediction model
for the PAH bioremediation of polluted soil. Chemosphere 161:145–150. https://doi.org/10.
1016/j.chemosphere.2016.07.003
22. Dutta M, Basu JK (2013) Application of artificial neural network for prediction of Pb(II)
adsorption characteristics. Environ Sci Pollut Res 20(5):3322–3330. https://doi.org/10.1007/
s11356-012-1245-x
23. Khoshsang H, Ghaffarinejad A (2018) Rapid removal of lead (II) ions from aqueous solutions
by saffron flower waste as a green biosorbent. J Environ Chem Eng 6(5):6021–6027. https://
doi.org/10.1016/j.jece.2018.09.020
24. Dil EA, Ghaedi M, Ghaedi A, Asfaram A, Jamshidi M, Purkait MK (2016) Application of
artificial neural network and response surface methodology for the removal of crystal violet by
zinc oxide nanorods loaded on activate carbon: kinetics and equilibrium study. J Taiwan Inst
Chem Eng 59:210–220. https://doi.org/10.1016/j.jtice.2015.07.023
25. Singh DK, Verma DK, Singh Y, Hasan SH (2017) Preparation of CuO nanoparticles using
Tamarindus indica pulp extract for removal of As(III): optimization of adsorption process by
ANN-GA. J Environ Chem Eng 5(1):1302–1318. https://doi.org/10.1016/j.jece.2017.01.046
An Evolutionary Stream Clustering
Technique Outlier Detection in Medical
Data
Abstract Clustering data streams from data mining become evident as one of the
most well-liked studies among the researchers correspond to their evolutionary field.
Numerous threats addressed by data streams on clustering like limited time, memory
and single scan data. In generic terms, data streams are defining an infinite sequence
of the element data and evolve without prior knowledge number of the clusters. Many
factors such as noise (outliers) have appeared periodically have a negative impact
on data streams environment. The density-based technique has demonstrated to be
an astonishing method in clustering data streams. It is computationally competent to
generate arbitrary shape clusters and detect noise instantaneously. The number of the
clusters are not required in advance to set as a parameter during the assessment begin.
In contradictory, traditional density-based clustering is not relevant to conduct in data
streams due to its own characteristics. Mostly all traditional density-based clustering
can be extended to the updated version of algorithms to achieve the objective of data
streams research. The idea is emphasizing on the density-based technique in the clus-
tering process dominate the restrain from data streams nature. The objective of this
paper intends a preliminary result on a density-based algorithm, named evoStream to
explore outlier detection on medical data sets, heart failure clinical records and gene
expression cancer RNA-seq. In due course, extensive evoStream later to develop to
optimize the model to detect outlier in data streams.
1 Introduction
Due to the fact, advance applications are practically implemented in related to various
field such as medical research. Often, data streams relate with as an enormous,
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 539
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_50
540 N. A. Supardi et al.
continuous and infinite data [1]. However, machine learning [2–4] address an issue
of interpretation data due to ambiguity from data streams. Data streams evolve with
time and complicate the stream processing. Data streams size increasing dispro-
portionately and leads to infinite data size [5]. Consequently, the transmission data
streams have influence memory space [6, 7]. Generally, the process to capture data
streams are extremely difficult to determine the number of the cluster and the shape
of the clusters where it considered unknown [8]. The outcome of clustering results
is changing consistently as the data arrives in streams [9–11]. The clustering process
and the shape of the clusters are mostly affected by outliers [11]. An outlier is defined
generally by [12, 13] which appears inaccurate information from the data.
This paperwork highlights the importance for clustering technique in outlier detec-
tion due to following factors [7, 14, 15]; (i) an outlier may distort data structures and
cluster result, (ii) the clustering methods can identify outlier detection as a small
cluster, (iii) the hybrid clustering algorithms are exemplifying the result for outlier
detection performs better than the existing algorithms.
In research studies, evoStream refers to data sets to conduct the clustering process
in the online step during idle periods and to optimize the outcome of the cluster
[16]. Thus, becomes a motivation to research problem. This research, however, was
primarily aimed at evaluating the outlier identification of evoStream implementation
with real datasets called clinical records as preliminary results.
The rest of this paper is arranged as follows: The related works is summarized in
Sect. 2. Methodology described in Sect. 3. The assessment is seen in Sect. 4, and the
result is outlined in Sect. 5.
2 Related Works
According to the recent study [18] found in the survey has made with distinct
approaches where density-based technique enhanced by practicing an efficient
density-based clustering along with maintaining the evolving medical data clusters.
evoStream algorithm from density-based clustering technique [16, 19] is imple-
mented on the following methodology in Sect. 3 to present the preliminary study
on outlier detection. The term of evoStream comprises from evolutionary algo-
rithms (EA) and DBStream [16]. evoStream algorithm brought together on pros
and cons from [19]. The advantage of evoStream are; (i) the performance of the final
cluster during online phase with high processing speed, (ii) removes computational
overhead. On the contrary, the disadvantage of evoStrem; (i) the application is not
suitable for multi-objective data stream, (ii) high dimensional datasets are impos-
sibly executed in evoStream. The role of evoStream exclusively focusing on the best
optimization solution for an online phase in the idle time to improve the final clus-
ters. Hence, a new approach to the purpose of this algorithm is mainly focusing on
investigating outlier detection from an inconsistency data set.
3 Methodology
An initial method to discover outlier detection in real data sets are implementing by
an algorithm named evoStream [16]. The purpose of this algorithm is mainly focusing
on investigating outlier detection from an inconsistency data object to differentiate
from a categorical problem.
Benchmarking data sets are retrieving collected from an open source UCI Machine
Learning Repository. The results yielded from evoStream are the preliminary result
for outlier detection. The process of evoStream is comprehensively scrutinizing
through preprocessing data, calculating centers and weight of clusters, profiling clus-
ters results, optimizing weight vector, develop as well as finalizing the validity of
optimizing evoStream model.
The beginning process is data collection and preprocessing. The data sets imported
into the R language and the data sets implemented in evoStream algorithm. The
radius, r has been set to r = 0.05 and k which represents as macro-cluster has been
fixed to k = 3 with incremental generations = 1 as well as re-cluster generations =
1000. Then, initialize evoStream package to insert observations, n = 1000 to process
with calculating centers and weight for micro-clusters and macro-clusters. The plot-
ting result is described according to the centers and weight of micro-clusters and
macro-clusters. Additional generations, n = 2000 examine to observe the different
to improve from the current plot result. The numbers assigned from micro clusters
to macro-clusters according to the nearest neighbor of data points to cluster.
542 N. A. Supardi et al.
Evaluation metrics are applied to evaluate the quality of statistical machine learning
model, well-known as cluster validation with known as internal measures and external
measures [8, 20]. The process cluster validation investigated under the measurement
of an algorithm to generate excellent results in data mining, particularly in clustering
techniques [20].
Seven (7) multitudes metrics SSQ and silhouette width are internal evaluation
metrics while purity, precision, recall, F-measure, NMI are external evaluation
metrics are selected to evaluate comparative analysis from [8, 10, 16]. Seven metrics
referring to Mathematically, the metrics are indicating with the following equations
except SSQ:
o Ci , o = o dist o, o
Silhouette = a(o) = (1)
|Ci | − 1
||.d
dh,l log dh clh,l
N M I = (2)
hdh log ddh lCl log cdl
1
k
F − Measur e = Scor eCi (3)
k i=1
1
k
Purity = Pr ecisionCi (4)
k i=1
Visum
PrecisionCi = (5)
nci
Visum
RecallCi = (6)
Vitotal
An Evolutionary Stream Clustering Technique Outlier Detection … 543
4 Assessment
4.2 Result
In this section, Figs. 1 and 2 show the final improved result from the preliminary
studies of two different data sets. Each of these benchmarking data sets employed
evoStream algorithm to report the studies outlier detection. The initial observations
of the online component n = 1000 required to complete at an early stage of the
investigation. During the idle time, the algorithm generates a macro-cluster when
there is no observation arrives, and the data point will move into their pertinent
clusters. Note that the performance of evolutionary step will not slow down the
processing time.
In general, evoStream did not employ re-clustering approach but applying re-
clustering steps when macro-clusters are requested and not incrementally Evalu-
ating generations does not affect underlying micro-clusters but does affect minor
improvement when implementing with 2000 generations into re-clustering. Hence,
5 Conclusion
The incoming data made stream processing is difficult to track down when data are
processed in real-time and brings the challenge of understanding the hidden patterns
from data. For that reason, the clustering technique introduced to unveil the nature
of data streams, especially outliers. Thereby, density-based clustering technique is
selected and has proven the most decent model in handling this investigation studies.
The proposed methodology will be introduced to optimize an evoStream model
to encounter the current problem. The purpose of initial investigation to analyze an
optimum weight vector by clustering the data according to their characteristics. Clus-
tering analysis outstandingly comes up with an accurate understanding of relevant
variables in the clusters. All in all, determine a clustering model is validating the
significance of the research question to achieve a valid result.
Acknowledgements The authors acknowledge the support of this research by Yayasan Universiti
Teknologi PETRONAS Fundamental Research Grant (YUTP-FRG) under Grant 015LC0-119.
References
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 547
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_51
548 R. Kumar et al.
1 Introduction
2 Background Study
SLA is legal documents that are defined and modeled on which CSP and consumers
must agree, approved, and signed after SLA negotiation [12]. SLA violations may
lead to monetary penalties, credibility, and reduced reliability for services. SLA
management describes complete assistance to configure SLA negotiation metrics
information, automatic negotiation, and monitoring, speedy recovery over issues,
and quick response over violation [2, 13].
Negotiation between CSP and a consumer based on metrics mainly price (P),
availability (Av), the response time (RT), reliability (R) [3]. Performance of cloud
services, negotiation strategy, and the process can be measure by a utility function
(UF), negotiation speed (NS), success rate (S) [14]. Factors to evaluate the strategy
of the negotiation process are the number of resources (NoR), Deadline (DL), degree
of satisfaction (DoS) [3, 6, 15, 16].
In the Negotiation process, consumers want to fulfill maximum requirements and
on the other hand, providers want to maximize sales of services. IANS focus on the
cycle of negotiation, negotiation matrices, and performance metrics [17].
Negotiation metrics: Cloud consumers must understand the key performance
indicators (KPI) for better negotiation. In Table 1, there are eight major high-level
cloud service attributes and under each level, some major KPIs to evaluate and
Analysis of Intelligent Agent and Its Components … 549
negotiate [4]. According to [3, 6, 11, 16], major negotiation metrics are price (P),
availability (Av), the response time (RT), reliability (R).
The proposed IANS design in Fig. 1. Follow the life cycle of SLA. IANS helps to finds
out the suitable CSP concerning the consumer requirements [7]. Agent components
allow us to understand the behavior of CSP by learning algorithms like reinforcement
learning [15], which help to speed up the process of negotiation.
There are five steps for SLA satisfaction between CSP and consumer [2, 5].
1. SLA Definition: CSP provides SLA with available offers and resources. A
CONSUMER SLA covers the requirements of the resource. Both types of SLA
can customize later [2].
2. Published and Discovery: Consumer published requirement to IANS, which
translates and analyzes the SLA and discover suitable CSP from its database.
Discovery of CSP based on rank-based that depends upon security, trust, perfor-
mance, financial, etc. IANS analyzes behavior property based on the functioning
of KPIs [2, 4].
3. Negotiation: Negotiation process has five major steps, (i) assessment of SLA,
(ii) allowance computation, (iii) achievable solution development, (iv) produce
offer, and (v) negotiation termination [16].
(a) Assessment of SLA: IANS assess the SLA and analyze the requirements
of resources with price, response time, and availability. IANS computes
UF, if the value of UF maximum, the agent gives positive feedback
[16, 19].
550 R. Kumar et al.
5. Termination: After the completion of the process, sharing of resources and all
activities ended [2].
In Fig. 2 [16], Workflow Diagram of Negotiation process showing communication
between consumer and CSP via IANS. IANS helps to make the process fast, secure,
and provide the best options for the consumer from a massive number of CSP [12].
In this research study, IANS functioning explored dependency factors of SLA nego-
tiation process for cloud services. It is defined that cloud SLA negotiation process
with IANS for negotiation will understand the requirement of the client and produces
perfect match from the massive number of CSPs with maintaining KPIs price, avail-
ability, response time and reliable. The descriptive analysis of the agent, its key factor,
and functioning is to develop an intelligent intermediary for the negotiation process.
IANS can focus on the dynamic environment of the cloud market and consumers’
requirements. Management of SLA described understanding the roles and respon-
sibilities of SLA. The framework of IANS discussed the workflow diagram of the
negotiation process. In future aspects, IANS may explore various learning algo-
rithms to design agents, decision-making algorithms for optimizing solution, security,
monitoring solution, and add more attributes for analysis.
552 R. Kumar et al.
References
1. Bahsoon R et al (2018) A manifesto for future generation cloud computing. ACM Comput
Surv 51(5)
2. Voorsluys RBW, Broberg J (2011) Cloud computing : principles and paradigms table of contents
3. Shojaiemehr B, Rahmani AM, Qader NN (2018) Cloud computing service negotiation: a
systematic review. Comput Stand Interfaces 55:196–206
4. Nadeem F (2020) A unified framework for user-preferred multi-level ranking of cloud
computing services based on usability and quality of service evaluation. IEEE Access 1–1
5. Iyer GN (2016) Cloud testing: an overview
6. Shojaiemehr B, Rahmani AM, Qader NN (2019) A three-phase process for SLA negotiation
of composite cloud services. Comput Stand Interfaces 64:85–95
7. Sim KM (2018) Agent-based approaches for intelligent interCloud resource allocation. IEEE
Trans Cloud Comput 1
8. Elhabbash A, Samreen F, Hadley J, Elkhatib Y (2019) Cloud brokerage: a systematic survey.
ACM Comput Surv 51(6):1–28
9. Rajavel R, Thangarathanam M (2016) Adaptive probabilistic behavioural learning system for
the effective behavioural decision in cloud trading negotiation market. Futur Gener Comput
Syst 58:29–41
10. Rajavel R, Iyer K, Maheswar R, Jayarajan P, Udaiyakumar R (2019) Adaptive neuro-fuzzy
behavioral learning strategy for effective decision making in the fuzzy-based cloud service
negotiation framework. J Intell Fuzzy Syst 36(3):2311–2322
11. Hsu CY, Kao BR, Ho VL, Li L, Lai KR (2016) An agent-based fuzzy constraint-directed
negotiation model for solving supply chain planning and scheduling problems. Appl Soft
Comput J 48:703–715
12. De la Prieta F, Rodríguez-González S, Chamoso P, Corchado JM, Bajo J (2019) Survey of
agent-based cloud computing applications. Futur Gener Comput Syst 100:223–236
13. Paputungan IV, Hani AFM, Hassan MF, Asirvadam VS (2019) Real-time and proactive SLA
renegotiation for a cloud-based system. IEEE Syst J 13(1):400–411
14. Mubeen S, Asadollah SA, Papadopoulos AV, Ashjaei M, Pei-Breivold H, Behnam M (2018)
Management of service level agreements for cloud services in IoT: a systematic mapping study.
IEEE Access 6:30184–30207
15. Rajavel R, Thangarathinam M (2017) ADSLANF: a negotiation framework for cloud manage-
ment systems using a bulk negotiation behavioral learning approach. Turkish J Electr Eng
Comput Sci 25(1):563–590
16. Li L, Yeo CS, Hsu CY, Yu LC, Lai KR (2017) Agent-based fuzzy constraint-directed negotiation
for service level agreements in cloud computing. Cluster Comput 1–15
17. El-Matary DM, El-Attar NE, Awad WA, Hanafy IM (2019) Automated negotiation framework
based on intelligent agents for cloud computing. In: Proceedings of the 2019 international
conference on innovative trends in computer engineering ITCE. pp 156–161
18. Vallejo D, Castro-Schez JJ, Glez-Morcillo C, Albusac J (2020) Multi-agentarchitecture for
information retrieval and intelligent monitoring by UAVs in known environments affected by
catastrophes. Eng Appl Artif Intell 87:103243
19. Pouyllau H, Carofiglio G (2013) Inter-carrier SLA negotiation using Q-learning. Telecommun
Syst 52(2):611–622
Early Detection of Myocardial Infarction
Using Machine Learning with Maximum
Accuracy
Abstract This paper presents a technique for detecting Myocardial Infarction (MI)
using Machine Learning through the ECG data. In today’s scenario, MI is one of the
major causes of demise worldwide. MI occurs due to coronary heart disease and if
detection or treatment is not done at appropriate time, the untreated MI may present
with serious late complications. For treating MI in earlier stage several methods have
been employed, but the parameters employed for those does not provide as much
accuracy. So, a system which detects the abnormality with maximum accuracy is
proposed. For that, Pan Tompkins algorithm is employed to filter and remove the
noise in the acquired ECG signal and S-T segment is extracted from it. The features
that are obtained from the extraction of S-T segment when given to the classifier
provided better results. Based on the application of the Machine Learning classifiers,
Naïve Bayes and Decision tree, the Decision tree classifier gave a higher accuracy
of 98.5% compared to Naïve Bayes of 93.9%.
1 Introduction
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 553
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_52
554 S. A. Manisa et al.
due to CVDs, which accounts for 31% of all global deaths. Out of these deaths, 85%
are due to heart attack and stroke [2].
It is observed that ECG has three main components: The atrial depolarization is
represented by P wave; the depolarization of the ventricles is represented by the QRS
complex; and the repolarization of the ventricles is represented by the T-wave. It is
shown in Fig. 1. In between the period of ventricular ion change, blood is supplied to
the heart muscles, via coronary arteries. When the blood flow decreases or stops to a
part of the heart, causes myocardial infarction. So, the salient point in early detection
of MI is the ST segment. The ST segment duration indicates the blood supply to the
heart. MI has three stages namely injury, ischemia and infarction [3]. An elevated
ST segment occurs at the stage of ischemia. For detecting more precisely, Machine
Learning is opted due to its less complexity and training time [4].
Previous methods [5–8], of MI detection are supervised learning approaches,
in which they took only the parameters like Age, BP etc., They did not provide
that much accuracy. Some methods [9–12] process the ECG recordings, but good
beneficial results are not yielded because only the difference between labels of the
entire ECG is taken. They did not take the difference between labels of heart beats
within the ECG. By using these methods, a large data cannot be classified and it is to
be overtrained. Even though these methodologies [13–15] gave good classification
results, it had less accuracy.
As this paper aims at detecting MI by extracting ECG from a noisy signal, it
requires accurate detection. For that, the filtering is done by Pan Tompkins algorithm
and the proper extraction of the S-T segment is done and, that is given to classifier
to detect the abnormality.
Early Detection of Myocardial Infarction … 555
2 Proposed Methodology
The structure of MI detection in our work consists of three main phases, namely:
Signal acquisition, ECG feature extraction and classification. The proposed method
is as shown in Fig. 2.
The noisy raw input is denoised using Pan Tompkins QRS Detection algorithms.
The extracted noiseless ECG signal is given to S-T segment detection and then the
features are extracted. The classification is done using naive bayes and decision tree
classifier and MI is detected.
The ECG signal is obtained using the ECG Sensor AD8232 and the Arduino Uno
Controller. The ECG sensor which senses the electrical activity of the human heart.
The three electrodes are attached to the left arm and right arm and right abdomen
to detect the heartbeat. The sensor is very small in size, portable and accurate to
measure the continuous heartbeat.
The Arduino controller gets the data from the ECG sensor and transmits the signals
to the signal processing through the COM port.
S-T
Segment
Detecon
Classifier
Feature
Model
Extracon
Raw ECG
BPF Differenator Squaring
Signal
Adapve Moving
Decision
window
Thresholding
Integrator
For a remote ECG monitoring, QRS detection is the preliminary step for detecting the
heartbeat for the subsequent rhythm classification [16]. It is the most significant part
of the ECG analysis algorithm. The Pan Tompkins algorithm uses the bandpass filter
having low pass filter and high pass filter which reduces the noises. It has derivative
function that highlights the R wave and then a quadratic function is applied to enhance
the high frequency features of the complexes (Fig. 3).
In ECG, initially the PQRST points are obtained by setting the threshold values and
from which the S-T segments are detected. This is the segment where the depression
or elevation of it distinguishes the normal and abnormal ECG. The isoelectric period
during which the ventricles are in between depolarization and the repolarization is
represented by the S-T segment.
The statistical features such as mean, standard deviation, skewness and kurtosis.,
are extracted from the MATLAB Tool Diagnostic Feature Designer [17] through the
data sets obtained from PTB Diagnostic Database, MIT-BH Arrythmia Database and
over 330 data were collected and features were extracted and tabulated as shown in
Fig. 4.
Early Detection of Myocardial Infarction … 557
The classifier model is built by using MATLAB classification learner app. The
extracted features are exported to the classifier model. Here, the classification is
done by using naive bayes and decision tree classification. The classifier is tested
and validated using the dataset obtained from the Physionet ATM database. Here the
hold out validation is given as 20% and then the dataset is trained.
The ECG Signal acquired through the sensor is plotted as shown in Fig. 5 using the
serial port, the system gets the analog value and plots that value using MATLAB.
The raw ECG signal after processed through the Pan Tompkins algorithm is found
as shown in Fig. 6 and the QRS peaks are detected as shown in Fig. 7.
558 S. A. Manisa et al.
The PQRST points detected are in the ECG Signal are shown in Fig. 8 and S-T
segment plotting from the detected PQRST points are shown in Fig. 9.
The deviation in the normal and abnormal ECG are shown in Fig.10 which serves
as deciding factor for the Myocardial Infarction. The abnormal ECG may have either
elevated or depressed S-T segment.
Early Detection of Myocardial Infarction … 559
The features extracted from the data sets are given to the Classification Learner App
and the scatter plot obtained for the Naïve Bayes and Decision Tree are shown in
Figs. 11 and 12. From which the accuracy for Naïve Bayes obtained as 93.9% and
Decision Tree as 98.5%
The confusion matrix for the both classifiers are as shown in Fig. 13.
Early Detection of Myocardial Infarction … 561
Myocardial Infarction needs an accurate and timely treatment for patients. So, in
machine learning, Naïve bayes and Decision Tree were used and compared, in which
562 S. A. Manisa et al.
the Decision Tree gave higher accuracy of 98.5%. In near future, we are trying
to improve the accuracy level and planning to give an additional support through
emergency assistance and transmit real-time data using IoT.
References
10. Subashini A, Sai Ramesh L, Raghuraman G (2019) Identification and classification of heart
beat by analyzing ECG signal using Naive Bayes. In: 2019 Third international conference on
inventive systems and control (ICISC). pp 691–694
11. Verma AR, Gupta B, Bhandari C (2020) A comparative study of ECG beats variability
classification based on different machine learning algorithms. Augment Hum Res 5:16
12. Mudasir MK, Syed IA (2016) Prediction of heart disease using decision tree a data mining
technique. IJCSN 5(6)
13. Chamuleau SA, van den Brink RB, Kloek JJ, Broekhuis E, de Beaumont EM, Koster RW (2005)
Complicaties van een niet herkend hartinfarct; ‘Complications of an unrecognized myocardial
infarction’. Ned Tijdschr Geneeskd. 2005 Nov 19; 149(47):2593–9. Dutch. Erratum In: Ned
Tijdschr Geneeskd. 2006 Dec 30; 150(52):2898. PMID: 16355569
14. Kshirsagar P (2020) ECG signal analysis and prediction of heart attack with the help of
optimized neural network. ACJ 9(4):497–506
15. Priyan MK, Usha DG (2018) A novel three-tier internet of things architecture with machine
learning algorithm for early detection of heart diseases. Comput Electri Eng 65:222–235
16. Pan J, Tompkins WJ (1985) A real-time QRS detection algorithm. IEEE Trans Biomed Eng
BME-32 (3):230–223
17. Mali B, Zulj S, Magjarevic R, Miklavcic D, Jarm T (2014) Matlab-based tool for ECG and
HRV analysis. Biomed. Signal Process Control 10:108–116
Design and Development of a Real Time
Mouse for Human Computer Interaction
Abstract In this paper there are two parts, Transmitter and Receiver. The transmitter
side is designed with Peripheral interface controller (PIC), Bluetooth, limit switches
and Micro electro mechanical systems (MEMS). In the receiver side there are Blue-
tooth and Personal Computer with Visual Basic 6.0. In the transmitter, the signals
sensed by the accelerometer and limit switches, then it is given input to the micro-
controller. The Analog to Digital Converter in the microcontroller converts these
sensor input to digital code and fed to Bluetooth module via Universal asynchronous
receiver/transmitter (UART) which triggers the module to transmit the signals to the
computer. At the computer, Bluetooth captures these signals. Thus, according to the
movement of hand, cursor is moved and according to limit switch press, functions
analogous to clicking the button is carried out in Personal Computer with the help
of Visual Basic 6.0.
1 Introduction
1.1 Overview
Nowadays the human computer interaction is a need for survival. Most of industries,
military, medical and domestic application require this type of interaction. Mainly
interactions such as speech and gestures have been used for communication. But
speech recognition lacks accuracy and probable to make errors [1] . So we require
a wireless real time user friendly device which can avoid such incidents or a system
that is accessible to gestures. In this paper therefore we go with the hand glove
gesture recognizing system, which can perform all the function that a real time
mouse does like dragging, selecting, scrolling the pointer to 3600 etc. A wearable
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 565
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_53
566 U. Vinothkumar et al.
wireless electronic glove that transforms hand and finger movement to real time
data for different applications. This paper replaces the standard computer mouse for
controlling a computer mainly used for power point presentations etc. Project uses
Microchip’s microcontroller IC named Peripheral Interface Controller PIC16F877A
and Microchip’s ‘Integrated Development Environment’, MPLAB, to simulate and
assemble the written code. Why we call such system user friendly? Because these
are some methods that we require to implement in a world where physical inabilities
of humans are a cause of concern the normal condition like backache, half paralyzed,
deaf and dumb etc. So we call them user friendly.
1.2 Objectives
• The objective of this paper is to design the wearable, real time and user friendly
device for human computer interaction.
• To mimic all the gestures performed in the real world wearing the data glove in
real-time with high accuracy.
The scope of the paper in future deals with the gaming sector which shall be further
more improvised for virtual environment applications. Another field where this can
be used is medical field, where paralyzed persons can be installed with this equipment
for their communications.
This paper replaces the standard computer mouse for controlling a computer
mainly used for power point presentations through wireless connection.
2 Existing Methods
In 2004 a social anthropologists ’Edward T Hall’ claimed that over 60% of human
interaction that we today is nonverbal in nature i.e. most of the communication can
be possible by gestures [2]. In this paper we use gestures which function indepen-
dently on speech i.e. we use autonomous gestures for communication. For a accurate
interaction the clarity of gestures is also a case of concern so we must select compo-
nents which can be user friendly can be accessible and reduce cost. Considering
the case where magnetic induction coils are used as sensors [3]. Here we require
separate sensors for each of the fingers respectively to improve degree of freedom
but looking for economic consideration we could go for better. Therefore we go for
TRI-axis MEMS sensor which have all the ability that magnetic induction coils and
have reduced cost [4].
Design and Development of a Real Time Mouse for Human Computer Interaction 567
The tri axis control can be obtained by using MEMS sensors. It gives opportunity for
free arm rotation. For each x, y, z axis program is written and it is being executed by
PIC using VB software. The tri axis control overcomes the disadvantages of magnetic
induction coil as it require less capital cost.
3 Methodology
Methodology section will served as a ‘technical blue print’ for this project to who
has an interest to upgrade or modify this project in the future. In this section, it is
important to know that all technical aspect regarding to this project will be included.
For this section, it will divided into two parts which consist of software and hardware.
For software sub-section, need MP-lab programming to program the programmable
integrated circuit and for interfacing the PIC module with pc visual basic 6.0 had
been used. In hardware sub-section, it will more on designing and developing the
circuit using strip board, fit in and solder the components on it and communicate
with microcontroller with the laptop through RS232 interface and Bluetooth. The
functional block diagram of Hand Data Glove: A New Generation Real-Time Mouse
for Human–Computer Interaction is shown in Fig. 1.
Here we uses a glove-type device which could detect hand position and movement.
In this approach user require to wear a glove like device, which uses sensors that
can sense the movements of hand and pass the information to the computer. These
approaches can easily provide exact coordinates of palm’s location and orientation,
and hand configurations. The main advantage of these approach are high accuracy
and fast reaction speed.
568 U. Vinothkumar et al.
The Bluetooth module are used for wireless transmission of data from wearable hand
glove device to personal computers. Bluetooth used here have a range of 10 m with
an available frequency range of 2.4 GHz. As per requirement we can set variable
distance by using amplifiers up to 100 m and so on. The Bluetooth is selected so that
it is low cost, low power, wireless mesh network standard (Fig. 2).
Fig. 3 MEMS
accelerometer
The limit switches used here to obtain different function done by mouse. Each
switches act for functions such as left button, right button and center click. In visual
basic 6.0 each of the functions are described as special character respectively. i.e.
for left button- dollar ($),right button-hash tag (#), center click-asterisk (*) char-
acters are used respectively. For each function when we click on limit switch the
following symbols are executed by PIC16F877A and the corresponding movements
are generated by windows application.
The MEMS accelerometer (Fig. 3) is placed above the wearable data glove. The
accelerometer measures acceleration forces. These forces may be static, like the
constant force of gravity, or they could be dynamic—caused by moving or vibrating
the accelerometer. Here we use tri—axis MEMS which recognize the movement of
arm in the form of respective co-ordinates of x, y, z respectively. Here the moment
of palm is recognized.
For each positions the function generated in visual basic 6.0 i.e. for x, y, z different
coding is written through visual basic 6.0 and is been executed in PIC16F877A.
See Fig. 4.
570 U. Vinothkumar et al.
Transmitter side:
Receiver Side:
Thus the Hand data glove is able to perform all functions of optical mouse with more
degree of freedom (Fig. 5).
5 Conclusion
The hand data glove can now mimic all the gestures performed by a real time mouse.
The added advantage given by degree of freedom of hand data glove is more than
that of mouse. The Bluetooth act as an interfacing device between the two hardware
sections. The detection of palm movement is sensed by using tri-axis accelerometer
Design and Development of a Real Time Mouse for Human Computer Interaction 571
MEMS. The results also indicates left, Right and Center click operations effectively
done by using accelerometer MEMS and limit switches.
6 Future Scope
The scope of the project in future deals with the gaming sector which shall be further
more improvised for virtual environment applications. Another field where this can
be used is medical field, where paralyzed persons can be installed with this equipment
for their communications.
This project replaces the standard computer mouse for controlling a computer
mainly used for power point presentations through wireless connection.
References
1. Piyush K, Jyoti V, Shitala P (2012) Hand data glove: a wearable real-time device for human
computer interaction. Int J Adv Sci Technol 43
2. Shiratori T, Hodgins JK (2008) Accelerometer-based user interfaces for the control of a
physically simulated character. ACM Trans Graph 27(5):1–9
3. Chin-Shyurng F, Herman S (2005) Development of a data glove with reducing sensors based on
magnetic induction. IEEE Trans Indus Electron 52(2)
4. PRIME Faraday Partnership” An Introduction to MEMS” January (2002) Wolfson School of
Mechanical and Manufacturing Engineering Loughborough University, Loughborough. ISBN
1–84402–020–7. http://www.primetechnologywatch.org.uk.
5. Pranathi N, Ahmed SM (2013) Tri-Axis motion detection using MEMS for unwired mouse
navigation system in the future generation machines. Int J Adv Res Comput Commun Eng 2(9)
Design and Implementation of Low Cost
Energy Meter Using MIT App Inventor
Abstract Rapid development of low-cost sensors and smart sensors in this twentieth
century makes the electronic based industries to get into the plenty of new solutions in
all the sectors. In this line software tools also support in nice a way to implement the
task with less effort. This proposed work going to enable such kind of an industrial
need application, namely low cost smart energy measurement is the popular area
in many accepts like Data control, management, saving and so on. Producing the
one unit of electricity is equal to saving half unit is the common thumb rule in the
electricity sector. Such a wonderful area utilizes all the advanced technologies to
implement its requirement.
Keywords Low cost energy measurement · Sensors and smart devices · Open
source softwares
1 Introduction
An open source hardware and software platform triggers the smart devices imple-
mentation with less effort. Every designer now-days tasting the open source tools
to implement their needs with short span of time [1]. There are two main reasons
behind that one is cost of overall implementation and second one is very important
P. Gajendran
Karpagam College of Engineering, Coimbatore, India
S. N. Deepa
Anna University Regional Campus, Coimbatore, India
N. Rajasingam (B)
Dr. N. G. P. Institute of Technology, Coimbatore, India
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 573
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_54
574 P. Gajendran et al.
one in this past decade this method accelerate the development work in the electronic
industries that is online community support. This one accelerate everyone simulation
work into real hardware implementation work [2]. This paper going to focus such
tools like Arduino and MIT APP inventor these two open source tools enable our
ideas to function in the real world environment.
A block diagram shown in Fig. 1 contains only two components in it. That the
Bi-directional communications between the server and client. In conventional energy
measurement topologies are defined as sensing and monitoring unit. Now these two
terms are slightly modified into server and client model [3]. Server going to plays
the vital role in thus proposed work. The functions are listed as follows.
• Sensing unit
• Data Manipulation
• Communication between Server and Client (Fig. 2)
Now the sever portion is clearly illustrated in the above block diagram. The heart of
the server unit is Arduino UNO, powerful hardware implementation. In this decade its
growth is tremendous in the area of embedded systems. Arduino community provides
hardware and software support to implement any low-cost embedded solutions.
Other major components are P.T. (Potential Transformer), C.T. (Current Trans-
former), ZCD (Zero Crossing Detector) and Bluetooth module. Using P.T. and C.T.
the important quantities required for the energy measurement in sensed and fed into
the microcontroller unit (Arduino UNO). Then the same voltage and current signal
are tapped from the output of the P.T. and C.T. and gives as the input to zero crossing
detectors [4, 5].
By receiving these signal ZCD provides the output as theta, that is phase angle
difference between V and I. To compute P (real power), Q (reactive power), Power
factor and energy. Finally, the computed signals were transmitted via Bluetooth
module (Fig. 3).
Client Unit is simply a smart phone using Bluetooth protocol the message packets
transmitted from the Server unit is received in the smart phone. Client Unit imple-
mentation is going to carried out with the help of MIT App Inventor. Again, this
one is for the open source cloud-based software support for the App Inventor. The
android app development is the one role in the client design. The Implementation
is clearly illustrated in the Fig. 4.This app has access security feature; to access the
particular “App Fingerprint Authentication” is added as the gateway to enter into
their application. Once the authentication is verified, it allows the user to get into the
network. In this way, server and client talks to each other. The calculated parameters
are then transmitted from server unit to receiver. Then the same will be displayed in
the dashboard window in the mobile App.
Hardware units are only present in the server client. They are,
1. P.T.
2. C.T.
3. ZCD
4. Bluetooth Module
576 P. Gajendran et al.
MIT App inventor tool is used to develop the android App required for this work. It
has two stages one is screen development and another one is the code blocks. There
are three screens used in this android app. First and second screen are used for user
verification using finger print sensor. Then third screen has main dashboard window
there user can take the readings of the energy meter.
Figure 5 clearly shows the components used in the screen-1 window. Here using
inbuilt fingerprint sensor in the smart phone is used to read the user input. Also the
Fig. 6 shows the code blocks required for finger print authentication.
Design and Implementation of Low Cost Energy Meter … 577
Fig. 6 Code blocks used in the screen-1 and Screen-2, Screen-3 view
578 P. Gajendran et al.
As shown in Fig. 6 Screen-2 will allow entering the next level once finger print
authentication is success. In screen-2 authentication confirmation is displayed then
provides the entry to reach the main dashboard window that is screen-3. Here in
screen-3 Bluetooth connectivity option is given to the user to pair with the server
Bluetooth module. Once it’s done it will start receive the transmitted parameters from
the server to display it.
At first download the App and install it in android based smart phone. One important
condition the phone must have fingerprint sensor. Then open the App the first screen
will appear as shown in the Fig. 7. Here the user needs to click the Scan Fingerprint
button followed by scan the finger. Once it’s done, it allows the user to enter in to the
second screen. There it will give Authentication Succeeded message. Then user need
to click the Go To Meter Dashboard button, immediately screen three will appear as
in the Fig. 8.
Finally the screen-3 window will shows the actual dashboard window of the low
cost energy meter. Here using Click to connect button user can pair the mobile
Bluetooth with server Bluetooth module that is HC-05. Once its paired successfully,
the dashboard will shows the various the electrical quantities.
4 Conclusion
In this way the proposed work is designed and implemented using Arduino and MIT
App inverter tools. This one serves the requested functions as mentioned in the initial
chapters. At present this work focuses on only single phase AC domestic loads. In
Design and Implementation of Low Cost Energy Meter … 579
future work the same need to be extended to three phase and DC measurements. Also
online cloud storage needs to implement to get in to the internet world.
References
Abstract The technology so called Internet of Things (IoT) has been knocking
around discussion circles since few years back. Internet of Things, the technology
behind the term is actually a simple idea, allowing the physical or dumb devices to
get communicate with one another via the internet. Although the Internet of Things
appears like it is just a concept at present, it will get to most of our lives very finely
and ever soon. Also various industries will be affected distinctly. Though massive
applications are there in IoT, data are getting transferred digitally in this modern world
every day which paves the way for hackers to steal our private information. Hence the
major grief is about digital security in todays’ existence. Enormous research in this
security area is in progress which includes privacy threats, vulnerabilities and secured
communication etc,. Among these issues the crucial aspect is security and privacy.
IoT network architecture get varies with respect to different applications which are
vulnerable to threats of different types used by the hackers to make use of the IoT
network. Sybil Attack is one of them which affect the communication module of the
whole network. In this paper, a speculation of detection and mitigation of the Sybil
attack has proposed. In contrast to other existing schemes, this proposed conception
will work better in each and every situation with less time consumption.
1 Introduction
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 581
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_55
582 C. Vinothini et al.
and sensor nodes to sense and monitor the environmental or physical conditions.
The characteristics of sensors are used to determine the performance in IoT network,
natch. Few characteristics are: i. Communication failures, if anyone of the node
in the network get fails while exchanging of information with others, which must
be informed without any retardation. ii. Utilization of sensors, the way the sensors
should produce the maximum performance. iii. Heterogeneity of sensor nodes means
the sensors of distinct varieties should work in a cooperative manner. iv. Scalability,
hundreds and thousands of sensors are getting employed in a network therefore the
IoT should be designed highly scalable.
In spite of many more advantages, the conjugations of IoT devices present a pivotal
challenge to industries in terms of intense security risks emerging from unmonitored
and unsecured devices connected to the IoT network. As applications of IoT network
grow continuously, data are getting transferred digitally in this modern world every
day which paves the way for hackers to steal our private information. Therefore, it
is understood that how IoT security is important in terms of security strategy and
protection schemes that significantly protect from the chances of security attacks
and even cyber-attacks on IoT devices that are interconnected to the network built
for specific applications and functionalities. Without strong and healthy security,
any connected physical device is vulnerable to getting compromised, breached and
controlled by an attacker.
Presently, though, there is not much cause for distress. Whereas very few people
are using IoT devices in a full-fledged manner, so there won’t be as many attackers
now to get access for them. This above-said statement is not to declare that this will
be the situation always. Soon, IoT will become a major part of our life in the near
future, and as days pass, it is essential to be monitored to the maximum extent and
protected even more carefully.
One of the crucial challenges faced with the evolving IoT applications is privacy
and security, and majority of the IoT smart devices are not structured to deal with
security and privacy issues [1]. Hence the increase in number of connected devices,
the chances for malicious attacks and other security threats is unavoidable.
Therefore the major grief is about digital security in todays’ existence. Here comes
the remaining section of this paper, Sect. 2 briefs about Sybil attack and its causes.
Related works are done relevant to the proposed work given in Sect. 3. Whereas the
Proposed work and Conclusion explained in Sects. 4 and 5 are respectively.
2 Sybil Attack
Physical and Network attacks, Software attacks and Encryption attacks are few
attacks that are vulnerable to Internet of Things. It includes DoS (Denial of Service)
attack, Node tampering, Node injection or Malicious code, Physical damage, etc.,
Under physical and network attacks, Sybil Attack is the one [2]. Earlier, the peer-
to-peer network category experiences this type of attacks in the beginning in which
Detection of Sybil Attack in Internet of Things 583
3 Related Works
As above-noted point, for the first time, the Sybil attack was identified in the peer-
to-peer network [4]. In a research work related to Sybil attack, the author framed
taxonomy for different formations of Sybil attack which targets any of the networks
to interrupt or damage that classifies into i. Direct and Indirect Communication ii.
Fabricated and Stolen Identities. iii. Simultaneity [1]. In all of these classifications,
this Sybil attack intrudes in the network catches a legitimate node and makes it into
an illegal node or a legal node by “malicious node” and gives identity to itself and
then releasing several forged IDs or fabricated IDs from other true nodes.
The fake or fabricated Identities will look similar to the legitimate node identities
but that is forged for non-existing nodes. This process will happen in the network is
getting affected by the Sybil attack. As a result of the Sybil attack, the legal nodes in
the network will think they have many legitimate peers [7, 8]. Once after misleading
the legitimate node into believing malicious nodes as their neighbor nodes this attack
breaks the condition of mapping between the identity and host of nodes [9].
Numerous works has been done for detecting and mitigating Sybil attacks in the
network. Few methodologies and schemes discussed here for reference. They are:
Message Authentication and Passing, Random Password Comparison (RPS),
Channel-based detection, Compare and Match Approach and still the progress goes
on.
Though these many techniques are there in existing, some limitations also to be
considered where the attack gets compromised.
584 C. Vinothini et al.
For Example in the RPS method there is no repair mechanism for route if the route
failure is the case. In Channel- Based Detection and RADS UWB methods focuses
only on stationary networks and the indirect Sybil attacks are get detected whereas
in the approach of Compare and Match has to know about complete information
about all the existing nodes else it fails. And in the RFID method the communication
overhead is high [10].
In a further Sybil attack detection mechanism proposed in [3] relies on the Ultra-
Wide Band (UWB) ranging capabilities in the physical layer and the distance between
each node and its neighbors are monitored periodically by every node which is a time-
consuming process and also the mobility of the nodes were not considered [3]. Also,
come across an algorithm named USAS (Unpredictable Software-based Solution)
for detecting compromised nodes. Here the administers attest few selected nodes
randomly to reduce the computation time of checksum, therefore there is a chance to
miss nodes that are far from the control node for attesting [4]. Another work proposed
in [5] for WSN architecture. The lightweight proposed method similar to message
authentication and passing algorithm employs watchdog nodes that overhear Hello
Packets exchanges between nodes. Then in the form of state diagram every watchdog
nodes produce detected information partially and generates the final Sybil nodes list
by using a detection rule defined already.
By using other personal nodes identities the Sybil activity can be created is well-
known to everyone and now almost 70% of research methods in existing for Sybil
attack detection are deals with Message authentication and passing method [3] that
requires some modification for time consumption and cost-effectiveness.
A secure RPL routing has been proposed in this paper by detecting and mitigating
Sybil attack the Rank attack. Based on the parameters used for the detection of
attackers in the IoT network, the trust values of direct and indirect for all parent and
child nodes were estimated [1].
One other author described the protocols named LEACH and LEACH-E. The
author states that the energy consumption is high when the cluster node and the BS
are getting communicated while comparing non-cluster nodes in WSN networks.
Therefore it is understood that the number of clusters-heads is directly proportional
to the energy consumption which leads shorten the lifetime of the corresponding
network. Hence, number of head nodes to be considered [4]. Here this concept can
be related to IoT network in terms of the number of sensors and actuators connected
to a host.
The light-weight identity certificate method is an authentication-based approach
to defeat Sybil attacks that use Merkle hash trees and a one-way key chain scheme.
Since Merkle hash tree is being used, this method consumes more memory storage
for information storage [5].
Interestingly another idea in [6] to mitigate Sybil attack in VPNETs also defined
and proposed a novel scheme called Local Sybil Resistance (LSR) for effective
defending against Sybil attack in privacy-preserving VPNETs. Here, in LSR scheme,
the sensitive information of all vehicles including identity and location privacy can
be well protected.
Detection of Sybil Attack in Internet of Things 585
4 Proposed Method
Our proposed algorithm is based on the authentication scheme for detecting and
mitigating Sybils and this work also uses the LEACH protocol [9]. This proposed
algorithm supports any kind of physical system. Here the network type chosen is IoT
as a physical system, sensor, and actuator nodes are get registered to the network and
generate their unique identities, then the credentials are shared to a particular host that
controls a network of devices in an application. Consider a high storage actuator as
a head node of the sensors which are connected to it. While sending and transferring
data to the host system via the actuator, the host will be using cryptography hash
function for the security purpose.
With reference to the KANGAROO CHF [8] scheme, the algorithm has been
proposed for the significant IoT network.
According to the proposed algorithm, If the sensor nodes and actuator nodes
are authenticated successfully once after the registration and generation of unique
Identities through the KANGAROO CHF scheme then the nodes are normal and
considered to be a legitimate and real node otherwise it must be a Sybil node.
Few parameters are considered here to calculate and by the result of it, the Sybil
attackers can be detected and mitigated [9]. The parameters are Threshold and Calcu-
lated Throughput of the receiver similarly the packet delivery ratio and total energy
consumed by the IoT network.
Once the Sybil attack is detected in the network then immediately the respective
Sybils should be mitigated from the IoT network. To mitigate it, first step is to
remove the identified malicious node by using any one of the existing algorithm like
RSSI. Let us assume a threshold values of throughput and energy are TDTh, TDEn
respectively for the IoT network. Calculate the total energy consumed by the whole
network at the deployment stage. Similarly calculate the Packet delivery ratio and
Throughput of the IoT network in the same manner. And once after designed IoT
application has been deployed for a certain need, Calculate again the parameters
throughput, packet delivery ratio and the energy consumption as said above. Now
to identify and to detect the Sybil attack, the calculated values of threshold, packet
delivery ratio and energy consumption at two different stages are to be compared.
Based on the difference in the values of these parameters calculated above deter-
mines the presence of Sybil attack and that will be shown as a major threat to the
entire system.
Thus, our proposed algorithm detects the malicious node and mitigates the same
from the IoT network. In our IoT network the Sybil attack attacks the communication
586 C. Vinothini et al.
layer or network layer of the IoT system. The procedure has been detail explained
in the below given algorithm.
5 Conclusion
References
1. Tandon F (2019) Trust-based enhanced secure routing against rank and sybil attacks in IoT. In:
2019 twelfth international conference on contemporary computing (IC3), pp 1–7
2. Priya J, Gunasekaran M (2019) Security-aware and privacy-sensitive of internet of things (IoT ):
a review. In: 5th International conference on advanced computing and communication systems
(ICACCS), pp 225–230
Detection of Sybil Attack in Internet of Things 587
3. Patel ST (2017) A review: sybil attack detection techniques in WSN. In: 4th international
conference on electronics and communication systems (ICECS), pp 184–188
4. Suriya U, Kumar R, Vayanaperumal R (2015) Detecting and preventing sybil attacks in wireless
sensor networks using message authentication and passing method. Scientif World J 1–8
5. Amuthavalli R, Bhuvaneswaran RS (2015) Genetic algorithm enabled prevention of sybil
attacks for LEACH-E. Mod Appl Sci 9(9):41–49. https://doi.org/10.5539/mas.v9n9p41
6. Zhang Q, Wang P, Reeves DS, Ning P, Carolina N (2005) Defending against Sybil attacks
in sensor networks. In: 25th IEEE international conference on distributed computing systems
workshops (ICDCSW’05), pp 1–7
7. Lin X, Member S (2013) LSR : mitigating zero-day Sybil vulnerability in privacy-preserving
vehicular peer-to-peer networks. IEEE J Selected Areas Commun/Supplement. 31(9):1–10
8. Kandhoul N, Dhurandher SK, Woungang I (2019) T _ CAFE : a trust based security approach
for opportunistic IoT. IET Commun 13:3463–3471. https://doi.org/10.1049/iet-com.2019.0657
9. Bertoni G, Daemen J, Peeters M, Van Assche G, Van Keer R (2018) In: KangarooTwelve : fast
hashing based on Keccak- p, Applied cryptography and network security book, pp 1–23
10. Lim J, Yu H, Gil J (2017) Detecting sybil attacks in cloud computing environments based on
fail-stop signature, symmetry. pp 1–12. https://doi.org/10.3390/sym9030035
Diagnosis of Brain Tumor Using ANN
with Spatial Fuzzy Clustering and Active
Level Set Contour
Keywords MRI brain images · Spatial Fuzzy-C means clustering · Active level set
contour · Anisotropic diffusion filter · Histogram of image · Artificial neural
network
1 Introduction
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 589
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_56
590 H. A. Munira and M. S. Islam
Image Histogram
2.1 Dataset
The required datasets for MRI images have been obtain from https://github.com/
sartajbhuvaji/brain-tumor-classification-dataset. Totally 699 MRI images of patients
comprised of glioma and meningioma tumor had been undertaken.
2.2 Pre-Processing
In order to get more assurance and simplicity in distinguishing the tumor [8, 9], the
motivation behind these methods is fundamentally to improve the image quality.
High Pass Filtering. It is used for edge detection and noise highlighting if any.
The algorithm steps for this pr-processing is given as: hp_fil = [−1 2 −1; 0 0 0; 1 −
2 1];
Median Filter. Median Filter is being utilized in the subsequent stage to eliminate
any undesirable clamor whenever recognized by the HPF.
Anisotropic Diffusion Filter. Anisotropic filter blurs the images without
removing any edge. This filter can be expressed as,
∂I
= div (c(x, y, t), ∇ I ) = ∇c · ∇ I (1)
∂t
where, c (x, y, t) denotes the coefficient of diffusion. Through this the rate of
diffusion is managed and is selected to preserve the image’s edges. The diffusion
co-efficient is suggested as,
||∇ I || 2 1
c(||∇ I ||) = exp − or c(||∇ I ||) = 2 (2)
K ||∇ I ||
1+ K
Figure 2 represents a filtered image using anisotropic diffusion filter which shows
that the image edges are not blur due to filtering.
Histogram of Images. The image was enhanced using histogram of images [9].
In Fig. 3, the enhanced image is shown.
Fig. 3 Pre-processed
enhanced image
2.3 Post-Processing
The improved image from pre-processing then reaches the post-processing phases.
Spatial Fuzzy C-means. In medical problems, fuzzy C-means clustering is widely
used because of its popularity [10, 11]. On the other hand, a membership function
u mn is used to denote the proportion of integration of nth object to the mth cluster in
fuzzy C-means. The cost function of FCM is given by:
N
C
J= μlmn [[in −Vm ]]2 (3)
n=1 m=1
C
N
μmn = 1; 0 ≤ u mn ≤ 1; μmn > 0 (4)
m=1 n=
where, μmn denotes the proportion of integration of nth object to the mth cluster
in FCM. The absence of spatial data [4, 5, 9] is the biggest issues of conventional
FCM algorithms for segmentation of image. The output of FCM segmentation will
decrease due to the image noise and artifacts. That’s why, spatial FCM algorithm is
used. A spatial FCM algorithm was suggested by Chuang [4] which can be defined
by:
p q
μmn h mn
μmn = C p q
(5)
k=1 μkn h kn
where, the respective controlling parameters are p and q. The spatial information
is carried by h mn , can be expressed as:
h mn = μnk (6)
k∈Nn
Active level set contour. Effective contrast boundaries for an image segmentation
is used in active level set contour [12, 13]. With the help of active contour level set,
image segmentation is a popular process [14]. Using the zero-level tracking t, the
evolution of active contours can be given by:
3 Results Analysis
The implementation of both spatial FCM algorithms and active level set contour has
been done in MATLAB 2020a. Through this proposed approach, initial segmentation
of images is possible with level set active contour and the accuracy of the trained
network on the 105 tested image is 96.7%. The network is trained on 489 MRI images
and validation and testing image are 105.
In Fig. 4, the segmented image after using spatial FCM (a, c) and active level set
contour image with edges (b, d) are shown.
In Fig. 5, the confusion matrix of the training is shown and Table 1 represents
the performance of the network based on sensitivity, accuracy and specificity. The
comparison of this proposed method with others is shown in Table 2.
Fig. 4 Spatial FCM segmentation (a, c) and active level set contour after 100 iterations (b, d)
594 H. A. Munira and M. S. Islam
Table 1 Classification
ANN classifier result
performance
Sensitivity Specificity Accuracy
97.38% 95.42% 96.7%
Diagnosis of Brain Tumor Using ANN with Spatial Fuzzy Clustering … 595
Thus, the improved spatial FCM segmentation is esteemed for trying to merge
intensity and spatial information altogether. This algorithm esteemed less noise and
it is perfect for initial level set contour in medical image processing.
4 Conclusion
For brain tumor detection, an automated image segmentation is proposed in this paper
using enhanced spatial fuzzy C-means with active level set contour. The proposed
approach contains the pre-processing steps where the images are being ready for
further segmentation using noise removal median filter, anisotropic diffusion filter
and histogram of images. Then, the images are sent to post-processing steps where
the spatial fuzzy C-means is used with active level set contour for better detection of
boundaries of abnormal tissues. The statistical extracted features are trained using
ANN where the accuracy is found 96.7% on test data which is far better than many
existing methods.
References
1. Loges Wari T, Karnan M (2010) In: International conference on signal acquisition and
processing, an enhanced implementation of brain tumor detection using segmentation based
on soft computing
2. Hemanth DJ et al (2009) Effective fuzzy clustering algorithm for abnormal MR brain image
segmentation. In: Advance computing conference 2009. IACC 2009, IEEE international. pp
609–614
3. Suganya R, Shanthi R (2012) Fuzzy C-Means algorithm—a review. IJSRP 2(11. ISSN 2250–
3153
4. Aslam A, Khan E, Beg MMS (2015) Improved edge detection algorithm for brain tumor
segmentation. In: Second international symposium on computer vision and the internet
(VisionNet’15), Elsevier
596 H. A. Munira and M. S. Islam
5. Lakra A, Dubey RB (2015) A comparative analysis of MRI brain tumor segmentation technique.
Int J Comput Appl 125:5–14 (0975–8887)
6. Sudharani K, Sarma TC, Prasad KS (2015) Advanced morphologicaltechniqueforautomat-
icbraintumordetectionandevaluationof statistical parameters. In: International conference on
emerging trends in engineering, science and technology (ICETEST)
7. Dhage PM, Phegade MR, Shah SK (2015) Watershed segmentation brain tumor detection. In:
International conference on pervasive computing (ICPC)
8. Kumbhar U, Patil V, Rudrakshi S (2013) Enhancement of medical images using image
processing in MATLAB. Int J En Res Technol 2(4):2359–2364
9. Chuang KS, Hzeng HL, Chen S, Wu J, Chen TJ (2006) Fuzzy c-means clustering algorithms
with spatial information for image segmentation. Comput Med Imaging Graph 30:9–15
10. Cai W, Chen S, Zhang D (2007) fast and robust fuzzy c-means clustering algorithms
incorporating local information for image segmentation. Pattern Recogn 40:825–838
11. Osher S, Fedkiw R (2003) Level set methods and dynamic implicit surfaces. Springer, New
York
12. Chan TF, Vese LA (2001) Active contours without edges. IEEE Trans Image Process 10:266–
277
13. Varuna shree N, Kumar TNR (2018) Identification and classification of brain tumor MRI images
with feature extraction using DWT and probabilistic neural network. In: Springer corporate
information
14. Havaei M et al (2017) Brain tumor segmentation with deep neural networks. Med Image Anal
35:18–31
Current Overview of Machine Learning
Application for Predicting Steam Huff
and Puff Injection Production
Performance
Abstract Thermal Enhanced Oil Recovery (EOR) is one of the main contributors
to EOR worldwide production. Steam huff and puff injection, one of its methods, is a
technique in which steam is injected in a cyclical manner alternating with oil produc-
tion. Reservoir simulation is considered as the most reliable solution to evaluate the
reservoir performance and designing an optimized production scheme. However, it
still remains time-consuming and expensive. Applying machine learning to build a
predictive proxy model is a suitable solution to deal with the issue. Presently, there
have been a limited number of studies covering the topic of proxy model development
to estimate production performance for this injection method. This study provides a
review of the machine learning implementations for estimating steam huff and puff
injection production performance, starting with an introductory explanation about
the method, followed by the currently deployed machine learning models along with
the challenges and future prospects.
Keywords Enhanced oil recovery · Machine learning · Oil and gas · Proxy
model · Reservoir engineering · Steam huff and puff
1 Introduction
As most of the giant oil fields are all currently at the mature stage, Enhanced Oil
Recovery (EOR) is considered a viable solution to recover the remaining oil after the
primary and secondary recovery stages. It involves the injection of a fluid or fluids
into a reservoir intending to supply the energy needed to displace oil. Furthermore,
the injected fluids interact with the reservoir rock and fluid to alter the physical
properties and create an advantageous condition to improve oil recovery [1].
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 597
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_57
598 M. G. Merdeka et al.
Thermal EOR is one of the main EOR methods applied worldwide operating
on the mechanism involving reduction of oil viscosity by the injection of thermal
energy. As reported by the International Energy Agency, it accounts for more than
40% of total EOR production in 2015 [2]. The target oil is a high-density and high-
viscosity oil. Heavy oils are such example for the target, and several reports on the
thermal EOR implementations showed that almost all of the reservoirs are heavy oil
reservoirs.
In evaluating the performance of a reservoir and designing an optimized produc-
tion scheme, reservoir simulation is conducted beforehand. It is considered as the
most reliable solution. However, the whole process of reservoir simulation still
remains time-consuming and expensive due to the complexity of a reservoir system
and the difficulty of building a reliable full reservoir model. Building predictive
proxy model using machine learning algorithms is considered a suitable solution to
deal with the issue, where proxy models will estimate the reservoir simulation results
using lighter mathematical models [3]. In reservoir engineering, the proxy model has
been mainly used for sensitivity analysis, risk analysis, history matching, produc-
tion forecasting and production optimization [4]. The advantages of the proxy model
over reservoir simulation are the computational speed and significantly economical
solution, at a tradeoff of slightly lower accuracy.
This paper will provide a review of the research works related to machine learning
applications in predicting steam huff and puff injection method. It is aimed to summa-
rize the findings and highlight the advantages and disadvantages of applying machine
learning approaches in this research area. The discussion will start with an introduc-
tory explanation of steam huff and puff method, including the how-to and past field
successes. Then, the currently deployed machine learning models for this method will
be explored. Finally, the challenges and future prospects to utilize machine learning
coupled with current reservoir engineering technologies will be discussed.
Steam huff and puff injection, also known as cyclic steam stimulation, is a method in
which steam is cyclically injected into the well. After injection, the well is shut-in,
allowing the steam to “soak” into the reservoir. The high temperature of the steam will
reduce oil viscosity near the steam-oil interface. This is called the soaking period.
The well is opened again after the soaking period, producing oil at a higher rate.
After some time, the heat dissipates, and the near-well oil viscosity returns to its
original value. The production declines and after reaching an economic limitation
rate, steam is reinjected into the well, starting a new injection cycle. This whole
process of injection, soaking and production is called the huff and puff method and
illustrated in Fig. 1. Several screening criteria of this method had been developed in
the past [5–7]. In general, a typical steam huff and puff process is well suited for
heavy oil reservoirs with high porosity and permeability, shallow depths and high oil
saturation.
Current Overview of Machine Learning Application … 599
Fig. 1 Steam huff and puff injection schematic by Green and Willhite [1]
Even though the recovery factor is still smaller in comparison to the other steam
injection methods, this injection method is still highly attractive due to small capital
investment and quick payout. It is one of the most implemented thermal EOR methods
with numerous reported commercial success in the past, for examples: Cold Lake
Field in Alberta, Canada [8, 9], Tia Juana Field in Venezuela [10, 11], Midway Sunset
Field in California [12], Liaohe Field in China [13] and Duri Field in Indonesia [14–
16]. Furthermore, it is often used as a precursor before conducting a full field scale
steam injection. To predict the performance of steam huff and puff injection, several
semi-analytical models had been developed in the past [12, 17–20]. However, the
drawbacks of those models are the limitations imposed by the assumption of the
underlying physical equation and the need to have a reasonable history matching
before using the models for prediction.
3.1 Introduction
Machine learning (ML) is currently gaining a lot of attention to the recent break-
through in data analysis. It is a tool that gives computers the ability to learn without
being explicitly programmed. In oil and gas, especially reservoir engineering, most
of the ML problem types were supervised learning problems, e.g. assisted history
matching and forecasting and optimizing production [4, 21]. Those problems can be
600 M. G. Merdeka et al.
3.2 Existing Proxy Models for Steam Huff and Puff Injection
Although there were many research works related to steam huff and puff performance
prediction, a very limited number related to the use of machine learning to develop
proxy model was found. In predicting steam huff and puff performance, the most
used machine learning algorithm was the Artificial Neural Network (ANN).
ANN is a model that emulates the biological neural system. It consists of nodes
which are similar to neurons in the human brain. A node receives signals from
adjacent nodes and processes them to produce an output. A generalized ANN model
structure is presented in Fig. 2.
It consists of an input layer, an output layer, and one or multiple hidden layers in
between. Each neuron from a layer is connected to the neurons from other nearby
layers with a connection called weight, which represents an influence of corre-
sponding input on the connected neuron. Then, to transfer the contained information,
an activation (or transfer) function is used. This is called forward propagation and
it is conducted for every connection in the network. After defining the model, the
backpropagation algorithm is used to readjust the weight and minimize the error
between the output and the database.
Several existing researches on ANN proxy models for steam huff and puff injec-
tion are presented below. They were presented as a general model and some are
the continuation and improvements from previous works over the years. ANN was
deployed to predict the production performance based on input parameters such as
reservoir properties and operating conditions.
Arpaci developed ANN models for steam huff and puff injection using horizontal
wells in naturally fractured reservoir [23]. The inputs of the model are reservoir rock
and fluid properties, operation design parameters and fracture design parameters.
The outputs are the number of cycles, cycle duration, oil flowrate, and cumulative
oil production. Synthetic data were generated using the help of a thermal reservoir
simulator. In addition, inverse models were developed to predict operation parameters
and fracture design parameters, using performance indicators such as oil production
and production period as additional inputs. The errors of each built models are less
than 5%. However, most of the reservoir parameters were assumed constant, such as
oil density, relative permeability, anisotropy and capillary pressures.
Sun and Ertekin developed a data-driven proxy model for steam-assisted gravity
drainage (SAGD) and steam huff and puff injection in naturally fractured reservoir
[24]. There are two kinds ANN models that were developed: a supervisory model
that predicted the number of cycles and a sub-model that predicted the reservoir
performance. It was set that the cycle automatically switches when the oil production
rate hits a certain economic limit, thus the total number of cycles in the end of
simulation period is not certain for every simulation case. The inputs will first visit
the supervisory model to determine the number of cycles, then the sub-model will
be predicting the reservoir performance, using both inputs from the database and the
result of the supervisory model. The developed models show a good match with the
database obtained from the numerical simulation. In addition to the ANN models,
a Graphical User Interface was developed to facilitate easier use of the models to
the end-users. Although this study had covered the disadvantage of less reservoir
parameters from the previous study, the effect of the heterogeneity, anisotropy and
capillary pressures were still neglected. The other limitations came from the fluid
model assumption where the fluid composition was dead oil and water, and the
number of predicted cycles in the sub-model were limited to 10 cycles.
Ersahin and Ertekin continued the previous study by Sun and Ertekin and devel-
oped ANN models for steam huff and puff injection in naturally fractured reservoir
[25]. In this study, there were two main performance indicators that became the
output of the model, which are cumulative production profile and viscosity contours.
The simulations were performed to understand the heat-exchange mechanisms in
the stimulation zone around the wellbore. Improvements were also carried out to the
input parameters, e.g., the relative permeability table was generated using correla-
tions specifically designed for sandstone rocks, as most of the thermal EOR projects
were performed in sandstone rocks [26]. The results were good for predicting the
outputs, with average error for cumulative production profile and viscosity contours
were 3.64% and 5.32 respectively.
602 M. G. Merdeka et al.
A limited number of research works encourage further explorations on this topic. The
challenges would be coming from the difficulties of building a reliable model itself
and finding the effective use of machine learning alongside conventional reservoir
engineering practices.
Current models were able to be developed due to the huge amount of available data.
Finding the data to train the model would be difficult if there are less implementation
of the methods themselves, especially in the future, where new innovations on steam
huff and puff injection method will emerge. The other challenge is the limitation of
the model. Proxy model is basically an interpolation model. Hence, it can only be
utilized when all the values of input parameters fall within the range of the training
data. Having more data with larger intervals of inputs will help in developing a better
general model.
Although proxy model provides a faster and cheaper computational solution, the
weakness of the technology should not be ignored. Conventional reservoir simulation
is still the most reliable tool in reservoir engineering and proxy model should not
be used as a replacement. Instead, it should be used as an assisting tool to achieve
reservoir engineering analysis more effectively. For example, as a starting guide in
creating a field or well development plan.
The future prospects of this research topic lie on the technological innovations of
steam huff and puff methods itself, as well as the advancements of machine learning
and its subset, deep learning. Thermal EOR is projected to be still one of the highest
contributors of EOR production in 2040, second only to miscible EOR [2]. There
have been several studies and field trials of the new modifications to steam huff and
puff injection, such as including additives to steam, conducting injection in different
well configurations and combining cyclic injection with fracturing [27]. Therefore,
one can expect more innovations in this method and their field implementations to
gather more data for modelling purposes. On the other side, improvements in machine
learning may highly contribute to determine the best approach to develop a proxy
model for performance prediction. For instance, choosing the best training algorithm
or determining the best adjustments for hyperparameters of the proxy model.
4 Conclusion
This study provided an overview of the steam huff and puff injection methods and
the application of machine learning to predict the production performance. Machine
learning helps to improve the reservoir simulation approaches in predicting the
production performance by developing a proxy model which is highly efficient in
reducing the required time and cost, at a tradeoff of slightly lower accuracy. However,
it should be noted that the proxy model is not a replacement for reservoir simulation,
but rather as an assisting tool that effectively helps in reservoir engineering practice.
Current Overview of Machine Learning Application … 603
For future prospective, the technological innovations of steam huff and puff method
and the advancements of machine learning are expected to explore further towards
more viable solutions.
References
1. Green DW, Willhite GP (1998) Enhanced oil recovery, 1st edn. Society of Petroleum Engineers
2. McGlade C, Sondak G, Han M (2018) Whatever happened to enhanced oil recovery?. Inter-
national energy agency (2018). https://www.iea.org/commentaries/whatever-happened-to-enh
anced-oil-recovery. Last Accessed 28 Oct 2020
3. Negash BM, Tufa LD, Ramasamy M et al (2017) System identification based proxy model of
a reservoir under water injection. Model Simul Eng 7645470
4. Zubarev DI (2009) Pros and cons of applying proxy-models as a substitute for full reservoir
simulations. In: SPE annual technical conference and exhibition, New Orleans, Lousiana, USA
5. Taber JJ, Martin FD, Seright RS (1997) EOR screening criteria revisited—part 1: introduction
to screening criteria and enhanced recovery field projects. SPE Reserv Eng J 12(3):189–197
6. Farouq Ali SM (1974) Current status of steam injection as a heavy oil recovery method. JPT
13(1):54–68
7. Sheng J (2013) Enhanced oil recovery field case studies. Gulf Professional Publishing
8. Buckles RS (1979) Steam stimulation heavy oil recovery at cold lake, Alberta. In: SPE
California regional meeting, 18–20 April, Ventura, California
9. Beattie CI, Boberg TC, McNab GS (1991) Reservoir stimulation in the cold lake oil sands.
SPE Reservoir Eng
10. De Haan HJ, Van Lookeren J (1969) Early results of the first large-scale steam soak project in
the Tia Juana. J Petrol Technol 21
11. Puig F, Schenk L (1984) Analysis of the performance of the M-6 area of the Tia Juana Field,
Venezuela, under primary, steam-soak, and steamdrive conditions. In: Society of petroleum
engineers AIME, SPE, vol 1. pp 263–276
12. Jones J (1977) Cyclic steam reservoir model for viscous oil, pressure depleted, gravity drainage
reservoirs. In: SPE California regional meeting California
13. Liu W-Z (1997) Steam injection technology to produce heavy oils. Petroleum Industry Press
Beijing, China
14. Gael BT, Gross SJ, McNaboe GJ (1995) Development planning and reservoir management in
the Duri steam flood. In: SPE western regional meeting, 8–10 March, Bakersfield, California
15. Pearce JC, Megginson EA (1991) Current status of the duri steamflood project Sumatra,
Indonesia. In: SPE International thermal operations symposium, 7–8 February, Bakersfield,
California
16. Bae WS, Masduki A, Permadi AK, Abdurrahman M (2017) EOR in Indonesia: past, present,
and future. Int J Oil and Gas Coal Technol 16(3):250
17. Boberg TC, Lantz RB (1966) Calculation of the production rate of a thermally stimulated well.
J Petrol Technol 18(12):1613–1623. Society of Petroleum Engineers
18. Gontijo JE, Aziz K (1984) A simple analytical model for simulating heavy oil recovery by cyclic
steam in pressure-depleted reservoirs. In: SPE annual technical conference and exhibition,
16–19 September, Houston, Texas
19. Sylvester NB, Chen HL (1988) Improved cyclic steam stimulation model for pressure-depleted
reservoirs. Soc Pet Eng AIME, SPE, pp 213–226
20. Gozde S, Chhina HS, Best DA (1989) Analytical cyclic steam stimulation model for heavy oil
reservoirs. In: SPE California regional meeting, 5–7 April, Bakersfield, California
21. Jaber AK et al (2019) A review of proxy modelling applications in numerical reservoir
simulation. Arabian J Geosci. Saudi Society for Geosciences
604 M. G. Merdeka et al.
22. Ertekin T, Sun Q (2019) Artificial intelligence applications in reservoir engineering: a status
check. Energies. MDPI
23. Arpaci B (2014) Development of an artificial neural network for cyclic steam stimulation
method in naturally fractured reservoirs. Master Thesis, The Pennsylvania State University
24. Sun Q, Ertekin T (2015) The development of artificial-neural-network-based universal proxies
to study steam assisted gravity drainage (SAGD) and cyclic steam stimulation (CSS) processes.
In: SPE Western regional meeting, 27–30 April, Garden Grove, California, USA
25. Ersahin A, Ertekin T (2019) Artificial neural network modeling of cyclic steam injection process
in naturally fractured reservoirs. SPE Reservoir Evaluation and Engineering. SPE
26. Alvarado V, Manrique E (2010) Enhanced oil recovery: an update review. Energies
27. Alvarez J, Han S (2013) Current overview of cyclic steam injection process. J Petroleum Sci
Res 2(3)
Enhanced Cluster Head Based Data
Gathering (ECHGS) Technique in IoT
Based Smart Irrigation System
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 605
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_58
606 K. Padmanaban et al.
1 Introduction
2 Literature
Wang et al. [5] proposed a path planning algorithm to reduce the mobile nodes count.
In this technique, an efficient algorithm is designed to reduce the nodes on the path by
path optimization. The results of the path planning algorithm are outperformed than
the existing technique CSWEEP in terms of travel length and number of nodes. Lee
et al. [6] presented a model based on Neural Network with input and output data as
working hours, temperature, weather, humidity and electrical consumption for IoT
applications. Model is verified with test data and performance is evaluated. Energy
consumption is optimized in this work. Mean squatted error method is applied to
the obtained data and simulation is done. Ke et al. [7] introduced a novel method
called Joint Optimization method for Data Offloading, Renewable energy-aware,
and Bandwidth allocation for IoT devices based on deep Reinforcement Learning
(JODRBRL). This method minimizes the total cost which includes data delay cost,
bandwidth cost and energy consumption cost by adaptive learning from the dynamic
IoT environment. The JODRBRL results are better than the existing techniques
Double DQN (DDQN), Dueling DQN and greedy policy.
Guo et al. [8] proposed a comprehensive energy estimation model for IoT
networks. This proposed model analyzes the insufficient random-access procedure
and utilize the power ramping and preamble picking for optimizing energy utilization.
Also, Multi-Agent Reinforcement Learning (MARL) algorithm based on Win-or-
Learn-Fast Policy Hill-Climbing (WoLF-PHC) is introduced in this work to decrease
the algorithm complexity. The results are evaluated and compared with the random
picking method. The proposed WoLF-PHC method results outperformed the random
picking method results. Padmanaban et al. [4] presented a method called Tree Cluster
based Data Gathering Scheme (TCDGS) in wireless sensor networks. In this work,
nodes are clustered and form a tree-like structure. Cluster head node is identified
for each cluster to maximize the residual energy. Data gathering is done by the data
gathering node in the network and data is sensed from the cluster head instead of
sensing all the nodes in the cluster and transfer to the base station. This particular
module increases the life time of the network. The TCDGS results are better than
the existing method of CIDT results.
All these existing works [9–17] have not considered residual energy and time
delay in the IoT based agriculture smart irrigation system. In this proposed work,
the residual energy and average time delay are considered. The residual energy is
maximized for all the nodes and average time delay is minimized when compared to
the existing methods CIDT and TCDGS.
608 K. Padmanaban et al.
The proposed architecture for the smart irrigation system in agriculture is shown in
Fig. 1. The agriculture lands are shown in the diagram with circles. The clusters are
formed based on the agriculture lands. Entire agriculture area is divided into lands.
Each land among the total agricultural area is considered as a cluster. There are many
sensors placed in each cluster. The sensors are placed in each agricultural land to
collect the water level information from the land. In this paper, the Enhanced Cluster
Head based Data gathering” (ECHGS) technique is used to collect the data from
sensors effectively by proposing a new module to identify the cluster head. In each
cluster of sensors, the cluster head is identified by conducting the poll based on the
load and energy parameters of sensor.
The poll is conducted by substation periodically and the sensor node which has
maximum energy and minimum load is considered as cluster head. The elected cluster
head is responsible for collecting water-related data and send it to the substation.
The substation forwards the data which is received from the cluster head to the base
station. The substations are placed in all the agriculture land. The base station which
sends the data that is received from the substation to the data server. The data server
stores the water data and calculates the irrigation level for each agriculture lands.
This processed information about water irrigation is given as output to the users who
are accessing the data server to know the level of water irrigation to their lands. Each
sensor’s load and energy values are varying in every time instance. Current cluster
head’s load and energy level is changed after transmission of data. To maintain the
efficiency of the IoT model after every data transmission from sensors, the proposed
algorithm suggests an idea to identify the alternate cluster head which has more
energy and less load is to be elected as cluster head by the substation.
The energy in a sensor is calculated based on transmitting and receiving the data
in Eqs. (1) and (2). The formula used to calculate the energy is given as,
T x E (m, l) = E ∗ m + λ ∗ m ∗ l 2 (1)
Rx E (m) = E ∗ m (2)
where, l is the length between two sensors, E is the consumption of energy factor
denoting the power per bit during transmission and receiving the data and λ is the
amplifier coefficient to send a single bit.
The proposed flowchart of ECHGS method is shown in the Fig. 2. Initially the
clusters are formed based on the land area among the total agriculture. Sensors are
placed equally in all the clusters. Cluster head is selected in each cluster based on
the high energy and less load values.
Data gathering is done by the cluster head and send to the substation. Substation
forwards the data to the data server for analyze the level of irrigation to each cluster.
These analytical reports can be given as output to the users when they use the IoT
based smart irrigation system.
4 Results
The proposed method results are analyzed using Network Simulator version-2 (NS2).
The nodes in the simulation environment are distributed in the communication
network. The channel type used in this simulation is a wireless channel, simula-
tion time considered is 100 s, the number of nodes considered are 50, MAC type
of the protocol is 802.11, traffic model used is CBR, transmission range is 250 m,
network interface type used is wirelessPhy and mobility model used is random way
model. In this simulation environment, the proposed method ECHGS results are
evaluated in terms of residual energy and throughput. The proposed ECHGS method
results are outperformed the results of TCDGS and CIDT.
The proposed model in the IoT environment aims to increase the network lifetime
by maximizes the throughput. The Residual Energy (RE) is calculated as the sum of
remaining energy available in all the nodes as shown in Eq. (3) Residual Energy in
each node is the amount of energy remaining in a sensor node at the current moment
of time.
610 K. Padmanaban et al.
M
RE = R Ei
i=1
(3)
Residual Energy is the remaining energy stored in the node. Figure 3 shows
the residual energy comparison of proposed system ECHGS and existing methods
TCDGS and CIDT. The proposed method ECHGS residual energy is higher than the
existing methods of TCDGS and CIDT. Residual is maximized gradually from the
Enhanced Cluster Head Based Data Gathering (ECHGS) Technique … 611
simulation time 0 s to 100 s. The proposed ECHGS method residual energy is higher
by 4.54 and 14.59% when compare to TCDGS and CIDT methods.
The proposed model in IoT environment aims to increase the performance by maxi-
mizes the throughput. Throughput is the total successful data delivered to the server
in the IoT environment. It is calculated Eq. (4) as the total successful data transferred
(RP) in a unit of time (T).
N
Th = R P j/T
j=1 (4)
Figure 4 shows the throughput of the proposed method. The throughput is eval-
uated for the proposed system ECHGS and compared with the existing methods
TCDGS and CIDT. Throughput of the ECHGS method is higher than the existing
methods TCDGS and CIDT. Throughput is increased gradually when the simulation
time 0 s to 100 s. The proposed ECHGS method throughput is higher by 11.57 and
50.00% when compare to TCDGS and CIDT methods.
612 K. Padmanaban et al.
Fig. 4 Throughput
5 Conclusion
In this paper, the IoT based smart irrigation system using ECHGS method is discussed
to address the solutions to increase the network lifetime and improve the network
performance. The sensor nodes are clustered and a cluster head for each node is
selected based on the essential parameters such as residual energy and load. The
effective way of selecting a cluster head in each cluster leads to the better results
from sensor node data collection step to data transfer up to the server. The proposed
ECHGS method results are evaluated and compared with the existing methods of
TCDGS and CIDT in terms of residual energy and throughput. The proposed ECHGS
method results are better than the existing methods. In future, IoT based smart irriga-
tion system can be implemented in real-time with other parameters packet delivery
ratio and packet loss ratio.
References
1. Rapate GS, Naveen NC (2018) Energy and routing efficiency in IoT: proposal for combined
approach. In: 2018 International conference on electrical, electronics, communication,
computer, and optimization techniques (ICEECCOT), Msyuru, India, pp 451–454
2. Atzori L, Iera A, Morabito G (2010) The internet of things: a survey. Comput Netw
54(15):2787–2805
3. Deshpande Niranjan R, Vadane Pandurang M, Sangle Sagar D, Dighe MS (2016) A IOT-based
modern healthcare system using body sensor network (BSN). Int J Innov Res Comput Commun
Eng 4(11):19540–19546
4. Padmanaban K, Jagadeesh Kannan R (2016) Tree cluster based data gathering scheme
(TCDGS) in wireless sensor networks. Int J Comput Technol Appl 9(61):2809–2818
Enhanced Cluster Head Based Data Gathering (ECHGS) Technique … 613
5. Wang C, Ma H (2011) Data collection in wireless sensor networks by utilizing multiple mobile
nodes. In: 2011 seventh international conference on mobile ad-hoc and sensor networks,
Beijing, pp 83–90
6. Lee S, Cha J, Kim KS (2019) Data gathering and application to building energy optimization
with sensitivity analysis for IoT applications. In: 2019 International SoC design conference
(ISOCC), Jeju, Korea (South), pp 184–185
7. Ke H, Wang J, Wang H, Ge Y (2019) Joint optimization of data offloading and resource allo-
cation with renewable energy aware for IoT devices: a deep reinforcement learning approach.
In: IEEE Access vol 7. pp 179349–179363
8. Guo Y, Xiang M (2019) Multi-agent reinforcement learning based energy efficiency optimiza-
tion in NB-IoT networks. In: 2019 IEEE globecom workshops (GC Wkshps), Waikoloa, HI,
USA, pp 1–6
9. Appala Raju V, Sri Harsha V, Bhanu Deepthi N, Prasanth N (2018) Zonal stable election
protocol for heterogeneous wireless sensor networks. Int J Eng Technol (UAE) 7:725–728
10. Dhage MR, Vemuru S (2018) A effective cross layer multi-hop routing protocol for heteroge-
neous wireless sensor network. Indonesian J Electri Eng Comput Sci 10(2):664–671. https://
doi.org/10.11591/ijeecs.v10.i2.pp664-671
11. Goutham Chand K, Sidhendra M, Hussain MA (2018) Soil nutrient measurement in paddy
farming using IoT. Int J Eng Technol (UAE) 7:356–358
12. Gupta P, Satyanarayan KVV, Shah DD (2018) Development and testing of message scheduling
middleware algorithm with SOA for message traffic control in IoT environment. Int J Intell
Eng Syst 11(5):301–313. https://doi.org/10.22266/IJIES2018.1031.28
13. Rao MV, Rama Krishna TV, Ganduri R, Roohi A (2018) An effective energy management
system for smart office cubicles using IoT. J Adv Res Dynam Control Syst 10(2 Special
Issue):338–347
14. Krishna MNV, Harsha NS, Kasula VDK, Swain G (2017) Optimization of energy aware path
routing protocol in wireless sensor networks. Int J Electri Comput Eng 7(3):1268–1277. https://
doi.org/10.11591/ijece.v7i3.pp1268-1277
15. Rajakumar R, Amudhavel J, Dhavachelvan P, Vengattaraman T (2017) GWO-LPWSN: grey
wolf optimization algorithm for node localization problem in wireless sensor networks. J
Comput Netw Commun. https://doi.org/10.1155/2017/7348141
16. Gupta P, Shah DD, Satyanarayana KVV (2016) An IoT framework for addressing parents
concerns about safety of school going children. Int J Electri Comput Eng 6(6):3052–3059.
https://doi.org/10.11591/ijece.v6i6.10448
17. Rao KR, Kumar TR, Venkatnaryana C (2016) Selection of anchor nodes in time of arrival for
localization in wireless sensor networks.https://doi.org/10.1007/978-81-322-2671-0_5
Estimation of State of Charge Using EKF
and SVR in Li-Ion Battery
1 Introduction
As air pollution due to fossil-fueled vehicle emissions has been raising at an alarming
rate, the automotive sector has been looking forward to finding other efficient ways
to power the engines. Electric Vehicles (EVs) seems to be a great solution to address
this problem. According to the International Energy Agency (IEA) in 2018 electric
car sales has been increased by 68% globally which is almost double the year-on-
year average sales growth. This shows the impact of EVs in the transportation sector.
However, the expert agrees on the need for the improvement in battery technology
which is a major bottleneck for EVs.
Proper operation of EVs depends on fuel cells (battery pack) and their management
systems. So for ideal energy transformation, stable and reliable power batteries are
very important [1]. SoC is considered as an important parameter in the Battery
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 615
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_59
616 P. Sannihith Reddy and R. Shanmughasundaram
Management System (BMS). The BMS monitors SoC and provides information
about power level of the battery. The inefficient control of Li-ion battery charging
process can affect the battery life and might also damage the battery. The SoC can
be defines as the level of charge measured in comparison to its total capacity.
Out of the many available methods to estimate the SoC, one simple and common
way is to build linear relation between battery open circuit voltage (OCV) and it’s
SoC. Despite, its simplicity this method is not preferred for dynamic applications.
Because OCV takes long time (few hours) to settle down after a charge or discharge
cycle [2]. The other notable methods include Ampere-hour, coulomb-counting, modi-
fied coulomb-counting method etc., that are used for estimating the SoC. However,
these methods suffer from accumulation error with increasing time. Few advanced
techniques like neural network method require huge training data and many complex
calculations are involved in the process [3].
The model proposed in this paper makes use of EKF and SVR for estimating
the SoC. EKF is one of the finest estimation methods available for a non-linear (but
differentiable) state space model of a physical system. In EKF, SoC is estimated by
comparing model outputs with the measurements acquired from the actual system.
The physical system is modelled in a way that Gaussian white noises affect state and
output equations [4]. The support vector regression model developed for estimating
state of charge is made from the utilization of the right battery data to train and
selection of suitable kernel function. The data set used in this model was collected
from the CALCE battery group [5].
2 Battery Model
Vt = V2 − V1 − R0I (1)
Estimation of State of Charge Using EKF and SVR in Li-Ion Battery 617
Substituting V2
From the above equations, the following system equations can be derived:
I
SoC
= kC2
(3)
Vt − R1IC1 Vt + 1
C1
I
V2 = k SoC + Vt + I R0 (4)
Equation (3) represents state equation and Eq. (4) represents observation equation
of the battery model. The discretization to get continuous equation and calculation
of SoC using EKF are explained next part.
3 EKF Implementation
From Eqs. (3) and (4), state equation of the system can be expressed as:
T
x = x1 x2 , x1 (t) = SoC, x2 = Vt (5)
Consider system input u(t) = I and output y(t) = Vt , then battery model will be
represented as below:
ẋ = f (x, u) + w (6)
ẏ = g(x, u) + v (7)
To linearize the model, Taylor series expansion is needed on Eq. (6) and (7). After
linearization:
ẋ = At x + Bt u + w (10)
ẏ = Ct x + Dt u + v (11)
∂ f (x, u)
At = (12)
∂x
∂ f (x, u)
Bt = (13)
∂u
∂g(x, u)
Ct = (14)
∂x
∂g(x, u)
Dt = (15)
∂u
The discretized model is expressed as follows:
xk+1 = Ak xk + Bk u k (16)
Estimation of State of Charge Using EKF and SVR in Li-Ion Battery 619
yk = Ck xk + Dk u k (17)
Pk = A Pk−1 A T + Q (19)
Pk = (I − L k Ck )Pk−1 (22)
Using the above formulae, EKF calculates the system state. The main step of
the EKF is to adjust the kalman gain according to error magnitude by comparing the
predicted value with the measured and this gain is used to compute the further values.
The bigger the error, greater the gain and greater the correction rate of estimated
value; lesser the error, smaller the gain, and lesser the correction rate of estimated
value. This provides a lucid reason as to why the EKF method is good in real-time
computation performance [7].
4 Simulation
The simulation of EKF algorithm has been carried out in Matlab Simulink based on
the battery model. The capacity of Li-ion battery used in this model is 2000mAh and
cut-off voltage is 4.2 V.
Simulation parameters are:
620 P. Sannihith Reddy and R. Shanmughasundaram
1
• Initial State:
0
10
• Initial Covariance:
01
• Q = 0.1
• R = 100
To validate the result of extended kalman filter algorithm, the simulation results
were compared with the experimental data collected from CALCE battery group.
Figure 2 shows the estimated SoC values by EKF to the actual values. Error in
the estimation can be observed from Fig. 3.
5 SVR Implementation
the information is not a direct task. However, the main idea remains same: to minimize
error, identify the hyperplane, which maximizes the margin, keeping in mind that
part of the error is tolerated [8, 9].
The accuracy of the SVR model majorly depends on the proper data set, kernel
function and model parameters. The model parameters should be chosen wisely or
else there is a chance that the model may not behave as it is intended to. Also, select
the input features that has considerable effect on the required target parameter. This
helps to avoid unnecessary complexity and improves the accuracy [10].
Here in this paper SVR is implemented using python sklearn library. The data
required to the model is collected from CALCE battery data group. The kernel func-
tion used is Radial Bias Function (RBF). Because, the optimal predictive outcome
is better for a nonlinear kernel compared to a linear one. The model parameter C,
which is the penalty parameter used to control the trade-off between training error
and a testing error. This helps in generalizing the classifier to unconsidered data. C
is set to 0.1 in this model. The value of C is selected by comparing the accuracy with
the various C values.
Figure 4 shows the variation in model prediction accuracy for various C values.
It is found that at C = 0.1 model is getting saturated. So C is set to 0.1 to avoid
further complexity by increasing the C value. The other model parameter gamma is
set to 1, gamma determines how quickly boundary dissipation happens; larger values
decrease the effect of any individual support vector. Figure 5 shows how variation in
c and gamma values effect the classification model.
The idea is to develop a model that has the ability to estimate SoC accurately in
real-time. It is known fact that EKF is highly dependent on the battery model. But
creating a battery model which can address all real-time scenarios is not an easy
task. So this might affect the EKF model estimation accuracy in actual situation. To
address this and improve the prediction accuracy, a kind of closed feedback SVR
model is developed by giving SoC estimated by the EKF algorithm as one of the input
parameter along with voltage, current and discharge capacity to the SVR model. This
acts as indirect closed feedback loop and help the model to correct the errors in EKF
model estimation, which results in better accuracy.
Figure 6 shows the variation in SoC values predicted by SVR algorithm vs actual
SoC values. The SVR model prediction accuracy is found to be 99%, which is 4%
more than EKF. The increase in accuracy may seem to be small but the model is
capable of handling real-time scenarios in a better way when compared to EKF
model alone.
6 Conclusion
A hybrid model for estimating the SoC based on EKF and SVR algorithm has been
proposed in this paper. Initially required battery data set for the model has been
obtained from CALCE battery group, and then a mathematical model of the battery
with the same specification as of data set collected is developed. Then using the model
Estimation of State of Charge Using EKF and SVR in Li-Ion Battery 623
system state equation and observation equation are obtained. Now using EKF algo-
rithm equations iteratively system state has been calculated and updated constantly.
Finally, an accurate estimation value is achieved from EKF. Now this estimated SoC
value along with other battery parameters is used to develop the SVR model, which
resulted in accurate SoC prediction.
References
1. Ding-Xuan Y, Yan-xia G (2003) SOC Estimation of Lithium-Ion battery based on kalman filter
algorithm. Appl Mech Mater 347–350
2. Ciortea F, Rusu C, Nemes M, Gatea C (2017) Extended Kalman Filter for state-of-charge
estimation in electric vehicles battery packs. In: 2017 International conference on optimization
of electrical and electronic equipment and 2017 Intl aegean conference on electrical machines
and power electronics, Brasov, pp 611–616
3. Muthumanikandan S, Shanmughasundaram R (2016) Estimation of state of charge of lithium
ion battery using artificial neural networks. Int J Control Theory Appl 9(10):4331–4338
4. Taborelli C, Onori S (2014) State of charge estimation using extended Kalman filters for battery
management system. In: IEEE international electric vehicle conference (IEVC), Florence, pp
1–8
5. https://web.calce.umd.edu/batteries/data.html
6. Krishnakumar A, Shanmughasundaram R (2018) Simplified SOC estimation by EKF in Li-Ion
cell. JARDC 3:616–622
7. Kitagawa M, Shibuya H, Takehara J (1992) Development of battery state of charge indicator
for electric vehicles. In: Proceedings of the 11th international electric vehicle symposium
8. Saimurugan M, Ramachandran KI, Sugumaran V, Sakthivel NR (2011) Multi component fault
diagnosis of rotational mechanical system based on decision tree and support vector machine.
Expert Syst Appl 38:3819–3826
9. Sabareesh GR, Sugumaran V, Ramachandran KI (2006) Fault diagnosis of a taper roller
bearing through histogram features and proximal support vector machines. In: IEEE inter-
national conference on signal and image processing, December, B V Bhoomaraddi college of
Engineering and Technology, Hubli
624 P. Sannihith Reddy and R. Shanmughasundaram
10. Haq IN, Saputra RH, Edison F, Kurniadi D, Leksono E, Yuliarto B (2015) State of charge
(SoC) estimation of LiFePO4 battery module using support vector regression. In: Proceedings
of the joint international conference on electric vehicular technology and industrial, mechanical,
electrical and chemical engineering, Surakarta, pp 16–21
11. https://scikitlearn.org/stable/_images/sphx_glr_plot_rbf_parameters_001.png
Evolving Spiking Neural Network
for Prediction Problems in Healthcare
System
Abstract This paper highlights the role of evolving spiking neural networks (an
enhanced version of SNN) for predicting medical diagnosis. This article aims to
focus on regression problems under a supervised learning strategy. In this paper, we
have trained and tested eSNN on benchmarking datasets. Among the three datasets,
one is the ICU Dataset which helps in predicting the recovery ration of patients who
stayed in ICU. Another dataset is Plasma_Retinol which predicts the risk of cancer-
related to certain carotenoids. Dataset pharynx is a part of a study conducted in the
USA to determine the success rate of two radiation types. The selected datasets are
those which were previously used for BioMedical Engineering related tasks. Later
the evaluation was conducted using Regression Metrics. From experiment results, it
is concluded that eSNN with standard parameters without optimization performed
well but there is still space available for improvement to achieve the highest possible
prediction scores.
1 Introduction
Artificial neural networks can be appraised as one of the vigorous classifiers. Because
of their ability to handle noisy data skillfully [1–4]. ANN solves problems that
are established on approved algorithmic techniques [5–7]. SNNs (Spiking Neural
Networks), the 3rd generation of ANNs, plays a significant role in the processing of
biological information [8]. The initial two prototypes of SNN are conspicuous which
are recognized as the Threshold and Conductance model [9]. There is one derived
model of SNN which is known as Evolving Spiking Neural Network (eSNN), which
was first presented by [10]. eSNN is the one that evolves through learning. For
encoding techniques, eSNN utilizes Population encoding. As for the training, the
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 625
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_60
626 T. Ibad et al.
one-pass learning method has been chosen and it fits well. The target of the training
is to establish an output neuron repository with class labels [11]. The objective
of eSNN is to serve as a fast-learning process that is suitable for both offline and
online tasks. eSNN is widely applied in classification tasks, such as face recognition
[12] ecological problems [13], taste recognition [14] and has accomplished better
results than conventional methods. But for regression tasks, the contribution of eSNN
is not highlighted. Therefore, this study aims to provide a brief review of eSNN
performance in Regression Problems.
2 Literature Review
In the literature study of eSNN, there is a wide range of its application for classifi-
cation tasks. Such as the Integration of eSNN with the Harmony search algorithm
[15]. Firefly Algorithm was used to optimized eSNN for classification problems [9].
Multi-Objective Differential Evolution algorithm is combined with eSNN in [16],
to achieve the best accuracy results for classification. Differential evolution (DE)
was integrated with eSNN in [19] to optimize it for Classification Problems. A new
hybrid K-DESNN for clustering issues was established by integrating D.E and K-
means with evolving spiking neural networks (K-means eSNN) in [17]. eSNN was
optimized with the Harmony Search Algorithm (HSA) in [18]. To tackle the issue
related to hyperparameter sensitivity of eSNN, an optimizer was needed by eSNN
for parameter optimization. Therefore, this article proposed the combination of HSA
as an optimizer and ESNN as a classifier for the optimization of parameters. There
were several attempts to solve the issue of manual parameter tuning. To compare
the results of the optimized eSNN model there is a standard eSNN with parameter
values: C: 0.9, Sim: 0.1, and Mod: 0.75 introduced in [19]. As stated earlier, mostly
optimized eSNN frameworks are for classification tasks. Therefore, this study aims
to evaluate the standard eSNN for Regression Problems. For this paper, we have
trained eSNN on popular datasets from literature, for Regression Problems.
Evolving Spiking Neural Network for Prediction … 627
3 Methodology
The training begins with the initialization of 3 eSNN parameters in the interval [0,
1]: C (proportion factor) and Sim (similarity value), and Mod (modulation factor).
Each sample i that is from a similar class l is encoded into numerous pre-synaptic
input neurons named as j. The weight wj is calculated using Mod i and or der j .
Where, Mod i represents modulation factor from Thorpe neural model. The or der j
defines the rank of an emitted spike from neuron j. The output neuron similarity is
estimated according to the Euclidean distance between the neuron weight vectors. If
a particular neuron is believed to be very similar to others, it will combine with the
truly similar one. Thus, the merging procedure involves the calculation of average
weight vectors along with the threshold value. N represents the figure of samples
formerly used to upgrade output neuron k.
3.2 Datasets
Benchmarking datasets from the OpenMl.org dataset repository, are used to train and
test eSNN see Table 1. The chosen datasets are related to health sciences. And were
collected to be used for a specific study based on regression analysis. The ICU dataset
contains information of 200 survived patients after staying in an adult intensive care
unit (ICU). Plasma Retinol, datafile comprises of 315 observations on 14 variables.
628 T. Ibad et al.
The pharynx data is a small portion of a large medical trial conducted in the United
States for patients with cancer (squamous carcinoma).
The validation of eSNN conducted using the following famous regression metrics.
• R-Squared: It is a statistical degree of indicating how well the regression
predictions estimated the real data values. It is also called the Coefficient of
Determination
• Mean Absolute Error: It calculates the average amount of the errors in a set of
calculations, without focusing their direction. Concisely, it is the average absolute
difference between X and Y.
• Root Mean Square Error: It is the measure to tell how the residuals (i.e., predic-
tion errors) are spread out over the regression line. It is the standard deviation of
the residuals.
This section demonstrates the findings of the analysis based on prediction accuracy
and other measures. Additionally, the experiments are carried out 10 times for both
pieces of training as well as testing on the whole data sets. From a total number of
samples, 80% of the data is the training dataset and the rest is used for testing. The
experiment was conducted on ICU, Plasma-Retinol, and Pharynx Datasets from the
field of Medical and Health Sciences. The results have been evaluated based on their
prediction performance. According to the results presented in Table 2 the average
R-square value is negative for Plasma_Retinol. It is conceivable to get a negative
R-square for calculations that do not comprise of the constant period. In the case of
negative results, R-square cannot be interpreted as the square of a correlation [20]. For
ICU and Pharynx datasets R-square values are positive but very low. In general, the
higher the R-squared, the better the model fits the data. Statistically, Mean Absolute
Error (MAE) is one of the several metrics for reviewing and evaluating the excellence
of a machine learning model. The unit of MAE value is the same as the target variable.
The average MAE value for ICU 0.49, for Plasma_Retinol it is 96, and for Pharynx
data it is 116. This value indicates that the prediction error for ICU is very low. But for
the rest of the two datasets, error values are high which means eSNN did not predict
accurately. RMSE tells us how focused the data is across the line of finest fit. If the
RMSE will be 0, it implies that the points lie on the regression line (and therefore
there are no inaccuracies). The unit of RMSE value is also the same as the target
variable. But the higher the value the greater the prediction error, which indicates the
poor performance of the model. According to average values from Table 2, it can be
seen that the RMSE value for Plasma_Retinol is 132, and for Pharynx data it is 186
which describes that eSNN did not predict well for these two datasets. As the rate of
error is high. But for the ICU dataset, the error rate of eSNN prediction performance
is low i.e., 1.13. This shows good eSNN performance. The overall results of eSNN
prediction performance is a combination of good and bad fit. In this study, eSNN
was with standard parameters, so it could be hypothesized that eSNN can improve
its performance with its optimized parameters (Fig. 1).
In this study, we have tested standard eSNN for regression problems. The model is
trained on widely used datasets from OpenML.org. During the literature analysis, it
was found that eSNN was widely used and tested for classification problems. But
regression problems are neglected except for a few stdies. So, this study aims to test
eSNN for the regression task. Here we used eSNN with standard parameter values.
As stated, earlier eSNN is sensitive to its parameters. Therefore, the selection of
parameter values is a crucial task. Here eSNN with standard values of parameters is
trained and tested. According to the results eSNN performed not so well for regres-
sion datasets. Hence there is wide room for improvement for the same problems.
Future work could be the optimization of eSNN with several well-known Optimiza-
tion Algorithms. So that eSNN would achieve the best possible results in solving
regression problems.
630 T. Ibad et al.
Acknowledgements The authors acknowledge the support of this research by the Yayasan Univer-
siti Teknologi PETRONAS Fundamental Research Grant (YUTP-FRG) under Grant 015LC0-119.
References
1. Morris A, Josifovski L, Bourlard H, Cooke M, Green P (2000) A neural network for classi-
fication with incomplete data: application to robust ASR. In: Proceedings ICSLP 2000 (No.
CONF)
2. Abdulkadir SJ, Suet-Peng Y, Foong OM (2013) Variants of particle swarm optimization in
enhancing artificial neural networks. Austral J Basic Appl Sci
3. Abdulkadir SJ, Yong S, Zakaria N (2016) Hybrid neural network model for metocean data
analysis. J Inform Math Sci 8(4):245–251
4. Abdulkadir SJ, Yong S-P, Marimuthu M, Lai F-W (2014) Hybridization of ensemble Kalman
filter and non-linear autoregressive neural network for financial forecasting. Mining Intell
Knowled Explor 72–81
5. Abdulkadir SJ, Yong SP (2014) Empirical analysis of parallel NARX recurrent network
for long-term chaotic financial forecasting. In: 2014 International conference on computer
and information sciences ICCOINS 2014—a conference of world engineering science and
technology congress ESTCON 2014—proceedings
6. Abdulkadir SJ, Yong SP (2015) Scaled UKF–NARX hybrid model for multi-step-ahead
forecasting of chaotic time series data. Soft Comput 19(12):3479–3496
7. Abdulkadir SJ, Yong S-P (2013) Unscented Kalman filter for noisy multivariate financial time-
series data. In: International workshop on multi-disciplinary trends in artificial intelligence
8. Ghosh-Dastidar S, Adeli H (2009) Third generation neural networks: spiking neural networks.
In: Advances in computational intelligence. pp 167–178
9. Roslan F, Hamed HN, Isa MA (2017) The enhancement of evolving spiking neural network
with firefly algorithm. J Telecommun Electron Comput Eng
Evolving Spiking Neural Network for Prediction … 631
10. The ECOS framework and the ECO learning method for evolving connectionist systems. J Adv
Computat Intell 2(6):195–202
11. Lobo JL, Laña I, Del Ser J, Bilbao MN, Kasabov N (2018) Evolving spiking neural networks
for online learning over drifting data streams. Neural Netw 1(108):1–9
12. Wysoski SG, Benuskova L, Kasabov N (2008) Fast and adaptive network of spiking neurons
for multi-view visual pattern recognition. Neurocomputing 71(13–15):2563–2575
13. Schliebs S, Platel MD, Worner S, Kasabov N (2009) Quantum-inspired feature, and parameter
optimization of evolving spiking neural networks with a case study from ecological modeling.
In: 2009 international joint conference on neural networks 2009 Jun 14 IEEE
14. Soltic S, Wysoski SG, Kasabov NK (2008) Evolving spiking neural networks for taste recogni-
tion. In: 2008 IEEE international joint conference on neural networks (IEEE World Congress
on Computational Intelligence) 2008 Jun 1. IEEE, pp 2091–2097
15. Yusuf ZM, Hamed HN, Yusuf LM, Isa MA (2017) Evolving spiking neural network (ESNN)
and harmony search algorithm (HSA) for parameter optimization. In: 2017 6th international
conference on electrical engineering and informatics (ICEEI) 2017 Nov 25. IEEE, pp 1–6
16. Saleh AY, Shamsuddin SM, Hamed HN (2015) Multi-objective differential evolution of
evolving spiking neural networks for classification problems. In: IFIP international conference
on artificial intelligence applications and innovations 2015 Sep 14. Springer, Cham
17. Saleh AY, Hameed HN, Najib M, Salleh M (2014) A novel hybrid algorithm of differential
evolution with evolving spiking neural network for pre-synaptic neurons optimization. Int J
Adv Soft Comput Appl 6(1):1–6
18. Saleh AY, Hamed HN, Shamsuddin SM, Ibrahim AO (2017) A new hybrid k-means evolving
spiking neural network model based on differential evolution. In International conference of
reliable information and communication technology 2017 Apr 23. pp 571–583
19. Saleh AY, Shamsuddin SM, Hamed HN (2016) A memetic harmony search algorithm based on
multi-objective differential evolution of evolving spiking neural networks. Int J Swarm Intel
Evol Comput 5(130):2
20. Kitchenham BA, Pickard LM, MacDonell SG, Shepperd MJ (2001) What accuracy statistics
measure. In: IEEE proceedings-software. Jun 1. vol 148(3). pp 81–5. Menard S. Applied logistic
regression analysis. Sage, 2002
Prototyping Pro-Active Wearable Gadget
for the Surveillance of Coal Miners
in Pakistan
Abstract After an advent of internet of things (IoT), the entire dynamics of this
world has been changed so rapidly. Being a student of science and technology, it is
very much essential for us to think about the potential alternatives for social issues.
In this regard, one may notice that there are thousands of people who die annually in
the coal mines of Pakistan and rescue team cannot help in any way. This is one of the
most emerging issues right now in other developing countries too where extraction
of minerals from in depth mines is going on. This paper identifies this issue and
suggests the prototyping of pro-active IoT-enabled wearable gadget for the smart
surveillance of coal miners within the mines of Pakistan. This wearable IoT device
not only monitor the hazardous gas, pulse rate of miner, humidity and temperature
conditions within and exact altitude where miner is working currently but it will also
provide a pro-active control to miners to generate any sort of query they face while
working within the dark and depth mines using error palette. This entire data will be
received at a specific Internet protocol (IP) using internet broker. This will also help
the rescue officers to decide their rescue flow accordingly.
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 633
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_61
634 R. Kumar et al.
1 Introduction
In the era, where internet of things (IoT) has evolved the dynamics of every part and
corner of the world, there are still some areas where the utilization of IoT-enabled
systems is still at the stage of infancy. One of such areas is the coal mining industry
where various Alliances of support workers and coal miners witnessed the loss of
their loved ones. In the list of recent incidents, one may see the tragedy of Sanjdi
coal mine of Baluchistan, Pakistan where more than 10 people died, and 8 people
were in critical situation because of the leakage of some poisonous gas.
So far, one may see the serious concerns of International Industrial Union for
coal miners and specifically the focus of Pakistan central mines labor federation
(PCMLF) to resolve such issues and they both are in favor of IoT enabled systems
i.e. wearable gadgets. The proposed IoT-enabled pro-active system in this paper
cannot be implemented within coal mine due to constraints of communication but if
managed on a wireless sensor nodes or network than thousands of the lives may be
secured easily. Even in the case of any uncertainty i.e. collapse of a mine, the rescue
team can perform their duties more efficiently without wasting unnecessary time.
The suggested system in this manuscript proposes an IoT-enabled system, that
comprises of several sensors as discussed in the section of methodology later and this
make this pro-active gadget as one of its kinds. There are most of the cases where these
rescue professionals start their operation at a very unknown place where there is very
less success probability. By utilizing the proposed system, these professionals can
save maximum coal miners because this system will not only be used for surveillance
purpose unlike in [1]. This gadget is pro-active in nature which means now the coal
miners can intimate their location to base station if they are fortunately alive and as
per altitude the rescue team initiate the operation.
2 Literature Review
This section brief some of the latest strategies that were previously used in the active
sensorization and detection of physical parameters. These same techniques will be
discussed in general to produce a statement that these sensors can be interfaced
with any sort of micro-controller to design a pro-active wearable gadget. This pro-
active wearable gadget will not only sense the physical parameters within the local
mines of Pakistan to update the base station [1] but this will also enable the miners
to generate a query too. The brief study of literature can be witnessed under this
section that can acquire the exact whereabouts of a miner and let us know about
his surroundings easily. Moreover, previously people used such techniques only for
monitoring and surveillance. One can see the ubiquitous and pervasive technology for
data acquisition [2]. This approach found to be suitable because of less computational
capability but less interaction between the system and user. Some of the researchers
have proposed the single dimension audio signal communication but this will increase
Prototyping Pro-Active Wearable Gadget … 635
the requirement of storage device. These techniques if utilized properly then the
rescue team will not dig at unknown place and instead of this, they will know exactly
the spot from where they can rescue the miner [3]. Researchers of today’s era are
still engaged in proposing the wearable and easy to carry devices, but all are limited
to surveillance [4].
After studying several research contributions, it has been noted that previously
proposed techniques i.e. An alert system design based on android applications [5]
were either limited to just monitoring or to generate the alert only. The one of the
major shortcomings of such strategies where they were entirely unable to share the
exact location of miner who is in trouble within the debris.
One may raise the questions regarding the accuracy of using such sensors; thus,
it is to share that these same sensors have been proposed in the design of various
sensitive incubators where exact data of babies are communicated using wireless
communication [6]. As a researcher, one may see various people have contributed
various alternatives like this to resolve many critical as well as sensitive issues [7].
One can see the use of such sensors in the design of sewer-bot that is designed to clear
the drain line and senses the temperature, humidity, and leakage of any hazardous
gas [8] (Table 1).
The leakage of hazardous gases results in various tragedies thus, there have been
many sensors and ways to identify the leakage such that average slope-based multipli-
cation [9]. In the catalogue of sensors that may detect the leakage of such hazardous
and explosive gases, one may find number of sensors that have made our work
easier to detect the leakage of such explosive or poisonous gases i.e. methane [10].
The main constraint for utilizing such sensors is to establish the communication
network within the underground environment. This type of communication is already
established using low frequency magnetic fields and their immediate transmission
approach [11]. While studying the literature related to such utilization of sensors,
one may come across with MEMS based sensors and devices that are very robust
when it comes to use them within mines or underground environment but this require
the digital communication followed by the standard IEEE 802.15.14 protocol [12].
This IEEE standard is proposed to communicate at low rate within a wireless area
network. Almost in all sort of research contributions so far, researchers are somehow
more tilted towards monitoring of an unknown area and generating the alerts with the
detection of some physical parameters. Some researchers have also proposed some of
the robots for this task i.e. the famous wolverine version 2 which was introduced for
the first time in the year of 2001 by mine health and safety administration (MHSA),
United States of America [13]. This is not the first time that robots are introduced in
such critical applications, they have been used before such that ground hog, Gemini
scout and numbat proposed for the similar issues. One may see the Table 1, that
highlights the summary of such techniques and systems proposed for the similar
cases.
The mentioned Table 1 presents a brief comparison among the proposed tech-
niques. One may see most of the proposed strategies are based generating either the
alert or just for monitoring purpose. In addition to this, majority of the techniques
cannot communicate their global positioning satellite (GPS) location, temperature,
humidity, exact depth and lastly, the status of miner whether he is alive or not.
3 Methodology
In this section, the hardware implementation for the pro-active wearable gadget is
proposed. This manuscript presents some of the efficient sensors to be embedded
with the system to work as per objectives as highlighted in the below mentioned
Table 2.
In this entire pro-active wearable gadget, the user will first turn on the system
by pressing the single pole single through switch (SPST). After just turning it on,
the sensors will start gathering the results of all physical parameters and send it to
Arduino board.
Once Arduino Mega board will receive it, this will further be processed it to send
it to static internet protocol (IP) using ESP8266 Wi-Fi module. One can see the
entire data i.e. temperature, humidity, pulse rate, altitude, and exact location on a
single IP-oriented screen. The role of error palette is to provide a pro-active access
to miner. This is because if anything happens in front of him for which he needs to
update the base station then this smart pro-active wearable gadget may provide him
this facility to communicate easily using this palette. This error palette comprises
of 01 switch and 01 Led dully connected with the same main system. This switch
Prototyping Pro-Active Wearable Gadget … 637
is there for generating an emergency query within the mine and to indicate for help
from base station. One may see the entire block diagram as mentioned in Fig. 1. The
entire prototype as shown in the Fig. 1 can be initiated in acrylic sheet [1] or may
be mounted on any cloth fabric or one may opt for a safety helmet and embed this
system with in it. In this manuscript the proposed system is not mounted on anything
to show that it can be adjusted into any form. The entire schematic diagram is shown
in Fig. 2. Whereas the hardware prototype is illustrated in Fig. 3.
The device has been tested for the results through different ways. In order to check out
the change in altitude and location, this device had been moved to different height and
positions and the entire results were updated accordingly on serial monitor of Arduino
integrated development environment shown in Fig. 4. Moreover, for checking the
pulse values, the sensor has been given an extended wire so that the miner may place
it on of his finger. The entire data has been sent to IP and one can see the display as
illustrated in Fig. 5.
In above figures, the altitude is shown in feet one may convert it into meters too.
The minus (−) sign shows the depth at which the miner is working.
After putting the values on LatLong.net, we come across the exact location of the
webinar as shown in Fig. 6 given below:
Whereas the user can also utilize the error pallet and may generate the query.
After generating the query, the base station may send the help or rescue team to help
638 R. Kumar et al.
the miner within the mines. The query will be also be generated on IP display unit
as shown below in Fig. 7.
5 Conclusion
Acknowledgement This is to acknowledge the support and facility provided for this project by
department of electrical and electronic engineering, Hamdard University, Karachi Sindh Pakistan.
642 R. Kumar et al.
References
1. Abro GEM, Shoaib AS, Safeeullah S, Ghulam A, Kun- dan K, Fiaz A (2018) Prototyping IOT
based smart wearable jacket design for securing the life of coal miners. In: 2018 International
conference on computing, electronics and communications engineering (iCCECE), IEEE, pp
134–137
2. Lukowicz P, Baker MG, Paradiso J (2010) Guest Editors’ introduction: hostile environments.
IEEE Pervasive Comput 9:13–15
3. Kwon GH, Smith-Jackson TL, Bostian CW (2011) Sociocognitive aspects of interoperability:
understanding communication task environments among different organizations. ACM Trans
Comput-Human Interac 18:1–21
4. Cernea D, Mora S, Perez A, Ebert A, Kerren A, Divitini M, Gil de La Iglesia D, Otero N (2012)
Tangible and wearable user interfaces for supporting collaboration among emergency workers.
In: Herskovic V, Hoppe HU, Jansen M, Ziegler J (eds) Collaboration and technology. Springer,
Berlin, Heidelberg, pp 192–199
5. Divitini M, Farshchian BA, Floch J, Mathisen BM, Mora S, Vilarinho T (2012) Smart jacket as
a collaborative tangible user interface in crisis management. In: Proceedings of the workshop
on ambient intelligence for crisis management
6. Chen W, Nguyen ST, Coops R, Oetomo SB, Feijs L (2009) Wireless transmission design for
health monitoring at neonatal intensive care units. In: Applied sciences in biomedical and
communication technologies, ISABEL 2009. 2nd international symposium on, IEEE, pp 1–6
7. Özmen A, Tekce F, Ebeoˇglu MA, Tas, altin C, Öztürk ZZ (2012) Finding the composition of
gas mixtures by a phthalocyanine-coated QCM sensor array and an artificial neural network.
Sensors Actuators B Chem 115(1):450–454
8. Markham A, Trigoni N (2012) Magneto-inductive networked rescue system (MINERS):
taking sensor networks underground. In: Proceedings of the 11th international con ference
on information processing in sensor networks, ACM, pp 317–328
9. Tanmoy M, Das PS, Mukherjee M (2011) Rescue and protection system for underground mine
workers based on ZigBee. Int J Adv Comput Eng Architect 1:101–06
10. Reddy AH, Kalyan B, Murthy CSN (2015) Mine rescue robot system–a review. Procedia Earth
Planet Sci 11:457–462
11. https://www.cdc.gov/niosh/mining/content/emergencymanagementan-dresponse/commtrack
ing/advcommtrackingtutorial3.html
12. https://pdfs.semanticscholar.org/d60c/0fd4782ab327ff61625f0863cba0e20bf67f.pdf
13. Jing L, Yingchun S, Wenhui C, Xufang B (2012) Application research on gas detection with
artificial olfactory system. Sensors Actuators 3:233–254
Design of a Self-Tuning PID Controller
for a Temperature Control System Using
Fuzzy Logic
Md. Tauhidul Islam, Ariful Islam, Rahul Kumar, Ghulam E Mustafa Abro,
Sourav Majumdar, and Vipin Kumar Oad
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 643
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_62
644 Md. Tauhidul Islam et al.
1 Introduction
Temperature control systems are being used in a wide variety of industries to manage
various operations or processing of manufacture. A temperature controller controls
temperature so that the process value matches the set point, but the response differs
due to the characteristics of the controlled object and the controlling method of the
temperature controller.
Various strategies can be exerted to the temperature control system, like PI, PD,
PID [1, 2], Artificial Intelligence (AI) [3], Fuzzy Logic [3], Genetic Algorithm (GA)
[1, 4], Fuzzy Self-tuning PID [3] etc. Recently, some strategies based on PID control-
ling and tuning method have been proposed in order to ameliorate system perfor-
mances. Gani et al. have proposed an ameliorated design using Genetic Algorithm
(GA) on optimal PID tuning of temperature system [1]. Salsbury proposed a feed
forward control method as a replacement of conventional PI feedback control [5].
Some has also done with self-tuning fuzzy PID controllers [3]. However those clas-
sical conventional PID controllers are less effective and some with modern strategies
like GA or AI, there is no doubt about their effectiveness but these strategies are too
hard to implement and costly too. Among these Self-tuning Fuzzy PID controller’s
algorithm is the easiest strategy to implement and it has better effectiveness [4,
6]. This paper predominantly focuses on designing a new self-tuning Fuzzy PID
controller to maintain required temperature.
The proposed design of self-tuning Fuzzy PID controller is mainly based on Fuzzy
Interference System (FIS). The procedure of fuzzy interference system includes three
phases: (i) Fuzzification: In this phase, crisp inputs are transposed into degrees of
membership. The degree of membership is discerned by plugging crisp inputs into
the membership function affiliated to the fuzzy set. (ii) Rule evaluation: In this phase,
each and every fuzzy rule is imposed with a strength value. The strength is discerned
by the degrees of memberships of the crisp inputs in the fuzzy sets of preceding
portion of the fuzzy rules. (iii) Defuzzification: In this phase the fuzzy outputs are
transposed into crisp outputs [7, 8] (Fig. 1).
K
G(s) = e−τ s (1)
Ts + 1
Design of a Self-Tuning PID Controller for a Temperature … 645
where, K = 1 is defined as the static gain and τ = 2.8 as pure lag time, T = 3 as
time-constant of the controlled object [9, 10].
PID controller is vastly used in every controlling system. The first letter of the name
of three controlling terms make up unitedly PID. It is actually a total combination of
proportional, integral and derivative terms of controlling system. Here PID controller
is used for controlling initially for the system. Equation of conventional PID can be
expressed as follows:
de(t)
u(t) = K P e(t) + K i e(t) + K d (2)
dt
The block diagram of the proposed system is shown in Fig. 3. The system is designed
in MATLAB through Simulink. Firstly, a Fuzzy Interference System is designed by
defining membership functions, universe of discourse and fuzzy rules in MATLAB
through Fuzzy Logic Toolbox. Then this FIS system has been implemented in the
de- signed block system in Simulink.
In this paper Mamdani method of Fuzzy Interference system (FIS) is used to
process the fuzzy system. In MATLAB, there is a tool named Fuzzy logic toolbox
which is used to create membership functions, universe of discourse and fuzzy rules.
Here, in this interference system, the crisp inputs are system error (e) and rate of
system error changing (ec) and the crisp outputs are proportional gain (K P ) integral
gain (K i ) and derivative gain (K d ). The range for the system error (e) is taken as
[−1 1], rate of change of error (ec) is [−1 1] and range for the output is considered
as proportional gain (K P ) at [1 5], integral gain (K i ) at [0 1] and derivative gain
(K d ) at [0 1.5]. These ranges are obtained from the observations of conventional PID
controller since the conventional tuning gives as K P 1.285, K i as 0.27 and K d as 0.9
(Fig. 4) and (Table 1).
The proposed design has been simulated in MATLAB through Simulink. Figure 5
shows the step response of both the conventional PID controller and proposed PID
controller. The step response is determined at the step input temperature of 500 °C.
The comparison of performance and robustness between these two controller is
shown in Table 2.
The analysis of the simulation result and comparison table exposes that the
proposed self-tuning controller overshoots 0.63% with peak value of 503.16 whereas
the conventional PID controller overshoots 13.82% with peak value of 569.13. The
proposed method reduced the overshooting problem as well as providing better
dynamic performance and robustness. So, It is clear that the proposed self-tuning
Fuzzy PID controller for temperature control system provides better efficiency in
controlling.
Design of a Self-Tuning PID Controller for a Temperature … 647
4 Conclusion
In this paper a self-tuning fuzzy PID controller has been designed for controlling
temperature system for industrial management. Fuzzy interference system (FIS)
has also been described elaborately in this paper. After simulating the system in
MATLAB, the results expose that this self-tuning fuzzy PID has almost zero over-
shoot with a better dynamic performance than conventional PID. Analyzing the
proposed algorithm and results of this paper, it is concluded that fuzzy PID is more
effective and it can be replaced with conventional PID. This system can be amelio-
rated by farther modification in the designing algorithm like membership functions,
rules and gains which impacts the controlling performance by extenuating both the
overshoot and dynamic response.
Design of a Self-Tuning PID Controller for a Temperature … 649
Acknowledgement This is to acknowledge the support provided from the state art of facility from
Power Electronic Engineering Lab of Hamdard University, Karachi Sindh Pakistan.
Sponsorship
This work is not sponsored by any of the organization.
Conflict of Interest
There is no conflict of interest among the authors.
References
1. Gani MM, Islam MS, Ullah MA (2019) Optimal PID tuning for controlling the temperature of
electric furnace by genetic algorithm. SN Appl Sci 1:880
2. Anwar AZ, Jabeen B (2019) Prototyping non-holonomic hovercraft for path planning and
obstacle avoidance. Sir Syed Univers Res J Eng Technol 9(1)
3. Baogang H, Hao Y (2001) Review of fuzzy PID control techniques and some important issues.
Automatic Sinica 27(4)
4. Rehman NU, Kumar K (2018) Implementation of an autonomous path planning and obstacle
avoidance UGV using SLAM. In: 2018 International conference on engineering and emerging
technologies (ICEET), IEEE, pp 1–5
5. Salsbury TI (1998) A temperature controller for VAV air-handing units based on simplified
physical models. HVAC and R Res
6. Abro GEM, Jabeen B, Ajodhia KK, Rauf A, Noman A (2019) Designing smart Sewerbot for
the identification of sewer defects and blockages
7. Hassan MMM (2013) Current studies on intrusion detection system, genetic algorithm and
fuzzy logic. Int J Distrib Parallel Syst (IJDPS) 4(2)
8. Mugisha JC, Munyazikwiye B, Karimi HR (2015) Design of temperature control system using
conventional PID and intelligent fuzzy logic controller. In: Proceedings of 2015 international
conference on fuzzy theory and its applications (iFUZZY) The Evergreen Resort Hotel (Jiaosi),
Yilan, Taiwan, November 18–20
9. Abro GEM, Bazgha J, Manan A (2019) Stabilization of Non-ho- lonomic 03 dof hovercraft
using robust RST control design. Sukkur IBA J Emerg Technol 2(1): 45–50
10. Diwakar TK, Vivek Y, Raut KH (2014) PID tuning rules for first order plus time delay system.
Int J Innov Res Electri Electron Instrument Control Eng 2(1)
11. Wu X, Wang X, He G (2020) A fuzzy self-tuning temperature PID control algorithms for 3D
bio-printing temperature control system. In: Chinese control and decision conference (CCDC)
by IEEE. ISSN: 1948–9447
12. Jiang W, Jiang X (2012) Design of an intelligent temperature control system based on the fuzzy
self-tuning PID. In: International symposium on safety science and engineering in China, 2012
(ISSSE-2012), Procedia Engineering, vol 43. pp 307–311
Design of a Low-Cost High-Speed
and Large-Capacity Data Logger Using
MicroSD Card
Abstract Data logging is an important process for evaluating the functionality and
performance of many electronic systems such as internet of things (IoT). Most of the
mixed-signal systems digitize the signals from the sensors to generate input signals
and then perform some digital signal processing on these input signals to generate
output signals. Data logging of these input and output signals helps to evaluate the
performance of the systems. Many of these systems contain an FPGA for digital
signal processing. The FPGA can also be used to facilitate data logging. This paper
presents a data logging and data extraction system which is a high capacity, low cost
and ensures data integrity over gigabytes of the data. The system has been evaluated
by logging data in a system, having four inputs and one output signal each having
a sampling rate of 1 kHz, for five hours. Data integrity has been observed for each
sample of the data.
1 Introduction
Sensors are an integral part of electronic systems. Systems such as IoT can have a
lot of sensors. They sense some physical quantities and convert them into electronic
signals. After digitization, these signals are passed to signal processing modules
as input signals. These modules perform digital signal processing on the input
signals and generate some output signals. Most of the electronics systems contain
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 651
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_63
652 M. Sajjad et al.
programmable logic devices like FPGA as a signal processing module. Data logging
of the input and the output signals is necessary for evaluating the performance of
the systems and verifying the functionality of the signal processing techniques in
real-time. In case of reliable data logging, the inputs can be passed to verified high-
level software routine to generate new signals. These signals can be compared to
the recorded output signals for verifying the functionality of the signal processing
modules.
One way for data logging is writing data to NOR-flash, NAND-flash, or EEPROM
which are normally present in the system electronics [1]. But these memories are
small in capacity and can hardly store a few megabytes of data [2, 3]. So these devices
are not suitable for data logging in high throughput systems. They also require erasing
before writing which reduces their data writing speed and add another task for the data
logging controller. Normally these devices are mounted on PCBs by the soldering
process and cannot be separated from the electronic systems. Because of this, the
electronics systems have to be moved in the lab to recover logged data for offline
analysis. Another way is using commercial off-the-shelf (COTS) data recorders [4].
These data recorders can store data in gigabytes and are interfaced with RS-232/RS-
422 ports [5]. They record data to file system of the attached USB flash or microSD
card. Main drawback of these COTS recorders is that they take electrical power and
some space and hence are not a suitable choice for critical missions.
The proposed data logging system consists of a microSD card connected to
four general-purpose digital pins (for implementing SPI) of a FPGA-based main
processing board. Raw data from the sensors and the processed data are written on
sectors of the microSD card. The digital design architecture ensures recording of
each sample, handling of contention arising from simultaneous read and write to the
buffers and matching of cyclic redundancy checks (CRC) during write operations to
the microSD card. Most of the existing data loggers use parity bit for error detection
which only applies to byte level errors in contrast to CRC which applies to the whole
data of the sectors. The data logging system has been tested for five hours duration
and no missing or corrupted data have been observed.
The paper is organized as follows. Section 2 discusses SD card Interfacing.
Section 3 presents implementation detail, followed by the Data and Information
Retrieval in Sect. 4. Results and Interpretation is presented in 5. Finally, the
conclusion is presented in Sect. 6.
2 SD Card Module
The Secure Digital card is widely used for data logging systems. It has flash memory
and a microcontroller embedded in it. The flash memory control operations such as
writing, reading, and erasing are performed inside the SD card. The data transfer
from the host system to an SD card is in the form of data blocks which are in
units of 512 bytes. Two modes can be used for data communication between a host
controller and an SD card i.e. SD mode and SPI mode. By default, the SD card is
Design of a Low-Cost High-Speed and Large-Capacity … 653
in SD mode. Both modes have their pros and cons. SPI communication protocol is
a bit simpler as compared to its default mode because the SD card can be attached
to any host controller using some GPIO pins. Therefore, this mode is most suited
for cost-effective embedded systems. Moreover, the SD card requires two different
clock frequencies for its initialization mode (100–400 kHz), and data transfer (up to
25 MHz) mode as specified in the SD card product manual whereas the commercially
available loggers record data from a RS422/RS485/RS232 which can hardly manage
data transfer up to 430600 Hz. Keeping in view all the challenges of the SD card, a
state machine has been designed to cater to the data archival on the microSD card.
After powering up the system, configure SPI clock frequency between 100–400 kHz
and put MOSI and CS pins to the high state. Send 74–80 clock pulses to the serial
clock pin of the SD card to make it ready for communication. Put CS to low and
send CMD0 to reset the card. The card will respond with 0 × 01 when it will be in
the idle state. After the idle state is achieved, send CMD8 with proper argument and
CRC before the start of the initialization process. The proposed data logging system
architecture comprises of a control unit, a microSD card module, two multiplexers,
and two block RAMs.
A Control unit is like a brain for the data logging system. A state machine is designed
to control operations. The flow chart for the state machine is shown in Fig. 1. At first,
the control unit will get the SD card ready flag from the SD card module to start its
operation. After the SD card is ready, the control unit will cater to the data flag from
sensors. After getting the data flag, the Control Unit will provide an 8-bit address
to a Buffer and store the sensors data on Block RAM. When block RAM is full,
the control unit will generate a RAM full flag to the SD card module to store the
respective RAM data on the sector of the SD card. The control unit also stops the
data writing process by providing an End data transmission signal to the SD card
module.
Single port Block Rams are implemented using IP cores. Each Block RAM has a size
of 500 bytes, having a 9-bit address, 8-bit data-in, 8-bit data-out, and its Read–write
operation is controlled by the control unit. The reason for its size of 500 bytes size is
that the sector length of the SD card is 512 bytes and the data packet after each 1 ms
is 25 bytes. The reason for using two block RAMs is that when the first RAM is full
and is being dumped on the SD card, the incoming data from sensors is stored on the
second RAM to avoid any data loss. Write address as it will take data from sensors,
and then write it on RAM. Whereas the address from the SD card is named as the
654 M. Sajjad et al.
Read address as it will read the stored data from RAM. The initialization process of
the card will start when CMD1 is received by the card. This initialization process
will end when the response is changed from 0 × 01 to 0 × 00. After initialization
data block length is set in units of 512 bytes by sending CMD16 to the card. CMD25
is then sent to the card to configure it for multiple block writing. After the command
is accepted by the card, the host can send data packets to the card.
3 Implementation
Data logging on a PC using serial cables is not possible in remote locations or for the
case of a drone flight. Acquiring real-time data in such a scenario is a tedious task.
The proposed data logging system is Implemented using four sensors connected
to the 4 channels of a 24-bit ADC, which is interfaced with an FPGA using SPI
protocol as shown in Fig. 2. The data from sensors is stored in an internal buffer of
FPGA. When the buffer is full, it generates a trigger to inform the control unit about
Design of a Low-Cost High-Speed and Large-Capacity … 655
data availability. The reason for using a buffer is that its access time is very small
as compared to external memories. Data is also available to actuators to perform
necessary actions and is then stored on the SD card for offline analysis. This data
from sensors can also be made available for further processing.
S1(n), S2(n), S3(n), and S4(n) represents the sensor outputs of first, second, third
and fourth sensor respectively (shown in Fig. 2), so, S(n) is the output of algorithm
which is running on these sensors’ outputs. The algorithm in this proposed system is
just an integration of all sensors. Hence, every sample of sensors data is used in this
integrator algorithm ensuring data integrity. The equation for this integration-based
algorithm is (1).
As ADC is a 24 bit, so each sensor’s data is a 24-bit data. After applying a scale
factor on the sensors data, it becomes a 32-bit data i.e. 4Bytes. So, there are 16 Bytes
of sensors data, and 4 bytes of last iteration summation result hence, making it a data
of 20 Bytes. This data is then framed by a start byte, 2 bytes of checksum, and 2
bytes of stop bytes to make it a data packet of 25 Bytes. After 1 ms, a data trigger
is generated that store these 25 bytes on Block RAM. Block RAM is 500 Bytes,
so after 20 ms Ram is full and is then ready to dump its data on the sector of the
SD card. The last 12 locations of each sector of the SD card are filled with 0xFF.
Sensors data is written in LSB to MSB format. Clock frequency for initialization
and data transfer mode has been set to 384.615 kHz in its native mode so it takes
approximately 1.33 ms to write 512 bytes on a sector. After every 20 ms, a sector
is written on the SD card. So, in 5 h duration, it writes on 900,000 sectors of an SD
card. Each sector is of 512 Bytes, so, the total data is 450 MB.
656 M. Sajjad et al.
For missions involving high data rates for a long duration, data logging in a single
file is not a feasible solution because the software used for post-processing or offline
analysis crashes due to limited memory. To avoid this situation, some computer
programs like hex editors can be used. With the help of MicroSD Card Reader, data
written on the sectors of the microSD card can be viewed and manipulated. A data
file can be created by specifying addresses for both the start sector and the end sector.
In this way, multiple files, having a manageable size for post-processing software,
can be generated through a hex editor. These files contain sector addresses and
bytes of data written by data logging setup. So, we cannot use them without further
processing. To make the files usable for analysis, these files can be further processed
by specially designed software routines on Integrated Development Environment
(IDE) like Microsoft Visual Studio or LabVIEW. Such routines can process these
files on the byte level. So, tasks like checksum matching, removing header or footer,
and concatenating relevant bytes to make a float, double, or integer type data values
can be performed easily.
For example, Fig. 3 shows a comparison of offline and online (or real-time) perfor-
mance of the algorithm and hence the electronic system. Figure 3 also shows outputs
of the same algorithm running in MATLAB (offline mode) and the FPGA (online or
real-time mode). This algorithm has been fed with the same data from the sensor and
ideally the graphs in Fig. 3 should track each other. But this type of mismatch does
not imply any data logging or algorithm error. This kind of deterministic error is due
to different type of truncation modes in double-precision floating-point computations
in MATLAB and FPGA. Since the algorithm consists of integration, which sums up
all the errors arising in the whole mission, so the error tends to increase with time.
6 Conclusion
This paper has presented a data logging setup for recording digitized input and
output signals in an electronic. The presented method ensures that every sample of
the signals is intact even for the missions with data size reaching to gigabytes. By
using the proposed data logging setup, large-capacity data can be recorded with full
integrity at a very low cost. This can help to speed up the evaluation and development
of IoT.
References
1. Kim GJ, Baek SC, Lee HS, Lee HD, Joe MJ (2006) LGeDBMS: a small DBMS for embedded
system with flash memory. In: VLDB, CiteSeer, pp 1255–1258
2. Lee SW, Moon B (2007) Design of flash-based DBMS: an in-page logging approach. In:
Proceedings of the 2007 ACM SIGMOD international conference on management of data,
pp 55–66
3. Kim J, Kim JM, Noh SH, Min SL, Cho Y (2002) A space-efficient flash translation layer for
compactflash systems. IEEE Trans Consum Electron 48(2):366–375
4. De-Leon HL, Quiros RE (2005) Self-contained flight data recorder with wireless data retrieval.
United States Patent 6,898,492
5. Yamawaki Y (1995) Traffic accident data recorder and traffic accident reproduction system.
United States Patent 5,446,659
Machine Learning in Healthcare:
Current Trends and the Future
1 Introduction
Identifying diseases rare or hard to diagnose also depends on detecting what are
known as’ edge cases. As this type of ML system relies on large datasets of raw
pictures (and different transformations) of these conditions, they are more precise
than humans for this type of detection [1]. They should improve automation produc-
tivity and intelligent decision making in primary/tertiary and public health care. This
would have the most significant effect on A.I. instruments since the quality of life
will increase for thousands of people worldwide [2].
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 659
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_64
660 U. A. Usmani and J. Jaafar
The Microsoft Project Inner Eye, which uses ML methods for segmentation and
detecting tumours with 3D images, is an excellent test case. It can assist in the pre-
case preparation, navigation and effective tumour contours for radiation prepara-
tion. M.R.I.s and other sophisticated imaging systems are fitted with ML algorithms,
increasingly used for early cancer detection. ML tools often offer significant value by
increasing the surgeon’s display with information such as cancer’s location during
robotic procedures and other image-led procedures. Therefore, the use of AI/ML
software/networks to help radiologists should be exponentially expanded. Unfortu-
nately, these details are messy and unstructured too. In particular, patient data can not
be used for basic computational and analytical modelling instead of regular business
transaction data [3]. The need for the hour is a powerful and agile AI-enabled plat-
form, which can link and analyze a diverse mix of data types (i.e. blood pathology,
genomics, radiological images, medical history) [4]. These systems should also dig
extensively and uncover the secret trends through the analyses [5]. Furthermore,
doctors and other health practitioners should understand and understand their find-
ings to achieve their success with great confidence and clearness. These bills are
appropriate for the interpretable A.I. and distributed ML systems and can fulfil these
systems’ requirements within a short period. Surgery robots can provide human
surgeons with unique assistance, enhance their vision and manoeuvrability during
operations and make precise and minimally invasive incisions [6]. The optimal stitch
geometry and wound cause less pain. There are inspiring opportunities to apply
AI/ML to such robots in digital operations.
should also use advanced AI/ML technologies and strategies in their organizational
aspects every day. Such programmes’ core purpose should be to build AI-supported
networks to improve healthcare outcomes for the most extensive section of ordinary
citizens. The ultimate aim of current processes in conventional organizations is to
maximize benefit. Robust A.I. methods for managing medical operations need to
differentiate themselves from traditional systems by mixing empathy with income.
Increasingly, important names in the pharmaceutical industry are selecting A.I. and
ML technologies to resolve the challenging issue of useful drug testing [11].
This article offers several famous examples—like Sanofi, Genentech, Pfizer. These
case studies cover all sorts of clinical fields—metabolism, cancer therapies,
immunooncology. Beyond the traditional long-term method, A.I. techniques are
being used more and more to speed up the fundamental processes of early selection
and mechanism discovery. The Berg biotechnology company uses its A.I. method for
study in its research. For start-ups, the evaluation of multi-channel data, including
research papers, patents, clinical trials and patients records, may include state-of-
the-art inferential approaches, Markov chain models, enhanced learning and natural
language (N.L.P.s). The key objectives are to identify patterns and establish high-
dimensional representations stored in the cloud and used during drug discovery [12,
13]. Also, A.I. instruments can help shape why and under what conditions diseases
are more likely to arise. Physicians can direct and plan to intervene (personalized)
even before they begin to display symptoms. The aim is extraordinarily complicated
and challenging. It is to identify exact treatment options for an individual based on
an ongoing personal medical history, decisions about behaviours, genetic data and
pathological tests. The ongoing COVID-19 crisis has shown the value of performing
hundreds of simultaneous development and therapeutic trials of vaccines. Traditional
statistical modelling approaches are almost impossible to use, designed for small
experiments to collect data and distinguish patterns from all these various sources,
often delivering high uncertainty levels. For such a planetary scale, A.I. techniques
must bear the challenge of solving. A wide range of exciting and forward-looking
approaches to AI/ML and platforms have been investigated in the field of health
care. Sujets from radiology assistants to intelligent medical management, person-
alized medicine and digital public health monitoring have been explored [14–16].
Data protection issues and legal frameworks are still recognized as barriers to the
complete adoption of these systems. The type of data that third parties can access and
use legally (for example, the owners of the A.I. and ML software, physical devices
or platforms) may be quite challenging to decide. As technologists and IA/ML prac-
titioners, it is essential to have a bright future in which the AI-algorithms support
ordinary citizens’ trillions to improve their primary health and wellbeing.
662 U. A. Usmani and J. Jaafar
With the rapid population growth, it seems impossible to record and evaluate large
patient information quantities. Machine learning gives us the ability to automatically
identify and process this data, making the healthcare system more complex and
resilient. Machine education in healthcare combines into a common thread two types
of fields: informatics and medical science [17–19]. Machine learning methodology
gives medical research development and analyses complex medical data for further
study.
Many researchers are working to develop new dimensions and features in this area.
Google recently developed an algorithm of learning machines to identify cancerous
tumour mammograms. Stanford also has a deep learning algorithm to assess cancer
of the skin [20, 21]. There are annual conferences, such as Machine Learning for
Healthcare, which seek to explore new automated medical technologies for better
service [22, 23]. Machine education aims to increase the machine’s prosperity, effi-
ciency and reliability. However, the machine learning method in a health system is the
doctor’s brain and experience what a patient requires a human touch and concentra-
tion [24, 25]. This can not be replaced by machine learning or any other technology.
An automated computer can provide better service. Here are the top 10 machine
learning frameworks for healthcare.
The heart is one of the primary organs of our body. We also have several heart
disorders like Coronary Artery Disease (CAD), CHD, etc. (Fig. 1).
Recent concepts have been used significantly in machine learning and data mining
to prevent liver disease. Disease with extensive medical records is challenging to
predict.
However, researchers use machine learning principles like grouping, clustering
and many more to solve this problem. Dataset (I.L.P.D.) can be used for hepatic
disease prediction. This dataset comprises ten variables or liver diseases. A classifier
can be used with the Vector Machine Support (SVM). You can use M.A.T.L.A.B. to
create a prediction system for liver disease.
One of the primary qualifications for healthcare training is robotic Surgery. This
application will soon be a promising field. This application can be divided into four
subcategories, including automatic suturing, surgical capacity assessment, surgical
materials, and surgical modelling. Automation of sutures minimizes machine life
and fatigue (Fig. 4).
The Raven Surgical Robot, for example. Researchers attempt to use a machine
learning approach to test the surgeon’s performance in the least invasive robot proce-
dure. University of California researchers, San Diego (UCSD), explore machine
learning technology to improve surgical robotics. Since robots can not function effi-
ciently in neurosurgery, the manual operational workflow requires time and can not
automatically provide input. The use of machine learning is a benchmark appli-
cation of machine learning in medicine in drug discovery [31]. The Microsoft
Project Hanover aims to implement precise medicine learning machine technolo-
gies. Many Organizations are currently using drug discovery machine learning tech-
nology. Benevolent AI, for example. It aims to use artificial intelligence (A.I.) to
Machine Learning in Healthcare: Current Trends and the Future 665
discover drugs. Machine learning has many advantages in this field, including accel-
erating the process and reducing the failure rate. Machine learning also optimizes
the development and expense of drug discovery (Fig. 5).
An intelligent electronic health record system can boost computer training, including
classification documents and optical character recognition. This programme aims
to create a plan to sort patient requests via email or convert a manual recording
system into an automated system. This application seeks to develop a secure and
easily accessible framework. Electronic health records’ rapid growth has increased
patient healthcare information storage and advancement. For example, duplicating
data removes data errors. It could be used as a classifier or artificial neural network
(ANN) to create an electronic health recorder managed learning algorithm, like the
Support Vector Machine (SVM). Aidoc provides radiologists with machine learning
software to speed up detection. The purpose is to interpret the medical picture to
recognize the body’s abnormalities understandably. The supervised algorithm for
machine learning is used mainly in this field. Machine learning technology is used
for medical image segmentation. Segmentation is the process by which structures
are represented in an image. The graph cutting method is used primarily for image
segmentation. For analysis, standard Radiology Text File Language Processing is
used. The use of computer radiology can also improve patient care. The clinical trial
may be a series of studies requiring responses for biomedical or pharmaceutical effi-
ciency and safety. This research aims to reflect on the ongoing progress of treatment
(Fig. 6).
It takes a lot of time for this clinical trial. In this area, machine learning has an enor-
mous impact. An ML system can provide full service and real-time monitoring. The
advantage of using machine learning technology is that it can be remotely controlled
in clinical trials and studies. Machine learning also provides patients with a healthy
clinical environment. The clinical trial’s efficiency can be improved with supervised
computer education in healthcare.
Genomics in October 2016, a joint project with Quest Diagnostics to support accurate
medicines using cognitive computing, genomic tumours, and a biopharmaceutical
company located in Berg. Current research programmes include intravenous tumour
dosing experiments and prostate cancer identification and management. In Google,
the Deep Mind Health disclosed several cooperation with the United Kingdom last
year, including Moorfields Eye Hospital in London, where macular degeneration
technology is developed in older eyes. Currently, supervised learning allows clini-
cians to choose from selective diagnoses or predict patient risk from symptoms and
genetic data, for example. I.B.M. Watson Oncology is a pioneering medical reform
and history institution to expand treatment possibilities. This personalized treat-
ment approach has significant consequences for patients to boost their wellbeing and
minimize overall healthcare costs. For instance, reductions in health care costs will
fluctuate and decrease as more people stick to medication and treatment plans.
Behavioural adaptation is also a significant driver in the prevention system,
addressed in December in an interview with Emerj by Catalia Health’s Cory Kidd.
And there are many start-ups with varying levels of success in cancer diagnosis,
prevention and care (for example). Somatix—a B2B2C software platform provider
with an ML database analysis that “recognizes mouth-tohand movement,” especially
when smoking stops. SkinVision is the first to build the software for ‘SkinVision’
which was DermCheck for ‘Just, Just C.E. Cert.’ The first time this app was devel-
oped, DermCheck sent images to dermatologists (people rather than machines) via
phone to return to a personalized treatment plan. In this area, the frontier appears to be
precision medicine with the detection of ‘multifactorial’ mechanisms and alternate
therapeutic routes. This study is unaccompanied, often mainly limited to detecting
data patterns without prediction (recently included in supervised learning). Machine
learning has many practical applications that can help develop and conduct clin-
ical trials. Advanced predictive analytics can identify clinical problems and thereby
contribute to lower, quicker, and cheaper testing, for example, for social media,
doctor visits and genetic information for target groups. According to McKinsey,
several more ML applications for improved clinical trial performance have been
developed, including determining sample sizes for better quality, addressing and
adjusting patient recruitment discrepancies, and using electronic health records to
remove information errors (e.g. double entry). DeepMind Wellness, Google, with
University College London Hospital, collaborated on machine-learning algorithms
to identify different levels of safe and cancerous radiation. DeepMind and U.C.L.H.
work to distinguish healthy and cancerous tissue variations on ML applications. More
knowledge is covered in our radiology machine learning applications. Classification
of documents (sorting patient requests by email, for example) with the aid of vector
machines and optical character recognition (transforming a digitized character of
cursive or another manuscript). The MATLAB ML and the handwritten character
recognition network of the Google Cloud Vision API provide only two examples of
creativity.
Intellectual and electronic health records are being developed over the next decade
by the M.I.T. Clinical Machine Learning Community. Diagnostics, clinical deci-
sions and personal care guidelines, including integrated ML/AI, where supported.
Machine Learning in Healthcare: Current Trends and the Future 669
The M.I.T. research site notes that ‘the need to build robust machine learning algo-
rithms will learn from smaller data on the labelling, interpretation and generalization
of natural languages in medical and institutional environments.’ ML and A.I. tech-
nology will monitor and predict all outbreak ovarian diseases through satellite-based
information, historical Web-based information and sociological information. Today’s
opioid crisis is a clear example of A.I. Vector support. Artificial neural networks were
used to predict malaria outbreaks based on temperature, mid-month precipitation,
total positive cases, and other data points. Epidemic severity predictions are urgently
needed, especially in third-world countries lacking in medical facilities, education,
and healthcare access. ProMED-mail is a software that monitors and detects emerging
diseases in real-time.
Significant issues still have to be addressed when integrating ML technology in
pharmaceuticals and medicinal products: data processing is one of the most pressing
problems today. Patient information is also sensitive and not easily accessible. It
seems safe to believe that most of the population is concerned with sharing infor-
mation rather than privacy concerns. It is noteworthy to note, in a survey carried out
in the United Kingdom by the Welcome Foundation in March 2016, that only 17%
of the respondents never decided to distribute their anonymized information to third
parties other than study.
The more open algorithms should comply with the strict drug development regu-
lations; the causes of machine findings must be seen and understood. It is essential in
the pharmaceutical industry to attract data science talent and develop a deep pipeline
of Skills. The breakdown of data silos and the promotion of a “data-centered perspec-
tive” across industries is critical in helping the industry to transform their thinking into
long-term progressive changes and to find value. Pharmaceutical firms have refused
to update or fund research programmes in history unless there is a considerable and
immediate monetary benefit. The relocation of already chaotic electronic records
through networks will be a first step towards creating tailor-made care solutions.
The two leading roles of technology in machine learning and N.L.P. in healthcare
are the most important. Over the years, N L.P. has advanced with validated R.O.I.
recognition, enabling clinics to transcribe notes with useful E.H.R. information using
natural language treatments in healthcare. Front-end speech recognition eliminates
medical records’ role rather than cure, while back-end technology works to identify
and correct transcription errors before moving them to human evidence. While the
market is almost overwhelmed by language recognition technology, many companies
disrupt the field with deep learning algorithms in mining applications that give more
significant opportunities. Clinical recording developments—Healthcare computer
training has resolved clinical recording and free doctors from a complicated E.H.R.
manual system. This was possible due to the dictation from speech to text and the
wording of the data entry capturing structured data while processing. As machine
learning progresses in healthcare, essential knowledge can be gathered from other
new sources, and P.H.M. and V.B.C. initiatives can be powered by analytics. Study of
data mining—The integration of data mining into healthcare systems allows organi-
zations to reduce decision-making subjectivity and provide valuable medical knowl-
edge. Data mining can be a cyclic technology until every HCO has started developing
a successful business plan for better patient care. Computer-assisted coding—NLP-
driven C.A.C. promises to prove coder accuracy im-. Computer-aided code collects
details and information on therapy to capture and optimize any code. The research
showed that current suppliers in the production of C.A.C. solutions on the market
must redirect their answers to a value-driven paradigm’s demands and ensure they
function as expected. Many health I.T. programmes have a regulatory burden because
metrics such as ejection fractions are not stored as separate values. Health systems
have to determine if an expulsion fraction (Fig. 8).
Health systems have to determine if an expulsion fraction is reported in a note and
save any value in a format which can be used for accurate reporting on the organi-
zational automated analytical reporting platform. New N.L.P. cases, such as clinical
study matching, N.L.P. and machine education, have an immediate impact. Some
companies are now trying to solve N.L.P. engine problems to suit the test. N.L.P.
simplifies and turns the test match process into a seamless one with innovations.
Prior approval—A report showed that the charge for doctors’ last approval needs is
growing. These standards improve overall activities and interfere with the delivery
of treatment. The question as to whether payers earned and approved reimbursement
was not rendered after a while as a result of language processing. I.B.M. Watson
and Anthem are now working on a payer N.L.P. module for quick prior authoriza-
tion assessment. Medical decision-making encourages physicians’ correct option
for natural languages production and engineering. Some healthcare areas require
extra monitoring, such as medical errors. Risk Adjustment and Hierarchical Cate-
gory Category Coding, a risk adaptation model, was initially developed to predict
patients’ potential costs. N.L.P. is also used for the control and evaluation of symp-
toms by physicians. H.C.C. coding is increasingly prevalent in value-based payment
models. H.C.C. uses ICD-10 coding to assign each patient risk ratings. The natural
language processing will allocate a risk factor to patients and estimate healthcare
costs.
Machine Learning in Healthcare: Current Trends and the Future 671
If your child has lost its wisdom teeth, individual doses of Vicodin are more likely
to be given. They may have Bactrim for an infection in the urinary tract (U.T.I.).
Machine Learning in Healthcare: Current Trends and the Future 673
Few patients, hopefully, will ever receive the same dose of any drug in the not
too distant future. Few people use the same medication if we know enough about
the patient’s biology and history. The promise of personalized medicine is a world
in which guidelines are customized to everyone’s wellbeing and disease remedies,
their medical history, genetic features, past ailments, nutrition, levels of stress and
much more. However, it might eventually expand to minor conditions (i.e. to give
somebody a smaller dose of bacteria for U.T.I., or an unparalleled version of bacteria
formulated for the Hooman Hakami from Medtronic, a Medtronic and I.B.M. diabetes
video) Medtronic would like to see insulin control pumps function independently,
monitor blood glucose levels and, if appropriate, administer insulin w While Western
Medicine has maintained its primary focus on the treatment and improvement of
diseases, proactive treatment and intervention are essential. This application is being
advanced by the first wave of IoT devices (particular FitBit). Machine training may be
used to track the worker’s performance or tension and seek beneficial improvements
in risk groups (not just to re-live or cure symptoms after reversals). Ethical concerns
regarding “increasing” human physical and (particularly) mental capacity are severe
and probably will become more critical as the technology of improvement in the next
15 years.
On the other hand, any human expert could easily defeat the narrower computer
vision application (if the model has been adequately trained). Federal “red tape”
or HIPAA also can make sports more “goliath” for the medical profession than
“david.” New social networks will catch teenagers up and beat Snapchat and Face-
book because of their virality, marketing and user interface. You only need a dozen
engineers at the right time, just like Instagram, but it is unlikely that a dozen engi-
neers have the industry, although they have earned ten million dollars. This labyrinth
can demand more money, contacts and know-how than any small Silicon Valley
company can collect. It seems like I.B.M. or Medtronic can have a healthy lead in
medical innovation because of these factors alone. In recent years, ML has contributed
to several areas, including the vision and processing of natural languages. Many
existing training data (e.g. images in vision or sentences in natural language) can
benefit from more complex models. We are now on the verge of a substantial change
in HE. Through applying ML to electronic health data, including genomic data, more
and more accessible, epidemiologists in health care can better understand the under-
lying risk for acquiring infectious diseases and transmission mechanisms and develop
specific strategies for reducing HAI. It is important to note that, while powerful, ML
can not recognize relationships that do not exist in the data.
Furthermore, ML does not substitute for regular statistical or random screening.
Instead, ML can be used to extend the existing HE toolbox. Interdisciplinary teams
would have the most significant influence on the future and work together to interpret
the data.
674 U. A. Usmani and J. Jaafar
References
1. Lipton RB, Scher AI, Steiner TJ, Bigal ME, Kolodner K, Liberman JN, Stewart WF (2003)
Patterns of health care utilization for migraine in England and the United States. Neurology
60(3):441–448
2. Islam MN, Inan TT, Rafi S, Akter SS, Sarker IH, Islam AKM (2020) A survey on the use of
A.I. and ML for fighting the COVID-19 pandemic. arXiv preprint arXiv:2008.07449
3. Lu C, Strout J, Gaudreau R, Wright B, Marcus FBDC, Buch V, Andriole K (2020) An overview
and case study of the clinical A.I. model development life cycle for healthcare systems. arXiv
preprint arXiv:2003.07678
4. Kaur J, Mann KS (2017) AI-based healthcare platform for real-time, predictive and prescriptive
analytics using reactive programming. J Phys: Conf Ser 933:012010
5. Tripathi G, Ahad MA, Paiva S (2020) S2HS-A blockchain-based approach for the smart
healthcare system. Healthcare 8:100391
6. Yoon JE, Suh CJ (2019) Research trend analysis by using text-mining techniques on the
convergence studies of A.I. and healthcare technologies. J Inf Technol Services 18(2):123–141
7. Gil-Lacruz M, Gracia-Pérez ML, Gil-Lacruz AI (2019) Learning by doing and training
satisfaction: an evaluation by health care professionals. Int J Environ Res Public Health
16(8):1397
8. Das S, Sanyal MK, Application of A.I. and soft computing in healthcare: a review and
speculation 8:21
9. Clifford GD (2020) The future A.I. in healthcare: a tsunami of false alarms or a product of
experts? arXiv preprint arXiv:2007.10502
10. Troncoso EL (2020) The greatest challenge to using AI/ML for primary health care: mindset
or datasets? Front Artif Intell 3:53
11. Lysaght T, Lim HY, Xafis V, Ngiam KY (2019) AI-assisted decision-making in healthcare.
Asian Bioethics Rev 11(3):299–314
12. Drysdale E, Dolatabadi E, Chivers C, Liu V, Saria S, Sendak M, Wiens J, Brudno M, Hoyt A,
Mazwi M (2019) Implementing A.I. in healthcare
13. Johnson SL (2019) A.I., machine learning, and ethics in health care. J Legal Med 39(4):427–441
14. Stanfill MH, Marc DT (2019) Health information management: implications of artificial
intelligence on healthcare data and information management. Yearb Med Inform 28(1):56
15. Pawar U, O’Shea D, Rea S, O’Reilly R (2020) Explainable A.I. in healthcare. In: 2020 interna-
tional conference on cyber situational awareness, data analytics and assessment (Cy- Bersa).
pp 1–2
16. Panesar A (2019) Machine learning and A.I. for healthcare. Springer
17. Farroha J (2019) Security analysis and recommendations for A.I./ML-enabled automated cyber
medical systems. Big Data: Learn Anal Appl 10989:109890
18. Adadi A, Berrada M (2020) Explainable A.I. for healthcare: from black box to interpretable
models
19. Reddy S, Allan S, Coghlan S, Cooper P (2020) A governance model for the application of A.I.
in health care. J Am Med Inf Assoc 27(3):491–497
20. Hernandez-Boussard T, Bozkurt S, Ioannidis J, Shah NH (2020) MINIMAL (MINimum Infor-
mation for Medical A.I. Reporting): Developing reporting standards for artificial intelligence
in health care. J Am Med Inf Assoc
21. Hunter P (2019) The advent of A.I. and deep learning in diagnostics and imaging: machine
learning systems have the potential to improve diagnostics in healthcare and imaging systems
in research. EMBO Rep 20(7):e48559
22. Greco L, Percannella G, Ritrovato P, Tortorella F, Vento M (2020) Trends in IoT based solutions
for health care: moving A.I. to the Edge. Pattern Recognit Lett
23. Henriksen A, Bechmann A (2020) Building truths in A.I.: making predictive algorithms doable
in healthcare. Inf Commun Soc 23(6):802–816
24. Halminen O, Tenhunen H, Heliste A, Seppälä T (2019) Factors affecting venture funding of
healthcare A.I. companies. ICIMTH 268–271
Machine Learning in Healthcare: Current Trends and the Future 675
25. Terry N (2019) Of regulating healthcare A.I. and robots. Available at SSRN 3321379
26. Srivastava SK, Singh SK, Suri JS (2020) State-of-the-art methods in healthcare text classifica-
tion system: A.I. paradigm. Front Biosci (Landmark edition) 25:646–672
27. Morley J, Machado C, Burr C, Cowls J, Taddeo M, Florida L (2019) The debate on the ethics
of A.I. in health care: a reconstruction and critical review. Available at SSRN 3486518
28. Tan Y, Jin B, Yue X, Chen Y, Vincentelli AS (2020) exploiting uncertainties from ensemble
learners to improve decision-making in healthcare A.I. arXiv preprint arXiv:2007.06063
29. Jiang F, Jiang Y, Zhi H, Dong Y, Li H, Ma S, Wang Y, Dong Q, Shen H, Wang Y (2017) Artificial
intelligence in healthcare: past, present and future. Stroke Vascular Neurology 2(4):230–243
30. Ellahham S, Ellahham N, Simsekler MCE (2020) Application of artificial intelligence in the
health care safety context: opportunities and challenges. Am J Med Qual 35(4):341–348
31. Fritchman K, Saminathan K, Dowsley R, Hughes T, Cock MD, Nascimento A, Tere- Desai
A (2018) Privacy-preserving scoring of tree ensembles: a novel framework for A.I. in health-
care. In: 2018 IEEE international conference on big data (Big Data), pp 2413–2422
Predictive Maintenance for a Turbofan
Engine Using Data Mining
Abstract Airplane safety remains one of the crucial areas that must have a robust
maintenance strategy due to its impact in the transportation of human beings and
goods. Predictive maintenance is a vital means of ensuring complex system such
as turbofan engines in airplane are being used safely and optimally. The advent of
information and communication technologies provide ways to collect useful data for
maintenance strategies and decision making. The acquired data are unstructured and
may contained incomplete information. Data mining transform the data to become
meaningful and useful for machine learning application. In this paper, data mining
techniques for predictive maintenance are presented, with machine learning algo-
rithms applied to predict the maintenance conditions of a turbofan engine and the
results are compared. The results show that support vector machine has a slightly
better accuracy than the other methods.
1 Introduction
Maintenance is a process that ensure tools, machine (simple or complex) are kept in
a good working condition and ensuring that it is repaired when the need arises as a
result of breakdown within the possible time. Thus, maintenance has been around
since the existence of machineries. It has also evolved with the changes that accom-
pany new innovations in machine systems. Traditionally, maintenance tends to be
reactive-respond to failures in equipment or devices. And as a result, when failure
occurs, it becomes very catastrophe. Human life and properties have been lost due
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 677
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_65
678 I. Mahmud et al.
2 Related Work
There have been some research works reported in the literature on predictive mainte-
nance. Researchers have proposed novel methodologies to produced high accuracy
in predictive maintenance. In Behera et al. [11], the authors proposed a data-driven
prognostic technique for predictive maintenance based on tree ensemble algorithms.
They used NASA turbofan engine dataset and found out that gradient boosted trees
performed better than random forest. Alberto et al. [12] proposed a data-driven
predictive maintenance process that perform data acquisition and processing for a
real machine process. The data was obtained from computer numerical control (CNC)
turning centre with two tool posts. Machine learning models were used in assessing
the remaining useful life of the machining tool. The concept shows that predictive
maintenance approach could be used in areal machine process where preventive
maintenance remain the feasible maintenance strategy adopted. Accorsi et al. [13]
developed a data mining framework for condition-based maintenance with different
machine learning classifiers. The models were applied to real-world case study, the
fault prediction accuracy was quantified and compared. The results show that random
forest having slightly better accuracy than other methods. In Bukhsh et al. [14], the
authors propose three tree-based classification methods for predictive maintenance
of railway switching systems. The models were able to trigger the activity type and
predict the maintenance need of the railway switches. Lee et al. [15] applied predictive
maintenance to two machine tool system elements using artificial intelligence algo-
rithms the two machine tool systems consist of spindle and cutting tools. The compo-
nent’s conditions were obtained from flank wear and bearing’s remaining useful life
for cutting tool and spindle respectively. Artificial neural network and support vector
machine were used in classifying the tool’s condition. Okoh et al. [16] proposed a
statistical techniques-based data-driven prognostic approach for predictive mainte-
nance. The model was conducted on a synthetic data validated by industrial experts.
In this paper, data mining based predictive maintenance is used for turbofan engine
with machine learning algorithms to predict the engine conditions.
3 Data Mining
Data mining provide means where hidden trends, patterns, and associations on
datasets are discovered. It involves many methods such as statistics and machine
learning to solve wide range of the problem at hand [17]. In a typical classification
problem, neural network uses gradient descent training to solve it, while a decision
tree uses greedy search algorithms to solve it. The main purpose of data mining is
model creation [18]. A framework to execute Data Mining similar to standard one
like Cross Industry Standard Process for Data Mining (CRISP-DM) [19] with modi-
fication to suit our classification model case is developed. The steps in the framework
includes; Data Preparation, Feature Selection and Prediction model.
680 I. Mahmud et al.
Row data are often not useful due to having presence of noise or too weak and in
some instance very large. Such data have to be process and reduced to a form, it could
be useful and easily be applied on predictive model. Some of the actions taking to
improve the signal quality include data quality assessment, operating condition oper-
ation discovery and data normalization. The tools used in data preparation include
filtering, amplification, data compression, data validation and de-noising.
The data goes to another form of process to make it more suitable for use. At this stage,
features are extracted and selected. The aim is to extract indictors or features with
fault or patterns and also having minimal subset features from problem domain while
maintaining original feature structure with high accuracy. The purpose of feature is
to select relevant and informative features effectively.
Decision tree, support vector machine, random forest and gradient boost tree are the
machine learning algorithms adopted for the prediction models in this paper. These
are supervised methods which build its models through establishing links between
input variables and target variables.
Decision Tree (DT). A decision tree is a machine learning algorithm that built
its classification model using a tree structure flowchart. Decision tree algorithm is
constructed with three set of nodes; root node, internal node and the leaf which
represents the class label. Among the algorithms used in developing decision tree
are C4.5 [20], classification and regression tree (CART) [21]. These algorithms are
implemented base on divide and conquer strategy where the training set is being
recursively partition.
Support Vector Machine (SVM). SVM is a statistical learning concept use to
perform classification through obtaining optimal hyperplane between two classes.
The hyperplane is the linear separator and in case of nonlinear classes, a kernel
function is used [22]. Kernel is a mathematical function that transformed nonlinear
data to be classified by a linear hyperplane and is given by
2
K x, x = exp −γ x − x (1)
Predictive Maintenance for a Turbofan Engine Using Data Mining 681
where ||x – x || is the Euclidean distance between two data points x and x . γ is
the control parameter that maintain error trade-off between variance and bias in the
model. Cost function is used in training SVM for error minimization
1 n
||w||2 + C ξk (2)
2 k=1
1
N
f (x) = Ti (x) (4)
N i=1
Gradient Boost Tree (GBT). Gradient boosted tree is also an ensemble of deci-
sion tree that produces weak tree classifiers and it use boosting approach to build
the base model. In the boosting approach, multiple models are built through resam-
pling of the training instance that previously could not be estimated and by having
some loss function (sum of squared errors or cross entropy) minimized [14]. If the
gradient overall ensemble function is F m (x) and it generates a new model F m+1 (x)
for the base-learner h(x). For a given squared error loss function, the algorithm is
express as:
N
ρm = argmin p ζ (yi , Fm−1 (xi ) + ρh(xi )) (5)
i=1
where N is the total number of training instances, ζ is the loss function and ρ is the
gradient step-size.
682 I. Mahmud et al.
4 Case Study
The prognostic and health management dataset used in this paper is from NASA
prognostic research center. It is a simulation of turbofan engine under several contin-
uous conditions. The data was captured using C-MAPSS (Commercial Modular Aero
Propulsion System Simulation) software. The turbofan engine architectural structure
with its primary components is shown in Fig. 1.
Air goes into the engine through the fan and low pressure compressor. Thereafter,
the air passes through high pressure compressor that mixed it with fuel and ignited
by a combustor. This results in lifting up discharge air velocity by the combust fuel
in the high pressure compressor (HPC) that controls the low pressure turbine (LPT)
and high pressure turbine (HPT). A 9000 lbs thrust class engine with various sensors
installed on it was used for the data readings collection. Several operating conditions
were used in gathering the engine information such as temperature variation (within
the range of − 51–39 °C) and the engine’s altitudes from sea level to 40,000 ft [11].
Table 1 provides some details of the four subsets of testing and training datasets in
the C-MAPSS dataset.
There is k-by-26 matrix-order stored in each subset where 26 is the number of
input attributes and k is the row count of the dataset. The three operational settings
have their combine data represented in every tuple of the dataset. The dataset is put
such that it start from engine number, followed by the cycle number, columns 3–5 are
the values of the operational settings and the remaining are the 21 different sensor
reading values attached on the aircraft engine [23]. It is considered that in each tuple,
the training and testing subset hold the total life-cycle time of the engine and the
last entry is the condition where the engine is on the brink of failure or declared
unhealthy [24, 25]. The major purpose of RUL estimation is to identify the condition
when the engine might need inspection or replacement. The RUL attribute targets
could be define as normal, warning and critical depending on the particular value in
the RUL attribute.
In this section, the FD001 dataset is used in building and testing the predictive
maintenance model for the turbofan engine. The four machine learning algorithms;
decision tree, support vector machine, random forest and gradient boosted tree are
trained to monitor and predict the conditions of turbofan engine.
Confusion matrix is used in evaluating the performance of the models’ prediction
ability. In the matrix, each row and column show the true and predicted values of the
model and can reveal which classifier is confused. A true positives are positive cases
correctly classified as positive class and true negatives are negative cases correctly
classified as negative class. Conversely, a false positives are negative cases incorrectly
classified as positive class and false negatives are positive cases incorrectly classified
as negative class. True positive rate is the proportion of correctly classified positives
while false negative rate is the proportion of misclassified positive cases (error rate
per class).
The input dataset for the models is extracted from FD001 dataset of NASA prog-
nostic and health management dataset. It is 20626 × 26 dataset with 26 input features.
21 features represent sensor signals employ in monitoring the conditions of the
turbofan engine. The first 5 attributes consisting of; engine numbers (100), number
of cycles (time) and three operating conditions. In the pre-processing of the dataset,
14 features are selected from the 21 sensor signals as being meaningful for the predic-
tion models. Principal component analysis (PCA) is applied to the selected features
to reduce the dimensionality of the features [15]. PCA is a mathematical process of
mapping multi-dimensional dataset to lower dimension and maintaining its original
information. The mapped dataset is used as the input for the models. The predic-
tion conditions are labelled as; critical, warning and normal which corresponds to
the remaining useful life of the dataset. Table 2 shows the accuracy obtained by
each model which measures the overall performance of the models. The results are
obtained using hold-out validation, by using 75% for training and the remaining 25%
for testing.
SVM had the better result with 88.6% accuracy while GBT had the least result of
83.6%. Figures 2, 3, 4 and 5 show the confusion matrices of the models.
From the Fig. 3, it shows that SVM correctly predict critical condition 87%,
warning condition 96% and normal condition 38%. The SVM maintained better
prediction among the models. Though, RF perform relative similar with the SVM in
all the indicators as shown in Fig. 4
6 Conclusion
In this paper, machine learning algorithms through data mining were employed in the
development of predictive maintenance of turbofan engine. Four models; decision
tree, random forest, gradient boosted tree and support vector machine were trained.
Three engine’s conditions (normal, warning and critical) represent the remaining
useful life of system. The results show that SVM had better overall prediction accu-
racy than the other three decision tree-based models. In the future, we intend to
enhance the performance of the decision tree-based model by having it hybridize
with a metaheuristic algorithm.
Predictive Maintenance for a Turbofan Engine Using Data Mining 685
References
1. Li Z, Wang K, He Y (2016) Industry 4.0-potentials for predictive maintenance. Adv Econ, Bus
Manag Res
2. Li Z, Wang Y, Wang K-S (2017) Intelligent predictive maintenance for fault diagnosis and
prognosis in machine centers: industry 4.0 scenario. Adv Manuf 5(4):377–387
3. Fitouri C, Fnaiech N, Varnier C, Fnaiech F, Zerhouni NJIP (2016) A decison-making approach
for job shop scheduling with job depending degradation and predictive maintenance. FAC-
PapersOnLine. 49(12):1490–1495
4. El Maraghy W, El Maraghy H, Tomiyama T, Monostori L (2012) Complexity in engineering
design and manufacturing. CIRP Ann 61(2):793–814
5. Khoshafian S, Rostetter C (2015) Digital prescriptive maintenance. In: Internet of things,
process of everything, BPM everywhere, pp 1–20
6. Colledani M et al (2014) Design and management of manufacturing systems for production
quality. CIRP Ann 63(2):773–796
7. Roy R, Stark R, Tracht K, Takata S, Mori M (2016) Continuous maintenance and the future–
foundations and technological challenges. CIRP Ann 65(2):667–688
8. Lee J, Wu F, Zhao W, Ghaffari M, Liao L, Siegel D (2014) Prognostics and health management
design for rotary machinery systems—reviews, methodology and applications. Mech Syst
Signal Process 42(1–2):314–334
Predictive Maintenance for a Turbofan Engine Using Data Mining 687
9. Saxena A, Goebel K, Simon D, Eklund N (2008) Damage propagation modeling for aircraft
engine run-to-failure simulation. In: 2008 international conference on prognostics and health
management. IEEE, pp 1–9
10. Kotsiantis SB, Zaharakis I, Pintelas P (2007) Supervised machine learning: a review of
classification techniques. Emerg Artif Intell Appl Comput Eng 160:3–24
11. Behera S, Choubey A, Kanani CS, Patel YS, Misra R, Sillitti A (2019) Ensemble trees learning
based improved predictive maintenance using IIoT for turbofan engines. In: Proceedings of the
34th ACM/SIGAPP symposium on applied computing, pp 842–850
12. Jimenez-Cortadi A, Irigoien I, Boto F, Sierra B, Rodriguez G (2020) Predictive maintenance
on the machining process and machine tool. Appl Sci 10(1):224
13. Accorsi R, Manzini R, Pascarella P, Patella M, Sassi S (2017) Data mining and machine learning
for condition-based maintenance. Procedia Manuf 11:1153–1161
14. Bukhsh ZA, Saeed A, Stipanovic I, Doree AG (2019) Predictive maintenance using tree-
based classification techniques: a case of railway switches. Transp Res Part C: Emerg Technol
101:35–54
15. Lee WJ, Wu H, Yun H, Kim H, Jun MB, Sutherland JW (2019) Predictive maintenance of
machine tool systems using artificial intelligence techniques applied to machine condition
data. Procedia CIRP 80:506–511
16. Okoh C, Roy R, Mehnen J (2017) Predictive maintenance modelling for through-life
engineering services. Procedia CIRP 59:196–201
17. Kovacs K, Ansari F, Geisert C, Uhlmann E, Glawar R, Sihn W (2019) A process model for
enhancing digital assistance in knowledge-based maintenance. In: Machine learning for cyber
physical systems. Springer, pp 87–96
18. Rokach L, Maimon OZ (2008) Data mining with decision trees: theory and applications. World
scientific
19. Shearer C (2000) The CRISP-DM model: the new blueprint for data mining. J Data Warehousing
5(4):13–22
20. Quinlan J (1993) C4. 5: programs for machine learning Morgan Kaufmann San Francisco. CA,
USA
21. Breiman L, Friedman J, Stone CJ, Olshen RA (1984) Classification and regression trees. CRC
press
22. Tong MT (2019) Using machine learning to predict core sizes of high-efficiency turbofan
engines. J Eng Gas Turbines Power 141(11)
23. Babu GS, Zhao P, Li X-L (2016) Deep convolutional neural network-based regression approach
for estimation of remaining useful life. In: International conference on database systems for
advanced applications. Springer, pp 214–228
24. Peel L (2008) Data driven prognostics using a Kalman filter ensemble of neural network models.
In: 2008 international conference on prognostics and health management. IEEE, pp 1–6
25. Wang T, Yu J, Siegel D, Lee J (2008) A similarity-based prognostics approach for remaining
useful life estimation of engineered systems. In: 2008 international conference on prognostics
and health management. IEEE, pp 1–6
Cyber Security Breaches
and the Long-Run Effect on Firms’
Market Value: A Conceptual Framework
Abstract Incidents of cyber security breaches are on the rise and pose a poten-
tial threat to businesses. Especially for publicly traded firms as they could create
a long-lasting influence on their financial performance and, thus, the market value
(MV) of a firm. Following the footsteps of the efficient market hypothesis, previous
studies have examined only the short-run impact on MV ensuing to security breach
announcements. Therefore, this study aims to conceptualize the impact of security
breaches on MV as manifested by long-run abnormal returns those firms. The study
is expected to provide a meaningful insight to investors and managers on the long-run
interconnection between cyber security breaches and firms’ market value.
1 Introduction
The Global Risk Reports by World Economic Forum (WEF) [1, 2] perceived a higher
risk for cyber-attacks with very high likelihood and impact following the enormous
risks associated with natural disasters. The increasing dependency on IoT and IR4.0
is making the business operations more vulnerable for incidents of cyber-attacks
such as spyware, malware, Denial of Service (DOS) attacks, phishing, ransomware,
and other vulnerabilities [3–5].
Measuring and analyzing the economic impact of security breaches has been
a point of interest from the scholars’ perspective. The effect of a security breach
on the trust and confidence of business stakeholders, including the investors in the
stock market [3, 6, 7]. A firms’ MV is one indicator to assess the change in the
confidence of investors. Therefore, various studies have studied the effect of security
breaches on the MV of a firm using theory for Efficient Markets [6, 8–14]. In most
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 689
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_66
690 S. E. A. Ali and F.-W. Lai
According to the Global Cost of Data Breach Report [17], the time to identify and
contain a data breach has been continuously increasing from 257 days in 2017 to 280
days in 2020, with estimated response cost ranging around $1 million for each firm.
The announcements such as security breaches indicate that the firms are reluctant to
disclose complete details of a breach on their first announcement. In recent times,
we have witnessed events of security breaches where the details concerning a breach
are disclosed months after the first breach announcement. For instance, in July 2019,
Equifax was penalized with US$ 700 million by the Federal Trade Commission
(FTC) and Consumer Financial Protection Bureau, after an enormous data breach
in 2017. From the investors’ perspective, they will probably have a close watch on
subsequent disclosures and announcements by the firm concerning a security breach
announcement that can influence their investment decision making. Hence, the impact
of a security event, especially concerning MV, cannot be judged by analyzing the
announcement effect only using the traditional event study methodology, as examined
by [7, 18]. Especially in cases where the announcement and economic impact vary.
Considering the above arguments, probably, the abnormal effect of a security
breach announcement by a listed firm can have a meaningful impression on future
cash flows, the required rate of return, financial distress, and credit rating. All these
concerns will be manifested on MV in the long run as well. Overall, it can be hypothe-
sized that the security breach announcement might lead to abnormality in the long-run
indicators of MV in the stock market, such as long-run abnormal returns. To the best
of our knowledge, the long-run impact on MV after a security breach has not been
addressed in the literature. An unfavorable abnormality might manifest that influence
in its long-run stock returns.
Cyber Security Breaches and the Long-Run Effect … 691
An influential research group has discussed various issues relating to the manage-
ment of security risks, such as security investments [19, 20], institutional stimulus
on innovation, and security [21]. Another line of research focuses on the market
implications of disclosures linked to the security of information [67, 68] and secu-
rity breaches [11, 22–28] by having an underpinning of ‘efficient market theory’ and
the methodology of an event study. Among all event studies on security, the current
paper is compatible with those scholarly works which have inspected the impact of
the security breach on the stock market. Among all event studies on security, the
current paper is compatible with those scholarly works which have inspected the
impact of the security breach on the stock market [18, 29–31]. It can be elucidated
from most of the studies that the security breach will have a significant unfavor-
able influence on the overall confidence of stock investors as signaled by negative
abnormal returns within 1 or 2 days of the breach announcement. All these system-
atic research works have based only on the short-run impacts of breaches while only
anecdotally exploring the long-run implications of security breaches.
The review of existing works reveals that the implications of events for the firm’s
long run prospects, especially the firms’ MV have become a common concern for
researchers. Long-run effects have been mostly explored with the BHAR model,
1–3 years after the incident. Within the information management avenue, the BHAR
model was used to address the long-run value of the company about the impact
of Capability Maturity Model (CMM) [32], Enterprise Resource Planning (ERP),
Supply Chain Management (SCM), and Customer Relationship Management (CRM)
[33, 34]. As far as we know, the impact of a security breach on the company’s long-
run MV has not yet been researched in the avenue of security. This research is the first
to conceptualize the effects of the security breach on the firms’ MV in the long-run.
Hypothesis Development and Conceptual Framework.
Long-run Abnormal Returns in the period before the breach announcement. Despite
breached firms’ intention to delay the announcement, there could be signs by which
security breaches can somewhat be anticipated. For example, the firm’s webpage
might become inaccessible for their customers and other stakeholders, a very sluggish
internet browsing, customers might face access issues on the firm’s website, pointless
popping up of messages at the time of customer’s log-in, abrupt changes to system
passwords or accounts, browser warning of errors. Moreover, the concerned users
might face infection alarms through antivirus toolkits installed in their systems. Thus,
these signals of system malfunction can negatively influence the trust and confidence
of customers. Likewise, the firms’ MV in the stock market might be influenced, and
692 S. E. A. Ali and F.-W. Lai
Fig. 1 A conceptual framework for analyzing the impact of security breaches on long-run MV
they might have allocated a likelihood of a security breach for the concerned firm.
It is also probable that the financial impact of a security breach might have already
incorporated by investors in their stock returns even before the actual security breach
announcement. Hence, it can be hypothesized that:
H1: In the period before the announcement, abnormal returns of security breached firms will
be negative.
3 Methodology
The methods and estimation techniques we have proposed in measuring the long-run
abnormal returns are different from those usually used in event studies when the short-
run impact of events is analyzed on the stock market. As the methodology of event
study frequently provides skewed approximations of both the ultimate economic
influence and the test statistics [36, 37]. Our proposed methodology is based on robust
Cyber Security Breaches and the Long-Run Effect … 693
and more accurate methods, which were newly established and used frequently in
studies [34, 38–42].
The literature argues as to how to measure long-run abnormal returns [35, 36].
The first problem is the relevant variables to be controlled for calculating abnormal
long-run returns. Previous studies on long-run stock valuation have been primarily
controlled for a company’s systemic risk (or beta). According to literature, the size,
market-to-book ratio, and previous performance are imperative predictors of stock
returns [43–45]. The present consensus, therefore, appears to be that abnormal returns
have to be determined after controlling size, market-to-book ratio, and previous
performance [38].
Buy-and-hold abnormal returns (BHARS) using one-to-one match samples:
Our proposed approach is one-to-one matching in which each sample firm is
compared to the appropriate control firm having similar size, market-to-book ratio,
and prior performance. Then it is recommended to create two individual samples of
one-to-one:
i. Choose a firm nearest in size to the sample firm (size-matched).
ii. Choose a firm nearest to the sample firm in terms of its previous performance
from the sample firm (performance-matched).
The abnormal return is the gap between the sample firm’s BHARs and its
matched control firm, ensuring consistency with the null hypothesis. Accordingly,
both parametric and non-parametric tests can be used for statistical inferences. This
paper provides an estimate of abnormal buy-and-hold-abnormal-returns (BHARs)
by daily return statistics. The raw returns and the return of the sample company
are compounded first through the frame for BHARs measurements. The abnormal
return is the disparity between the sample firm’s cumulative returns and its benchmark
(Fig. 2). BHARs are determined as:
T
T
B H A Ri = (1 + Rit ) − (1 + Rbt ) (1)
t=1 t
1
N
B H A Rit = B H A Ri (2)
N i=1
To collect findings over time, we conceptualize to map the calendar to date of the
occurrence for each entity in our sample. The day of the announcement is day 0, the
next day for trade is day one, and the day before the announcement is day-1, etc. It
is advised to forecast abnormal returns for at least 18 months, beginning six months
prior announcements through two periods of six months each after the security breach
notification. It consists of 250 trading days each year.
Furthermore, it contemplated deducting from both sides a 2-week duration (10
trading days) while measuring abnormal changes in stock returns. Moreover, security
breaches may, in the period before the disclosure, have already impacted stock prices,
and thus, the MV. Therefore, before the actual announcement, abnormal changes
may occur (Fig. 3). In the following three periods of six months each, we degree
buyand-hold abnormal returns and standard deviation from daily stock returns.
• Six months pre-announcement: trading days -135 to -11,
• Six months post-announcement: trading days 11–135,
• Twelve months post-announcement: trading days 136–260.
4 Conclusion
Findings in future studies based on the proposed framework might emphasize the
significance of security and determine that it cannot be disregarded. Events of security
Cyber Security Breaches and the Long-Run Effect … 695
breach do not occur unwittingly, and over time, users have a lower level of tolerance
for such events. Also, the news or media attention exposes the presence of flaws in
the firm’s security, rendering the firm more likely to be the victim of such assaults in
the future. Taking SONY as an example, this reveals that SONY’s loss of MV of 293
million dollars a year after the incident and total loss of 937 million two years after
the event would result from a security breach event. It implies that investors would
reevaluate the worth of the company in the stock exchange as they face tangible and
intangible risks from security breach incidents.
Acknowledgements The authors would like to acknowledge the efforts of the Department of
Management & Humanities-Universiti Teknologi PETRONAS (UTP) and Faculty of Business &
Economics-Universitas Islam Indonesia (UII), under grant cost center: 015MEO-114.
References
1. Collins A (2019) The global risks report 2019. World economic forum, Geneva, January 15
2019. Available: https://www.weforum.org/reports/the-global-risks-report-2019
2. Collins A (2018) The global risks report 2018. World Economic Forum, Geneva, January 17
2018. Available: https://www.weforum.org/reports/the-global-risks-report-2018
3. Smith KT, Jones A, Johnson L, Smith LM (2019) Examination of cybercrime and its effects
on corporate stock value. J Inf Commun Ethics Soc 17(1):42–60
4. Jansen J, Junger M, Montoya L, Hartel P (2013) Offenders in a digitized society. In: Cybercrime
and the Police. Eleven International Publishing, Hague, ND, pp 45–59
5. Lukonga I (2018) Fintech, inclusive growth and cyber risks: focus on the MENAP and CCA
regions. In: IMF working papers, vol 18. Available: https://doi.org/10.5089/978148437490
0.001
6. Cavusoglu H, Mishra B, Raghunathan S (2004) The effect of internet security breach announce-
ments on market value: capital market reactions for breached firms and internet security
developers. Int J Electron Commer 9(1):70–104. https://doi.org/10.1080/10864415.2004.110
44320
7. Bose I, Leung ACM (2014) Do phishing alerts impact global corporations? A firm value
analysis. Decis Support Syst 64(August):67–78
8. Ettredge M, Richardson VJ (2002) Assessing the risk in e-commerce. In: Presented at the
annual Hawaii international conference on system sciences. Hawaii, US
9. Campbell K, Gordon LA, Loeb MP, Zhou L (2003) The economic cost of publicly announced
information security breaches: empirical evidence from the stock market. J Comput Secur
11(3):431–448. https://doi.org/10.3233/JCS-2003-11308
10. Hovav A, D’Arcy J (2004) The impact of virus attack announcements on the market value
of firms. Inf Syst Secur 13(3):32–40. https://doi.org/10.1201/1086/44530.13.3.20040701/830
67.5
11. Kannan K, Rees J, Sridhar S (2007) Market reactions to information security breach announce-
ments: an empirical analysis. Int J Electron Commer 12(1):69–91. https://doi.org/10.2753/JEC
1086-4415120103
12. Malhotra A, Kubowicz Malhotra C (2011) Evaluating customer information breaches as service
failures: an event study approach. J Service Res 14(1):44–59
13. Sinanaj G, Muntermann J (2013) Assessing corporate reputational damage of data breaches:
an empirical analysis. In: Presented at the 26th Bled E-Conference—E-Innovations challenges
and impacts for individuals, organizations and society. Bled, Slovenia
696 S. E. A. Ali and F.-W. Lai
36. Barber BM, Lyon JD (1997) Detecting long-run abnormal stock returns: the empirical power
and specification of test statistics. J Financ Econ 43(3):341–372. https://doi.org/10.1016/
S0304-405X(96)00890-2
37. Kothari S, Warner JB (1997) Measuring long-horizon security price performance. J Financ
Econ 43(3):301–339
38. Lyon JD, Barber BM, Tsai CL (1999) Improved methods for tests of long-run abnormal stock
returns. J Financ 54(1):165–201. https://doi.org/10.1111/0022-1082.00101
39. Lee I, Loughran T (1998) Performance following convertible bond issuance. J Corp Finan
4(2):185–207
40. Hendricks KB, Singhal VR (2014) The effect of demand–supply mismatches on firm risk. Prod
Oper Manag 23(12):2137–2151
41. Hendricks KB, Singhal VR (2005) An empirical analysis of the effect of supply chain disrup-
tions on long-run stock price performance and equity risk of the firm. Prod Oper Manag
14(1):35–52
42. Ali SEA, Khurram S (2017) Impact of demographic and health factors on GDP growth of South
Asian Countries. Int J Acad Res Bus Soc Sci 7(3):2222–6990
43. Carhart MM (1997) On persistence in mutual fund performance. J Financ 52(1):57–82
44. Fama EF, French KR (1996) Multifactor explanations of asset pricing anomalies. J Financ
51(1):55–84
45. Jegadeesh N, Titman S (1993) Returns to buying winners and selling losers: implications for
stock market efficiency. J Financ 48(1):65–91
46. Ali SEA, Lai F-W, Dominic PDD, Brown N, Lowry PB, Ali RF (2021) Stock market reactions to
favorable and unfavorable information security events: A systematic literature review. Comput
Secur 110(November):102451
47. Ali SEA, Lai F-W, Hassan R (2020) Socio-economic factors on sector-wide systematic risk of
information security breaches: Conceptual framework. Paper presented at the 9th International
Economics and Business Management Conference, Melaka, Malaysia
Towards Early Distribution
of Container-Based Microservices
in Cloud Computing Environment
Abstract Microservices gained the popularity for use in scalable cloud application.
The application based on microservices uses intensive network communication to call
other microservices or to pass on the messages to hundreds of microservices running
inside the cloud/edge nodes. This study addressed the early distribution of container-
based microservices and proposed two distribution strategies named as Random
Distribution and Design Pattern Distribution. In random distribution approach the
microservices are assigned arbitrarily to the available data centers. While in design
pattern distribution the microservices are grouped together on the basis of behavioral
design patterns which identifies common communication patterns among objects.
The proposed solution was tested using custom built simulation environment and the
results showed that the early distribution of microservices according to the design
pattern of the application resulted in significant reduction of network calls to the
microservices hosted at other network nodes or data centers.
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 699
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_67
700 A. Saboor et al.
1 Introduction
In the last decade cloud computing has gained interest of the industries and research
institutes. The Cloud Computing enables ubiquitous computing and provides the
convenient on-demand access to the shared pool of configurable resources, such as
storage, network, server, applications and services [1]. The cloud data centers mainly
relies on the concept of virtualization to provide access to the shared resources and
cloud services [2]. For the last few years there had been a shift towards the container-
ization. The concept of containers creates lightweight and flexible environment by
allowing applications to share operating system. Recent advancement in container
technology have encouraged organization and researchers to use the containers in the
cloud environment. The adoption of containers helped in realization of microservices
at cloud level. Microservices offer number of advantages such as scalability, relia-
bility, re-usability, fast response time, reduced cost, and many more [3, 4]. Microser-
vices also comes with number of challenges. The microservices adoption challenges
include security vulnerabilities [5, 6], finding the right size and number of services
[3].
As microservices are independent units and often needs to interact with other
microservices. The large number of microservices and their interaction thus increases
the communication overhead. One way to provide reliable and fast communication
is by decreasing the communication among physical machine hosted at multiple data
centers. This study suggests different ways to distribute the microservices on physical
machines so that the microservices have to use lesser communication network. The
microservices are ordered/managed in such a way that they mostly interact with the
microservices available at same physical machine and do not have to interact most
of time with the microservices hosted on other physical machines.
2 Literature Review
In cloud computing one of the most critical factor is to provide quality of service
(QoS) within the defined constraints. Traditionally, heuristic scheduling algorithms
had been used to provide QoS by minimizing the processing time and schedule tasks
efficiently [7], but the design shift of cloud services from monolithic applications to
loosely coupled micro-services brought in new challenges such as increased make-
span, heavy network utilization, and increased cost. These challenges are addressed
by the researchers time to time. According to Zahang [8] container placements on
VM is an issue. The VM placement and container placement should not be addressed
individually, therefore Zahang proposed a solution named Container-VM-PM. The
solution resulted in optimization of the placement of new containers and minimized
the number of physical machines and resource wastage.
In a study by Lin [9], the container-based microservice scheduling was done based
on multi-objective optimization method using an ant colony optimization algorithm.
Towards Early Distribution of Container-Based Microservices … 701
The solution took into account the utilization of computing and memory resources of
physical machines However, the study considered microservice scheduling in single
container only, and did not considered the producer consumer relationship of the
microservices. Similarly, Lv et al. [10] proposed the container distribution strate-
gies to address the problem of heavy communication workloads among containers.
For container placement they proposed the worst fit decreasing algorithm, and
for container assignment proposed a two stage Sweep and Search algorithm. The
experimental results showed the reduction in communication overhead among the
containers.
Zhou et al [11] proposed the optimal placement schema for containers in the
cloud environment. They introduced one-shot algorithm which work as placement
schema for the container cluster, and an online algorithm which breaks down the
online decision making into on-spot decision depending on resource price.
Recently another study by Sampaio et al [12] proposed an adaption mechanism,
named as REMaP (Run-time Microservices Placement). In this model the automatic
run-time placement of microservices is based upon usage history of resources and
mciroservices affinity.
The literature review showed one common concern that heavy network commu-
nication and latency is one of the major issue in container based cloud environment.
Especially the decomposition of a particular application into many small microser-
vices results into large round trips between service calls. The possible solution could
be aggregating multiple services or batch service call in a single round trip, or
replacing expensive inter-process communication call with language-level function
calls [13]. This study propose an early distribution of container-based microservices
in cloud environment such that the microservices are grouped in such a way that it
results in lower number of network calls. The solution addressed the distribution of
microservices and containers at early stages of application deployment.
3 Methodology
For this paper the distribution of microservices is considered in such a way that
one container hosts one and only one microservice in the cloud/edge environ-
ment. This section describes two proposed microservice distribution approaches i.e.
Randomized Distribution, and Design Pattern Distribution.
The random distribution and assignment approach does not take into account any
performance parameters while assigning the containers to randomly chosen phys-
ical machine of the data centers, refer Fig. 1. The services are deployed arbitrarily
702 A. Saboor et al.
Microservice 4
Pool 1
Microservice 1
Microservice 7
Microservice 2
Microservice 3
Microservice 3 Cloud Data
Center
Pool 2
Microservice 4 Microservice 11
Microservice 5 Microservice 8
Microservice 6
Microservice 6
Pool 3
Microservice 7 Cloud Data
Microservice 2
Microservice 8 Center
Microservice 9
Microservice 9
Microservice 10
Pool 4
Microservice 10
Microservice 1
Microservice 11 Cloud Data
Microservice 5
Center
The term pattern used by software designers describes the software abstraction, thus
patterns used across other procedures, subroutines, and objects may also combine
for more common abstraction [14]. According to Erl [15] it is set of related patterns
which provides the building blocks and these building blocks are built upon the other
blocks thus defining the pattern application sequence. For this study we considered
the behavioral patterns as these design patterns are concerned with communication
between objects.
Towards Early Distribution of Container-Based Microservices … 703
Microservice 1 Microservice 2
Microservice 2 Microservice 4
Microservice 3 Microservice 7
Cloud Data
Microservice 4 Microservice 9 Center
Microservice 5 Microservice 1
Microservice 6 Microservice 6
Microservice 7 Microservice 3
Cloud Data
Microservice 8 Microservice 8 Center
Microservice 9 Microservice 5
Microservice 10 Microservice 10
Microservice n Microservice n
Cloud Data
Center
To assess the proposed techniques, we designed the simulation program using Java
to produce the microservices utilization data sets. To get the insights of the produced
data sets we used an online tool called plotly [16]. Plotly provided online analytic
and visualization of the data sets. The data sets were produced for one hour of
microservices calls, and for the assessment of the proposed system it was considered
that maximum of 1000 calls can be originated for the microservices at any given time.
The anatomized simulation process however reached maximum of 729 simultaneous
microservices calls at a given time. The number of service calls originated from the
system are shown in Fig. 3. They are the simulated calls for microservices which will
be used for testing the random and design pattern distribution. To get the test results
for classical random approach, the service calls were originated for the microservices
distributed randomly at different location i.e. data centers. The number of network
calls placed by the microservices in random distribution are shown in Fig. 4. It clearly
704 A. Saboor et al.
Fig. 3 Number of microservices calls originated from the system over a period of one hour
shows that heavy network traffic is generated by such a distribution. On average 189
network calls were placed at a given time.
In the second simulation the design pattern distribution strategy was tested and the
results showed that the number of network calls reduced significantly (refer Fig. 5).
On average 84 network calls we placed at a given time, which are very low when
compared to random distribution. The comparison of both distribution strategies is
shown in Fig. 6.
Fig. 6 Comparison of random distribution (RD) and design pattern distribution (DPD) network
calls
During the multiple simulation run of the system, it was noted that for couple
of time the design pattern based distribution of network calls exceeded the random
distribution network calls, but it was rare. This was due to the ideal match of the
microservices calls and the distribution done during the random distribution. Thus
less network calls placed in such a configuration were by coincidence.
The design shift of cloud services from monolithic applications to loosely coupled
micro-services brought in many challenges. This study addressed the communication
issues and proposed two microservices distribution strategies to reduce the network
calls. The random distribution and design pattern distribution were suggested, and
the simulation tests showed that the early distribution of microservices according
to the design pattern of application significantly reduce the network call to other
microservices.
The proposed system was tested using a custom built data sets and the system
further needs to be tested using real cloud traces. In future the study will also
be extended by defining other early distribution strategies such as priority based
distribution and weighted mean average distribution.
References
4. Malavalli D, Sathappan S (2015) Scalable microservice based architecture for enabling dmtf
profiles. In: 2015 11th international conference on network and service management (CNSM).
IEEE, pp 428–432
5. Newman S (2015) Building microservices: designing fine-grained systems. ”O’Reilly Media,
Inc.”
6. Yu D, Jin Y, Zhang Y, Zheng X (2019) A survey on security issues in services communication of
microservices-enabled fog applications. Concurrency Comput: Pract Experience 31(22):e4436
7. Juarez F, Ejarque J, Badia RM (2018) Dynamic energy-aware scheduling for parallel task-based
application in cloud computing. Futur Gener Comput Syst 78:257–271
8. Zhang R, Zhong AM, Dong B, Tian F, Li R (2018) Container-vm-pm architecture: A novel
architecture for docker container placement. In: International conference on cloud computing.
Springer, pp 128–140
9. Lin M, Xi J, Bai W, Wu J (2019) Ant colony algorithm for multi-objective optimization of
container-based microservice scheduling in cloud. IEEE Access 7:83088–83100
10. Lv L, Zhang Y, Li Y, Xu K, Wang D, Wang W, Li M, Cao X, Liang Q (2019) Communication-
aware container placement and reassignment in large-scale internet data centers. IEEE J Sel
Areas Commun 37(3):540–555
11. Zhou R, Li Z, Wu C (2019) An efficient online placement scheme for cloud container clusters.
IEEE J Sel Areas Commun 37(5):1046–1058
12. Sampaio AR, Rubin J, Beschastnikh I, Rosa NS (2019) Improving microservicebased
applications with runtime placement adaptation. J Internet Services Appl 10(1):1–30
13. Richardson C (2019) Microservices patterns: with examples in Java. Manning publications
14. Coplien JO (1998) Software design patterns: common questions and answers. In: The patterns
handbook: techniques, strategies, and applications, vol 13, pp 311
15. Erl T (2008) SOA design patterns (paperback). Pearson Education
16. Sievert C, Parmer C, Hocking T, Chamberlain S, Ram K, Corvellec M, Despouy P (2017)
Plotly: create interactive web graphics via ‘plotly. js’. R package version 4(1):110
An AI-Based Chicken Disease
Management System
Abstract Food stability has always received worldwide attention, especially in the
development of the poultry industry. However, the poultry’s diseases have caused the
loss of the poultry population and direct income of the owners. In this work, a system
comprising of a website and a mobile application has been developed to support
disease identification in chickens; the disease spread locations can be managed and
traced with maps. In addition, the system allows farmers to make contact and get
support from experts through chat, voice calls, and video calls. It also has some
e-commerce functions for improving sales of poultry’s owners. The system uses the
improved ResNet-50 model, with an accuracy of about 93.56%.
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 707
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_68
708 L.-D. Quach et al.
1 Introduction
2 Methodology
3 Application
To build the system, research based on the requirements, the process of finding
out management problems in practice, and building the model include the Main
functional model of the system (use case model), relational database model, business
model, and application development.
The study collected 492 images, divided into four disease classes: avian pox, bird-flu,
Marek, and infectious laryngotracheitis. Details of the number and illustrations of
the disease are shown in Table 1 and Fig. 1.
After obtaining the data, the study conducted steps aimed at data transformation,
including the following steps:
• Step 1: Use the SURF feature [12] to perform characteristic extraction from
diseased chickens. Each SURF feature has 64 vectors with parameters 5000 and
10,000. The image is depicted as Fig. 2 and the quantity in Table 2.
• Step 2: Research using the K-Means algorithm [13] with parameter k = 4 to filter
the pixels from the obtained feature vector. As a result, the study obtained a result
An AI-Based Chicken Disease Management System 711
Fig. 1 Illustrative image of chicken diseases, from left to right: Bird-flu, Avian pox, Infectious
laryngotracheitis, Marek
Fig. 2 Extract the SURF feature with parameters 5000 and 10,000
Fig. 3 Examples of samples selected after step (ii), from left to right, are diseases: Bird-flu, Avian
pox, Marek, Infectious laryngotracheitis
set of 1393 characteristics of the four obtained diseases described in the table. The
results obtained include 1393 characteristics to train with four different diseases
with the number as Table 3 and illustrations Fig. 3.
• Step 3: The research using the ResNet-50 model [14] to train the disease identifi-
cation dataset. The study conducted experiments, selected parameters and input
size 64 × 64 × 3 with default. The reseach got the best result was 93.56%, with
a learning rate of 0.01, and the worst result on our test was 90.93, with a learning
rate of 1. Efficien-cy The best yield of our adjusted model on training and test
sets, learning rate = 0.01, is shown in Fig. 4. Research results are used to illustrate
the process of disease identification for the system.
The use case diagram is used to represent the relationship between the use case
and the system agent. The actors in the diagram are users (people), some internal
or external applications. Use cases are a major function of the system [15]. In the
system of research (see Fig. 5), parts of the use case diagram include:
• Target users: Administrators, owners of livestock products shops, veterinary
establishments, and smallholder farmers.
• Use cases that describe the system’s main functions are disease identification in
chickens, contact advice with experts from veterinary facilities, e-commerce, map
zone management, and user decentralization.
An AI-Based Chicken Disease Management System 713
Fig. 4 Training results with learning rate = 0.01, epochs = 1000, input size = 64 × 64 × 3
The relational database model is the model that builds a digital database based on
data relationships, as proposed by the E. F. Codd in 1970. A software system will
714 L.-D. Quach et al.
use this database to avoid data redundancy and ensure data consistency is described
[17]. Therefore, research building a basic management system includes the entities.
The system functions to meet the requirements set out include store management,
products, users, etc.
Based on the relational database model and use case diagram, the research team has
built up the system’s operation diagram with essential functions on two platforms.
Dissipation includes:
1. Business Model of the Website System
The system includes a home page to login; disease and density zoning maps;
consulting contact by message, call, or video call. On the management website
of the system, the research team used the streets-v9 map from Mapbox GL-JS
[17] with the map center coordinates of [16] and the zoom level of 4.8 to allow
the territory of Vietnam to be displayed immediately when users access and
direct consulting function on the website (see Fig. 6).
2. Business Model of Mobile Application:
The mobile phone’s application includes authentication of users through
registration and login, homepage, image-based diagnostics through library or
camera, contact consultants such as system website, purchase, and farm manage-
ment. Based on analytical models, the research team has developed a mobile
application that includes functions for small poultry farmers (see Fig. 6).
4 Conclusion
With the available data, research has shown the feasibility of identifying the disease
in chickens. With the combination of the SURF feature determination method, the
K-means algorithm has helped the ResNet-50 model improve the training and recog-
nition accuracy and avoid the limitations of the CNN model, such as saturation and
overfitting 93.56% accuracy with four common diseases. The demo video of the
application is shown in the QR code (Fig. 6). Through the application of the image
recognition model and the deployed system, the system can develop in the following
directions:
• Collecting data directly from users: The purpose is to increase data, improve
model accuracy, and apply other CNN models.
• By zoning the data using the K-means algorithm, the model found that an indi-
vidual chicken can suffer from many diseases. Therefore, the application of this
model helps to detect any problems on the individual body.
• Improved model accuracy through a combination of image recognition and
descriptive (text) data.
• Develop a farm management model according to world standards such as
AsianGAP, GlobalGap, etc. from here; it is possible to query the origin of clean
poultry meat to serve consumer needs.
• Develop an ecosystem to link farmers with veterinary facilities, livestock product
suppliers to make e-commerce easier.
• The disease management system helps trace the origin of the disease and helps
the regulator timely detect and localize the infection to avoid spreading to other
areas and affecting people.
References
1. Kuberka L, Vuillemin J, Mezoughem C (2020) Livestock and poultry: world markets and trade.
US Department of Agriculture Foreign Agricultural Service, Washington, DC, p 16
2. OECD (2016) Table 3.A1.4—World meat projections. In: OECD-FAO Agricultural Outlook
2016–2025. OECD Publishing, Paris. https://doi.org/10.1787/agr_outlook-2016-table102-en
3. World Organisation for animal health (2020) Highly pathogenic avian influenza (HPAI)
report. No. 12 https://www.oie.int/fileadmin/Home/eng/Animal_Health_in_the_World/docs/
pdf/OIE_AI_situation_report/HPAI_asof30072020.pdf. Accessed 08/2020
4. Shriner SA, Root JJ, Lutman MW, Kloft JM, VanDalen KK, Sullivan HJ et al (2016) Surveil-
lance for highly pathogenic H5 avian influenza virus in synanthropic wildlife associated with
poultry farms during an acute outbreak. Sci Rep 6:1–11. https://doi.org/10.1038/srep36237
5. Mellata M (2013) Human and avian extraintestinal pathogenic Escherichia coli: infections,
zoonotic risks, and antibiotic resistance trends. Foodborne Pathog Dis 10(11):916–932
6. Astill J, Dara RA, Fraser E, Sharif S (2018) Detecting and predicting emerging disease in
poultry with the implementation of new technologies and big data: a focus on avian influenza
virus. Front Vet Sci 5:263. https://doi.org/10.3389/fvets.2018.00263
716 L.-D. Quach et al.
7. Hemalatha MS, Maheswaran R (2014) Recognition of poultry disease in real time using extreme
learning machine. In: Procceedings of the internation conference on inter disciplinary research
in engineering & technology 2014 [ICIDRET 2014], pp 44–50
8. Rizwan M, Carroll BT, Anderson DV, Daley W, Harbert S, Britton DF, Jackwood MW (2016)
Identifying rale sounds in chickens using audio signals for early disease detection in poultry.
In: 2016 IEEE global conference on signal and information processing (GlobalSIP). USA, pp
55–59. https://doi.org/10.1109/GlobalSIP.2016.7905802
9. Astill J, Dara RA, Fraser ED, Sharif S (2019) Corrigendum: detecting and predicting emerging
disease in poultry with the implementation of new technologies and big data: a focus on avian
influenza virus. Front Vet Sci 5:337
10. Quach L-D, Quoc NP, Tran DC, Hassan MF (2020) Identification of chicken diseases using
VGGNet and ResNet models. In: EAI INISCOM 2020—6th EAI international conference on
industrial networks and intelligent system, springer publishing (presented)
11. Quach L-D, Quoc ND, Thi NN, Tran DC, Hassan MF (2020) Using SURF to improve ResNet-50
model for poultry disease recognition algorithm. In: International conference on computational
intelligence 2020. Malaysia
12. Bay H, Tuytelaars T, Gool L (2006) SURF: speeded up robust features. In: Leonardis A, Bischof
H, Pinz A (eds) ECCV 2006, vol 3951. LNCS, pp 404–417
13. Steinley D, Brusco MJ (2007) Initializing k-means batch clustering: a critical evaluation of
several techniques. J Classif 24(1):99–121
14. He K, Zhang X, Ren S, Sun J (2016) Identity mappings in deep residual networks. European
conference on computer vision. Springer, Cham, pp 630–645
15. Khurana N, Chhillar RS, Chhillar U (2016) A novel technique for generation and optimization
of test cases using use case, sequence. Act Diagram Genet Algorithm. JSW 11(3):242–250
16. Paredaens J, De Bra P, Gyssens M, Van Gucht D (2012) The structure of the relational database
model. vol 17. Springer Science & Business Media. https://scholar.google.com/scholar?hl=
en&as_sdt=0%2C5&q=The+structure+of+the+relational+database+model&btnG=
17. Eriksson O, Rydkvist E (2019) An in-depth analysis of dynamically rendered vector-based
maps with WebGL using Mapbox GL JS
Automatic Polyp Segmentation
in Colonoscopy Images Using Single
Network Model: SegNet
Abstract Colorectal cancer is the third most common diagnosed cancer world-
wide. Early detection and removal of adenoma during the colonoscopy examination
may increase the survival probability. A novel computer-aided tool for automated
polyp segmentation in colonoscopy images is described in this work. SegNet, a deep
convolutional neural networks has been chosen to map low resolution features with
the input resolution for automated pixel-wise semantic polyp segmentation. Publicly
available databases, CVC-ClinicDB, CVC-ColonDB, and ETIS-LaribPolypDB were
used to train and to test the model. The outcome demonstrated the proposed method
is feasible as it attains an average of 81.78, 92.35% for mean intersection over union,
and dice coefficient, respectively for testing on a combination of the aforementioned
datasets.
1 Introduction
Accounting all types of cancer, colorectal cancer is the third most common diag-
nosed cancer worldwide [1]. As the name indicates, colorectal cancer is a malignant
tumor found in colon or rectum and often due to the specific type of polyp [2].
C. Y. Eu
Department of Electrical and Electronic Engineering, Universiti Teknologi PETRONAS, 32610
Seri Iskandar, Malaysia
T. B. Tang
Centre of Intelligent Signal and Imaging Research (CISIR), Universiti Teknologi PETRONAS,
32610 Seri Iskandar, Malaysia
C. Y. Eu (B) · C.-K. Lu
Smart Assistive and Rehabilitative Technology (SMART), Institute of Health and Analytics
(IHA), Universiti Teknologi PETRONAS, 32610 Seri Iskandar, Malaysia
e-mail: [email protected]
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 717
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_69
718 C. Y. Eu et al.
Currently, optical colonoscopy is the gold standard to screen colon for polyp detec-
tion and prevention of colorectal cancer. However, the polyp miss rate in colonoscopy
screening is approximately 10–30% [3]. Missed and misdiagnosed polyp are due to
various polyp structure [4]. Thus, computer-aided diagnosis (CAD) tool needs to
be developed to help medical expert in enhancing the accuracy and promptness of
medical diagnosis.
Automated polyp segmentation task can be presented in various methods which
essentially divided into two domain, traditional handcrafted method and modern deep
learning method. Those traditional methods often use features for polyp segmenta-
tion. For example, Sánchez-González et al. [5] used texture and color features for
training classifiers. The author utilize image windowing to increase generalization
and effectiveness. Active contour segmentation method was proposed by Dutta et al.
[6], where the active contour framework was modified to work without edges.
Deep learning techniques have been massively initiated to the image analysis
domain and outperform the old traditional ways consequently in current era. In
Vázquez et al. [7], they used Fully Convolutional Network—8 (FCN-8) to up-sample
the image to recover spatial resolution for per pixel prediction task. Huang et. al. [8]
proposed and compared two deep neural networks, DeepLab_v3 with Long Short-
Term Memory networks (LSTMs) and Segnet with LSTMs. The LSTMs are utilized
for keeping localization information that could slowly vanish along the encoder
networks. Another two proposed methods namely Dilated ResFCN and SE-Unet
were also compared in Guo and Matuszewski [9].
In this paper, a SegNet based CAD tool was proposed for polyp segmentation. The
proposed method has been validated in publicly available datasets, CVC-ClinicDB,
CVC-ColonDB and ETIS-Larib. The rest of the paper presented as follows. The
method used in this algorithm are demonstrated in Sect. 2, and the results obtained
are further discussed in Sect. 3. Lastly, Sect. 4 concludes this paper.
Fig. 1 Left image represents original image and right image represents pre-processed image
Feature extraction through any image processing is greatly affected by the image
quality itself. In this research work, thresholding was performed on image datasets
as a method for pre-processing. Image pixel threshold values were set for both red
channel and green channel as 50 and 15 respectively. Those pixel values that lower
than pre-defined threshold values were set to zero as the non-polyp background pixel.
The model network trained can be more biased to the polyp features rather than the
non-polyp background features since the image was mostly contributed by non-
polyp pixels. Next step is to normalize the image size to meet the input requirement
of network. The image datasets have been resized to 360 × 480. Figure 1 shows the
original image and the pre-processed image.
SegNet with VGG-19 framework architecture was used as our model. Datasets in
RGB image format were pre-processed, followed by model network training and
in the end test the model with testing images. The general steps involved in this
automated system are presented in Fig. 2.
SegNet is an architecture that was designed to map low resolution features with
the input resolution for pixel-wise semantic segmentation [10]. Encoder network
of SegNet is essentially the same as typical Convolutional Neural Network (CNN).
Input Layer
3×3 Conv + BN + ReLU
Max-pooling
Max-unpooling
Softmax
Pixel Classification
Layer
Encoder network consists of input layer, convolution layer (Conv), activation layer
and pooling layer. In convolution layer, a filter will be convolving around receptive
field [11]. Each of these filters can be thought of as feature identifiers. The filters on the
first layer convolve around the input image will be getting high values through convo-
lution when a filter found its desired feature from the input image. The same thing will
be iterated through all the location of input layer to form feature map. We included
batch normalization (BN) after each convolutional layer to normalize the output,
allowing for a larger learning rate that accelerate the training procedure. Trainable
parameters that remain untrained previously are able to start training. The following
layer is Rectified Linear Unit (ReLU), an activation layer which can limit the gener-
ated output. Pooling layer which usually comes after activation layer, is applied to
decrease the spatial size of the convolved feature. By doing this, the computational
power required to process the data can be greatly reduced. Moreover, it is useful for
extracting dominant features which are rotational and positional invariant. There is
a decoder network in SegNet which corresponded to the encoder network. It utilizes
max-pooling indices from previous encoder network to up-sample the input feature
map encoded from the encoder network [10]. At the final layer before SegNet last
output layer, there is one layer called Softmax. This layer works to segment the polyp
by activates output neurons via generating discrete probability distribution across all
the output neurons with the help of cross-entropy loss calculation. Overview of the
entire model network has been shown in Fig. 3.
In this project, pre-trained VGG-19 network was used as the framework architecture.
The network has been modified into SegNet by adding deconvolution network to meet
this project objectives. The training datasets were fed into the network together with
ground truth images for machine learning process. Testing images will be used to
evaluate the performance of the network in terms of mean Intersection over union
(mIoU) and dice coefficient. The network was trained with training options set to
Automatic Polyp Segmentation in Colonoscopy Images … 721
100
80
Accuracy
60
40
20
0.9 momentum, 0.001 learning rate, 20 epoch and mini batch size of 4. The network
was trained using single CPU, Intel® Core TM i7-2600 CPU @ 3.40 GHz, 8 GB
DDR3 RAM. Figure 4 shows the graph of training process. Our method results are
detailed in Table 1. This method was evaluated by comparing the performance with
other proposed work and the results surpass them.
According to Fig. 5a, the results show that entire polyps were successfully
segmented. Some over-segmented polyps can be seen in these results prove that
this model is poor in determining the edges of polyps. From Fig. 5b, the results
obtained were moderate. The outcome shows more over-segmented polyp regions
than the good results from Fig. 5a. The model did missed segment some true polyp
region and false polyp segmentation on light-reflected region. Some poor performed
outcomes also determined and presented in Fig. 5c. The model totally missed the
Fig. 5 First row of images show segmentation results, second row of images show ground truth. a
Good segmentation results. b Moderate segmentation results. c Poor segmentation results
location of polyp and most of segmented non-polyp pixels were found. These poor
results indicate that the model is not intelligent enough and unable to tackle more
complicated structure polyp and poor condition colon.
Our proposed model obtained 81.87% mIoU, lower than [4] and slightly lower
than [14] by 1.53%. Our dice coefficient reaches 92.35%, surpassed all previous
works that are all under 90% except [4] which is 98.00%. Our model outperforms
[14] on dice coefficient by 3.25%. Our model performed poorer than [4] due to our
model contains only one single network, while [4] proposed an ensemble of three
different networks into one final model. Our model is weak in classifying edges
of polyps, possibly due to some of polyp features being eliminated across deeper
encoder path especially convolution and max-pooling operations.
4 Conclusion
Acknowledgements This work was supported in part by Ministry of Higher Education, Malaysia,
under Grant FRGS/1/2020/TK0/UTP/02/23, and partially supported by YUTP-Fundamental
Research Grant (015LC0-002).
Automatic Polyp Segmentation in Colonoscopy Images … 723
References
Abstract Respiratory failure cases are many flowing and filling the ICUs of the
hospitals at any point of time. In cases of respiratory failure, either short-term or long-
term ventilation is required for the patient for survival. Maintaining of O2 saturation
in patients is of prime objective in mechanical ventilation which in turn contributes
for their recovery. Patients connected to the ventilator are constantly monitored for
O2 saturation. Any fall in O2 saturation requires immediate intervention of clinician
to introspect the problem and sort out the issues. One of the main reasons for low
O2 saturation is low value FiO2 settings on the ventilator. This FiO2 setting has to be
increased or decreased manually as per monitored SpO2 analyzer. There are always
possibilities that exist, that patient on ventilators may go unnoticed or get less atten-
tion during emergencies cases entering the ICU. In this project we approached to
automatically blend Air/O2 mixer of an ICU ventilator to maintain a target SpO2
Saturation on the patient, say 98%. Algorithm mechanism is designed in such a way,
continuous FiO2 changes will lead to achieve targeted SpO2 level of the patient.
S. Prabakar (B)
Sona College of Technology, Salem, India
J. Samson Isaac
Biomedical Engineering, Karunya University, Coimbatore, India
R. Karthikeyan
Dr. N.G.P. Institute of Technology, Coimbatore, India
K. Porkumaran
Sri Sairam Engineering College, Chennai, India
e-mail: [email protected]
R. Kannan
EEE, Universiti Teknologi Petronas, Perak, Malaysia
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 725
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_70
726 S. Prabakar et al.
1 Introduction
Respiratory failure cases are flowing and filling the ICUs of the hospitals every now
and then. This respiratory distress cases range from road accidents to snake bite and
suicide cases. All the above cases require either short-term or long-term ventilation
for their survival. For the fast recovery of patients, maintaining O2 saturation level
is the prime objective of mechanical ventilation.
Normally, Alveolar Ventilation is unconsciously regulated to maintain constant
arterial blood gas tensions, despite variable levels of oxygen consumption and CO2
production. Many drugs and techniques used in anesthesia interfere with control
or mechanics of ventilation, and it is the Anesthetists responsibility to ensure the
adequacy of ventilation during the preoperative period. Equipment related to venti-
lation is consequently of great importance to the Anesthetist and Anesthetic Techni-
cian. Correct use of the equipment relies on a good understanding of basic respiratory
physiology as well as how the individual ventilator operates.
Patients connected to the ventilator are constantly monitored for O2 saturation.
Any fall in O2 saturation requires immediate intervention of clinician to introspect
the problem and to sort out the issues. The main reason for Low O2 Saturation is
low FiO2 settings on the Ventilator. So there are always some possibilities that exist,
such as patients under ventilation are left unnoticed or given less attention that may
lead to even death for patients under ventilation.
FiO2 is Fractional Inspired Concentration of Oxygen. It is a mixer of Air and
Oxygen, blend at a specific proportion, delivered to the patient through ventilator.
Air/O2 Mixer or Blender is one of the main components in any ventilator that mixes
Air and O2 at a particular proportion that is delivered to the patient to maintain his/her
O2 Saturation. The range of mixing is from 21 to 100% of Oxygen as atmospheric
air which always contains 21% Oxygen by nature.
A complete design and approach for the automatic intelligence and feedback system
are described. It includes the block diagram and algorithmic approach in depth.
Figure 1 shows the input, processing and output modules of the automatic FiO2
regulation in ventilators.
After the analysis of the deviation in the O2 saturation, FiO2 value is newly
computed according to the algorithm designed. The computed FiO2 value is feedback
to the ventilator as an input. Depending on that new value, the SpO2 value will
gradually increase and gives relief to the patient. This is done repeatedly until the
patient attains the normal O2 saturation level.
The SpO2 sensor detects the oxygen saturation from the patient. That percentage
output as a feedback as well as the set FiO2 from the ventilator are given as inputs to
728 S. Prabakar et al.
the ARDUINO hardware via the input channels. The flow control algorithm is fed
to the ARDUINO hardware to give the relevant output. The algorithm as follows:
Step 1: Set the FiO2 value in ventilators and sense the SpO2 level from the patient
using sensors.
Step 2: Calculate the difference between the optimal and measured SpO2 and
determine the “Half the difference” for blender correction mode.
Step 3: Compute the new FiO2 value for the ventilator to supply on patient.
Step 4: Then again sense the SpO2 and repeat the same steps to attain the desired
SpO2 value on the patient.
According to the above algorithm a new FiO2 value is given as a output from the
ARDUINO hardware. That FiO2 value is given as a feedback to the ventilator. Thus
the SpO2 value is automatically modified according to the new FiO2 value.
The algorithm is developed from the SpO2 deviation on the patient under ventilation
in order to compute the desired FiO2 automatically. Figure 2 provides the flow of
FiO2 computation from the obtained values.
Initially we will consider two values such as FiO2 and SpO2 . FiO2 is the fractional
inspired concentration of oxygen, which is the set value in the ventilator (for say 50%)
and SpO2 is saturation of O2 %, which is measured from the patient through the sensor.
An Agile System for FiO2 Regulation in Ventilators 729
The deviation between the standard SpO2 value, which is the normal level of oxygen
saturation i.e. 98% and the measured SpO2 , which is theme assured value from the
patient through sensor is determined. The determined value is the difference between
the standard SpO2 value and the measured SpO2 value. According to the difference,
the FiO2 will be changed according to the at present SpO2 value. A new FiO2 value
will be computed as a result it is given to the ventilator as a feedback input. According
to that FiO2 value which is newly computed, the SpO2 value gets modified. So this
progress is done continuously and the corrections are done gradually in loops (Table
1).
We are here taking a situation that a patient connected to a Ventilator with FiO2
set value of 50% results in the patient’s saturation of O2 80%. But, normal SpO2 to
be achieved is 98%.
Here there is a difference between normal and achieved saturation (98–80%)
which is 18%. This is an indication that the FiO2 set on the ventilator is in adequate
and it has to be increased. This increase has to be systematic and it should tend
to increase the O2 saturation towards the normal level of SpO2 . Hence the FiO2
(Air/O2 mixer) is increased to a level that is “Half of the Difference” between target
and achieved SpO2 , i.e., 9%. Now the new (Air/O2 mixing) FiO2 setting has to be
increased from 50 to (50 + 9) 59%.
730 S. Prabakar et al.
Let us assume that this increase in FiO2 is taking the patient to a higher O2
saturation of 88% from 80%. Still the Patient’s SpO2 is below normal by a value of
10% (98–88%), that is deficit. This monitored value gives a feedback to the Air/O2
mixer to increase the FiO2 to another “Half of the Difference” that is 5%.Now the
new FiO2 on the ventilator would be 59 + 5% i.e., 64%. This process will tend to
continue till the patient achieves the targeted SpO2 level of 98%.The above process
is the systematic way of maintaining a patient’s O2 saturation through “Intelligent
Feedback System”.
require any manual adjustments of clinician and nurses. Whenever there is a fall in
O2 saturation occurs the ventilator automatically blends the FiO2 settings according
to the patient’s oxygenation (Fig. 3).
Automating the Blending process gives a great relief to the clinician inpatient
management. They can also over-ride this mechanism in case they want to have
direct control of the patient. This mechanism allows the clinicians to concentrate on
other areas that need more time. This is a very safe method gives mental relief to
doctors and nursing staffs and at the same time patient is also ensured of minimum
O2 Saturation for his good life.
4 Conclusion
Thus the proposed work concludes that automating the Blending process gives a great
relief to the clinician in patient management. Thus patient needs not require contin-
uous monitoring and manual correction. They can also over-ride this mechanism in
case they want to have direct control of the patient. The automatic regulation can be
done not only for FiO2 but also for all the blood gas values in different algorithmic
procedures like PaO2 , PaCO2 etc. Thus the system will lead to complete automatic
operation on ventilator without clinicians’ periodical check and also ensures the
patients good care. This mechanism allows the clinicians to concentrate on other
areas that need more time. This is a very safe method gives mental relief to doctors
and nursing staffs and at the same time patient is also ensured of minimum O2
Saturation for his good life.
References
1. Zhou F (2009) Component-based design of software for embedded control systems: the medical
ventilator case study. In: Guan W, Sierszecki K, Angelov C (eds) Embedded software and system,
ICESS’09, international conference. Mads Clausen Institute for Product Innovation, University
of Southern Denmark, Soenderborg, pp 157–163
2. Ionescu CM (2009) Flow controlled artificial ventilation of a COPD patient. In: De Keyser R
(ed) Control conference (ECC). System and Automation Department, Ghent University, Gent-
Zwijnaarde, Belgium pp 2475–2481
3. Jafari MM (2005) Senior member, IEEE, Control systems engineering group, Tyco Healthcare
Puritan Bennett R&D—“Robust Feedback Design for Proportional Assist Ventilation-System
Dynamics and Problem Definition” decision and control, 2005 European Control Conference,
CDC-ECC’05, 44th IEEE Conference, pp 4389–4844
4. An P (2010) High-precision ventilator with compensation algorithm and automatic regulation.
In: Song JT (ed) Database technology and application (DBTA), 2nd international workshop.
Schema of Electronics and Information Engineering, Ningbo University of Technology, Ningbo,
China, pp 1–4
5. Tehrani FT (2007) A new decision support system for mechanical ventilation. In: Engineering
in medical and biology society, 29th annual international conference of the IEEE. California
State University, Fullerton, pp 3569–3572
An Agile System for FiO2 Regulation in Ventilators 733
6. Tzavaras A (2011) Locating of the required key-variables tobe employed in a ventilation manage-
ment decision support system. In: Weller PR, Prinianakis G, Lahana A (eds) Engineering
in medicine and biology society EMBC, Medical Instrumentation Technology Department,
Technological Educational Institute (TEI) of Athens, Greece, pp 112–115
7. Yi W (2009) Fuzzy control for regulating level of pressure support ventilation. In: Zhang Q,
Wang Y, Xiong J (eds) Measuring technology and mechatronics automation, ICMTMA’09,
international conference, vol 2. Mechatronics and Automatic Schema, Natural University of
Defense Technology, Changsha, China, pp 562–565
8. Li Y (2005) Investigation on dynamics of the flow control unit in ventilator systems and its funda-
mental performance limitations. In: American control conference, vol 3. VIASYS HealthCare,
Yorba Linda, CA, USA, pp 2181–2186
Computer-Aided Diagnostic Tool
for Classification of Colonic Polyp
Assessment
Abstract Colorectal cancer is the third most common malignancy and the fourth
leading cause of cancer-related deaths worldwide. This paper presents a combination
of techniques (e.g., pre-processing, transfer learning, principal component analysis,
and support vector machine) to detect the polyp during colonoscopy. In particular, we
carefully choose the pre-trained deep convolutional neural networks (i.e., AlexNet,
GoogLeNet, ResNet-50, and VGG-19) according to their performance extracting
features. A publicly available database, Kvasir, is used to train and test the detection
model. The result indicates that to use ResNet-50 as a pre-trained network provides
the best results among the rest. Our proposed model achieves an accuracy of 99.39%,
and its sensitivity and specificity are 99.39, 99.41, and 99.38%, respectively.
1 Introduction
Colorectal cancer (CRC) is one of the leading causes of cancer deaths based on
the analysis of the American Cancer Society (ACS) as the number of cancer deaths
was increased year by year from 1997 to 2015 [1]. The small and tiny polyp might
be missed [2], which caused the chance of getting converted into a giant neoplastic
polyp. Hence, the discovery of colonic polyps is essential as the polyp can be inhibited
by removing it and reduce the incidence of CRC, whereas the detection of malignant
polyps at an early stage may result in a curative intervention [3]. However, it is not
as difficult to detect polyp as before [1] because of the elegant automated computer-
aided diagnosis (CAD). Those CAD tools can detect the polyp and provide polyp
location depending on the curvature of polyps. Several CAD systems have been
developed for various cancer diseases such as prostate cancer, brain cancer, lung
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 735
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_71
736 W. S. Liew et al.
cancer, etc. These CAD systems obtain the images through the hardware renders and
save them as digital movies using image processing software. Therefore, doctors
can accurately carry out the diagnosis and decision based on the CAD system’s
information.
With the CAD scheme’s rapid development, the detection system is continuously
optimized and improved with AI techniques. AI is considered as a machine emulation
with human thinking. Thus, deep learning (DL) applications in AI allow the systems
to become intelligent, learning, and self-organizing [4]. Pre-processing is vital for
every detection system, as every image has different sizes and features (i.e., color,
density, shape, and texture) [5]. The images need to be pre-processed to reduce the
image’s noise and degradation; normalization is required to resize the images into the
same size. Besides, the images need to undergo feature extraction and selection to
carry out classification task with a different type of classifiers such as ensemble clas-
sifier [6], support vector machine (SVM) [7], Naïve Bayes classifier [8, 9], random
forest [9] and K-Nearest Neighbors (KNN) [9].
In the research, a publicly available database, Kvasir, is used to train, validate, and
test the detection model. All images are color colonoscopy images in JPG format
with a resolution of 720 × 576 pixels.
A total number of 2000 images in the dataset consisted of 1000 normal cases (non-
polyp) and 1000 abnormal cases (polyp). The colonoscopy images are split into 70%
for training and 30% for testing. The model is developed through the MATLAB
platform on a Windows 10 64-bit operating system with Intel® Core i5-2430 M
CPU @ 2.40 GHz, 8 GB RAM, and NVIDIA GeForce GT 525 M GPU. For this
research work, each step is illustrated in Fig. 1.
2.1 Pre-processing
During colonoscopy screening, the images will be affected due to the visual condition
in the bowel. This is because there is no ambient light in the bowel, but only the
capsule on-board light as the illuminating source. Under this condition, the images
captured by the camera will have some artifacts, vignettes, and illuminations [5, 10].
To enhance the polyp feature in the image, the thresholding method is used to separate
the polyp from the background, which indicates the image consists of two classes of
pixels (foreground and background) [11]. A color image has three channels, which
are red (R), green (G), and blue (B); the threshold values of R and G channels are
scaled to 0 if the values are below 50 and 15, respectively. On the other hand, the B
channel is not utilized for thresholding due to pixel values of polyp and background
are almost the same.
Furthermore, the images needed to resize into the same scale through normaliza-
tion to feed into the neural networks for training. Nevertheless, different networks
have different input sizes; AlexNet can only process the images that are 227 × 227
pixels, while GoogLeNet, ResNet-50, and VGG-19 process images with 224 × 224
pixels. The normalization equation is computed as follows:
y−y
y (Resi zed I mage) = (1)
σ
where y is the original pixel vector, y is the mean of the pixel vector, and σ is its
standard deviation.
Deep learning (DL) is potent as it helps us a lot in complex tasks. However, it is not
easy to create a DL model because it requires generous data to train the deep neural
network from scratch. To resolve this issue, transfer learning (TL) is a powerful and
effective solution to limited data, especially in the medical research field [12].
In addition, TL can improve the accuracy of the network while reducing its
training time [13]. The original architecture of pre-trained networks such as AlexNet,
GoogLeNet, ResNet, and VGG can be used and fit into own task by replacing the last
three layers with the new specific task from the database. Besides, the neural networks
can learn from training to progress their proficiency [14]. Nevertheless, different
neural networks have various performances in terms of accuracy and training time
due to the number of deeper layers and the training database’s size.
738 W. S. Liew et al.
2.3 Classification
From the view of machine learning, one of the classification methods is supervised
learning. Support vector machine (SVM) is a classifier with supervised learning [15],
and it provides remarkable performance and accuracy with less computation power.
SVM divides data into two classes (normal and abnormal cases) using hyperplanes
in a high dimensional feature space.
However, SVM has a higher dimensionality, which may lead to model overfitting.
Therefore, principal component analysis (PCA) is utilized to reduce data dimen-
sionality [16] to the maximum extent. To achieve dimensionality reduction, PCA’s
feature elimination is one of the techniques as it remains interpretability of variables
[17].
The models are evaluated based on the confusion matrix with true positive (TP),
true negative (TN), false positive (FP), and false negative (FN), where TP: a polyp
is detected in a frame that contains a polyp; TN: No polyp is detected in a frame
without a polyp image; FP: a polyp is detected in a frame without a polyp image;
FN: A polyp is missed in a frame that contains a polyp. The accuracy, sensitivity,
and specificity are calculated from the confusion matrix as follow:
TP +TN
Accuracy( ACC) = (2)
T P + T N + FP + FN
TP
Sensitivit y(T P R) = (3)
T P + FP
TN
Speci f icit y(T N R) = (4)
T N + FP
The summary of results is provided in Table 1. Within these four deep neural
networks, ResNet-50 is better than the rest of the networks with an accuracy of
99.39%, sensitivity of 99.41%, and specificity of 99.38%.
In Table 1, the area under receiver operating characteristics (AUROC) is measured
based on the receiver operating characteristics (ROC) curve in Fig. 2. The ROC curve
is a graph of true positive rate (TPR) versus false positive rate (FPR) to show a clas-
sification model’s performance at all the classification thresholds. TPR is sensitivity;
while FPR is calculated as follows:
Although every neural network has different computation time for data pre-
processing and training, but the exact training iteration time is still depending on
the hardware or the size of mini-batch set. Figure 3 shows the graph of accuracy
versus training time for four neural networks. AlexNet is the fastest network, but it
has the lowest accuracy. In contrast, VGG-19 takes the longest time for training and
its accuracy is still lower than ResNet-50.
Due to the network complexity, the number of convolution layer directly relates to
the performance of networks. ResNet-50 has the most convolution layers, and it tends
to overfit the training data. As a result, it has the highest accuracy for the colonic polyp
detection task and higher training time than the AlexNet and GoogLeNet. In contrast
with training time, the number of learnable parameters plays an important role.
ResNet-50
Accuracy (%)
GoogLeNet
VGG-19
AlexNet
Fig. 3 Accuracy against relative training speed for each neural network
VGG-19 takes the longest time for the training because it has the biggest learnable
parameters among the four networks.
Based on the discussion above, ResNet-50 is chosen for this research work. The
results obtained by using the proposed method are compared with the state-of-the-art
techniques. The comparison results have been made with previous works in Table 2.
Our proposed method has better performance evaluation scores than other studies. In
this way, the proposed method increases CAD systems’ success rate with the most
effective medical image detection method for the colonic polyp.
After the training process, the network’s parameters are saved as a MAT file so that
there is no training required again during the testing of the image. For the testing,
a graphical user interface (GUI) is created to load the images, and it is designed as
in Fig. 1. There are two buttons available in the GUI, which are machine learning
and load image. First, the user requires to choose a model trained by four different
CNN architectures. Once a network is chosen, all the parameters will be shown in
the GUI, and it is ready to load the image for testing. Then, the classification result
will be shown in the identified class column of the GUI.
3.3 Dataset
From the dataset, there are two categories of images, which are with polyp and
without polyp. These 2000 colonoscopy images have different visual quality and
features such as the brightness, distribution of pixel intensity, position, and polyp
size. To prove the diversity of data, we have various types of images in Fig. 4, which
allows the detection model to learn with multiple features.
4 Conclusion
Acknowledgements This work was supported in part by Ministry of Higher Education, Malaysia,
under Grant FRGS/1/2020/TK0/UTP/02/23, and partially supported by YUTP-Fundamental
Research Grant (015LC0-002).
References
17. Xu J-W, Suzuki K (2014) Max-AUC feature selection in computer-aided detection of polyps
in CT colonography. IEEE J Biomed Health Inform 18(2):585–593
18. Fan L, Song B, Gu X, Liang Z. Semi-supervised graph embedding-based feature extraction
and adaptive kernel-based classification for computer-aided detection in CT colonography. In:
2012 IEEE nuclear science symposium and medical imaging conference record (NSS/MIC),
pp 3983–3988
19. Fiori M, Musé P, Aguirre S, Sapiro G (2010) Automatic colon polyp flagging via geometric and
texture features. In: 2010 Annual international conference of the IEEE engineering in medicine
and biology, pp 3170–3173
Live Obstacle Detection Audio System
for Visually Impaired People
Abstract Object detection is the complex task for the visually impaired people
when they are on move. This paper proposes smartphone’s capability more useful
for visually impaired people to overcome day to day problems with live navigation.
Depth sensing camera is used to differentiate foreground objects from background
objects which is used to place real world objects in different layers. As soon as the
user moves to the particular layer the fore coming objects are detected and notified
to the user by audio instructions. Information from the internet can be collected
using the mobility services. The headphone with mic is used by the system for
receiving user’s voice input. Emergency tasks like sending SMS and location are
sent to the subject’s guardian while the subject is in danger with the help of player.
This concept allows visually impaired people to walk both indoors and outdoors
independently. Experimental results obtained show that the navigation system eases
the user’s travelling experience in complex indoor and outdoor environments. Hence,
it serves as challenged people-friendly device for helping the visually impaired people
to travel hassle-free and safely. The goal is to provide an effective means of navigation
for the visually-impaired.
1 Introduction
The number of visually impaired people is on the raise over the recent years. Report of
the world health organization (WHO) claims about 253 million people live with vision
impairment. Among them, 36 million people are visually impaired and 217 million
have moderate to severe vision impairment. Most of the visually impaired people
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 745
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_72
746 P. Anitha et al.
is aged 50 or older and live in the developing countries struggling when they are
alone. Every day these visually impaired people face problems in understanding and
interacting with the surroundings, particularly those that are unfamiliar. It is hard for
a visually impaired person to go out alone and there are not many available products
that can assist them. However, research has been going on for decades for developing
an effective device for visually impaired people. In this paper, a device with new
features that can aid in multiple tasks while maintaining low development cost.
The device can easily guide the visually impaired people and is able to give proper
directions. The most valuable thing for a disabled person is gaining independence. A
blind person can lead an independent life with some specifically designed adaptive
things for them. There are lots of adaptive equipment that can enable a blind person
to live their life independently but they are not easily available in the local shops or
markets. A blind person needs to hunt and put much effort to get each equipment
that can take them one step closer towards independence.
2 Related Works
Various methodologies have been proposed for the welfare of vision-impaired people
using the concept of AI and ML. Here are some of them in brief.
Li et al. [1] proposed an approach pioneering effort of indoor robotic naviga-
tion, NavChair used sonar sensor to detect objects. A vector field histogram (VFH)
algorithm and later Minimal VFH were proposed to provide effective sonar-based
obstacle detection and avoidance. For updating a local map with detected obstacle
in the form of uncertainty grids, VFH used sonar readings. It was used to calcu-
late a polar histogram to represent the geometric information of the obstacles. At
last, NavChair found a direction with minimum VFH cost and followed it to bypass
obstacles. The RGB-D indoor navigation system. An emerging research focus for the
robotics community to improve the mobility of blind and visually impaired people is
Intelligent assistive navigation. Numerous studies have been carried out for indoor
navigation on mobile devices.
Bai et al. [2] proposed a system that includes a depth camera for acquiring
the depth information of the surroundings, an ultrasonic rangefinder consisting of
an ultrasonic sensor, a MCU (Microprogrammed Control Unit) for measuring the
obstacle distance, an embedded CPU (Central Processing Unit) board acting as main
processing module, which does such operations as depth image processing, data
fusion, AR rendering, guiding sound synthesis, etc., pair of AR glasses to display
the visual enhancement information and an earphone to play the guiding sound.
Feng Lan, Shanghai Jiao Tong University, Shanghai, China [3] proposed a smart
glass prototype system based on Intel Edison. The Intel Edison is a tiny computer
offered by Intel as a development system for wearable devices. The board’s main SoC
is a 22 nm Intel Atom “Tangier” (Z34XX) that includes two Atom Silvermont cores
running at 500 MHz and one Intel Quark core at 100 MHz (for executing RTOS Viper
OS). The SoC has 1 GB RAM integrated on package. There is also 4 GB eMMC
Live Obstacle Detection Audio System for Visually Impaired People 747
flash on board, Wi-Fi, Bluetooth 4 and USB controllers [4]. Edison is powerful in
computing but small in shape that makes it perfect to build a wearable device. Public
signs in cities can act as guides for people. Following the public signs, people can find
public infrastructure, such as public toilets, bus stations, subway stations, hotels and
so on. So finding and recognizing the public signs outsides would help someone who
is visually impaired or visually impaired gain increased independence and freedom.
Therefore, they have implemented an application of public signs recognition in this
developed smart glass system. This application can automatically detect, analyse and
recognize all kinds of public signs around the visually impaired and visually impaired
and give corresponding voice hints through wireless bone conduction headphones.
With the help of this smart glass system, the visually impaired may find bus stations,
subway stations, restaurants, hotels and so on. The visually impaired may even go
for a trip alone.
Bourne et al. [4] proposed a project to solve instrumentation problems encountered
by visually impaired employees or potential employees who are not suited to (nor
do they require) complex microcomputer or synthetic speech circuitry for practical
solutions. A common problem encountered by custom cabinetmakers, as well as
individual homemakers, is finding the structural members within walls (“studs”)
onto which heavy shelves and wall hangings can be mounted. There have been
many devices for finding “studs,” but these are traditionally visual instruments. They
have successfully modified a commercial stud finder to present auditory feedback
to the visually impaired cabinetmaker and handyman”. The Zircon “Stud Sensor”
(also marketed by Radio Shack as the “Archer Stud finder”) is a relative capacitance
meter that senses hidden structures by their capacitance effect; it presents these
relative indications to a sighted user via a column of LEDs. By reverse engineering, a
“relative capacitance” analog signal was found which could be used to drive a voltage-
controlled oscillator. Thus, with their modified instrument the visually impaired user
hears a distinct rise in pitch when the capacitance increases—which happens when
a stud is found. Four prototypes are being made for field trials, and several visually
impaired individuals who have tested the unit have found it effective and easy to use.
Rentschler et al. [5] proposed a system to test the walker when run on a two-drum
test machine. The walker was then run through the functionality test and visually
inspected for cracks or damage. This approach helps the visually impaired navigate
hassle free through streets.
Elaiyaraja et al. [6] proposed a system in which the prototype consists of an optical
glass with head-mounted camera that captures images and offloads the computation
to the database. The database is built using Microsoft Azure API and consists of
details of all known persons to the visually impaired. Based on the computations,
the database performs facial recognition and identification. If it is a face that already
exists in the database, then the person’s name and face features will be returned.
If it is an unknown face, then it returns that the person is unknown to the visually
impaired. With the help of identification information obtained from the unknown
person, he/she can be added to the database. The Text to Speech API converts the
input text into human-sounding custom speech and returned as an audio file.
748 P. Anitha et al.
Some of the implemented guide systems like smart guide proposed by Manufali
[7] gives the user a prototype of an intelligent guide for the blind person is successfully
designed, implemented and tested. The prototype device is facilitating the movement
of blind person by warning him/her about any nearby obstacles in order to help
him/her during daily activities. The guidance will be provided in the form of audio
instructions through the headset and based on real-time situation for both indoor and
outdoor environments. But it cannot detect what is against the user and it is also low
battery storage which cannot long last for long duration.
3 Proposed Framework
3.1 Overview
The Audio navigation system proposed can be implemented in the smart phone
itself. It is more battery efficient and portable. A dedicated voice assistant directs
the visually impaired with help of google maps. Notifications will be given through
headphones via voice commands so that it cannot confuse the user. It will guide the
person during walk so one can make sure he is stable while stepping. This requires
the internet connection; however, most of the processing is done offline. This system
also allows the user to do basic mobile functionalities such as calling and messaging
through voice commands. Figure 1 represents the overview of the flow in the system.
Obstacles in the visually impaired’s pathway are captured with live relay using the
phone’s camera system shown in Fig. 2. The data is then processed and used in
guiding the visually impaired via an obstacle free path. This is implemented with
the help of TensorFlow powered API. Initial process of this audio navigation system
is to find the obstacles before the user. It is required to guide the user how to avoid
them without encountering. A depth-sensing camera with IR sensor in a mobile can
provide us the data how far an object is away from the camera. Input data is processed
through tensor flow and open-cv. Detected objects with its accuracy of detection will
be the outcome of this module.
Depth sensing camera helps with identification of objects in layers which conveys
which object is to encountered first. The camera also detects the floor using the floor
level deduction which helps to identify pits and peaks. Floor detection is done by the
barometer which is installed default in the phone.
Audio navigation is required to notify the person “How to move” and “What is
before them”. First objective of this module is filter what to be notified and what
must be neglected.
Live Obstacle Detection Audio System for Visually Impaired People 749
The input to this module will be from the text output of camera vision. The detected
objects distanced detected by depth cam, which is shown in Fig. 2, and the object
itself will be formed, as a sentence through a program. It will be played as audio to
the user. The text to audio conversion is done by gTTS engine service. In terms of
storage, data is stored internally.
750 P. Anitha et al.
Objective of this module is to get users audio input and direct them to the places
where they like to move. Audio input is processed as text by gTTS module. User’s
location is obtained from the device as input. The input is given to “DIRECTIONS
API” to find the directions for their given place from the user’s location. The reply
from the API is used by the user to guide them all the way. Nearby API locates the
nearby location to their location.
A visually impaired person downloads the developed application in his/her smart-
phone and performs the initial setup with the guardians help. The person then connects
the headphones to the smars application helps to navigate to the desired destination
with the help of our assistant providing audible directions. In addition to this, the
assistant also provides recommendations for nearby eateries, shops, hospitals etc. In
order to use these features the person has to long press the headphone button.
3.3 Applications
The main task of the application is to detect lively the obstacles like roadcrosses,
pits, persons and other objects. It can also identify an object when the user captures
it. Once it captures the objects it conveys the user by the voice commands. If an user
needs to know the nearby locations like hospital it can be provided by asking the
system. It provides the hospitals nearby. If the user needs any other locations like
hotels or shops it provides them those location details within 1 km. Basic mobile
functionalities such as call, message and checking battery status. It locates one to
their home with live routing. It says the way to their home in steps. In an emergency
it can send SMS to a friend or doctor with the current location details. It can also
make call to their Friend or doctor immediately once the system hears the keywords.
It allows the visually impaired navigate through cities independently. It can be used
for live tracking globally with some constraints. This system will be trained for an
individual user which gives them more accessibility.
4 Experimental Result
The camera analyzes and recognize objects. It then conveys this information to the
visually impaired via the headphone. As an example consider the above two images.
Figure 3 shows how the camera recognize the human faces along with the certainty
that the object is a human. Similarly, Fig. 4 shows how camera recognize the bottle.
Figure 5 shows the voice assistant interface which will converted as background
process. The interface consist of five parts: Results, Response, Get_gps, Live tracking
and input. In input field the user’s voice is taken as input. Get_gps gives the user
location. Input can be given through voice or text. Live tracking can give the obstacle
detection lively when the user points at any direction. Audible reply is given by the
Live Obstacle Detection Audio System for Visually Impaired People 751
assistant and also the response field will carry the reply. The app logo within an app
drawer is shown in Fig. 6.
Most of the audio comments are given by google assistant’s default voice with
the pitch rate and speed of 20%.It can be adjust in accordance with google account
maintenance (Table 1).
On real time comparison between the existing systems and our system we came
up with a prediction amount shown in the below matrix. The system detection rate
of persons is most accurate as it uses object detection principle for classification of
objects. The existing system 1 is down in accuracy because it uses ultra-sonic sensors.
These sensors may miss the person’s body when there gaps in between them. The
existing system 2 uses an object detection which is less accurate to recognize a
person. Our system and existing system 2’s detection rate of glass objects is slightly
less because of the light diffraction whereas ultrasonic sensors don’t miss glass
object because of light. Pits detection is completely missed by existing system 1 as
it considers it as far away object. Our system considers it by using depth sensing
752 P. Anitha et al.
which detects floor level and varies a pit. Whereas in system 2 it misses the accuracy
without using floor level detection. We use TensorFlow most accurate algorithm and
online api for detecting road crosses where as other systems lacks it.
5 Conclusion
A majority of the proposed system involve the use of the 3d glass or an obstacle
detection system to assist the visually impaired for navigation. The systems are often
complex to operate and expensive to implement. This paper describes an approach
similar to ones discussed but differs in terms of simplicity and cost. The system
proposed requires basic smartphone, dedicated service with internet connectivity
and a headset to achieve live navigation for visually impaired effectively. Obstacle
detection and audible guidance using voice assistant gives an easier way of access.
The system developed is portable in the sense that it requires a smartphone with
internet connectivity. Thus, a cost effective and simplistic solution is made possible.
Live Obstacle Detection Audio System for Visually Impaired People 753
Table 1 Comparative analysis on performance of various supportive systems for visually impaired
Obstacles Predicted (%)
Smart audio system Smart guide [7] Smart guiding glasses [2]
Actual (%) Persons 95 75 85
Glass 60 80 60
Pits 85 2 80
Road cross 92 0 20
Average (%) Overall 83 39.25 61.25
6 Future Enhancement
Live detection is to be made more accurate and fully offline capable. This is similar
to a person walking beside the visually impaired person and providing direction.
This approach will be implemented with machine learning. The image processing is
done using TensorFlow. The ultimate goal is to guide the visually impaired through
cities effectively and hands-free.
References
1. Li B, Munoz JP, Rong X (2019) Vision-based mobile indoor assistive navigation aid for blind
people. IEEE Trans Mobile Comput 18(3):702–714
2. Bai J, Lian S, Liu Z, Wang K, Liu D (2018) Smart guiding glasses for visually impaired people
in indoor environment. IEEE Trans Consum Electron 63(3):258–266
3. Lan F, Zhai G, Lin W (2015) Lightweight smart glass system with audio aid for visually impaired
people. In: TENCON, IEEE, region 10 conference
4. Bourne RRA, Flaxman SR, Braithwaite T, Cicinelli MV, Das A, Jonas JB, et al (2017) Vision
Loss Expert Group. Magnitude, temporal trends, and projections of the global prevalence of
visually impairedness and distance and near vision impairment: a systematic review and meta-
analysis. Lancet Glob Health 5(9):e888–e897
5. Rentschler AJ, Cooper RA, Blasch B, Boninger ML (2003) Intelligent walkers for the elderly:
performance and safety testing of VA-PAMAID robotic walker. J Rehabil Res Dev 40(5):423–
432
6. Elaiyaraja, Aishwarya, Akshaya, Monikha M (2019) AJNA—a smart assistance for the visually
and cognitively impaired. Int J Adv Res Electric, Electron Instrument Eng
7. Manoufali M, Aladwani A, Alseraidy S, Alabdouli A (2011) Smart guide for blind people
international conference and workshop on current trends in information technology
Reducing the Aleatoric Uncertainties
of Failure Prediction Using Singular
Value Decomposition
Ahmad Kamal Mohd Nor, Srinivasa Rao Pedapati, and Masdi Muhammad
1 Introduction
Long term exploitation of aerospace assets is one of the goals sought by the industry.
Maintenance efforts to ensure operational longevity have evolved steadily from
simple corrective maintenance, to preventive measures and more recently to a more
systematic framework called Prognostic and Health Management (PHM). PHM aims
to optimize maintenance activities by accurately predicting failure event. It has the
effect of minimizing maintenance operations, reducing workload and decreasing
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 755
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_73
756 A. K. M. Nor et al.
2 Related Work
SVD has become a standard tool in denoising data in general due to its ability to
eliminate singular values residing in noisy subspaces and retaining only the clean
ones. In Wang et al [18] a combination of Wavelet Transforms and SVD is used
to denoise image data. The wavelet transform is used to divide the image into
low and high frequencies divisions. Then SVD is employed to denoise the high
frequency parts to enable a clean image reconstruction. In Zhang et al. [19], a two-
division recursion SVD is proven to perform best in improving contaminated signals
amongst the various SVD model variants. The methods also outperform other signal
decomposition techniques such as Empirical Mode Decomposition (EMD), ensemble
EMD (EEMD) and Variational Mode Decomposition (VMD). SVD is deployed on
matrix of image patches in He et al. [20], and by doing so, separating the signal
subspace from the noise subspace. Then, SVD denoising is performed on each group
of similar patches that were previously combined as training samples, producing
leading singular vectors candidates that represent the corresponding patch. Singular
vectors coming from noise populated samples and noise subspace are then elimi-
nated, thus leaving the clean singular vectors that form the local basis of each patch.
In Wang et al. [21], the Singular Value Difference (SV) between the noisy and clean
image is exploited to denoise image in wavelet domain. SVD is applied to denoise
SVs in high frequency parts of the noisy image.
Apart from image denoising, SVD also performs well with general data. In Liu
et al. [22], seismic data is denoised by SVD in the frequency domain. Raw seismic
data is first presented in the frequency domain by Fourier transform. Then SVD is
employed to decompose the frequency spectra into eigen images, which will be then
cleaned by selecting clean singular values. In Zhao et al. [23], a reweighted Singular
Value Decomposition (RSVD) framework which has the advantage of improving
weak features while denoising is proposed. The periodic modulation intensity, an
indicator that represents the quality of Singular Components (SCs) information level,
is introduced to rank the SCs. Finally, the clean signal is reconstructed based on the
information level presented by the SCs.
3 Methodology
The training thus intends to minimize the gap between the mean of the predicted HI
distributions with the HI target values while at the same time predicting variable
standard deviations.
In LSTM, gating functions acts as data flow regulator, preserving or forgetting
important or unimportant memory, respectively. LSTM consists of Input Layer,
Hidden Layer and Output Layer with a recurrent mechanism enabling the output
of the Hidden Layer to be added with new input. This made LSTM particularly
powerful in modelling sequential data. Below is the mathematical representation of
the model.
LSTM gating equation:
f t = σ W f .X t + R f .h t−1 + b f (1)
Ct = f t ∗ Ct−1 + i t ∗ Mt (5)
h t = ot ∗ ∅(Ct ) (6)
W ∗ = Input Weight.
R∗ = Recurrent Weight.
b∗ = Bias.
σ, ϕ, ∅ = Activation Functions.
Looking at the equations above, Eq. (1)–(3) are Forget Gate, Input Gate and
Output Gate equations. Equation (4) is used to calculate the candidate value that
will be added to the new state “t” (5) together with the old state “t-1”, regulated
by the Forget Gate. Equation (6) represents the final output of LSTM. “.” is matrix
multiplication while “” is point to point multiplication [25].
Reducing the Aleatoric Uncertainties of Failure Prediction … 759
A ∈ RT ×D
with T , the time steps and D the number variables, i.e., column of features
and with the condition that T < D.
2. A can be expressed in a reduced singular value decomposition form as:
T
A = U SV = u k sk v k (7)
k=1
 ∈ RT ×D
and E is the matrix of noise containing random noise with standard deviation, ∈
that can be expressed as:
E ∈ RT ×D
with their respective SVD representations and the same descriptions as mentioned
before:
T
 = Û Ŝ V̂ = ûk ŝk v̂ k (8)
k=1
T
E = USV = uk sk vk
k=1
760 A. K. M. Nor et al.
 = A + E (9)
Lower SVD modes in noisy data stay quite clean, thus filtering the higher effected
ones could give a cleaner signal. We assume the noise to be independent, identically
distributed (i.i.d.) Gaussian noise.
1. We project noisy time series data à as SVD matrix to obtain ũ k ,
sk and ṽk .
2. By fitting a Marchenko-Pastur distribution to the tail of noisy singular values
s̃, we can calculate the measurement error ∈. For full calculation, please refer
to [14].
3. We can then estimate the root mean square error (RMSE) of the modes with:
⎧ ⎡ ⎤ 21 ⎫
⎪
⎪ ⎪
⎪
⎪
⎪ ⎥ ⎪ ⎪
⎪
⎪ ⎢ ⎪
⎪
⎨ ⎢ ⎥ ⎪ ⎪
2 ∈⎢ D − w w T λ̃ m 3λ̃ k − λ̃ m ⎥ ⎬
R M S E(ṽk ) ≈min , ⎢ + ⎥ ⎥
⎪
⎪ D s̃k ⎢
⎢ D D λ̃m − λ̃k ⎥ ⎪ ⎪
⎪
⎪ ⎣ m=1 ⎦ ⎪ ⎪
⎪
⎪ ⎪
⎪
⎪
⎩ ⎪
⎭
m = k
4. For i.i.d Gaussian noise, the spatial correlation parameter, f and the effective
smoothing window width, w, should be 1 [26]:
f =w=1
5. Once RMSE of the modes obtained, we can calculate the parameter tk , which
represent the cleanness of a mode.
log(R M S E(ṽk )) − log 2
D
tk = (11)
log(R M S E(ṽ1 )) − log 2
D
6. After that, we estimate the rank for minimum loss data reconstruction with the
rank rmin as:
9. Finally, the clean signal can be reconstructed with rmin and s k obtained
previously:
rmin
Ar = ũk s k ṽ k (15)
k=1
4 Case Study
Turbofan failure dataset from NASA Commercial Modular Aero Propulsion System
Simulation (CMAPSS), published by Nasa Prognostic Centre (PCoE) of Ames
Research Centre is used in this study [30]. The simplified diagram of the studied
turbofan is shown in Fig. 1.
The chosen training and testing data consist of 100 recorded turbofan degrada-
tions. Each Training and Testing data correspond to a turbofan whose health condition
deteriorate after certain cycle [30]. Each fleet training and testing data are time series
sequence comprising of Time (Cycle), 3 Operating Conditions (OC) and 21 sensors
measurements as in Table 2. The OC refers to different operating regimes combina-
tion of Altitude (O-42 K ft.), Throttle Resolver Angle (20–100), and Mach Number
(0–0.84) [31]. High levels of noise are incorporated, and the faults are hidden by the
effect of various operational conditions [31]. However, no RUL information is given
in the training dataset. Only the ground truth RUL for each turbofan, are given. Data
pre-processing is thus applied to select the appropriate sensors, visualize the sensor
trends, and extract each cycle HI target.
Only strictly monotonic sensors are selected [32]. These sensors are useful as
they best represent single trending degradation contrary to irregular and unchanged
signals. 14 sensor signals, corresponding to sensors 2, 3, 4, 7, 8, 9, 11, 12, 13, 14,
15, 17, 20, 21 are used.
Reducing the Aleatoric Uncertainties of Failure Prediction … 763
outlet
S3 T30 Total temperature at HPC 0R
outlet
S4 T50 Total temperature at LPT 0R
outlet
S5 P2 Pressure at fan inlet psia
S6 P15 Total pressure in psia
bypass-duct
S7 P30 Total pressure at HPC psia
outlet
S8 Nf Physical fan speed rpm
S9 Nc Physical core speed rpm
S10 Epr Engine pressure ratio N/A
(P50/P2)
S11 Ps30 Static pressure at HPC psia
outlet
S12 Phi Ratio of fuel flow to Ps30 Pps/psi
S13 NRf Corrected fan speed rpm
S14 NRc Corrected core speed rpm
S15 BPR Bypass ratio N/A
S16 farB Burner fuel–air ratio N/A
S17 htBleed Bleed enthalphy N/A
S18 Nf_dmd Demanded fan speed rpm
S19 PCNfR_dmd Demanded corrected fan rpm
speed
S20 W31 HPT coolant bleed lbm/s
S21 W32 LPT coolant bleed lbm/s
To obtain RUL sequence labels for training, piece-wise linear degradation model
is employed [33, 34]. Each fleet health is thus considered stable in the beginning
until the start of deterioration which is the failure start point that initiate a linear
degradation until failure.
Each fleet is associated with time series corresponding to the total operational
duration of the fleet and the last cycle indicate the final instant before failure. Thus
initially, we model the RUL of a fleet to start with the value of the last cycle and
degrade linearly until 0 as shown in Fig. 2.
764 A. K. M. Nor et al.
The failure start point in each sensor is obtained by using Cumulative Sum
(CUSUM) anomaly detection technique, which returns the first index of the upper
or lower cumulative sums of each sensor signal that have drifted beyond 5 standard
deviations from the target mean. This indicates the initiating points of degradation
[35]. Then, the mean of the failure start points for each sensor can then be calculated.
From these means, we can calculate the mean of the means of the sensor’s detection,
which is our failure start point. Combining the linear degradation obtained earlier
and the failure start point, the transformed Fleet 1 RUL sequence is presented in
Fig. 3.
We show the sequence length plot of the training data in Fig. 4. From this plot, we
derive the RUL plot as shown in Fig. 5 following steps mentioned in Sect. 73.3.
In Fig. 6 we show the sequence length of testing and data its associated expected
RUL in Fig. 7 following our previous reasoning in Sect. 73.3.
However, as we can see, the ground truth RUL, shown in Fig. 8 are different from
the expected ground truth. Some testing data have long sequence length while at the
same time having very short RUL. We thus anticipate that the model to perform more
poorly on these data.
Reducing the Aleatoric Uncertainties of Failure Prediction … 765
Window size 2 was found appropriate for MOVAVG while 3 for EWMA. The
uncertainty measurement according of the full dataset and respective methods are
presented. For illustration purposes, we take the case of Fleet 1 as an example (Fig. 9).
As we can see from the illustrations given, the lowest degree of uncertainty is accorded
to SVD denoising in Fig. 10 while the highest uncertainty come from EWMA and
EMD, shown respectively in Fig. 12 and 14. The curve of rolling standard deviation
for SVD is also smoother than the rest of the methods. As for RMSE, again, SVD
surpasses other denoising method in improving the probabilistic LSTM performance
(Table 3).
MOVAVG and EWMA, as respectively shown in Fig. 11 and 12, cannot denoise
the data properly as we are applying the same simplistic rule (window k and fixed
weights) to denoise the whole data. As for Wavelet Transform and EMD, presented
here in Fig. 13 and 14, the usage of default parameter setting, due to absence of
domain expert, is obviously not sufficient to fulfil its intended function. To execute
effective denoising in Wavelet Transform, manual manipulations need to be done first
to determine the mother wavelet and decomposition level [9]. EMD denoising, on
Reducing the Aleatoric Uncertainties of Failure Prediction … 769
the other hand, possess various decomposition modes that need deep understanding
by the user to be successful [9]. SVD however treat each data sites individually by
removing higher modes more prone to contamination, thus it denoise more effec-
tively. We show the complete HI prediction for the original data and after SVD
treatment in Fig. 15 and 16, respectively.
Reducing the Aleatoric Uncertainties of Failure Prediction … 771
6 Conclusion
In this paper, SVD denoising treatment is applied to gas turbine failure dataset
to reduce the Aleatoric uncertainty of failure prediction. A probabilistic LSTM is
employed to predict the RUL distribution of the gas turbines based on sensor data. Our
proposed method show superiority in reducing the prediction uncertainty compared
to various other denoising techniques such as simple moving average, exponential
weighted moving average, empirical mode decomposition and wavelet transform.
772 A. K. M. Nor et al.
The results also indicate that SVD treatment improves the performance of prediction
better than the other mentioned techniques.
Acknowledgements We would like to thank Yayasan Universiti Teknologi Petronas (YUTP) for
funding this research
References
8. Alaa A, Schaar M (2020) Frequentist uncertainty in recurrent neural networks via blockwise
influence functions
9. Kim M, Liu K (2020) A Bayesian deep learning framework for interval estimation of remaining
useful life in complex systems by incorporating general degradation characteristics. IISE Trans.
https://doi.org/10.1080/24725854.2020.1766729
10. Peng W, Ye Z, Chen N (2020) Bayesian deep-learning-based health prognostics toward prog-
nostics uncertainty. IEEE Trans Industr Electron 67(3):2283–2293. https://doi.org/10.1109/
TIE.2019.2907440
11. Shaker MH, Hüllermeier E (2020) Aleatoric and epistemic uncertainty with random forests.
In: Berthold M, Feelders A, Krempl G (eds) Advances in intelligent data analysis XVIII. IDA
2020. Lecture notes in computer science, vol 12080. Springer, Cham. https://doi.org/10.1007/
978-3-030-44584-3_35
12. Ståhl N, Falkman G, Karlsson A, Mathiason G (2020) Evaluation of uncertainty quantification
in deep learning. In: Lesot MJ et al (eds) Information processing and management of uncer-
tainty in knowledge-based systems. IPMU 2020. Communications in computer and information
science, vol 1237. Springer, Cham. https://doi.org/10.1007/978-3-030-50146-4_41
13. Hüllermeier E, Waegeman W (2019) Aleatoric and epistemic uncertainty in machine learning:
a tutorial introduction
14. Deng W, Ye B, Wu J, Wang X, Bao J, Chen C (2019) Sparse denoising of eddy current signals
from conductive material defects based on K-SVD dictionary learning. In 2019 IEEE 8th data
driven control and learning systems conference (DDCLS). Dali, China, pp 656–660. https://
doi.org/10.1109/DDCLS.2019.8909047
15. Seyedebrahim M, Mansouri A (2017) Non-local means denoising based on SVD basis images.
In: 2017 3rd international conference on pattern recognition and image analysis (IPRIA).
Shahrekord, pp 206–210. https://doi.org/10.1109/PRIA.2017.7983047
16. Yang Y, Rao J (2019) Robust and efficient harmonics denoising in large dataset based on random
SVD and soft thresholding. IEEE Access 7:77607–77617. https://doi.org/10.1109/ACCESS.
2019.2921579
17. Zhao Z, Wang S, Wong D, Guo Y, Chen X (2020) The sparse and low-rank interpretation of
SVD-based denoising for vibration signals. In: 2020 IEEE international instrumentation and
measurement technology conference (I2MTC). Dubrovnik, Croatia, pp 1–6. https://doi.org/10.
1109/I2MTC43012.2020.9129272
18. Wang M, Li Z, Duan X, Li W (2015) An image denoising method with enhancement of the
directional features based on wavelet and SVD transforms. Math Probl Eng 2015:1–9. https://
doi.org/10.1155/2015/469350
19. Zhang G, Xu B, Zhang K, Hou J, Xie T, Li X, Liu F (2020) Research on a noise reduction
method based on multi-resolution singular value decomposition. Appl Sci 10:1409. https://doi.
org/10.3390/app10041409
20. He Y, Gan T, Chen W, Wang H (2011) Adaptive denoising by singular value decomposition.
IEEE Signal Process Lett 18:215–218. https://doi.org/10.1109/LSP.2011.2109039
21. Wang M, Yan W, Zhou S (2018) Image denoising using singular value difference in the wavelet
domain. Math Probl Eng 2018:1–19. https://doi.org/10.1155/2018/1542509
22. Liu B, Liu Q (2020) Random noise reduction using SVD in the frequency domain. J Petrol
Explor Prod Technol 10:3081–3089. https://doi.org/10.1007/s13202-020-00938-w
23. Zhao M, Jia X (2017) A novel strategy for signal denoising using reweighted SVD and its
applications to weak fault feature enhancement of rotating machinery. Mech Syst Signal Process
94:129–147. https://doi.org/10.1016/j.ymssp.2017.02.036
24. Sherstinsky A (2018) Fundamentals of recurrent neural network (RNN) and long short-term
memory (LSTM) network
25. Zheng S, Ristovski K, Farahat A, Gupta C (2017) Long short-term memory network for
remaining useful life estimation pp. 88–95. https://doi.org/10.1109/ICPHM.2017.7998311
26. Epps B, Krivitzky E (2019) Singular value decomposition of noisy data: noise filtering. Exp
Fluids 60:1–23. https://doi.org/10.1007/s00348-019-2768-4
774 A. K. M. Nor et al.
27. Epps B, Krivitzky E (2019) Singular value decomposition of noisy data: mode corruption. Exp
Fluids 60:1–23. https://doi.org/10.1007/s00348-019-2761-y
28. Wang Q, Zheng S, Farahat A, Serita S, Gupta C (2019) Remaining useful life estimation using
functional data analysis. In: 2019 IEEE international conference on prognostics and health
management (ICPHM). San Francisco, CA, USA, pp 1–8. https://doi.org/10.1109/ICPHM.
2019.8819420
29. Ge Y, Sun L, Ma J (2019) An improved PF remaining useful life prediction method based on
quantum genetics and LSTM. IEEE Access 7:160241–160247. https://doi.org/10.1109/ACC
ESS.2019.2951197
30. Ramasso E, Saxena A (2014) Performance benchmarking and analysis of prognostic methods
for CMAPSS datasets. Int J Prognostics Health Manag 5:1–15
31. Saxena A, Goebel K, Simon D, Eklund N (2008) Damage propagation modelling for aircraft
engine run-to-failure simulation. Int Conf Prognostics Health Manag. https://doi.org/10.1109/
PHM.2008.4711414.M
32. Gao Y, Zhou J, Wu K, Zhao G, Hu C (2019) Construction method of turbine engine health
indicator based on deep learning. In: 2019 prognostics and system health management confer-
ence (PHM-Qingdao). Qingdao, China, pp 1–6. https://doi.org/10.1109/PHM-Qingdao46334.
2019.8943055; Wang Q, Zheng S, Farahat A, Serita S, Gupta C (2019) Remaining useful life
estimation using functional data analysis. In: 2019 ieee international conference on prognostics
and health management (ICPHM). San Francisco, CA, USA, pp 1–8. https://doi.org/10.1109/
ICPHM.2019.8819420
33. Li J, Li X, He D (2019) Domain adaptation remaining useful life prediction method based
on AdaBN-DCNN. In: 2019 prognostics and system health management conference (PHM-
Qingdao). Qingdao, China, pp 1–6. https://doi.org/10.1109/PHM-Qingdao46334.2019.894
2857
34. Detect small changes in mean using cumulative sum. MATLAB MathWorks® https://www.
mathworks.com/help/signal/ref/cusum.html
35. Heimes F (2008) Recurrent neural networks for remaining useful life estimation. In: IEEE
international conference on prognostics and health management
Reducing Uncertainty in Failure
Prediction Using Singular Value
Decomposition Feature Selection
Ahmad Kamal Mohd Nor, Srinivasa Rao Pedapati, and Masdi Muhammad
1 Introduction
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 775
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_74
776 A. K. M. Nor et al.
Uncertainty related to the quality of input data is called Aleatoric uncertainty. This
uncertainty occurs when sensor data, used as input to deep learning model, is
subjected to measurement error or noise contamination [10–13]. This uncertainty
cannot be reduced by collecting more data [12, 13].
In the real-world applications, where ground truth RUL is not available, uncer-
tainty is often the main indicator for users to trust the model’s prediction. Security,
safety, and investment strategies depend on prediction uncertainty.
Since Aleatoric uncertainty concerns the quality of input data, it is natural to assume
that certain features contribute more to uncertainty than others. However, there have
been no notable work dedicated to study this effect. Available feature selection tech-
niques are suitable only to improve frequentist point estimates prediction without
Reducing Uncertainty in Failure Prediction Using Singular Value … 777
considering the impact it can cause uncertainty. It is obvious that in the real-world
deep learning applications, the lowest uncertainty in prediction is preferred.
In this paper, a feature selection technique to reduce Aleatoric uncertainty is
proposed. Noisy data is denoised using SVD. Then, the Signal to Noise Ratio (SNR)
for each clean feature is calculated and ranked. Signals with low SNR, i.e., high noise
are eliminated according to certain percentage.
We demonstrate our approach in a gas turbine Remaining Useful Life (RUL)
prediction problem. A probabilistic LSTM is employed to predict the gas turbine’s
RUL sequence distributions. We compare the impact of our approach on prediction
uncertainty with full feature dataset. Comparison is also done with several others
commonly used feature selection techniques such as Pearson and Spearman correla-
tions and F Regression. We believe that this is the first such work concerning deep
learning uncertainty reduction using feature selection. This paper is very important
as uncertainty management is crucial for real-world deep learning application.
This paper is organized as follows: In Sect. 2, we mention the related literature to
our work. In the next section, the methodology of SVD and SNR feature selection is
explained. Then in Sect. 4, we describe the case study before we present and discuss
our findings in Sect. 5. Finally, we present our concluding remarks in the last section.
2 Related Work
Most of the works concerning uncertainty in deep learning relate to its quantification.
Very few work touches about uncertainty management. One such publication is [16]
where Fuzzy Deep Neural Network (FDNN) is proposed to reduce data ambiguity
and noise for classification task through the fusion of fuzzy rules and neural network
logic representation. In Boskos et al. [17], Wasserstein ambiguity sets are employed
in dynamic data collection process. Ambiguity sets enables the characterization of
the true random variable’s probability distribution’s uncertainty, thus exploiting high
confidence guarantees to minimize uncertainty. In Du et al. [18], to minimize the
uncertainty in domestic solar energy harvesting, only predictions of photovoltaic
(PV) output with large probability are chosen. These probabilities help to construct
plausible scenarios enabling accurate home energy scheduling. In Rodemerk et al.
[19] context-based indicators are used to reinforce the prediction confidence, i.e.,
minimize the uncertainty of car driver’s behavior at intersections. This is done by
combining context information, such as head pose and gaze direction with vehicle
data to predict the intention of the driver.
778 A. K. M. Nor et al.
SVD is a popular tool for denoising data without the intervention of domain expert.
It is widely used in denoising image data. In Yang et al. [20] SVD, together with
non-local self-similarity technique are used to denoise image data. Block matching
techniques construct matrix based on similar patches that will be denoise by SVD.
The denoised patches will be then be combined to reconstruct the clean image.
In Leal et al. [21], sparse representation and SVD are employed to denoise MRI
image for enhancing medical diagnosis. Sparse representations are obtained from
the image subdivisions. Then, new subdivisions based on the representations are
produced through SVD denoising process. These clean subdivisions will be used
later to reconstruct the clean image. The same concept is applied in Yang et al.
[22] where SVD is used to separate and denoise Computed tomography (CT) image
components.
SVD is also used to denoise general noisy data. In Zhu et al. [23], an algorithm
based on generalized S transform (GST) and SVD is employed to denoise echo
signals in ultrasonic pulse-echo testing. The GST’s function is to transform the echo
signal to time–frequency matrix before going through SVD denoising treatment. In
Zhang et al. [24], noisy signal is presented by SVD in Henkel matrix form where
clean subspaces and noisy subspaces are separated. By only taking singular values
residing in clean subspace, clean signal can be reconstructed. In Schanze [25] a signal
compression and denoising method via objective transformation and singular values
denoising is presented. The noisy signal is first mapped into matrix form where
denoising and compression process are done. Then, inverse mapping is executed to
reconstruct the clean signal. The application of this technique is demonstrated using
biomedical signals.
3 Methodology
Ct = f t ∗ Ct−1 + i t ∗ Mt (5)
h t = ot ∗ ∅(Ct ) (6)
A ∈ RT ×D
780 A. K. M. Nor et al.
with T , the time steps and D the number variables, i.e., column of features and
with the condition that T < D.
2. A can be expressed in a reduced singular value decomposition form as:
T
A = U SV = uk sk v kT (7)
k=1
 ∈ RT ×D
and E is the matrix of noisy data containing random noise with standard deviation,
∈ that can be expressed as:
E ∈ RT ×D E ∈ RT ×D
with their respective SVD representations and the same descriptions as mentioned
before:
T
 = Û Ŝ V̂ = ûk ŝk v̂ k (8)
k=1
T
E = U S V = uk sk vk
k=1
 = A + E (9)
Reducing Uncertainty in Failure Prediction Using Singular Value … 781
Lower SVD modes in noisy data stay quite clean, thus filtering the higher effected
ones could give a cleaner signal. We assume the noise to be independent, identically
distributed (i.i.d.) Gaussian noise.
1. We project noisy time series data à as SVD matrix to obtain ũ k ,
sk and ṽk .
2. By fitting a Marchenko-Pastur distribution to the tail of noisy singular values
s̃, we can calculate the measurement error ∈. For full calculation, please refer
to [14].
3. We can then estimate the root mean square error (RMSE) of the modes with:
⎧ ⎡ ⎤ 21 ⎫
⎪
⎪ ⎪
⎪
⎪
⎪ ⎥ ⎪ ⎪
⎪
⎪ ⎢ ⎪
⎪
⎨ ⎢ ⎥ ⎪ ⎪
2 ∈⎢ D − w w T λ̃ m 3 λ̃ k − λ̃ m ⎥ ⎬
R M S E(ṽk ) ≈min , ⎢ + ⎥ ⎥
⎪
⎪ D s̃k ⎢
⎢ D D λ̃m − λ̃k ⎥ ⎪ ⎪
⎪
⎪ ⎣ m=1 ⎦ ⎪ ⎪
⎪
⎪ ⎪
⎪
⎪
⎩ ⎪
⎭
m = k
4. For i.i.d Gaussian noise, the spatial correlation parameter, f and the effective
smoothing window width, w, should be 1 [14]:
f =w=1
5. Once RMSE of the modes obtained, we can calculate the parameter tk , which
represent the cleanness of a mode.
log(R M S E(ṽk )) − log 2
D
tk = (11)
log(R M S E(ṽ1 )) − log 2
D
6. After that, we estimate the rank for minimum loss data reconstruction with the
rank rmin as:
8. Finally, the clean signal can be reconstructed with rmin and s k obtained
previously:
rmin
Ar = ũk s k ṽ k (15)
k=1
For amplitude-based signal, the SNR can be calculated by Eq. (16) with Sclean as the
clean signal and Snoise is the noise signal.
Sclean
S N R S = 20log10 (16)
Snoise
4 Case Study
The data we use is from CMAPPS (Commercial Modular Aero Propulsion System
Simulation) Turbofan run-to-failure dataset, published by Nasa Prognostic Centre
(PCoE) of Ames Research Centre [31]. The diagram of the studied turbofan is shown
in Fig. 1.
Run-to-failure time series data from 100 recorded turbofan degradations is used
for training and testing. Each set of data corresponds to a turbofan whose health
condition begin to deteriorate after certain cycle [31]. Each set comprises of Time
(Cycle), 3 Operating Conditions (OC) and 21 sensors measurements as in presented
in Table 2. The OC refers to different operating regimes combination of Altitude
(O-42 K ft.), Throttle Resolver Angle (20–100), and Mach Number (0–0.84) [32].
High levels of noise are incorporated, and the faults encountered are hidden by the
effect of various operational conditions [32]. Training data RUL targets are absent,
only the ground truth RUL for each turbofan, are given. We thus need to do data
preparation to select the appropriate sensors, visualize the sensor trends and extract
each cycle RUL target.
From the 21 sensors mentioned above, only 14 strictly monotonic sensors, namely
sensor 2, 3, 4, 7, 8, 9, 11, 12, 13, 14, 15, 17, 20 and 21 are selected [33]. Mono-
tonicity of these sensors represent single trending degradation contrary to irregular
and unchanged sensor signals.
To obtain RUL sequence labels for training, piece-wise linear degradation model
is employed [34, 35]. Each fleet health is thus considered stable in the beginning
until the start of deterioration which is the failure start point that initiate a linear
degradation until failure.
784 A. K. M. Nor et al.
outlet
S3 T30 Total temperature at HPC 0R
outlet
S4 T50 Total temperature at LPT 0R
outlet
S5 P2 Pressure at fan inlet psia
S6 P15 Total pressure in psia
bypass-duct
S7 P30 Total pressure at HPC psia
outlet
S8 Nf Physical fan speed rpm
S9 Nc Physical core speed rpm
S10 Epr Engine pressure ratio N/A
(P50/P2)
S11 Ps30 Static pressure at HPC psia
outlet
S12 Phi Ratio of fuel flow to Ps30 Pps/psi
S13 NRf Corrected fan speed rpm
S14 NRc Corrected core speed rpm
S15 BPR Bypass ratio N/A
S16 farB Burner fuel–air ratio N/A
S17 htBleed Bleed enthalphy N/A
S18 Nf_dmd Demanded fan speed rpm
S19 PCNfR_dmd Demanded corrected fan rpm
speed
S20 W31 HPT coolant bleed lbm/s
S21 W32 LPT coolant bleed lbm/s
Each fleet is associated with time series corresponding to the total operational
duration of the fleet and the last cycle indicate the final instant before failure. Thus
initially, we model the RUL of a fleet to start with the value of the last cycle and
degrade linearly until 0 as shown in Fig. 2.
The failure start point in each sensor is obtained by using Cumulative Sum
(CUSUM) anomaly detection technique, which returns the first index of the upper
and lower cumulative sums of each sensor signal that have drifted beyond 5 stan-
dard deviations above and below a target mean, indicating the initiating points of
degradation [36]. Then, the mean of the failure start points for each sensor can then
be calculated. From these means, we can calculate the mean of the means of the
sensor’s detection, which is our failure start point. Combining the linear degradation
obtained earlier and the failure start point, the transformed Fleet 1 RUL sequence is
presented in Fig. 3.
See Tables 3, 4, 5, 6.
Reducing Uncertainty in Failure Prediction Using Singular Value … 787
The noisy data uncertainty is presented in Fig. 9. Here we present the result of our
test. For illustration purposes, we only show the uncertainty result for Fleet 1.
790 A. K. M. Nor et al.
The result of SVD + SNR with 85% features is shown in Fig. 10. As we can see
in Fig. 11, SVD + SNR is the only technique that achieves the lowest uncertainty
measurement in the order of 1e-3 with 75% features. Pearson and Spearman corre-
lation feature selection with 85% features also produces lower uncertainty than the
original data as shown in Fig. 12.
Looking at the RMSE result in Table 7, we can also see that the prediction perfor-
mance is generally better with SNR + SVD than correlation-based feature selection
methods. We can understand this phenomenon as only SVD + SNR that has the effect
of denoising the data, thus improving the prediction uncertainty. In addition, SVD
is known to be amongst the best denoising technique available. Thus, the effect of
SVD denoising is really apparent with 25% less noisy features. Pearson and Spearman
correlations only optimize the features in term of linear and nonlinear correlation,
avoiding redundancy between correlated features. F Regression on the other hand,
only evaluate the correlation between features and target. This also indicates that
commonly used feature selection techniques such as Pearson and Spearman Corre-
lation as well as F Regression could result to a higher prediction uncertainty like
shown in Figs. 13 and 15. To have a clearer picture, we present in Figs. 16 and 17 the
difference in uncertainty between prediction with noisy data and with SVD + SNR
feature selection.
792 A. K. M. Nor et al.
6 Conclusion
In this paper, we present a feature selection technique based on SVD + SNR where
more noisy features are eliminated to reduce the Aleatoric uncertainty in an RUL
prediction problem. A probabilistic LSTM that predicts RUL distributions of gas
turbines is employed. From the evaluations that have been done, we demonstrate
the superiority of this method compared to correlation-based feature selection tech-
niques in reducing deep learning Aleatoric uncertainty. We also show that our feature
Reducing Uncertainty in Failure Prediction Using Singular Value … 793
selection method yields a generally more accurate failure prognostic than the other
tested techniques. Finally, this work illustrates that usual feature selection techniques
popularly used such as Pearson and Spearman as well as F Regression could result
in higher deep learning prediction uncertainty.
794 A. K. M. Nor et al.
Acknowledgements We would like to thank Yayasan Universiti Teknologi Petronas (YUTP) for
funding this research
References
1. Tsui KL, Chen N, Zhou Q, Hai Y, Wang W (2015) Prognostics and health management: a
review on data driven approaches. Math Prob Eng 2015(793161):17. https://doi.org/10.1155/
2015/793161
2. Xu M, Han M (2012) Research of prognostics and health management for Aircraft Propul-
sion System. In: Proceedings of the IEEE 2012 prognostics and system health management
conference (PHM-2012 Beijing). Beijing, pp 1–4. https://doi.org/10.1109/PHM.2012.6228891
3. Zhang A, Cui L, Zhang P (2013) State prognostics based health management for complex
systems. In: 2013 international conference on information technology and applications.
Chengdu, pp 458–461. https://doi.org/10.1109/ITA.2013.111
4. Das S, Hall R, Herzog S, Harrison G, Bodkin M, Martin L (2011) Essential steps in prog-
nostic health management. In: 2011 IEEE conference on prognostics and health management.
Montreal, QC, pp 1–9. https://doi.org/10.1109/ICPHM.2011.6024332
5. Sutharssan T, Stoyanov S, Bailey C, Yin C (2015) Prognostic and health management for
engineering systems: a review of the data-driven approach and algorithms. J Eng 7:215–222.
https://doi.org/10.1049/joe.2014.0303
6. Wilkinson C, Humphrey D, Vermeire B, Houston J (2004) Prognostic and health management
for avionics. In: 2004 IEEE aerospace conference proceedings (IEEE Cat. No.04TH8720), vol
5. Big Sky, MT, pp 3435–3447. https://doi.org/10.1109/AERO.2004.1368149
7. Shin I, Lee J, Lee JY et al (2018) A framework for prognostics and health management appli-
cations toward smart manufacturing systems. Int J Precis Eng Manuf-Green Tech 5:535–554.
https://doi.org/10.1007/s40684-018-0055-0
8. Wang J, Wen G, Yang S, Liu Y (2018) Remaining useful life estimation in prognostics using
deep bidirectional LSTM neural network. In: 2018 prognostics and system health management
conference (PHM-Chongqing). Chongqing, pp 1037–1042. https://doi.org/10.1109/PHM-Cho
ngqing.2018.00184
Reducing Uncertainty in Failure Prediction Using Singular Value … 795
9. Wang H, Liu Y, Liu Z, Wang Z, Huang H (2013) Remaining useful life estimation for degrada-
tion and shock processes. In: 2013 international conference on quality, reliability, risk, mainte-
nance, and safety engineering (QR2MSE). Chengdu, pp 1762–1764. https://doi.org/10.1109/
QR2MSE.2013.6625917
10. Ghoshal B, Tucker A, Sanghera B, Wong WL (2019)Estimating uncertainty in deep learning
for reporting confidence to clinicians when segmenting nuclei image data. In: 2019 IEEE 32nd
international symposium on computer-based medical systems (CBMS). Cordoba, Spain, pp
318–324. https://doi.org/10.1109/CBMS.2019.00072
11. Shaker MH, Hüllermeier E (2020) Aleatoric and epistemic uncertainty with random forests.
In: Berthold M, Feelders A, Krempl G (eds) Advances in intelligent data analysis XVIII. IDA
2020. Lecture notes in computer science, vol 12080. Springer, Cham. https://doi.org/10.1007/
978-3-030-44584-3_35
12. Kendall A, Gal Y (2017) What uncertainties do we need in Bayesian deep learning for computer
vision? In: Proceedings of the 31st international conference on neural information processing
systems (NIPS’17). Curran Associates Inc., Red Hook, NY, USA, pp 5580–5590
13. Depeweg S, Hernández-Lobato JM, Doshi-Velez F, Udluft S (2018) Decomposition of
uncertainty in bayesian deep learning for efficient and risk-sensitive learning. ICML
14. Epps B, Krivitzky E (2019) Singular value decomposition of noisy data: noise filtering. Exp
Fluids 60:1–23. https://doi.org/10.1007/s00348-019-2768-4
15. Epps B, Krivitzky E (2019) Singular value decomposition of noisy data: mode corruption. Exp
Fluids 60:1–23. https://doi.org/10.1007/s00348-019-2761-y
16. Deng Y, Ren Z, Kong Y, Bao F, Dai Q (2017) A hierarchical fused fuzzy deep neural network for
data classification. IEEE Trans Fuzzy Syst 25(4):1006–1012. https://doi.org/10.1109/TFUZZ.
2016.2574915
17. Boskos D, Cortes J, Martinez S (2020) Data-driven ambiguity sets with probabilistic guarantees
for dynamic processes. IEEE Trans Autom Control 1–1. https://doi.org/10.1109/TAC.2020.301
4098
18. Du Z, Wang W, Zhang J, Zhang Y, Xu X, Liu J (2020) A data-driven home energy scheduling
strategy under the uncertainty in photovoltaic generations. IEEE Access 8:54125–54134.
https://doi.org/10.1109/ACCESS.2020.2980850
19. Rodemerk C, Winner H, Kastner R (2015) Predicting the driver’s turn intentions at urban
intersections using context-based indicators. In: 2015 IEEE intelligent vehicles symposium
(IV). Seoul, pp 964–969. https://doi.org/10.1109/IVS.2015.7225809
20. Yang G, Wang Y, Xu B, Zhang X (2019) An image denoising algorithm based on singular value
decomposition and non-local self-similarity. In: Vaidya J, Zhang X, Li J (eds) Cyberspace safety
and security. CSS 2019. Lecture notes in computer science, vol 11983. Springer, Cham. https://
doi.org/10.1007/978-3-030-37352-8_44
21. Leal N, Zurek E, Leal E (2020) Non-local SVD denoising of MRI based on sparse
representations. Sensors 20:1536
22. Yang W, Hong JY, Kim JY, Paik SH, Lee SH, Park JS, Lee G, Kim BM, Jung YJ (2020) A novel
singular value decomposition-based denoising method in 4-dimensional computed tomography
of the brain in stroke patients with statistical evaluation. Sensors (Basel) 20(11):3063. https://
doi.org/10.3390/s20113063 PMID:32481740;PMCID:PMC7309118
23. Zhu Y et al (2019) Denoising ultrasonic echo signals with generalized s transform and singular
value decomposition. Traitement du Signal 36:139–145
24. Zhang X, Tang J, Zhang M, Ji Q (2016) Noise subspaces subtraction in SVD based on the
difference of variance values. J Vibroeng 18:4852–4861. https://doi.org/10.21595/jve.2016.
16745
25. Schanze T (2018) Compression and noise reduction of biomedical signals by singular value
decomposition. IFAC-PapersOnLine 51:361–366. https://doi.org/10.1016/j.ifacol.2018.03.062
26. Ruben M (2014) BayesOpt: a Bayesian optimization library for nonlinear optimization,
experimental design and bandits. J Mach Learn Res 15(1):3735–3739
27. Sherstinsky A (2018) Fundamentals of recurrent neural network (RNN) and long short-term
memory (LSTM) network
796 A. K. M. Nor et al.
28. Zheng S, Ristovski K, Farahat A, Gupta C (2017) long short-term memory network for
remaining useful life estimation 88–95. https://doi.org/10.1109/ICPHM.2017.7998311
29. Zhang X et al (2019) Remaining useful life estimation using CNN-XGB with extended time
window. IEEE Access 7:154386–154397. https://doi.org/10.1109/ACCESS.2019.2942991
30. Bruneo D, De Vita F (2019) On the use of LSTM networks for predictive maintenance in
smart industries. In: 2019 IEEE international conference on smart computing (SMARTCOMP).
Washington, DC, USA, pp 241–248. https://doi.org/10.1109/SMARTCOMP.2019.00059
31. Ramasso E, Saxena A (2014) Performance benchmarking and analysis of prognostic methods
for CMAPSS datasets. Int J Prognostics Health Manag 5:1–15
32. Saxena A, Goebel K, Simon D, Eklund N (2008) Damage propagation modelling for aircraft
engine run-to-failure simulation. Int Conf Prognostics Health Manag. https://doi.org/10.1109/
PHM.2008.4711414.M
33. Gao Y, Zhou J, Wu K, Zhao G, Hu C (2019) Construction method of turbine engine health
indicator based on deep learning. In: 2019 prognostics and system health management confer-
ence (PHM-Qingdao). Qingdao, China, pp 1–6. https://doi.org/10.1109/PHM-Qingdao46334.
2019.8943055; Wang Q, Zheng S, Farahat A, Serita S, Gupta C (2019) Remaining useful life
estimation using functional data analysis. In: 2019 IEEE international conference on prognos-
tics and health management (ICPHM). San Francisco, CA, USA, pp 1–8. https://doi.org/10.
1109/ICPHM.2019.8819420
34. Li J, Li X, He D (2019) Domain adaptation remaining useful life prediction method based
on AdaBN-DCNN. In: 2019 prognostics and system health management conference (PHM-
Qingdao). Qingdao, China, pp 1–6. https://doi.org/10.1109/PHM-Qingdao46334.2019.894
2857
35. “Detect small changes in mean using cumulative sum” MATLAB MathWorks® https://www.
mathworks.com/help/signal/ref/cusum.html
36. Heimes F (2008) Recurrent neural networks for remaining useful life estimation. In: IEEE
international conference on prognostics and health management
Unsupervised Classification of Acoustic
Emission Signal to Discriminate
Composite Failure at Low Frequency
Abstract The use of acoustic emission (AE) for damage assessment and detection
technique in structural engineering is widely used and has earned a reputation as
one of the reliable non-destructive techniques. AE source is produced based on the
elastic wave propagation through the specimen which converted into the electrical
AE signal by the AE sensors. Certain AE features belong to the signal allow their
use to discriminate mode of damage in a composite material. However, the challenge
encountered during analysis of AE signals attributed to the techniques like pattern
recognition and classification method. In this paper, several orientation of laminated
fiber specimens are undergoing tensile test. According to the information from tensile
load test, significant features were monitored namely amplitude and energy in order to
execute the classification method. The AE data are successfully cluster by k—means
algorithm.
1 Introduction
Carbon fiber laminates composite are widely used in many structure and engi-
neering application due to their excellent properties such as good corrosion resistance,
high modulus elasticity, light weight accompanied by the high strength. Despite all
these advances, the monitoring and prediction of the failure experience by a fiber
composite remain major problems. Understanding damage on fiber composite is very
challenging due to the complex process had occurred during the process.
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 797
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_75
798 N. A. A. Rahman et al.
second part of the process. There are many cluster validity techniques for determina-
tion the cluster. However, in this paper the mechanical failure of composite structure
will be verified by the load diagram which obtain from the tensile testing.
2 Experimental Procedure
Composite laminates were composed using 2 mm High Strength Carbon Fibre Sheet
which a cured carbon fiber plate supplied by Easy Composites Ltd. Staffordshire,
United Kingdom. High Strength Carbon Fibre Sheet is produced from in-house
manufacturing proses with specially design platen press of 3 layers of XPEG XC130
from the same company. XPEG XC130 is epoxy based 2 × 2 twill weave prepreg
carbon of Pyrofil TR30S high strength 3 k carbon. The specimens were prepared into
three types of lay-up which were 30°, 45° and 90° lay-ups with final dimensions of
(270 × 250 × 2.5) mm using a diamond coated abrasive cutter. Each lay-out was
cut by referring to the initial 0° of the vertical direction of the first corner of 2 × 2
twill in weave. Prior to the tensile test, all specimens underwent pre-conditioning at
23 °C for 16 h in a desiccator to stabilize the humidity to satisfy the ASTM standard.
Table 1 summarizes the lay-ups, dimension and abbreviated name used in the rest of
paper.
The mechanical testing was performed using a tensile test following the ASTM
D3039/D, standard testing method to generate damage modes of composite material.
The specimens were fixed between the jaws of the machine to undergo testing.
These tests were conducted using a universal tensile machine (UTM) manufactured
by Gotech Testing Machine Inc. (AI-7000L). The testing was used for inducing
external force to the specimen at 15 kN capacity with a constant speed 2 mm/s, at
room temperature. For each type of lay-up, at least five specimens were tested. The
average value and the standard deviation were then recorded.
800 N. A. A. Rahman et al.
2.3 AE Equipment
3 Methodology
A big data was collected during acquisition from the hardware. The hardware was
set at 40 dB threshold value in order to cut off low amplitude signal related to
environmental noise. The methodology used in the discrimination of failure modes
in post composite laminates is by using unsupervised clustering techniques. The
software used in implementation this methodology is KNIME Analytics Platform
version 4.0.2. KNIME is increasingly used by researchers in various areas of data
mining and machine learning research work. It is use graphical sequence based to
add the node for each algorithm. The pattern recognition techniques selected for the
present study is k-means. The data were normalized prior to the clustering procedure.
Fig. 1 Typical AE
waveform generated on
carbon fiber composite
during acquisition
Unsupervised Classification of Acoustic Emission Signal … 801
3.1 k-means
The k-means algorithm aims to minimize the sum of the squared distances between
all the vectors of a cluster and its center. The cluster number k is specified based
on the mechanical test information. Process for executing the k-means algorithm in
KNIME is shown in the following diagram (Fig. 2).
The k-means performs a crisp clustering that assigns a data vector to exactly
one cluster. Thus, the algorithm terminates when the cluster assignments do not
change anymore. Basically, the clustering algorithm uses the Euclidean distance on
the selected attributes. The AE features information were normalized prior to the
clustering process.
The investigation of the current effort has been carried out entirely experimentally
by laboratory tests. The results are segregated to subsections to make it easier for
understanding the insight of each effective variables.
Amplitude, dB
each specimen 90
0.2
Stress, kN
80
0.15
70
0.1
60
0.05 50
0 40
0 50 100 150 200 250
Time, sec
0.3 5000
0.25 T30
4000
0.2
Energy, eu
Stress, kN
3000
0.15
2000
0.1
0.05 1000
0 0
0 50 100 150 200 250
Time, sec
0.3 110
T45 100
0.25
Amplitude, dB
90
Stress, MPa
0.2
80
0.15
70
0.1 60
0.05 50
0 40
0 100 200 300
Time, sec
0.3 5000
0.25 T45
4000
Stress, MPa
0.2
Energy, eu
3000
0.15
2000
0.1
0.05 1000
0 0
0 100 200 300
Time, sec
Unsupervised Classification of Acoustic Emission Signal … 803
Amplitude, dB
90
Stress, MPa
20
80
15
70
10 60
5 50
0 40
0 50 100 150 200 250
Time, sec
30 5000
T90
25 4000
Stress, MPa
20
Energy, eu
3000
15
2000
10
5 1000
0 0
0 50 100 150 200 250
Time, sec
the T90 specimen followed by T45 later. Figure 3 also show the evolution of AE signal
energy and amplitude during tensile load carried out on each specimen. Generally,
the matrix cracking and some interfacial failures are recorded. Since the interested
frequency (< 100 dB) are recorded. This is logical because, as the load increase, not
only is there much more extensive matrix cracking but also many more delaminations
occur at the previous loading steps. Moreover, as the loading continued, the crack
propagation spread to the top, rupturing of the rest of specimen, while fiber pull-out
gradually occurred.
According to Fig. 3, as expected the highest energy value recorded for all spec-
imen are right before the specimen were break. Generally, the release elastic energy
produced i.e. by matrix cracking and delamination is detected from the beginning
of the tensile load test. Moreover, these failures producing a very ‘loud’ sound to be
listened by the sensor. The highest recorded energy before fiber break for specimen
T30, T45 and T90 are 4051 eu, T45 3094 eu and T90 4508 eu respectively. When
the load increased, trends of the AE energy of the specimen were similar regardless
the fiber orientation of the composite. Similar pattern is recorded by the amplitude
804 N. A. A. Rahman et al.
4.2 Clustering
AE cluster behavior were assessed and compared to each of the test specimen exposed
to tensile loading. AE clusters were based on the normalized amplitude and energy
feature of each AE event. Practically, the clustering is based on the damage evolution
during tensile load testing. If one damage type is the exclusive source of an AE
cluster, both damage and the accumulation of event cluster would start together [1].
In this study, the k value is based on the type of mechanical during tensile load.
According to stress diagram, the mode of failures can be divided into two major
cases namely delamination and matrix cracking (occurred at the beginning of test)
and fiber breakage fiber pull-out towards end of experiment. As mention previously,
only a low frequency of damage is counted, based on the type of AE sensor. Variations
in the damage sources associated with the same mechanism most probably exist. The
clustering results assigning the AE features (amplitude and energy) data as shown in
Fig. 3 (Fig. 4).
5 Conclusion
T30
Amplitude
Energy
Peak frequency Peak frequency
T45
Amplitude
Energy
Peak frequency
T90
Amplitude
Energy
References
1. Sawan HA, Walter ME, Marquette B (2015) Unsupervised learning for classification of acoustic
emission events from tensile and bending experiments with open-hole carbon fiber composite
samples. Compos Sci Technol 107:89–97
2. Bakhtiary Davijani AA, Hajikhani M, Ahmadi M (2011) Acoustic emission based on sentry
function to monitor the initiation of delamination in composite materials. Mater Des 32(5):3059–
3065
3. Calabrese L, Galeano M, Proverbio E, Di D, Cappuccini F, Donato A (2016) Monitoring of
13% Cr martensitic stainless steel corrosion in chloride solution in presence of thiosulphate by
acoustic emission technique. Corrosion Science
4. Ech-Choudany Y, Assarar M, Scida D, Morain-Nicolier F, Bellach B (2017) Unsupervised
clustering for building a learning database of acoustic emission signals to identify damage
mechanisms in unidirectional laminates. Appl Acoust 123:123–132
5. Koruk H, Genc G (2015) Investigation of the acoustic properties of bio luffa fiber and composite
materials. Materials Letter 157:166–168
6. Gutkin R, Green CJ, Vangrattanachai S, Pinho ST, Robinson P, Curtis PT (2011) On acoustic
emission for failure investigation in CFRP: Pattern recognition and peak frequency analyses.
Mech Syst Signal Process 25(4):1393–1407
7. Ramirez-Jimenez CR, Papadakis N, Reynolds N, Gan TH, Purnell P, Pharaoh M (2004) Identi-
fication of failure modes in glass/polypropylene composites by means of the primary frequency
content of the acoustic emission event. Compos Sci Technol 64(12):1819–1827
8. De Groot PJ, Wijnen PA, Janssen RB (1995) Real time frequency determination of acoustic
emission for different fracture mechanisms in carbon/epoxy composites. Compos Sci Technol
55:405–412
9. Gholizadeh S, Leman Z, Baharudin BTHT, Inayatullah O (2018) Acoustic emission analysis
for characterisation of damage mechanisms in glass fiber reinforced polyester composite. Aust
J Mech Eng 16(1):11–20
Finding the EEG Footprint of Stress
Resilience
Rumaisa Abu Hasan, Syed Saad Azhar Ali, Tong Boon Tang,
and Muhamad Saiful Bahri Yusoff
Abstract Work stress faced by adults can lead to decreased job performance,
reduced mental and physical wellbeing, and other detrimental health problems.
Researchers are reporting resilience as a key factor in determining a person’s vulner-
ability towards mental stress disorders. Psychosocial measures of resilience conven-
tionally use the self-assessment approach which is susceptible to potential biases
caused by self-reporting and concerns of social stigma. With increasing emphasis
of its role in mental health, researchers are using fMRI modality to identify the
brain activity of stress resilience. But this approach is costly and lack practicality
when evaluating stress resilience in daily tasks. The EEG modality provides a cost-
efficient alternative with better practicality and high temporal resolution in studying
the brain activity of stress resilience. However, EEG-based literatures on stress
resilience are limited to brain activity during resting state. With reference to the cogni-
tive affective conceptual stress model, we define stress resilience as an adaptation
process, involving cognitive appraisal, physiological arousal and coping behaviour,
that utilizes individual resources to cope with stress. This paper proposes an approach
to identify the features of EEG-neural correlates of stress resilience through brain
rhythms, hemispheric asymmetry and brain network.
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 807
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_76
808 R. A. Hasan et al.
1 Introduction
In year 2002, the World Health Organization published a monograph concerning the
alarming mental health problem as one of the leading causes of disability and dis-
ease world-wide [1]. The increase in work-related stress among adults is a raising
concern throughout the world [2–4]. This problem has been recognized as detrimental
in both poor and rich countries, and persist in population regardless of age, gender and
social strata. In Malaysia, the percentage of clinically-diagnosed adults with stress-
disorders have increased [5], with a 29.9% overall prevalence of work-related stress
[2]. This indicates that one of three adults are suffering from work-related stress.
Despite being an “unseen” condition, work-related stress has its pathogenicity [2]
and leads to mental health deterioration [6].
As resilience is the adaptation of mental and physical responses to stress, the brain
also exhibits neural activities that reflect this mechanism. When continuously ex-
posed to stress, a region within the brain limbic system known as amygdala under-
goes structural changes through dendritic growth and spine synapses formation,
whereas medial prefrontal cortex has been observed to undergo dendritic short-
ening [14]. Neural activity within the prefrontal cortex has also been reported to
influence the reactivity of amygdala in emotional arousal [15]. The hypothalamic–
pituitary–adrenal (HPA) axis is another neural circuitry which plays a key role in
determining the degree of individual stress resilience from the gestational stage [8].
Stress stimuli alter the neuroendocrine activity of this axis, triggering the cascading
effects of neuronal modulation and hormonal secretion that affects the brain. More-
over, the reciprocal communication of both the HPA axis and medial prefrontal
cortex with the dorsal raphe nucleus forebrain region highlights the functional impli-
cation of prefrontal cortex on the HPA axis [8]. Franklin et al. (2012) highlighted
an interesting hypothesis on the role of ventral medial prefrontal cortex in acquiring
resilience through the cognition of progressive learning. These findings on the func-
tional connectivity of frontal activity and both amygdala and the HPA axis emphasize
the relevance for a brain-based measure for stress resilience as this characteristic has
been strongly associated to vulnerability to mental illness [8, 9].
The role of resilience in stress coping is evident [8, 16–18]. However, the EEG-based
literature investigating the stress resilience is still at its infancy despite the increasing
findings from fMRI studies highlighting the role of prefrontal cortex in resilience.
Findings from studies that focus on cohorts with specific life experience may not be
generalized to normal healthy populations. The association of EEG-brain waves with
resilience at resting state of the brain may also differ when the brain is responding
to a stressful event. These limitations highlight the need to identify EEG-neural
correlates for resilience in a stress-inducing state. These findings can further be used
in resilience training program for healthy population.
2 Proposed Methodology
presented to subjects in the fMRI-based literatures [13, 17, 18]. The stimuli trigger
the stress resilience mechanism which attempts to cope with the brain activity
changes induced by stress. Although many studies on stress resilience using the
EEG modality evaluated the brain activity at resting state [11, 22, 23], these findings
provide directions for further analysis in identifying the features for stress resilience.
Compared to the literature on stress responses, the EEG brain markers of stress
resilience are less studied [23]. Paban et al. (2019) has observed significant corre-
lation between resilience level and brain network flexibility in the delta, alpha and
beta waves. Flexibility index is a measure of how often one node (i.e. a brain region)
within a network changes its affiliation with a module (i.e. a set of strongly connected
brain regions performing specific function [24]) over time [23, 25, 26]. Flexibility
of delta, alpha and beta brain networks during 10 min of resting with eyes closed
was negatively correlated with the psychological resilience score [23]. Network flex-
ibility of specific brain regions significantly decreased with higher resilience score,
suggesting that the brain core network is less flexible in resilient subjects.
Zhang et al. (2018) also investigated the association of resilience of network at
resting-state of EEG signal between patients with major depressive disorders and
healthy participants [22]. In the study, network resilience reflected the degree of
tolerance when the functional brain network is subjected to unstable events. Patients
with MDD had lower inter-nodal mean and positive correlation coefficients in the
alpha wave of the functional brain network, which suggests weaker density of node-
pairing connections compared to healthy participants. This finding strengthens the
postulation by Paban et al. (2019) that people with higher psychological resilience
level have a more stable rich-club network.
In a study investigating biomarker for attentional control using theta/beta ratio,
Putman et al. observed that higher power density of the ratio is significantly associated
with lower self-reported trait attentional control [27]. Findings from this study was
concluded to contribute in predicting an individual’s resilience to effects of stress-
inducing cognitive performance anxiety on the self-reported state attentional control.
Although this conclusion is unclear as to its association with stress resilience, the
beta band may be linked to the brain activity reflecting stress resilience as observed
by [23].
Curtis and Cicchetti studied the resilience and its associated hemispheric EEG
asymmetry during resting eyes-open and closed in maltreated and non-maltreated
children [11]. Findings suggested that resilient children had greater left hemispheric
activity in the central cortical area, C3 and C4. Asymmetrical activity in the parietal
cortices, favoring P3 location, was also observed in non-maltreated children.
812 R. A. Hasan et al.
Sit uat ional Dem and Cognit ive Appraisal Physiological Arousal Inst rum ent al and coping behaviour
intervention conditions [31]. The sample size estimated for this experiment is N =
29, calculated as follows [32]:
2
2 z α + z β (1 + (n − 1)ρ)
N=
nδ 2
N = 29
where
N = sample size.
Zα = 1.96 at level confidence of 95%
Zβ = 0.842 at 80% power.
n = 3, number of time points.
ρ = 0.5, assumed correlation of repeated measure.
δ = 0.6, effect size.
To address our hypothesis, the flow of EEG-signal analysis shown in Fig. 2 is
proposed for our study. We aim to identify the prominent EEG bands and hemispheric
asymmetries that may reflect the adaptation process of stress resilience. We also
aim to identify the brain networks between stress-inducing event and stress-induced
training using EEG source-space connectivity method.
End
814 R. A. Hasan et al.
4 Summary
Prolonged stress in working environment can lead to reduced quality of job perfor-
mance and quality of life. Literatures have found that mental health is associated
with lower stress level and higher resilience. As resilience is defined as the ability to
cope and adapt to stressful situation, there is a need to identify the EEG-neural corre-
lates of this psychological parameter when stress in being induced. In addition to
its advantages of being non-invasive, having high temporal resolution, low cost and
ease-of-portability, the EEG modality has been extensively used in monitoring stress
responses. Through this study, we would be able to identify the features of stress
resilience and correlate them to the conventional psychological measures. Similar to
the application of EEG-feature of stress response, these potential features may be
used in clinical practices for risk assessment, diagnosis and interventions.
References
1. Harnois G, Gabriel P (2002) Mental health and work: impact, issues and good practices. World
Health Organization
2. Kassim MSA, Ismail A, Ismail R (2018) A review of occupational stress prevalence and
its predictors among selected working populations in Malaysia. Malays J Publ Health Med
18(2):1–6
3. Fink G (2016) Stress: Concepts, Cognition, Emotion, and Behavior, ser. In: Handbook of Stress
Series. Academic Press, vol. 1
4. Feltoe G, Beamish W, Davies M (2016) Secondary school teacher stress and coping: insights
from Queensland, Australia. Int J Arts Sci 9(2):597–608
5. Menon J (2016) Depression on the rise as Malaysians burn out from stress, expert warns.
MalayMail. [Online]. Available: https://www.malaymail.com/news/malaysia/2016/03/24/dep
ression-on-the-rise-as-malaysians-burn-out-from-stress-expert-warns/1086087
Finding the EEG Footprint of Stress Resilience 815
6. Thye TSLL (2016) Work stress and mental health. New Straits Times. [Online]. Available:
https://www.nst.com.my/news/2016/04/142074/work-stress-and-mental-health
7. Smith RE, Ascough JC (2016) Promoting emotional resilience: cognitive-affective stress
management training. The Guilford Press, New York
8. Franklin T, Saab B, Mansuy I (2012) Neural mechanisms of stress resilience and vulnerability.
Neuron 75(5):747–761. [Online]. Available: http://www.sciencedirect.com/science/article/pii/
S0896627312007532
9. Davydov DM, Stewart R, Ritchie K, Chaudieu I (2010) Resilience and mental health. Clin
Psychol Rev 30(5):479–495. [Online]. Available: http://www.sciencedirect.com/science/art
icle/pii/S0272735810000437
10. Connor KM, Davidson JR (2003) Development of a new resilience scale: the connor-davidson
resilience scale (CD-RISC). Depression Anxiety 18(2):76–82. [Online]. Available: https://onl
inelibrary.wiley.com/doi/abs/https://doi.org/10.1002/da.10113
11. Curtis WJ, Cicchetti D (2007) Emotion and resilience: a multilevel investigation of hemispheric
electroencephalogram asymmetry and emotion regulation in maltreated and nonmaltreated
children. Dev Psychopathol 19(3):811–840
12. Burton NW, Pakenham KI, Brown WJ (2010) Feasibility and effectiveness of psychosocial
resilience training: a pilot study of the ready program. Psychol Health Med 15(3):266–277,
pMID: 20480432. [Online]. Available: https://doi.org/10.1080/13548501003758710
13. Keynan JN, Cohen A, Jackont G, Green N, Goldway N, Davidov A, Meir-Hasson Y, Raz
G, Intrator N, Fruchter E, Ginat K, Laska E, Cavazza M, Hendler T (Jan 2019) Electrical
fingerprint of the amygdala guides neurofeedback training for stress resilience. Nat Hum Behav
3(1):63–73. [Online]. Available: https://doi.org/10.1038/s41562-018-0484-3
14. McEwen BS (2006) Protective and damaging effects of stress mediators: central role of the
brain. Dialogues Clin Neurosci 8(17290796):367–381. [Online]. Available: https://www.ncbi.
nlm.nih.gov/pmc/articles/PMC3181832/
15. Goodman RN, Rietschel JC, Lo L-C, Costanzo ME, Hatfield BD (2013) Stress, emotion regu-
lation and cognitive performance: the predictive contributions of trait and state relative frontal
EEG alpha asymmetry. Int J Psychophysiology 87(2):115–123. [Online]. Available: http://
www.sciencedirect.com/science/article/pii/S0167876012005934
16. Kong F, Wang X, Hu S, Liu J (2015) Neural correlates of psychological resilience and their
relation to life satisfaction in a sample of healthy young adults. NeuroImage 123:165–172.
[Online]. Available: http://www.sciencedirect.com/science/article/pii/S1053811915007296
17. Peres JF, Foerster B, Santana LG, Fereira MD, Nasello AG, Savoia M, Moreira-Almeida A,
Lederman H (2011) Police officers under attack: resilience implications of an fMRI study. J
Psychiatr Res 45(6):727–734. [Online]. Available: http://www.sciencedirect.com/science/art
icle/pii/S0022395610003225
18. Reynaud E, Guedj E, Souville M, Trousselard M, Zendjidjian X, Khoury-Malhame ME, Fakra
E, Nazarian B, Blin O, Canini F, Khalfa S (2013) Relationship between emotional experi-
ence and resilience: an fmri study in fire-fighters. Neuropsychologia 51(5):845–849. [Online].
Available: http://www.sciencedirect.com/science/article/pii/S0028393213000134
19. Meir-Hasson Y, Keynan JN, Kinreich S, Jackont G, Cohen A, Podlipsky-Klovatch I, Hendler
T, Intrator N (May 2016) One-class fmri-inspired EEG model for self-regulation training. PLOS
ONE 11(5):e0154968. [Online]. Available: https://doi.org/10.1371/journal.pone.0154968
20. Scarapicchia V, Brown C, Mayo C, Gawryluk JR (2017) Functional magnetic resonance
imaging and functional near-infrared spectroscopy: insights from combined recording studies.
Frontiers Human Neurosci 11:419. [Online]. Available: https://www.frontiersin.org/article/
https://doi.org/10.3389/fnhum.2017.00419
21. Thibault RT, Lifshitz M, Birbaumer N, Raz A (2015) Neurofeedback, self-regulation, and brain
imaging: clinical science and fad in the service of mental disorders. Psychother Psychosom
84(4):193–207. [Online]. Available: https://www.karger.com/DOI/https://doi.org/10.1159/000
371714
816 R. A. Hasan et al.
22. Zhang M, Zhou H, Liu L, Feng L, Yang J, Wang G, Zhong N (2018) Randomized EEG
functional brain networks in major depressive disorders with greater resilience and lower rich-
club coefficient. Clin Neurophysiol 129(4):743–758. [Online]. Available: http://www.scienc
edirect.com/science/article/pii/S1388245718300324
23. Paban V, Modolo J, Mheich A, Hassan M (2019) Psychological resilience correlates with
EEG source-space brain network flexibility. Netw Neurosci 3(2):539–550. [Online]. Available:
https://doi.org/10.1162/netn_a_00079
24. Bassett DS, Wymbs NF, Porter MA, Mucha PJ, Carlson JM, Grafton ST (2011) Dynamic
reconfiguration of human brain networks during learning. Proc National Acad Sci 108(18):
7641–7646. [Online]. Available: https://www.pnas.org/content/108/18/7641
25. Mattar MG, Betzel RF, Bassett DS (July 2016) The flexible brain. Brain 139(8):2110–2112.
[Online]. Available: https://doi.org/10.1093/brain/aww151
26. Braun U, Schäfer A, Walter H, Erk S, Romanczuk-Seiferth N, Haddad L, Schweiger JI, Grimm
O, Heinz A, Tost H, Meyer-Lindenberg A, Bassett DS (Sept 2015) Dynamic reconfiguration
of frontal brain networks during executive cognition in humans. Proc National Acad Sci US
Am 112(26324898):11678–11683. [Online]. Available: https://www.ncbi.nlm.nih.gov/pmc/art
icles/PMC4577153/
27. Putman P, Verkuil B, Arias-Garcia E, Pantazi I, van Schie C (Jun 2014) EEG theta/beta ratio as
a potential biomarker for attentional control and resilience against deleterious effects of stress
on attention. Cogn Affect Behav Neurosci 14(2):782–791. [Online]. Available: https://doi.org/
10.3758/s13415-013-0238-7
28. Lazarus RS, Folkman S (1984) Stress. Springer Publishing Company, Appraisal and Coping
29. Kyriacou C (2001) Teacher stress: directions for future research. Educ Rev 53(1):27–35
30. Research H (2015) Supporting teachers during times of change and stress. Hanover Research
31. Jhangiani RS, Chiang I-CA, Price PC (2015) Research Methods in Psychology. BCcampus
Open Education
32. Hedeker D, Gibbons RD, Waternaux C (1999) Sample size estimation for longitudinal designs
with attrition. J Educ Behav Stat 24:70–93
33. Seo S-H, Lee J-T (2010) Stress and EEG. In: Crisan M (ed) Convergence and hybrid information
technologies. Intech, ch 27, pp 413–426
34. Brouwer A-M, Neerincx MA, Kallen V, van der Leer L, ten Brinke M (2011) EEG alpha
asymmetry, heart rate variability and cortisol in response to virtual reality induced stress. J
Cyberther Rehabil 4(1):27–40
35. Flo E, Steine I, Blågstad T, Grønli J, Pallesen S, Portas CM (2011) Transient changes in frontal
alpha asymmetry as a measure of emotional and physical distress during sleep. Brain Res
1367:234–249. [Online]. Available: http://www.sciencedirect.com/science/article/pii/S00068
99310021566
Modeling and Tuning of PID Controller
for Continuous Stirred Tank Reactor
1 Introduction
Continuous stirred tank reactor which is widely used equipment mainly in the chem-
ical industry, exhibits highly nonlinear behavior. A CSTR is a exothermic irreversible
reaction, a fluid stream is continuously fed to the reactor. Since the reactor is perfectly
mixed the exit stream has the same concentration and temperature as the reactor fluid.
The jacket surrounding the reactor has feed and exit streams. The jacket is assumed
perfectly mixed, energy passes through the reactor walls into the jacket, removes
the heat generated by reaction. The mathematical model is developed from material
balance equation [1]. The two state variables are the reactor concentration and reactor
temperature. PID control has three modes i.e., proportional, integral and derivative.
The proportional term is proportional changes for error (the difference between the
set point and process variable) to the control output. The integral term examines
the process variable over time and it corrects the output by reducing the offset from
A. Suguna
Government College of Technology, Coimbatore, India
S. N. Deepa
Anna University Regional Campus, Coimbatore, India
N. Rajasingam (B)
Dr. N. G. P. Institute of Technology, Coimbatore, India
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 817
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_77
818 A. Suguna et al.
process variable. Derivative control mode monitors the rate of change of the process
variable and therefore changes the output when there are unusual variations. Each
parameter of the three control functions adjust to obtain the desired performance
from the process [2]. Easy implementation and maintenance due to their simple struc-
ture, PID controllers are the most extensively used controllers in the process control
industry. Different types of controllers are available like sliding mode control, robust
control. But the PID controllers still used in industries, Conventional tuning methods
of PID controller used for this CSTR process.
The mathematical model for the Continuous Stirred Tank Reactor is derived from
mass balance and energy balance equation [3]. Mass balance equation is given by
the rate of accumulation for material is equal to difference between rate of material
inflow and outflow.
dV C A
= FC A0 − FC A − r V (1)
dt
Energy balance equation is given by
d v P C p (T − T r e f ) /dt = Fρ C p T f − T r e f − Fρ C p (T − T r e f )
+ (−H )Vr − U A(T − Ti ) (2)
dC A F
f 1 (C A , T ) = = CAf − CA − r (3)
dt V
dT F −H UA
f 2 (C A , T ) = = Tf − T + r− (T − Ti ) (4)
dt A pC p vpC p
F −H −E UA
f 2 (C A , T ) = 0 = Tf − T + K 0 ex p CA − (T − Ti )
V pC p RT vpC p
(7)
3 PID Controller
This paper presents ZN closed loop tuning method. It requires the determination of
the ultimate gain and ultimate period [5] By adjusting the controller gain (Ku) till the
system goes sustained oscillations (ultimate gain) this can be achieved, while main-
taining the integral time Constant (Ti) at infinity and the derivative time constant
(Td) at zero. Note down ultimate gain-Ku (gain at oscillation occurs) and ultimate
period—Pu. The tuning parameters Proportional gain, Integral Time and Derivative
Time are calculated from prescribed ZN tuning rules as mentioned in the Table.1 [6].
820 A. Suguna et al.
32+ 6τ
1 T 4 τ T
Cohen and coon K τ 3 4T τ τ 4
13+ 8τ
T 11+ 8τ
T
(τ +0.5T ) τT
IMC based PID K p (λ+0.5T ) τ + 0.5T 2τ +T )
B.D Tyreus and W.I Luyben in 1997, proposed this method. To calculate the ultimate
gain Ku and ultimate period Pu the same Ziegler Nichols tuning procedure have to
be followed. Controller parameters calculated from the rules given below.
The other name of Cohen and Coon method is Process reaction curve method, the
step input is given to the open loop response of the system, the S shaped curve
obtained. From this curve gain, dead time and time constant calculated, Substituting
this parameters into the tuning rules controller tuning parameters computed.
3.4 IMC
To model CSTR process by fitting the open-loop step test data as a first order function
with time delay, the internal model control-PID (IMC-PID) tuning rules have the
advantage of using a single tuning parameter to achieve a clear tradeoff between
closed-loop performance and robustness to model inaccuracies.
The IMC-PID controller provides good set-point tracking but has a sluggish distur-
bance response [7]. The different tuning rules for PID controller as shown in Table
1.
2. The unit step response for the close loop control of CSTR is given to get the output
response.
Closed loop response parameters are analyzed using the time domain specifica-
tions of all PID tuning methods and shown in Table 3. To use these empirical tuning
rules to CSTR process, we need to fit the data to a first order with dead time. First
order with dead time system obtained the approximation of the CSTR by giving step
signal input.
Figures 1, 2, 3 and 4 represents the closed loop response of the CSTR process based
on the PID tuning parameters of Ziegler-Nichols, Tyreus—Luyben, Cohen-Coon and
IMC based PID methods.
Fig. 1 Response of
CSTR-Ziegler Nicholas
822 A. Suguna et al.
Fig. 2 Response of
CSTR-Tyrus Luyben
Fig. 3 Response of
CSTR-Cohen and Coon
5 Conclusion
Fig. 4 Response of
CSTR-IMC based PID
analyze IMC based PID controller provide good transient response for the CSTR
process.
References
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 825
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_78
826 M. I. Beer Mohamed and M. F. Hassan
1 Introduction
This section briefs on the rationale and motivation behind the conducted research
study. The security concerns and the potential vulnerabilities of the cognitive robotic
process automation are reviewed in terms of available literature, and the mandatory
need for a proactive security model is highlighted as the traditional reactive security
model is inadequate in this modern era of cognitive automation with the advent of
related software and hardware technologies.
In addition to the common enterprise application integration risks, the potential risks
for cognitive RPA are listed below [5, 6].
• Privilege escalation: It is the risk of abusing the given access privilege given to a
bot.
• Data tampering: It is the risk of disclosing sensitive data that are tampered with
during inter-process communication.
• Denial of Service: It is the risk of stressing with the bot with a greater number of
fake requests so that it will deny or delay in servicing to even legitimate requests.
• Application vulnerabilities: It is the risk of enacting the security threat using the
existing vulnerabilities in the application components and/or supported software
modules that are deployed on bots.
• Abusing the digital identity and access control: It is the risk of illegitimately
applying digital federated identity and access privileges that are given to specific
bots.
• Data identification and protection: It is both business and operations risk of sensi-
tive data handling by robotics and automation tools adhering to compliance with
the governance body.
828 M. I. Beer Mohamed and M. F. Hassan
The potential security threats for RPA are increasing as the attack surface is widening
due to the advent of the latest automation and supported technologies. Hence, the
security for RPA became a mandatory element in the business operating model. The
traditional reactive way of security threat identification and perimeter defense strate-
gies are obsolete right now [9–11]. It should be the paradigm shift with continuous
monitoring and response to threats, shifting from reactive to a proactive approach to
handling and preventing security attacks with the support of Artificial Intelligence
(AI) and related technologies.
An adaptive security approach applies a layered way of detecting and preventing
security threats. Hence it is focused to sense, detect, and respond to the prevention
of security threats before the potential vulnerabilities are being used by the attackers
on performing attacks for the exploitation of enterprise assets.
The adaptive security approach focuses on security aspects before, during, and
post-execution of automated processes, targeted for (i) reducing the attack surface
and attack velocity, (ii) real-time event-based monitoring on identifying security
breaches, (iii) assurance of reliability and availability of automation resources, (iv)
ensuring reliability in automation processes, (v) prevention of data and identity loss,
and (vi) anticipated security thereat for earlier prevention.
The artificial neural networks (ANN) approach differs from conventional compu-
tation techniques in problem-solving as the conventional techniques apply the stan-
dard procedural and algorithmic approach with a pre-defined set of instructions where
the ANN is being used to solve the problems which cannot be expressed exactly in
terms of algorithmic steps. Security for cognitive-based solutions is one kind of
such problem that can be effectively solved only through AI-based analytics and
implementation [12].
Adaptive Security for Cognitive Robotic Process Automation … 829
signals meet the threshold limit in a short period. The transmitted electrochemical
signal from dendrites to an axon via the nucleus will act as an input dendrite signal
for another neuron.
The artificial neuron collects input signals as values that are being multiplied by
the factor, named weights which represent the strength of the respective input signals.
Then, these values are fed to the mathematical function (aka transfer function) which
determines whether to activate the neuron or not. If the weights to the input signals
are high, then the corresponding neuron is strongly connected. The mathematical
function is equivalent to the nucleus in a biological neuron, and the weights are similar
to synapses which can hold positive, negative, or neutral (zero) values. The learning
in ANN is the process of adjusting these weights for desired functionality such as
threat identification and pattern matching. For this research study, the hyperbolic
tangent sigmoid mathematical model is taken as a transfer function which has the
representation as below.
ex − e−x
f(x) =
ex + e−x
∂E
The gradient of each weight concerning the desired output is measured as, ∂w ij
where E represents the sum of input values which are multiplied by their corre-
sponding assigned weights, and wij represents the actual assigned weight on an edge
which is connecting node i to node j. The backpropagation performs this gradient
measurement on each edge and adjusts the weights incrementally using stochastic
∂E
gradient descent as wi j = wi j − η ∂w ij
for all the nodes in the defined neural network.
The process flow of the proposed AI-powered threat prevention solution involves
a systematic approach as outlined in Fig. 4.
The proposed security component “ISE” as depicted in Fig. 2 embeds this process
model as given in Fig. 4. This software component can be easily added with any
RPA processing nodes through an interceptor design pattern with a simple config-
uration where the input/output messages will be intercepted for threat monitoring
and prevention. Though this model is good enough on learning the existing security
vulnerabilities and threats, the training part will require a significant amount of time
to build the knowledge base. Instead, the data mining techniques can be used to
construct the initial knowledge base where it will be further refined by neural nets.
on-boarding journey from due diligence checking, fraud monitoring, KYC updates,
opening up the banking accounts, and registration for debit/credit cards, up to instruct
the customer upon the account is ready for operation. The cheque clearance bot
performs an intelligent check on signature matching, verification of the given details
in the cheque, validation of transactions, and perform the fund transfer.
The artificial neural network model with back-propagation is chosen for the
learning part and implemented for fraud and security risk assessment purposes, as
part of PoC implementation. This ANN-based model produces reliable results when
compared to other forecasting mechanisms.
As obtained from the conducted PoC, the statistical data on security attack and
prevention ratio with existing (without AI) and proposed (AI-powered) security solu-
tions on the chosen attack vector on cognitive RPA is given in Table 1. The ratio is
represented as a failure of protection to protect from security attacks, for example, the
ratio of 0.6000: 0.4000 represents 60% of failure on protection and 40% of protection
from the underlying security attacks.
By analyzing the PoC results, the observed sensitivity rate falls above 75% for
the proposed AI-powered security solution. Except for the security protection for
broken access control, the protection rate of the proposed security solution is at least
35% higher than the existing security solutions. Moreover, a higher protection rate
for broken access control can be achieved through customized machine learning
processes. The graphical representation of the obtained result as given in Table 1
is portrayed in Figs. 5, 6 and 7. With this analysis, it is clear that the proposed AI-
powered security solution for cognitive RPA supersedes the existing non-AI based
security solutions with a higher protection rate.
5 Conclusion
The cognitive robotic process automation is the modern trend in enterprise computing
and automation targeted for cost reduction and higher throughput in a minimum
amount of time. With cognitive RPA, the automated tasks are getting completed
with more accuracy and expected scalability. However, the security boundaries are
getting expanded due machine to machine interfacing and end-to-end automation
of business processes without any human interventions. In this paper, the major
security risks on cognitive RPA are studied in the view of available literature, a navel
security software component is constructed and presented in an adaptive way forward
approach. This security approach is equipped with artificial intelligence and related
technologies and best suited for this current era of business process automation. The
proposed approach is practically implemented and tested as part of proof-of-concept,
the obtained results are analyzed, and proved that the proposed AI-powered security
solution provides at least 35% higher security than the existing non-AI based security
solutions.
References
1 Introduction
Image inpainting is the method of restoring the target regions that are scratched,
missed or damaged. This process was first introduced to reconstruct the deteriorated
old pictures or artworks. It is also called as image disocclusion or completion. Image
inpainting has become a dynamic research subject in the area of computer vision and
image processing since the recent advancement of digital image usage. The ultimate
goal of image inpainting is to restore the lost region of a damaged picture in an
unnoticeable manner. Nowadays image inpainting are widely used for reconstructing
old pictures and restoring digital images.
B. Vidhya (B)
Dr. N. G. P Institute of Technology, Coimbatore, India
M. N. Madhav
Sri Shakthi Institute of Engineering and Technology, Coimbatore, India
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 837
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_79
838 B. Vidhya and M. N. Madhav
The main classification of image inpainting methods are: the exemplar based,
the sparse based and the diffusion-based techniques. Partial Differential Equations
(PDE) are used in the diffusion based approach where the missing regions are filled
by the diffusion of image details from the source regions. The filling of missing
regions was achieved by the diffusion of the image details along the isophote direc-
tion by Bertalmio et al. [1]. Later Bertalmio et al. [2] presented the Navier–Stokes
equivalence in fluid dynamics into the process of image inpainting. T. F. Chan et al [3]
utilized the Total Variation (TV) model for reconstructing the damaged regions. Then,
T. F. Chan et al [4] anticipated the curvature-driven diffusion (CDD) equivalence to
improve the TV model that underperformed in the connectivity principle.
P-Harmonic energy minimization was introduced in the TV model for filling the
target regions by Zhang et al. [5], while other methods normally use surrounding
pixels to fill the target pixels. Telea used first marching method to complete the
target region by filling the known pixels of the neighbour from exterior to interior
region [6]. Weickert et.al [7] proposed a unified framework for interpolation built on
elliptic partial differential equations. Image was restored by Takeda et al. [8] through
the association of the nonparametric statistics field and expanded kernel regression
concepts. Many examinations were conducted by the researchers in the process of
image inpainting to attain an improved restoration result.
Reconstruction of non textured missing regions that are smaller in size achieved
by the diffusion-based algorithms have proven results. But its performance is poor
for smaller textured or structured portions because of inadequacy of semantic
texture/structure combination, and are unsatisfactory for larger missing regions.
The issues in image inpainting methods were addressed by the image sparse
representation. Here the illustration of the image is performed by the sparse union
of an over complete dictionary using DCT, wavelet, Contourlet, wavelet and so on
[9–13] and the approximation of the missing pixels are corrected adaptively by the
sparse representation. As an illustration, the damaged picture is separated into the
texture layers and cartoon layers, that are correspondingly sparsely represented by
Curvelet transforms and the Discrete Cosine Transform [9].
The problems in image inpainting can be addressed by the sparse representation
of the image. Guleryuz [14] and Fadili et al. [15] characterized an image by its
sparse combination using DCT, wavelets, Contourlet, Curvelet in the overcomplete
dictionary approach. Pixels that are missing are calculated dynamically by restoring
these sparse representation periodically. The target region or a damaged region in
an image is segregated into textural regions and structural regions that are character-
ized sparsely into Curvelet Transform and Discrete Cosine Transform respectively
[13]. Fadili et al. [16] presented an approach of expectation maximization (EM)
for image inpainting in Bayesian framework for restoring the coefficients by the
indorsing of prior penalty sparsity. All the above methods have limited generalization
of algorithms, that are improved by the adaptive dictionary learning approach.
The proposed approach concentrates on the image inpainting technique based on
matching performed by the patch sparsity. Here, multi-direction feature (MDF) using
patch sparsity for inpainting the image is proposed for preserving the clarity of the
textures and coherence of the structures. Patch sparse representation and structure
Multidirection Features Based Image Inpainting with Color … 839
sparsity are used to represent the patch sparsity in the images. Based on the concept
of algorithms used in the image inpainting algorithms, super-wavelet transform is
implemented for extracting the multi-direction feature of an image.
It will be blended with the color details to form a weighted color-direction distance
(WCDD) to estimate the homogeneity of patches. The value of WCDD calculated
between the known and target patches also focuses on preserving the structure coher-
ence through the color-direction structure sparsity (CDSS) function. For yielding
subspace for patch sparse representation, numerous reference patches are checked
using WCDD.
Additionally, an enhancement calculation with local patch uniformity limitations
in both the color and the multi direction spaces is created to attain sparse linear
combination factors of reference patches. The ultimate aim of this technique is to
preserve the consistency of the inpainted regions with respect to the neighboring
patches and to enhance the clarity of the structures and the textures of the images.
The key contributions of this article are: (i) to present the multi-direction feature
to the image inpainting algorithm; (ii) to implement weighted color-direction interval
to explore appropriate reference patches; (iii) to claim CDSS to get a strong order
of filling; and (iv) to integrate the multi-direction and color limitations into the
optimization measure for getting superior inpainting outcomes. This paper is ordered
in the following order. Section 2 explains the method of extracting multi-direction
characteristics of an image and the suggested exemplar based inpainting technique.
Experimental results are shown in Sect. 3 and possible future work is proposed in
the conclusions.
2 Proposed Algorithm
Image inpainting is the process of filling the target regions based on the details of
neighboring regions. Image inpainting is mainly performed to preserve the structural
consistency and surrounding consistency. If the details of the source regions known
is higher, then the outcomes of the inpainting process will be better. For corrupted
images, gradient features or divergence characteristics were used by the existing
methods of image inpainting. Image consists of varied direction characteristics, so
the gradient function that defines only 2 directions will not be able to completely
define the details of an image.
The multidirectional and multiscale characteristics can be estimated using super-
wavelet transform. Hence in the proposed algorithm, for obtaining the multidi-
rectional attributes of an image, the super-wavelet transform is used. Among the
different types of super-wavelet transforms, Curvelet transform [17], is selected
840 B. Vidhya and M. N. Madhav
in the proposed system as it has proven results of extracting valid details accu-
rately than other transforms like Contourlet transform [18] (CT), Non-subsampled
Contourlet transform [19] (NSCT). The method of extracting eight direction matrices
is explained with an illustration as given below.
Images are considered in YUV color space compared with RGB color space for
coding efficiency. Curvelet forward transform is taken for the Y component of the
image (IY) for obtaining the multi direction and multi scale decomposition. The
coefficient matrix set of an image IY, is given by the Curvelet forward transform
(T + ) of the Y component of the image.
Q = T+ (IY) (1)
where Q = {Qs, d } gives the coefficient matrix set of image IY, and s and d in the
matrix indicates the scale and the direction of the image respectively. If the multi
scale curvelet coefficient is illustrated, it will resemble the structure as shown in the
Fig. 1, where the first scale square matrix is present at the centre of the figure.
The subsequent squares observed from inside to outside of the structure represents
the second to fifth scale coefficient matrices in the increasing order. Varied number of
Curvelet coefficient matrix can be observed for all scales of coefficient matrix except
for first scale. This is because of the reason that first scale coefficient matrix represents
the low frequency characteristics whilst other scale coefficients are high frequency
components. The second to fifth scale coefficient matrices comprises of the direction
matrices of the size sixteen, thirty two, thirty two and sixty four respectively. As a
starting point to figure out the coefficient matrix of every scale layer, 45° is considered
as a reference point and is measured in anti-clockwise direction. Figure 2 shows
the representations of fourth scale layer of the image consisting of 32 coefficients
matrices.
{Qs, d} values as {Q4, 1}, {Q4, 2}, {Q4, 3},… denotes the distinct coefficient
matrices for varied directional characteristics of an image. These coefficient matrices
are used to calculate the Multi Direction Features of an image. The second to fifth
scale layers of the Curvelet coefficient matrices are divided into eight sets based on
the direction of angles given by 157.5°, 135°, 112.5°, 90°, 67.5°, 45°, 22.5° and 0°
in anti- clockwise direction that are represented by Z 1 , Z 2 ,..., and Z 8 respectively.
By using the following rule, nth directional multi-direction matrix An is given by
An = T− (Cn(Q)) (2)
{Qsd}, Qs, d ∈ Z n
Cn(Q) = (3)
0, otherwise
This equation provides eight directional matrices An for eight directions. Image
size IY.
and nth directional multi-direction matrix An are of same size and every point An
(i, j) represents the direction intensity corresponding to the image pixel IY (i, j) that
is observed in the nth direction. If the value of An (i, j) is, the higher, then the value
of the pixel IY (i, j) will have richer feature details in the nth direction.
According to the direction partition rule based on the Fig. 1, the direction for every
coefficients set is assigned with the higher limit of the angle interval. Example, if
the coefficients set is given by (0°–22.5°), then the direction is assigned to be 22.5°.
Similarly for (22.5°–45°) and (157.5°–180°) coefficient set direction is 45° and 0°
respectively. Eight directional matrices are obtained by applying the above rule as
842 B. Vidhya and M. N. Madhav
illustrated in the Fig. 3b. The multi-direction matrices An of the first row of Fig. 2b
illustrates the images obtained through the direction features at 0°, 22.5°, 45° and
67.5° from left to right and second row of illustrates the images obtained through
the direction features at, the image at 90°, 112.5°, 135° and 157.5° from left to right
respectively. For understanding the direction partition rule, consider the first image
A1 , where the angle of the upper white stripe is around 175° but the angle of the
lower white stripe is nearly 160°. This image is assigned with direction 0° according
to the partition rule and only distinguished direction features are obtained.
The direction features that are extracted from the previous step are applied to
the image inpainting technique. Initially the multi directional features are integrated
with the color details to obtain the WCDD for estimating the difference between
neighboring patches.
Then the color direction structure sparsity is obtained through the patch selection
procedure for calculating the filling order. Using WCDD procedure, the best candi-
date patches are selected and multi directional features are combined with color
consistency constraints for the representation of patch sparsity in the optimization
equation.
Multidirection Features Based Image Inpainting with Color … 843
Accurate process of image inpainting can be obtained only if the distance between
the two patches is captured correctly. Distance measurement is a very important step
because it is used to select the best matching patch as well as to find the filling order
through the estimation of CDSS value. These two steps are necessary to perform
image inpainting through patch matching process.
To estimate the difference of two patches, the multi color features are combined
with the color details. Consider a corrupted image I that has a missing region . If
p and q are the two patches that are centred at pixels p and q respectively. Using
super wavelet transform, An can be estimated for sparse representation coefficients
and the preliminary image Ip that is repaired can be obtained by the considering the
size of the neighborhood N(p) as 25 × 25. A local search criterion is used to find the
multiple candidate patches. The size of the patch p that is centred at pixels p and
the size of the multi directional patch Gp that is centred at pixels p in An are same.
Similarly the size of the multi directional patch Gq and q are same.
To estimate the difference between the patches p and q, a weight color direc-
tion distance (WCDD) is performed which is denoted by dw that is given by the
following expression.
where η represents the weight coefficient that balances the details of the color
and direction features. The function d (p, q) represents the mean square distance
between p and q. To estimate WCDD, only the known values of the patches p,
q and Gp, Gq are considered in the Eq. (4) which makes the estimation to be precise
for measuring the difference of color and direction features and WCDD better.
To illustrate the performance of the proposed WCDD algorithm in image
inpainting, the test results were compared with SSD technique as shown in Fig. 4c,
d. In this demonstration, all the same procedures were followed except the applica-
tion of SSD and proposed WCDD algorithm for image inpainting to fill the missing
region by the best matching patches. Figure 4c shows some structures in the inpainted
image and WCDD method give better results of image inpainting. Similarly, these
tests were conducted over wide number of images for analysing the performance of
the proposed system with SSD method. Table 1 gives the PSNR value of the five
different test images for the SSD and WCDD method. From the results, it is proven
that the proposed WCDD algorithm has higher PSNR value and reproduces the image
that is more similar to the original image than SSD method. From the results of Table
1, it is inferred that the proposed WCDD method provides better PSNR value and
used for filling the filling order through CDSS function.
844 B. Vidhya and M. N. Madhav
Fig. 4 Analysis of Image inpainting techniques, a Original image, b corrupted image, c SSD, d
WCDD
The structure patch must be filled first and texture of the patches must be preserved
whist filling the missing patches from the neighboring pixels. To accomplish this,
a sturdy filling order must be calculated which is usually defined by the product of
color direction structure sparsity S(p) and the confidence term C(p). This product is
termed as the priority function V(p).
This procedure is performed to make sure that priority of filling the patch is given
to the structure patch compared to other patches by S(p) parameter. S(p) parameter
is used to preserve the texture and smooth patches in the patch filling.
A. Criminisi et.al [16, 17] defined the confidence term C(p) as a degree of reli-
able details neighboring the pixel in an image. Z. Xu and J. Sun [20] measured
the confidence of a patch present at the structure by introducing the structure spar-
sity. This produced a better filling order as that of the method explained in [16, 17,
21, 22]. But few constraints were identified like patches in complex structural and
Multidirection Features Based Image Inpainting with Color … 845
textural regions or the missing regions with less or absence of nearby patch details
or consistent confidence value is not provided by the structure sparsity.
To overcome the above limitations, a color direction structure sparsity (CDSS)
based on weighted color-direction distance (WCDD) is used to isolate the structures
and the textures of an image. A weighted color-direction similarity function built on
WCDD is given by the following expression
1 dw
X p,k = exp − 2 (6)
Z ( p) σ
where θ is a correcting factor that can take up the value between 0 and 1. In the
proposed system, value of θ is assigned as 0.21 after evaluating using different values.
θ is mainly used to rise the minimum value of S(p) to fill the target patches exactly as
texture patch propagates poorly. CDSS uses color and MDS details to estimate the
similarity of the patches like the structure sparsity techniques and also more accurate
results are obtained.
3 Experimental Results
Fig. 5 Comparative results of image inpainting by wavelet transforms with directions. a Four
direction feature extracted by Curvelet Transform and. b Contourlet transform –CT. c Eight direction
feature extracted by Curvelet Transform and. d CT
which has the traces of objects in the restored images. But the proposed method,
produced an image similar to the original image with higher PSNR values.
Multidirection Features Based Image Inpainting with Color … 847
Fig. 6 Analysis of results of image inpainting techniques for object removal. a Original image. b
corrupted image. c Exempler method [23]. d Proposed method
4 Conclusion
An image inpainting algorithm is proposed in this article to preserve the structural and
textural coherence characteristics along with the neighboring consistency through
process of adopting color and multi directional features in the process of image
inpainting. The proposed method can be used for various applications like object
removal, text and scratch removal. First, the multidirectional features are introduced
in the process of image inpainting and WCDD is estimated to measure the distance
848 B. Vidhya and M. N. Madhav
and similarity of the patches. Then to find the filling order of the patches, CDSS
which is a function of structure sparsity is used. To achieve sharp inpainting results,
limitations of the color and multi-direction features are combined in the optimization
measure. It has been proved that the proposed method provides better structural
coherence over the other existing inpainting techniques.
References
1. Bertalmio M, Sapiro G, Caselles V, Ballester C (2000) Image inpainting. In: Proceeding 27th
Annual Conference ACM SIGGRAPH, pp 417–424
2. Bertalmio M, Bertozzi AL, Sapiro G (Dec 2001) Navier–Stokes, fluid dynamics, and image
and video inpainting. In: Proceedings IEEE Computer Society Conference on Computer Vision
and Pattern Recognition, pp I-355–I-362
3. Chan TF, Shen J (2001) Mathematical models for local nontexture inpaintings. SIAM J Appl
Math 62(3):1019–1043
4. Chan TF, Shen J (2001) Nontexture inpainting by curvature-driven diffusions. J Vis Commun
Image Represent 12(4):436–449
5. Zhang HY, Wu B, Peng QC, Wu YD (2007) Digital image inpainting based on p-harmonic
energy minimization. Chin J Electron 3(3):525–530
6. Telea A (2004) An image inpainting technique based on the fast marching method. J. Graph.
Tools 9(1):23–34
7. Weickert J, Welk M (2006) Tensor field interpolation with PDEs. Visualization and processing
of tensor fields. Springer, Berlin, Germany, pp 315–325
8. Takeda H, Farsiu S, Milanfar P (2007) Kernel regression for image processing and reconstruc-
tion. IEEE Trans Image Process 16(2):349–366
9. Miyoun J, Bresson X, Chan TF, Vese LA (2011) Nonlocal Mumford-Shah regularizers for
color image restoration. IEEE Trans Image Process 20(6):1583–1598
10. Wen Y-W, Chan RH, Yip AM (2012) A primal–dual method for total-variation-based wavelet
domain inpainting. IEEE Trans Image Process 21(1):106–114
11. Mainberger M et al (2012) Optimising spatial and tonal data for homogeneous diffusion
inpainting. Scale space and variational methods in computer vision. Springer, Berlin, Germany,
pp 26–37
12. Yu G, Sapiro G, Mallat S (2012) Solving inverse problems with piecewise linear estimators:
from Gaussian mixture models to structured sparsity. IEEE Trans Image Process 21(5):2481–
2499
13. Elad M, Starck J-L, Querre P, Donoho DL (2005) Simultaneous cartoon and texture image
inpainting using morphological component analysis (MCA). Appl Comput Harmonic Anal
19(3):340–358. 9
14. Guleryuz OG (2006) Nonlinear approximation based image recovery using adaptive sparse
reconstructions and iterated denoising—Part I: theory. IEEE Trans Image Process 15(3):539–
554, 11 Mar 2006
15. Fadili MJ, Starck JL (Sep 2005) EM algorithm for sparse representation- based image
inpainting. In: Proceedings IEEE International Conference Image Processing, pp II-61–II-64.
10
16. Criminisi A, Perez P, Toyama K (Jun 2003) “Object removal by exemplar- based inpainting. In:
Proceedings IEEE Computer Society Conference on Computer Vision and Pattern Recognition,
pp II-721–II-728
17. Criminisi A, Perez P, Toyama K (2004) Region filling and object removal by exemplar-based
image inpainting. IEEE Trans Image Process 13(9):1200–1212
Multidirection Features Based Image Inpainting with Color … 849
18. Wu J-Y, Ruan Q-Q (2009) A novel exemplar-based image completion model. J Inf Sci Eng
25(2):481–497
19. Do MN, Vetterli M (2005) The contourlet transform: An efficient directional multiresolution
image representation. IEEE Trans Image Process 14(12):2091–2106
20. Xu Z, Sun J (May 2010) Image inpainting by patch propagation using patch sparsity. IEEE
Trans Image Process 19(5):1153–1165, CT
21. Guleryuz OG (2006) “Nonlinear approximation based image recovery using adaptive sparse
reconstructions and iterated denoising—Part II: adaptive algorithms. IEEE Trans Image Process
15(3):555–571, 12 Mar 2006
22. Wong A, Orchard J (Oct 2008) A nonlocal-means approach to exemplar-based inpainting. In
Proceeding 15th IEEE International Conference Image Processing, pp 2600–2603
23. Zhang Q, Lin J (2012) Exemplar-based image inpainting using color distribution analysis. J
Inf Sci Eng 28(4):641–654
Classification of Sub-frequency Bands
Based Two-Class Motor Imagery Using
CNN
Abstract EEG has been primarily used in both clinical and research applications.
Brain-computer system (BCI) is one of the leading EEG research applications that
offer special users a new means of communication. Previous studies have reported
the occurrence of MI patterns in mu and beta rhythms, but that does not provide
in-depth knowledge of the frequency range. This paper focuses on the classification
of 2-class Motor Imagery using several frequency sub-bands in the mu and beta
range. “EEG motor imagery dataset from the Physionet database,” has been used for
validation purposes. Although this data includes both imagery and real movements,
we have just used the imagination data. Data is collected from 109 healthy subjects,
but we have only used the first 15 subjects in the study. The study aims to divide
the data into multiple frequency bands to study the motor imagery classification
behaviour over different frequencies. Afterward, a CNN-based deep learning model
with two convolutional layers has been used to classify the left and right classes for
different types of same data. The study seeks to compare the results from various
sub-frequency bands.
1 Introduction
The number of people with disabilities who have lost their mobility is substantial.
In addition to their damaged body, they have a working sound brain in their bodies.
The brain should be used to allow them the freedom to move easily by using their
minds to power the wheelchair. The brain-computer interface offers a forum for
people to create a communication relationship between users and computers [1]. It
helps us to interact with the outside world using our brains independently. The BCI
framework does not require any actual muscle movement commonly involved in the
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 851
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_80
852 M. A. Awais et al.
2 Methodology
The research will develop a brain-computer interface system which aims to introduce
the use of several sub-bands of motor imagery-based EEG data. CNN with frequency
sub-bands seeks to achieve better performance, reliability and flexibility.
2.1 Dataset
EEG MI dataset from the Physionet database has been used in the proposed system
[9]. In producing these records, the BCI2000 instrumentation system has been used.
The data collection comprises above 1500 recordings of one and two minutes from
109 subjects. The dataset incorporates both actual and motor imagination tasks
from the subjects. The experimental paradigm comprised 14 runs, where subjects
performed different tasks including rest state, the right fist/left fist movement, the
imagination of right fist/left fist movement, both fists/ both feet movement, and the
imagination of both fists/ both feet movement. The EEG signals were recorded from
64 electrodes as per the international 10–10 system. The data was sampled at 160 Hz
for all subjects. In this study, only the MI trials (imagination of left and right fist
movement) of the first 15 subjects were used. For motor imagery, subjects were
instructed to perform the imagination task as the target stimulus appears on the left
side of the screen or the right side of the screen. The subjects imagine the fist to
be opened and closed before the target stimulus disappears. Then the subject goes
to the rest state. Every subject recorded three sessions for each type of task. The
single-session comprises 7–8 random trials of each class, i.e., left or right movement
imagery. Each trial is carried out for 4 secs, followed by the rest period of 4 s ± 5%.
2.2 Pre-processing
It is essential to remove the noise or unwanted components from the signal. The
undesired signals in our case can be EMG signals (real hand movement) and the
frequency components of the EEG signal other than the motor-imagery task.
The 15 subject’s data has been filtered using BrainStorm in MATLAB. Initially,
the data has been passed through 3 steps including, DC offset correction, notch filter
at 60 Hz in order to eliminate the electrical interference, and the bandpass filter from
8 to 32 Hz, to eliminate other frequencies like EMG (50–250 Hz) and other unwanted
signals.
854 M. A. Awais et al.
The pre-processed data from the previous stage has been divided into nine sub-
frequency bands using the same platform, i.e., Brainstorm, in order to study the
data in detail. The sub frequency bands include (8–12 Hz), (12–16 Hz), (16–20 Hz),
(20–24 Hz), (24–28 Hz) and (28–32 Hz), (8–16 Hz), (16–32 Hz) and (8–32 Hz).
In this study, CNN based deep learning approach has been adopted to identify the
effect of MI pattern on different frequency ranges.
CNN is a Deep Learning algorithm that can take in any input and process it
efficiently to classify the different classes. CNN uses a system that has been designed
for reduced processing requirements. The layers of a CNN consist of an input layer,
an output layer, and a hidden layer containing multiple convolutional layers, pooling
layers, fully connected layers and normalization layers.
The overall 15 subject’s data is divided into training (10 subjects) and validation
(5 subjects).
The CNN model (Fig. 1) proposed in this research comprises 2 convolutional
layers with 5 kernels per layer. 1st layer performs convolution and the output data is
passed through batch normalization. ReLU activation function is used in the model
and the output is reduced in the max-pooling layer. The same process is applied
to the 2nd layer as well. The output of the 2nd layer is passed through the dropout
layer in which we have used the default value, i.e., 50%. The data is fed to the fully
connected layer after flattening it into 1-D. The output is generated by a softmax
layer with many neurons that fit the number of categories in the data to be classified.
In our case, the output is classified as the left and right class.
The CNN model is composed of 2 convolutional layers and the model is trained
for 10 epochs. The initial learning rate is set to 0.001, while the validation frequency
is 100.
The stated CNN model has been used for two different cases. In case 1, the model
has been implemented on all the 64 channels using 9 sub-frequency bands stated in
Sect. 80.3. Whereas in Case 2, all 9 sub frequency bands have been implemented
using only 3 channels (i.e., C3, Cz and C4) from the central region of the brain.
The CNN model has been implemented on the basis of two cases. Table 1 illustrates
the results derived from both cases (i.e., use of 64 channels data and the use of only
3 EEG channels). The accuracy of the model is calculated over 10 epochs.
Classification of Sub-frequency Bands Based Two-Class Motor … 855
2D Convoluon
Batch Normalizaon
Acvaon Funcon: ReLU
2D Max Pooling
2D Convoluon
Batch Normalizaon
Acvaon Funcon: ReLU
2D Max Pooling
Dropout
Fully Connected
Somax
In case 1, the maximum accuracy, i.e., 77.93%, was reported by the 8–12 Hz
frequency band. In contrast, the frequency band 16–32 Hz resulted in the lowest
validation accuracy, i.e., 61.11%. Whereas in case 2, the maximum accuracy, i.e.,
77.82% has been reported by 8–12 Hz. On the other hand, the minimum validation
accuracy for case 2 has been obtained by the sub-band 16–20 Hz. Figure 2 provides
a graphical representation of the classification comparison of cases 1 and 2.
Among 9 different frequency bands, the 8–16 Hz main band and its two sub-
bands, including 8–12 Hz and 12–16 Hz, has shown better results for 2-class MI
prediction. Figure 2 also demonstrates that frequency bands ranging from 16–32 Hz,
including sub-bands and the main band, has given the lowest classification accuracies.
Thus, it shows that there is no effect of dividing the frequency if the main band is not
giving a better performance. One of the major limitations of getting low classification
accuracies might be the use of a small dataset. Deep learning algorithms are enhancing
efficiency by reducing complexity. But still, it lacks the ability to handle the small
datasets with the same level of efficiency being addressed in the case of large EEG
datasets.
4 Conclusion
The objective of this research is the analysis of sub-frequency bands for the devel-
opment of the BCI system using 2-class motor imagery. CNN has been used for the
classification of two classes using several sub-frequency bands for 15 subjects’ data.
The maximum classification accuracy does not exceed 77.93% as the data is consid-
ered to be a small dataset in order to be processed in a deep learning algorithm. In
Classification of Sub-frequency Bands Based Two-Class Motor … 857
the future, we will be using a sufficient dataset for deep learning algorithms in order
to achieve the maximum classification results effectively.
Acknowledgements This research is supported by the Ministry of Education Malaysia under the
Higher Institutional Centre of Excellence (HICoE) Scheme awarded to the Centre for Intelligent
Signal and Imaging Research (CISIR).
References
1. Abiri R, Borhani S, Sellers EW, Jiang Y, Zhao X (2019) A comprehensive review of EEG-based
brain-computer interface paradigms. J Neural Eng 16(1):011001
2. Biasiucci A, Franceschiello B, Murray MM (2019) Electroencephalography. Curr Biol
29(3):R80–R85
3. Awais MA, Yusoff MZ, Yahya N, Qamar MU(2020) Brain controlled wheelchair: a smart
prototype. In: J Phys Conf Series 1529:042075
4. Andrade MK et al.,(2020) An EEG brain-computer interface to classify motor imagery signals.
In: Biomedical signal processing. Springer, pp 83–98
5. Aoh Y et al (2019) Event-related desynchronization/synchronization in Spinocerebellar Ataxia
Type 3. Front Neurol 10:822
6. Padfield N, Zabalza J, Zhao H, Masero V, Ren J (2019) EEG-based brain-computer interfaces
using motor-imagery: techniques and challenges. Sensors 19(6)
7. Tacchino G, Coelli S, Reali P, Galli M, Bianchi AM (2020) Bicoherence interpretation, in
EEG, requires signal to noise ratio quantification: an application to sensorimotor rhythms. IEEE
Transactions on Biomedical Engineering
8. Tariq M, Trivailo PM, Simic M (2020) Mu-Beta event-related (de) synchronization and EEG
classification of left-right foot dorsiflexion kinaesthetic motor imagery for BCI. Plos one
15(3):e0230184
9. EEG Motor Movement/Imagery Dataset https:/www.physionet.org/physiobank/database/eeg
mmidb/. Last Accessed 25 Oct 2020
Prediction of Methane Hydrate
Formation Rate in Multiphase System
using Artificial Neural Network
Wan Adli Nuaim Bin Wan Ahmad Nadzri, Omar Nashed, Bhajan Lal,
Khor Siak Foo, and Khalik Mohd Sabil
Abstract The research of the hydrate formation has advanced over the last few
decades. Several models have been developed to understand the hydrates formation
kinetics and conditions. In this study, computer based model is used to predict the
gas hydrate formation rate. Since the hydrae formation is stochastic phenomenon
and it is common to get inconsistent data, Artificial Neural Network (ANN) model
has potential to outstand other conventional kinetic models. ANN used to predict the
methane hydrate formation rate in multiphase system. The liquid phase composed
of water + drilling oil + nonionic surfactants was used to form methane hydrates
at pressure 8.80 MPa and temperature of 274.15–277.15 K. This research would
essentially assess the effectiveness of ANN model for the kinetic modeling of the
formation of gas hydrate from the acquired regression analysis. The result of this
research revealed that ANN model with 16 number of hidden neurons had a better
prediction as the highest regression value R was found to be 0.9956.
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 859
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_81
860 W. A. N. B. W. A. Nadzri et al.
1 Introduction
Gas hydrates are an ice-like inclusion compound, which mainly consist of low
molecular weight gas molecules entrapped into hydrogen bonded water cages [1,
2]. Gas hydrate formation in gas production may hinder the flow and eventually lead
to serious operational problem and huge financial loss. Despite imposing negative
impacts in the oil and gas industries, gas hydrate has some potential application in
some industries [3, 4]. Gas hydrate-based technology could be applied for water
desalination, gas separation, transportation, and storage [5]. Not to forget that, gas
hydrates are available naturally at deep marine and permafrost area [6, 7].
Several modeling studies available are focusing mainly on the thermodynamics
properties of gas hydrate, e.g. phase equilibria and molecular structures [8]. However,
less papers are available in the literature discussing and predicting the time-dependent
behaviour of gas hydrate formation and dissociation [9]. This is due to the naturally
occurring intrinsic stochastic phenomena of gas hydrate which lead to difficulty in
developing a well-established model [2, 10, 11]. Some of the hydrate growth kinetic
models available in the literature include reaction kinetics based model [12], mass
transfer based model [13, 14], combination of reaction kinetics and mass transfer [15,
16], heat transfer based model [17] as well as model based on combination of heat
transfer and reaction kinetic [18]. Gas hydrate formation kinetics are very complex
process, and it is almost difficult to display a definite outcome from the kinetic model
as hydrates form are very uncertain and formation is based on various driving forces
[19]. In addition, the gas hydrate growth rate depends on the phases composition as
well as testing conditions. Recently, artificial neural network (ANN) has attracted a
massive attention due to its ability to deal with scattered and big data to solve complex
problems. ANN is a technique used nowadays for prediction of data set which use
the structure and function of neuron [9, 20–22]. In gas hydrate studies, most of the
papers utilizing ANN focusing on the prediction of the phase equilibrium. However,
less research conducted on the kinetic of the hydrate formation.
Zahedi et al. used two different methods with 203 experimental data points from
literature to predict the hydrate formation temperature. It was found that the best
predictor of hydrate formation temperature was MLP with seven neurons [21].
Moreover, Chapoy et al. estimated hydrate dissociation pressure in the presence of
inhibitor. In their studies, 19 input variables with 35 neurons in one hidden layer were
trained, but various numbers of hidden layers and neurons have not been examined
to find the best structure in the neural network [23].
In this work, ANN was employed to predict methane hydrates formation rates in
multiphase system in the presence of nonionic surfactants. The HLB values were used
to identify the surfactant and used as input along with temperature and surfactants’
concentration.
Prediction of Methane Hydrate Formation Rate in Multiphase … 861
2 Methodology
The data used in this work were collected from experimental study of the gas hydrate
formation kinetics in multiphase system (water + oil + gas) by using 6 different
chemicals: Span 20, Span 40, Span 80, Tween 20, Tween 40, and Tween 80. The
experimental data is presented in Table 1. The network was trained to minimize the
errors between the desired target values and the values computed from the model
based on iterative gradient using MATLAB. The inputs are the types of non-ionic
surfactant represented by Hydrophilic-Lipophilic Balance (HLB), temperature, and
concentration. Meanwhile, the output is the rate of CH4 hydrate formation. The data
set were split into two subset which were for training as well as for validation/testing.
By default, the percentage of each set were 70% training, 15% testing and 15%
validation. For the training algorithm, it was selected according to the quantity of
the data set so that the network could simulate a task with minor error. Usually, the
Levenberg–Marquardt algorithm is the most used algorithm to automatically adjusted
the biases and weights of the neural network [9].
However, for this model, Bayesian Regularization algorithm was selected because
this algorithm could result in good generalization for difficult, small, and noisy
datasets which suits our input data sets. Next, the number of hidden layers were
manipulated to get the least error. The number of hidden layers were determined
through a process of trial an error [22]. In this paper, the number of hidden layers
were decided through the optimization of regression. A correlation coefficient (R) was
chosen to evaluate the prediction as 1 means a perfect correlation while 0 indicates
for random correlation.
Dividing the data to training, validation, and testing data set was the first step for
applying ANN prediction. This ANN model was built from 19 number of data of
six chemicals, each taken from different temperature, HLB value and concentration.
In ANN model, the number of hidden neurons were manipulated starting from 1 to
20 hidden neurons. Trial-and-error results showed that optimal hidden neurons 16.
Figure 1 illustrates the ANN architecture with respect to performance function.
Regression analysis was carried out to determine the correlation between the
experimental and predicted data by the ANN models. The highest R is found to be
0.9956 for the ANN model with 16 number of neurons. It is evident there is excel-
lent agreement between the predicted and experimental data for the same training
conditions as shown in Fig. 2. However, ANN model used in this work could have
insufficient number of data with only 3 inputs. Additional data might be required to
get more practical ANN model to predict the gas hydrate formation rate. Addition-
ally, another limitation of ANN model prediction is that the model valid only within
the studied range of HLB, concentration and temperature for the prediction of gas
hydrate formation rate. Consequently, ANN model is not applicable for extrapolation.
Lastly, it is worth mentioning that using too many hidden neurons increases the
complexity of the model which lead to over-fitting. In other hands, different problem
occurs when the network is assumed to has very few hidden neurons causing the
network unable to fit or in other words, under-fitting. The number of hidden neurons
in the hidden layers determines to a network’s capability to compute the output
results.
Fig. 2 Correlation coefficient between experimental and predicted data by ANN with 16 number
of neurons
4 Conclusion
ANN was used to perform the numerical study of kinetic behavior of methane gas
hydrate formation rate in the presence of nonionic surfactants. It was found that
the predicted results by ANN with 16 number hidden neurons is the closest to the
experimental data. The correlation coefficient value for 16 hidden neurons is 0.99556.
Therefore, ANN could be an efficient method for predicting the hydrate formation
rate.
References
1. Bavoh CB, Lal B, Keong LK (2020) Introduction to gas hydrates, in chemical additives for
gas hydrates. In: Lal B, Nashed O (eds). Cham, Springer International Publishing, pp 1–25.
https://doi.org/10.1007/978-3-030-30750-9_1
2. Yin Z, Khurana M, Tan HK, Linga P (2018) A review of gas hydrate growth kinetic models.
Chem Eng J 342:9–29. https://doi.org/10.1016/j.cej.2018.01.120
3. Yin Z, Chong ZR, Tan HK, Linga P (2016) Review of gas hydrate dissociation kinetic models
for energy recovery. J Nat Gas Sci Eng 35:1362–1387. https://doi.org/10.1016/j.jngse.2016.
04.050
864 W. A. N. B. W. A. Nadzri et al.
4. Yaqub S, Lal B, Mellon NB, Sufian BS (2018) Effect of the natural green materials on methane
hydrate formation kinetics. IOP Conf Ser: Mater Sci Eng, 458:012074. https://doi.org/10.1088/
1757-899x/458/1/012074
5. Nashed O, Lal B, Shariff AM, Sabil KM (2020) Gas hydrate promoters. In: Lal B, Nashed O
(eds), Chemical additives for gas hydrates. Cham, Springer International Publishing, pp 47–65.
https://doi.org/10.1007/978-3-030-30750-9_3
6. Khan MS, Lal B, Bustam MA (2020) Gas hydrate inhibitors. In: Lal B, Nashed O (eds),
Chemical additives for gas hydrates. Cham, Springer International Publishing, pp 27–46. https://
doi.org/10.1007/978-3-030-30750-9_2
7. Yaqub S, Lal B, Shariff AM, Mellon NB (2019) Unraveling the effect of sub-cooling temper-
atures on the kinetic performance of biopolymers for methane hydrate. J Nat Gas Sci Eng
65:68–81. https://doi.org/10.1016/j.jngse.2019.03.002
8. Manakov AY, Penkov NV, Rodionova TV, Nesterov AN, Fesenko EE Jr (2017) Kinetics of
formation and dissociation of gas hydrates. Russ Chem Rev 86:845–869. https://doi.org/10.
1070/rcr4720
9. Foroozesh J, Khosravani A, Mohsenzadeh A, Mesbahi AH (2014) Application of artificial
intelligence (AI) in kinetic modeling of methane gas hydrate formation. J Taiwan Inst Chem
Eng 45:2258–2264. https://doi.org/10.1016/j.jtice.2014.08.001
10. Naeiji P, Varaminian F, Rahmati M (2019) The kinetic modeling of methane hydrate growth
by using molecular dynamic simulations. Int J Heat Mass Transf 142:118356. https://doi.org/
10.1016/j.ijheatmasstransfer.2019.07.006
11. Sayani JK, Pedapati SR, Lal B (2020) Phase behavior study on gas hydrates formation in gas
dominant multiphase pipelines with crude oil and high CO2 mixed gas. Sci Rep 10:1–2. https://
doi.org/10.1038/s41598-020-71509-6
12. Vysniauskas A, Bishnoi PR (1985) Kinetics of ethane hydrate formation. Chem Eng Sci
40:299–303. https://doi.org/10.1016/0009-2509(85)80070-1
13. Englezos P, Kalogerakis N, Dholabhai PD, Bishnoi PR (1987) Kinetics of formation of methane
and ethane gas hydrates, vol 42, pp 2647–2658. https://doi.org/10.1016/0009-2509(87)870
15-X
14. Skovborg P, Rasmussen P (1994) A mass transport limited model for the growth of methane and
ethane gas hydrates. Chem Eng Sci 49:1131–1143. https://doi.org/10.1016/0009-2509(94)850
85-2
15. Hashemi S, Macchi A, Servio P (2007) Gas hydrate growth model in a semibatch stirred tank
reactor. Ind Eng Chem Res 46:5907–5912. https://doi.org/10.1021/ie061048+
16. Henning RW, Schultz AJ, Thieu V, Halpern Y (2000) Neutron diffraction studies of CO2
clathrate hydrate: formation from deuterated ice. J Phys Chem A, Article 104:5066–5071.
https://doi.org/10.1021/jp0001642
17. Freer EM, Selim MS, Sloan ED Jr (2001) Methane hydrate film growth kinetics. Fluid Phase
Equilib 185:65–75. https://doi.org/10.1016/S0378-3812(01)00457-5
18. Bollavaram P, Devarakonda S, Selim MS, Sloan ED Jr (2000) Growth kinetics of single crystal
sII hydrates. Elimination of mass and heat transfer effects, Annals of the New York Academy
of Sciences 912:533–543. https://doi.org/10.1111/j.1749-6632.2000.tb06808.x
19. Partoon B, Sahith J, Lal B, and Maulud ASB (2020) Gas hydrate models. In: Lal B, Nashed O
(eds), Chemical additives for gas hydrates. Cham: Springer International Publishing, pp 67–85.
https://doi.org/10.1007/978-3-030-30750-9_4
20. Sahith J, Pedapati SR, Lal B (2019) Application of artificial neural networks on measurement
of gas hydrates in pipelines. Test Eng Manag 81:5769–5774
21. Zahedi G, Karami Z, Yaghoobi H (2009) Prediction of hydrate formation temperature by both
statistical models and artificial neural network approaches. Energy Convers Manage 50:2052–
2059. https://doi.org/10.1016/j.enconman.2009.04.005
Prediction of Methane Hydrate Formation Rate in Multiphase … 865
22. Ghavipour M, Ghavipour M, Chitsazan M, Najibi SH, Ghidary SS (2013) Experimental study of
natural gas hydrates and a novel use of neural network to predict hydrate formation conditions.
Chem Eng Res Des 91:264–273. https://doi.org/10.1016/j.cherd.2012.08.010
23. Chapoy A, Mohammadi AH, Richon D (2007) Predicting the hydrate stability zones of natural
gases using artificial neural networks. Oil Gas Sci Technol 62:701–706. https://doi.org/10.
2516/ogst:2007048
Chemometrics Analysis and Wavelength
Biomarker Identification Using Fourier
Transform Infrared Spectroscopy
for Lard Adulteration
Abstract Food authenticity is a major concern that is closely linked to public health
and safety in the period in which the long food supply chain is the norm. One of the
key problems facing customers in meat markets is the detection of non-halal meat.
While several studies have been carried out to identify biomarkers for adulterated
meats such as lard, still these studies are at an early stage, and there is no authenticated
biomarker available which can produce specific identification as a result. The purpose
of this research is to utilize the technique of Fourier Transform Infrared Spectroscopy
(FTIR) to analyze pig, cow, lamb, and chicken fat to identify potential biomarkers
for the identification of lard. FTIR studies have shown that chicken and lard fat have
particular peaks relative to lamb and beef fats at wavenumbers 1600, 1750, 2750 and
2920 cm–1 . Furthermore, PCA shows promising clustering and correlation between
the species which can lead to development of a model for successful adulteration
detection.
1 Introduction
In the food industry around the world, food authenticity is a significant concern. Food
fraud is still a problem because of the glut of packaged food with a long supply chain
on the market. First coined by Spink and Moyer which stated that food fraud is defined
as alteration of the true food labeling ingredients in which costly, fewer accessible raw
materials are substituted by cheaper alternatives [1]. Some of these adulterations may
impact the public mass financially only, but others may be more seriously affected
by some adulteration, i.e. food allergy [2], food poisoning, religious views [3], etc.
Although food tampering is not a recent issue, some of these tampering was quite
poisonous. For instance, incidences such as addition of sawdust to make white bread
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 867
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_82
868 M. A. Siddiqui et al.
[4], formula milk adulteration with melamine [5], adulteration of consumable oil with
non-consumable [6]. Not only dangerous to health, but some incidences also involved
misrepresentation of food ingredients, such as adulteration of meat with horse meat
in 2013 in the UK [7]. Because the price difference between pork and other red
meats, halal and kosher consumers are always concerned with the authentication of
halal and kosher foods [8]. Global halal market is worth billions, and it is projected
to increase to Trillions by 2021 [9]. Therefore, the industry is very interested to
strengthen consumer’s trust in the halal brand. Market confidence is necessary to
protect this [10]. Several authentication techniques can be employed to ensure halal
and kosher brand food products adhere to the standards, such as DNA based methods
[11]. Majority of approaches still require considerable time for preparing sample or
getting vague results due to extreme sensitivity to adulterants which results in overall
poor accuracy. Nowadays, spectroscopy based on vibrations is popularized due to
its effectiveness for adulteration detection in various food [12]. This is because the
simplicity of this nondestructive method reduces time and cost for overall experiment.
FTIR spectroscopy can separate, easily and with high accuracy. For instance, pork
meat in meatball broth [13], chocolate [14], vegetable oils [15], etc. Some research
starts to veer to Near Infrared (NIR) spectroscopy mainly because the feasibility
of this would open possibilities to make food authentication instrumentation setup
to be portable [16]. However, the research done on this reported lower accuracy in
discrimination of adulterants. Although FTIR is quick and relatively inexpensive
with easier sample preparation and non-destructive process [17], it is still limited in
portability due to its bulky equipment. The aim of this research paper is to improve
the accuracy of NIR spectrometer by means of optimizing post-processing analysis,
and to quantify the level of lard adulteration in mixtures.
In food safety, the production of food authentication is important since this is the
mechanism that will verify whether a certain food follows its label definition. [18].
This can include the origin of the food (species, gene, geographical), production
method (conventional, organic, free range, etc.), and processing method (frozen,
etc.). Vibrational spectroscopy is based on the idea that inside molecules, atom-to-
atom bonds vibrate at frequencies that can be represented and are therefore subject
to measurement by the laws of physics. In molecules, during the compression step,
the respective electron clouds of the two bound atoms, as well as the charges on the
nuclei, limit the approach of the nuclei, forming an energy barrier. The bond will
inevitably split at the extension of the stretch when the vibrational energy exceeds
the energy of dissociation. Table 1 shows the different stretching frequencies.
Chemometrics Analysis and Wavelength Biomarker Identification … 869
Table 1 Important IR stretching corresponding to their respective frequency and bond [19]
Chemical bond Spectrum location (Cm−1 ) Absorbance strength
C≡N 2260–2220 Medium
C≡C 2260–2100 Medium to weak
C=C 1680–1600 Medium
C=N 1650–1550 Medium
1500–1430 Strong to weak
1.2 Chemometrics
A = ε cd (1)
Samples were purchased from a local market at Teronoh, Malaysia. After that, the
meat was cut into small element (1 cm × 1 cm) and keep it at –2C°.
Lard was adulterated with fats of lamb, Beef and Chicken to obtain a series of standard
sets of ten pure and thirty samples containing 10–50% v/v of lard in Lamb, Beef,
Chicken samples. This method follows according to Rohman et al. [18]. The mixture
samples are mixed according to Table 2.
The data analysis was performed using MATLAB and Spectrograph. Spectrograph
1.1 has been used to extract information from spectrum results, where the data is pre-
processed as needed. To further examine the findings from pre-processing, MATLAB
R2017b was used. To analyze the consistency of lard adulteration, the Principle
Components Analysis (PCA) technique was used.
The FTIR Spectra of pure fats is shown in Fig. 1. This spectrum consists of four
regions: the first region from 4000 to 2500 cm–1 , the second region from 2500 to
2000 cm–1 , the third region from 2000 to 1500 cm–1 , and finally the fingerprint
region from 1500 to 800 cm–1 . For second region, there are no peaks showing no
functional group present in this range. Moreover, 1st and 3rd region have the potential
to differentiate between the sample species.
Figure 2 represents PCA results for all the samples, at 1-D PCA projection, some
of the samples such as Chicken and Pig are already starting to cluster together, even
though there is still overlap between samples of different species. PCA results at 2-D
projection shows that Chicken and Pig samples clustered together without overlap,
Lamb samples have one outlier, but the beef samples proved challenging since results
are still scattered. The 3-D PCA shows better results though with all pig samples
lined around third PCA Component.
Fig. 1 FTIR Spectrum of all the meat species with region identification
872 M. A. Siddiqui et al.
Fig. 2 1D, 2D and three-dimensional representation of chicken, beef, lamb, and lard samples using
principal components
4 Conclusion
Identification of lard discrimination from chicken, lamb, and cow fats was demon-
strated in this paper. FTIR studies have shown that by having a peak at a wavenumber
1600 cm–1 , while other species of fats do not have peak at this stage, lard differenti-
ates itself from other species of fats. In addition, shoulder peak was seen for lard in
the range of wavenumber 2950–2990 cm–1 , while cow and lamb fats showed sharp
peaks, and double peaks showed chicken. In comparison to the usual peaks seen for
the other fats, at wavenumber 2750 cm–1 a reversed peak was detected for lard. For
future work we can further develop this method by analyzing the adulterated samples
as well.
References
1. Spink J, Moyer DC (2011) Defining the public health threat of food fraud. J Food Sci 76:R157–
R163
2. Guardian T (2017) Allergic teenager’s death after eating kebab was accidental, rules coroner.
The Guardian.
3. Li DK (2017) Toddler allergic to dairy dies after pre-school serves him grilled cheese. In: New
York Post, ed
4. Barlass T (2015) Child aged 10 dies after drinking coconut drink as importer admits label
charges. In: The Sydney Morning Herald ed
5. FSA (2013, 7/2/2018) Timeline on horse meat issue. Available: http://webarchive.nation
alarchives.gov.uk/20150403184406/, http://www.food.gov.uk/enforcement/monitoring/horse-
meat/timelinehorsemeat
6. Rohman A, Che Man YB (2012) Analysis of pig derivatives for Halal Authentication Studies.
Food Rev Int 28:97–112
7. Tähkäpää S, Maijala R, Korkeala H, Nevas M (2015) Patterns of food frauds and adulterations
reported in the EU rapid alert system for food and feed and in Finland. Food Control 47:175–184
Chemometrics Analysis and Wavelength Biomarker Identification … 873
8. Abaitua Borda I, Philen RM, Posada de la Paz M, Gomez de la Camara A, Diez Ruiz-Navarro
M, Gimenez Ribota O et al (1998) Toxic oil syndrome mortality: the first 13 years. Int J
Epidemiol 27:1057–1063
9. Regenstein JM, Chaudry MM, Regenstein CE (2003) The kosher and halal food laws. Compr
Rev Food Sci Food Saf 2:111–127
10. Chuah L-O, He XB, Effarizah ME, Syahariza ZA, Shamila-Syuhada AK, Rusul G (2016)
Mislabelling of beef and poultry products sold in Malaysia. Food Control 62:57–164
11. Reuters T (2016) State of The Global Islamic Economy Report 2016/17
12. Barnett J, Begen F, Howes S, Regan A, McConnon A, Marcu A et al (2016) Consumers’
confidence, reflections and response strategies following the horsemeat incident. Food Control
59:721730
13. Schmutzler M, Beganovic A, Böhler G, Huck CW (2015) Methods for detection of pork
adulteration in veal product based on FT-NIR spectroscopy for laboratory, industrial and on-site
analysis. Food Control 57:58–267
14. Vlachos A, Arvanitoyannis IS, Tserkezou P (2016) An updated review of meat authenticity
methods and applications. Crit Rev Food Sci Nutr 56:1061–1096
15. Danezis GP, Tsagkaris AS, Camin F, Brusic V, Georgiou CA (2016) Food authentication:
techniques, trends & emerging approaches. TrAC Trends Analyt Chem 85:123132
16. Kurniawati E, Rohman A, Triyana K (2014) Analysis of lard in meatball broth using Fourier
transform infrared spectroscopy and chemometrics. Meat Sci 96:94–98
17. Meza-Márquez OG, Gallardo-Velázquez T, Osorio-Revilla G (2010) Application of mid-
infrared spectroscopy with multivariate analysis and soft independent modeling of class
analogies (SIMCA) for the detection of adulterants in minced beef. Meat Science 86:511–519
18. Rahmania H, Sudjadi, Rohman A (2015) The employment of FTIR spectroscopy in combination
with chemometrics for analysis of rat meat in meatball formulation. Meat Sci 100:301–305
19. Bruice PY (2016) Organic Chemistry, 8th edn. PEARSON
Deep Learning Approach for Divergence
Behavior Detection at High Density
Crowd
Abstract At high-density crowd gatherings, people naturally escape from the region
where any unexpected event happens. Escape in high-density crowds appears as a
divergence pattern in the scene and timely detecting divergence patterns can save
many human lives. In this paper, we propose to physically capture crowd normal
and divergence motion patterns (or motion shapes) in form of images and train a
shallow convolution neural network (CNN) on motion shape images for divergence
behavior detection. Crowd motion pattern shape is obtained by extracting ridges
of Lagrangian Coherent Structure (LCS) from the Finite-Time Lyapunov Exponent
(FTLE) field and convert ridges into the grey-scale image. We also propose a diver-
gence localization algorithm to pinpoint anomaly location(s). Experimentation is
carried out on synthetic crowd datasets simulating normal and divergence behav-
iors at the high-density crowd. Comparison with state-of-the-art methods shows
our method can obtain better accuracy for both divergence behavior detection and
localization problems.
1 Introduction
Divergence detection at the high-density crowd is a tough task due to several chal-
lenges involved in high-density crowd videos e.g., few pixels available per head,
extreme occlusion, cluttering and noise, and perspective problems, etc. If crowd
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 875
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_83
876 M. U. Farooq et al.
N1
N2
(a) (b)
N1
D1
N2
D2
(c) (d)
Fig. 1 Demonstration of crowd divergence at Love parade 2010: a crowd walking under normal
conditions with low density b crowd walking paths N1, N2 under normal conditions c high-density
crowd within the same region d crowd diverging through paths D1 and D2
divergence is not detected earlier at its development stage, it may lead to larger
disasters like a stampede. Figure 1 shows an example of high-density crowd diver-
gence behavior (Love parade 2010 musical festival [1]) where divergence eventually
leads to disastrous stampede. Figure 1a, b demonstrate high-density crowd normal
behavior following paths N1 and N2, whereas in Fig. 1c, d, a critical situation is
shown where the incoming crowd is blocked by a stationary crowd and diverge
through D1, D2 paths. Such divergence situations are common in mass gatherings
where the whole crowd is marching towards a common destination and with an
increase of density, ends up with a half-stationary half-moving crowd segments that
result in divergence behavior.
Previous divergence detection methods [2, 3] learn manual motion features for
every individual in the crowd from optical flow (OPF) including location, direction,
magnitude, etc. An inherent problem with such methods is with an increase of crowd
density, it is almost impossible to capture individual-level motion information and
one must learn global crowd features. Later several methods have been developed to
capture global crowd motion information e.g., optical flow with pathline trajectories
[4–7], pathlines with Lagrangian particle analysis [8], streakflow [9–13], etc. These
methods performed well in capturing crowd global motion information under normal
Deep Learning Approach for Divergence … 877
behavior scenes only. Unfortunately, no results are reported in the literature for
abnormal behavior detection at very high-density crowd levels.
In this work, we solve divergence detection in the high-density crowd by directly
capturing crowd global motion in form of images and learn crowd normal and diver-
gent motion shapes through a neural network that predicts crowd behavior for the
unknown scene. We also propose a novel divergence localization algorithm to pin-
point divergence location with the help of a bounding box. Finding a source of
divergence can help to efficiently deploy crowd management staff right at the critical
locations.
2 Related Work
Motion is one of the key ingredients in the crowd scene analysis and the success of the
behavior prediction scheme greatly relies on the efficiency of the motion estimation
(ME) method. Therefore, we provide a comprehensive review of ME techniques and
the corresponding abnormal behavior detection methods with emphasis on their capa-
bilities for ‘high’ density crowded scenarios. OPF is considered to be one of the most
fundamental motion flow model [14–17] that has been widely employed for motion
estimation [18, 19], crowd flow segmentation [20], behavior understanding [21–23]
and tracking in the crowd [24]. However, OPF methods suffer from various prob-
lems like motion discontinuities, lack of spatial and temporal motion representation,
variations in illumination conditions, severe clutter and occlusion, etc.
To overcome problems of OPF ME, researchers employ particle advection
concepts from fluid dynamics into the computer vision domain [8] and obtain long-
term “motion trajectories” under the influence of the OPF field. We et al. [7] employ
chaotic invariants on Lagrangian trajectories to determined either the behavior of
the crowd is normal or not. They also perform localization of anomaly by deter-
mining the source and size of an anomaly. Unfortunately, no results were reported
for the high-density crowd. Similarly, Ali et al. [8] obtain Lagrangian Coherent
Structures (LCS) from particle trajectories by integrating trajectories over a finite
interval of time termed as Finite-Time Lyapunov exponent (FTLE). LCS appears as
ridges and valleys in the FTLE field at the locations where different segments of
the crowd behave differently. Authors perform crowd segmentation and instability
detection in the high-density crowd using LCS in FTLE, however actual anomalies
of the high-density crowd like crowd divergence, escape behavior detection, etc.
are not performed. Similarly, authors in [10, 11] obtain particle trajectories using
high accuracy variational model for crowd flow and perform crowd segmentation
tasks only. Mehran et al. [9] obtain streakflow by spatial integration of streaklines
that are extracted from particle trajectories. For anomaly detection, they decompose
streakflow field into curl-free and divergence-free components using the Helmholtz
decomposition theorem and observe variations in potential and streak functions used
with SVM to detect anomalies like crowd divergence/convergence, escape behavior,
878 M. U. Farooq et al.
etc. However, results are reported for anomaly detection and segmentation at low-
density crowd and efficacy is still questionable for anomalies at the high-density
crowd. Eduardo et al. [25] obtain long-range motion trajectories by using the farthest
point seeding method called streamline diffusion on streamlines instead of spatial
integration.
Behavior analysis is performed by linking short streamlines using Markov
Random Field (MRF). However, only normal behavior detection and crowd segmen-
tation results are reported. Although particle flow methods discussed above are
better candidates for ME of the high-density crowd, but they are rarely employed
for abnormal behavior detection at high density crowded scenes. Figure 2 provides
a comparison of ME methods for high-density crowd performing Tawaf around
Kabbah. Conventional object tracking based ME methods [26, 27] (Fig. 2b, c) works
best at low crowd density but completely fails at high crowd density. The OPF method
from Brox et al. [15] can estimate motion at high density but motion information is
short-term. SFM [28] method can provide better motion estimation in low-density
crowd areas but at high density, the performance of SFM also degrades. Streakflow [9]
method also performs similarly to the SFM method at a high-density crowd. Unfor-
tunately, all these methods are unable to provide a clean motion-shape for the crowd.
FTLE method [9] (Fig. 2g) produce clear ridges at cowd boundaries and can be best
to describe high-density crowd motion. Therefore, in this work, we utilize the FTLE
method to obtain crowd motion-shape and translate it to a single channel greyscale
image (Fig. 2h) for both normal and abnormal behavior analysis. Our framework for
divergence detection is shown in Fig. 3 (top portion). It consists of two main phases:
Phase 1: low-level FTLE feature extraction and conversion into a grey-scale motion
shape image; Phase 2: behavior classification using a CNN. Motion shape images
are also used for divergence localization process.
Deep Learning Approach for Divergence … 879
A deep CNN network developed for normal and divergence classes in the high-density
crowd is shown in Fig. 6.
880 M. U. Farooq et al.
Fig. 4 Synthetic crowd data for Love parade 2010 disaster: a, b are camera-top views; c, d, e- same
crowd with perspective views
Fig. 5 Synthetic crowd data for Kabbah Tawaf—Normal behavior and divergent crowd
Fig. 7 Top row: Undesired motion-shape variations due to crowd oscillatory motion; Bottom row:
Real shape change due to divergence
882 M. U. Farooq et al.
Fig. 8 Baseline blob extraction pipeline for normal and divergence behaviors
Baseline blob extraction pipeline extracts a baseline blob from the normal and
divergence motion shapes and input to divergence localization algorithm is shown in
Table 1. The divergence localization algorithm indicates divergence location(s) with
the bounding box.
4 Experimentation Results
For divergence behavior detection at two scenarios, the crowd is simulated to diverge
from 25 different locations in each scenario and 1000 motion-shape images are
captured (total images for 25 divergence locations = 25 × 1000 = 25,000 divergence
images for each scenario). One thousand images for each divergence location are
generated to train CNN with minor local motion changes contributed by crowd
oscillatory motion. Similarly, 2500 images are generated for normal crowd behavior.
The dataset for each scenario is split into two parts: randomly 20 divergence locations
data (20 × 1000 = 20,000 images) are used for training/validation purposes, whereas
the remaining random 5 divergence locations data (completely unseen to CNN) is
used for prediction. Figure 9 provides a confusion matrix of divergence behavior
detection for both scenarios and performance is compared with the OPF method.
For both the Love parade and Kabbah scenario, our method can achieve 100%
accuracy. However, in both scenarios, OPF was able to detect approx. 50% of diver-
gence behaviors only. Motion-shapes obtained through the OPF method are not
as smooth and consistent as produced by our method; hence OPF performance
degradation is evident.
Deep Learning Approach for Divergence … 883
Table 1 (continued)
Step 3 FTLE field search for refined divergence location bounding box (local features anal-
ysis)
FOR ROI=1:length(ROI_list)
- Spatial Averaging: For each ROI boundary pixel, perform 8-connected spatial
averaging of FTLE field (for all frames from j-s_n to j+s_n frames)
- Temporal averaging: For each ROI boundary pixel in j-t_n to j+t_n frames,
perform temporal averaging.
- FOR all ROI boundary spatio-temporal averaged values
IF spatio-temporal averaging value >= st_thr,
keep pixel as valid
ELSE
discard pixel
END
- Obtain bounding box coordinates of current ROI by calculating maximum and
minimum (x,y) coordinates from selected valid pixels list in above step
END
Return bounding box coordinates for all ROIs in ROI_list
Fig. 9 Confusion matrices for divergence behavior detection: Love parade scenario a Proposed
method b OPF method; Kabbah scenario c Proposed method d OPF method
Area of overlap
IoU = (1)
Area of union
Generally, an IoU score greater than 0.5 (50% overlap) is considered a good
prediction by any bounding box (b. box) detection algorithm [33]. In this work,
the IoU score is calculated for N post i_t frames. IoU score of six selective frames
(out of N = 50 post i_t frames) for the Love parade scenario is shown in Fig. 10.
The green color b.box represents ground truth and the red color b.box represents
prediction by our algorithm. The Final IoU score is obtained by averaging N frames
IoU scores. The average IoU score of our algorithm for the Love parade scenario
is 0.501 (50% overlap). We also perform divergence region detection using OPF
motion images. The average IoU score with the OPF method is found to be 0.15
(15% overlap) which proves our method performs well than OPF for divergence
i_t+1 i_t+5
i_t+10 i_t+15
i_t+20 i_t+25
Fig. 10 Divergence region localization with our proposed method. IoU scores shown for six post
i_t frames
886 M. U. Farooq et al.
region localization. Similarly, the average IoU score for the Kabbah scene with our
algorithm is 0.63 (63% overlap) and 0.18 (18% overlap) for the OPF method.
5 Conclusion
References
1. Helbing D, Mukerji P (2012) Crowd disasters as systemic failures: analysis of the Love Parade
disaster. EPJ Data Sci 1(1):1–40. https://doi.org/10.1140/epjds7
2. Chen CY, Shao Y (2015) Crowd escape behavior detection and localization based on divergent
centers. IEEE Sens J 15(4):2431–2439. https://doi.org/10.1109/JSEN.2014.2381260
3. Wu S, Wong HS, Yu Z (2014) A bayesian model for crowd escape behavior detection. IEEE
Trans Circuits Syst Video Technol 24(1):85–98. https://doi.org/10.1109/TCSVT.2013.2276151
4. Andrade EL, Blunsden S, Fisher RB (2006) Modelling crowd scenes for event detection. Proc
Int Conf Pattern Recognit 1:175–178. https://doi.org/10.1109/ICPR.2006.806
5. Courty T, Corpetti N (2007) Crowd Motion Capture. Comput Animat Virtual Worlds
18(September 2007):361–370
6. Nam Y, Hong S (2014) Real-time abnormal situation detection based on particle advection in
crowded scenes. J Real-Time Image Process 10(4):771–784. https://doi.org/10.1007/s11554-
014-0424-z
7. Wu S, Moore BE, Shah M (2010) Chaotic invariants of lagrangian particle trajectories for
anomaly detection in crowded scenes. In: Proceedings of the IEEE computer society conference
on computer vision and pattern recognition, pp 2054–2060
Deep Learning Approach for Divergence … 887
8. Ali S, Shah M (2007) A Lagrangian particle dynamics approach for crowd flow segmentation
and stability analysis. In: IEEE conference on computer vision and pattern recognition, pp 1–6
9. Mehran R, Moore BE, Shah M (2010) A streakline representation of flow in crowded scenes.
Lect Notes Comput Sci (including Subser Lect Notes Artif Intell Lect Notes Bioinformatics)
LNCS 6313(PART 3):439–452. https://doi.org/10.1007/978-3-642-15558-1_32
10. Wang X, Gao M, He X, Wu X, Li Y (2014) An abnormal crowd behavior detection algorithm
based on fluid mechanics. J Comput 9(5):1144–1149. https://doi.org/10.4304/jcp.9.5.1144-
1149
11. Wang X, Yang X, He X, Teng Q, Gao M (2014) A high accuracy flow segmentation method in
crowded scenes based on streakline. Opt Int J Light Electron Opt 125(3):924–929. https://doi.
org/10.1016/j.ijleo.2013.07.166
12. Wang X, He X, Wu X, Xie C, Li Y (2016) A classification method based on streak flow for
abnormal crowd behaviors. Opt Int J Light Electron Opt 127(4):2386–2392. https://doi.org/10.
1016/j.ijleo.2015.08.081
13. Huang S, Huang D, Khuhro MA (2015) Crowd motion analysis based on social force graph
with streak flow attribute. J Electr Comput Eng 2015. https://doi.org/10.1155/2015/492051
14. Horn BK, Schunck BG (1981) Determining optical flow. Artif Intell 17(1981):185–203
15. Brox T, Papenberg N, Weickert J (2004) High accuracy optical flow estimation based on a
theory for warping. In: Computer Vision - ECCV 2004, vol 4, no May, pp 25–36. https://doi.
org/10.1007/978-3-540-24673-2_3
16. Lucas BD, Kanade T (1981) An iterative image registration technique with an application to
stereo vision. Proc Imaging Underst Work 130:121–130
17. Fortun D, Bouthemy P, Kervrann C, Fortun D, Bouthemy P, Kervrann C (2015) Optical flow
modeling and computation: a survey. Comput Vis Image Underst 134:1–21
18. Lawal IA, Poiesi F, Anguita D, Cavallaro A (2016) Support vector motion clustering. IEEE
Trans Circuits Syst Video Technol X(X):1–1. https://doi.org/10.1109/TCSVT.2016.2580401
19. Cheriyadat AM, Radke RJ (2008) Detecting dominant motions in dense crowds. IEEE J Sel
Top Signal Process 2(4):568–581. https://doi.org/10.1109/JSTSP.2008.2001306
20. Ali S, Shah M (2007) A lagrangian particle dynamics approach for crowd flow simulation and
stability analysis
21. Hu MHM, Ali S, Shah M (2008) Learning motion patterns in crowded scenes using motion
flow field. In: 2008 19th international conference pattern recognit, pp 2–6. https://doi.org/10.
1109/ICPR.2008.4761183
22. Solmaz B, Moore BE, Shah M (2012) Identifying behaviors in crowd scenes using stability
analysis for dynamical systems. IEEE Trans Pattern Anal Mach Intell 34:2064–2070. https://
doi.org/10.1109/TPAMI.2012.123
23. Chen DY, Huang PC (2011) Motion-based unusual event detection in human crowds. J Vis
Commun Image Represent 22(2):178–186. https://doi.org/10.1016/j.jvcir.2010.12.004
24. Hu W, Xiao X, Fu Z, Xie D, Tan T, Maybank S (2006) A system for learning statistical
motion patterns. IEEE Trans Pattern Anal Mach Intell 28(9):1450–1464. https://doi.org/10.
1109/TPAMI.2006.176
25. Pereira EM, Cardoso JS, Morla R (2016) Long-range trajectories from global and local motion
representations. J Vis Commun Image Represent 40:265–287. https://doi.org/10.1016/j.jvcir.
2016.06.020
26. Dalal N, Triggs B (2005) Histograms of oriented gradients for human detection. In: EEE
computer society conference on computer vision and pattern recognition (CVPR’05), pp 886–
893
27. Barnich O, Van Droogenbroeck M (2011) ViBe: a universal background subtraction algorithm
for video sequences. IEEE Trans Image Process 20(6):1709–1724. https://doi.org/10.1109/TIP.
2010.2101613
28. Mehran R, Oyama A, Shah M (2009) Abnormal crowd behavior detection using social force
model. In: IEEE computer society conference on computer vision and pattern recognition work.
CVPR Work 2:935–942. https://doi.org/10.1109/CVPRW.2009.5206641.
29. https://www.oasys-software.com
888 M. U. Farooq et al.
30. Shadden SC, Lekien F, Marsden JE (2005) Definition and properties of Lagrangian coherent
structures from finite-time Lyapunov exponents in two-dimensional aperiodic flows. Phys D
Nonlinear Phenom 212(3–4):271–304. https://doi.org/10.1016/j.physd.2005.10.007
31. Zeiler MD et al (2013) On rectified linear units for speech processing New York University,
USA Google Inc ., USA University of Toronto , Canada. In: IEEE international conference on
acoustic speech and signal processing (ICASSP 2013), pp 3–7
32. Krausz B, Bauckhage C (2012) Loveparade 2010: automatic video analysis of a crowd disaster.
Comput Vis Image Underst 116(3):307–319. https://doi.org/10.1016/j.cviu.2011.08.006
33. Ahmed F, Tarlow D, Batra D (2015) Optimizing expected intersection-over-union with
candidate-constrained CRFs. In: Proceedings of the IEEE international conference on computer
vision, vol 2015, pp 1850–1858. https://doi.org/10.1109/ICCV.2015.215
Anomaly Localization at High-Density
Crowd Using Motion Shape Image (MSI)
1 Introduction
It is well known that automated anomaly detection in surveillance videos has attracted
a lot of attention due to efficient methods developed in the past. However, anomaly
detection alone is less effective without localizing it. Anomaly localization is critical
information that can help to efficiently deploy rescue-resources at the right place to
save many precious lives.
There have been several methods developed in the past focusing anomaly local-
ization at low to medium density crowd [1–3], however, localizing anomaly at high-
density is challenging and there are not many methods available in the literature [4–7].
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 889
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_84
890 M. U. Farooq et al.
change in motion pattern from the normal straight or circular patterns in any direc-
tion is considered as divergence. In the next subsection, the localization method for
divergence anomaly at the high-density crowd is discussed in detail.
2 Proposed Method
A detailed block diagram of our anomaly detection and localization framework for
divergence behavior at a high-density crowd is shown in Fig. 1. The top portion of the
block diagram shows an anomaly detection network using an MSI. The performance
evaluation of the proposed anomaly detection network is discussed in detail in [8]. The
proposed anomaly detection network outputs scores for two classes i.e., normal and
anomalous (divergent). A divergent class score greater than a pre-defined threshold
triggers the divergence localization process, shown at the bottom of Fig. 1.
Fig. 2 Divergence localization method for source/starting point detection and divergence mask
Anomaly Localization at High-Density Crowd … 893
Fig. 3 Top row: undesired motion-shape variations due to crowd oscillatory motion; Bottom row:
real shape change due to divergence
Global motion shape variation occurs if any crowd segment stops that reduce
crowd velocities to zero, results in no OPF/ FTLE field in that crowd region. Thus,
LCS ridges in the FTLE field vanishes from stationary crowd segments and only
appear at the moving crowd segments, causing an undesired global motion shape
variation. Due to the undesired changes in motion shape, raw MSI’s cannot be utilized
for further analysis and we need to extract baseline motion shape for both normal
and divergence behaviors. A baseline motion shape represents a shape that is most
common throughout all the specific behavior frames. Figure 4 shows the pipeline for
baseline motion shape extraction that extracts baseline motion shape for both normal
and abnormal video frames.
Baseline motion shapes obtained for normal and divergent motion patterns differ in
the shape in the way that divergent motion shape will contain extra blobs representing
diverging crowd. The reference normal baseline motion shape and baseline divergent
motion shape are then subtracted that generates smaller blobs. The resulting small
Fig. 4 Pipeline for baseline motion shape extraction for normal and divergence behaviors
894 M. U. Farooq et al.
Fig. 5 Left side: Crowd circulatory normal motion with crowd diverging in the top second image
(lower row shows corresponding MSI’s); right side images for synthetic love parade dataset crowd
diverging from different locations (lower row are corresponding MSI’s)
Fig. 6 Top row: MELAN dataset with crowd diverging from left and right gates; bottom row:
PILGRIM dataset with two normal scenarios and last image showing divergent crowds for two
normal crowd scenarios
blobs contain the actual divergent region(s) and many noisy blobs. The noisy blobs
are filtered out by applying temporal and spatial filtering on OPF and FTLE fields
on current and past N filter frames.
The blobs obtained from the above process represent divergent regions or diver-
gence masks around the diverging crowd. However, we need to find an exact starting
point or source of divergence from the divergent mask detected above. Initially, the
Anomaly Localization at High-Density Crowd … 895
contour of the divergence mask is obtained and the Euclidean distance of contour
pixels from normal baseline motion shape pixels is computed. The contour pixels
having a distance less than a threshold is/are marked as divergence source or starting
point(s). The experimentations have shown that a threshold of 10 pixels is most
optimal for the majority of scenes and can represent correct source points. Euclidean
distances more than 10 pixels are generally far from the actual source point and can
represent false source point detections.
The diverging crowd direction and the angle of divergence are critical information for
first responders to effectively manage the crowd. To quantify the crowd divergence
information, crowd “dominant motion” needs to be computed. Dominant motion
represents the motion of the major segments of the crowd in the video. In the case of
divergence, we have to compute dominant flows for both normal and divergent crowd
segments to show how far the divergent crowd moves away from the normal walking
crowd. In this work, dominant motion flow is computed using the method from [12].
Summarizing the method of [12] for dominant motion flow, initially, video is divided
into several segments and particle advection is performed to obtain small size raw
tracklets in each segment. Raw tracklets provide short duration motion information
and crowd starting (source) and ending (sink) points cannot be determined. Local
long tracks are identified in each video segment by using an unsupervised hierar-
chical clustering algorithm that utilizes the Longest Common Sub-sequence (LCS)
similarity measure to combine smaller tracklets into local large trajectories. To find
dominant flow throughout the video, local trajectories from different video segments
are also combined using the same clustering and similarity method to produce global
896 M. U. Farooq et al.
dominant flows. This method is also able to identify the source and sink point of a
dominant motion.
The above algorithm is run to obtain dominant motion flows in both normal and
divergence masks. To quantify the divergent dominant motion from normal dominant
motion, we need to compute the mean angle for both dominant flows and the angle
difference between the two flows would show how far the divergent crowd moves
away from the normal crowd. A dominant motion trajectory represents the positions
of a particle Pi it takes from source to sink in video and there is associated angle
information (∅i ) at every position of a particle. The particle angle (∅i ) of each particle
Pi in the dominant flow trajectory is computed using OPF velocities given in Eq. 1.
v
∅i = tan−1 (1)
u
where u and v are optical flow horizontal and vertical velocities respectively.
It is important to notice here that the OPF used in the particle advection process
is computed at an evenly spaced grid whereas particle positions of dominant flow
trajectory are at off-the-grid indexes. And we do not have OPF at off-grid particle
positions. Therefore, linear interpolation is implemented to obtain OPF at off-grid
particle positions and ∅i is computed from interpolated OPF u and v velocities. The
angle ∅i computed above lies in the range of [0, π]. To know the exact quadrant in
which a particle lies, ∅i is mapped to a unit circle with the angle from [0, 2π]. We
use the method in [13] to map the linear angle into a circular angle on the unit circle.
The circular angle shows the actual quadrant of a particle it lies in. A similar process
is repeated for all particles by mapping all on the unit circle. Then a mean circular
angle is computed that represents the mean direction of a dominant flow. The mean
angle for both normal and divergent dominant flows is computed and the divergence
angle is calculated by taking the difference of two mean angles.
Divergence class score value from CNN lies between 0 and 1 and the threshold
at the class score is set to 0.65 (65%) to trigger the localization detection process.
Residual blobs are filtered out through temporal averaging filtering and spatially
through connected neighbors filtering. The number of images for temporal averaging
N filter is set to 10 i.e., averaging previous 10 frames OPF and FTLE fields. Spatial
filtering is applied by performing an 8-connected neighbor on residual blob(s).
Fig. 7 Divergence localization comparison with state-of-the-art methods. Legends: Red triangle
[5], Green triangle [1], Orange triangle our method, Blue triangle Ground truth. The top two rows
show divergence localization at low-density scenes. False divergent center detections by our method
at 2nd , 3rd image at the first row and 1st image, second row. Existing art failed to detect divergent
centers at high density (last two images in the third row) while our method can accurately detect
divergent centers
velocity. Similarly, the method in [1] shows poor performance at high-density diver-
gence. Assumption of three moving objects and solving three straight-line equations
is possible at low density but, at the high-density crowd. As there can be hundreds
of people in the neighbor of a divergent center and solving straight line equations
for hundreds of moving objects is computationally very expensive. The method with
existing model settings is unable to detect actual divergent center locations at high
density. However, the model can be improved by obtaining intersections of a large
number of neighboring moving objects with reasonable compute. False detections by
existing art on the high-density crowd can be seen in the last two images of the third
row in Fig. 7. Our method perform well at high density scenario and can idenfity
divergence points at correct locations.
As mentioned earlier our localization algorithm also identifies divergence regions
(or divergence mask) evolves temporally. Figure 8 shows three examples of diver-
gence mask detected by our algorithm. The first row is a crowd divergence sequence
taken from the PETS2009 dataset, the second row shows people diverging from the
normal path of walking over the zebra crossing, and the last row shows people
900 M. U. Farooq et al.
Fig. 8 Divergence region mask changes over time. Divergence regions: first row PETS2009;
second-row UCF; and third-row Synthetic Kabbah datasets. The first column is the region at the time
of divergence detection (td ) by CNN. The second column represents divergence region detection
n_frames after td , the third column represents divergence region detection 2*n_frames after td, and
so on. The value of n_frames is different for each sequence
diverging from circular Tawaf after completing seven laps of Tawaf. Images in
columns show the temporal progression of divergence where images are taken every
n_frames after the time instant (t d ) divergence is detected by CNN. Divergence mask
size can be seen increasing over time indicating variations in the size of divergence
and also shows the direction in which divergence is leading.
As mentioned earlier our divergence localization algorithm not only identify diver-
gence source points but also detect region(s) of divergence. To qualitatively evaluate
the performance of our algorithm, we compute Intersection over Union (IoU) between
the predicted divergence region and ground truth divergence region. Ground truth
regions are obtained by hand labeling divergence regions at each abnormal frame.
IoU score is calculated using Eq. (2),
Area of Overlap
IoU = (2)
Area of Union
Anomaly Localization at High-Density Crowd … 901
where,
Area of overlap is the overlap area between the predicted region and ground truth
region.
Area of union is the area encompassed by both the predicted bounding box and
ground truth bounding box.
IoU score greater than 0.5 (50% overlap) is generally considered a good prediction
by algorithm [14]. Two samples of divergent regions detected overlaid with ground
truth divergence region are shown in Fig. 9. Again it is clear from Table 2 that at low-
density datasets (UMN, PETS2009, NGSIM), the divergence shape is not smooth
and is broken, resulting in a low IoU score, whereas the IoU score at high densities is
better meaning our algorithm is better able to detect divergent regions at high crowd
densities. IoU scores for divergent datasets used in this work are provided in Table
2.
Fig. 9 Examples of divergence region by our method (yellow) compared to ground truth (green)
We use pre-trained CSRNet from [10] on high-density crowd images from Kabbah
and Hajj and perform experimentation on benchmark datasets. Figure 10 shows a
crowd density map and crowd count inside divergence masks for various datasets.
The crowd count inside the divergence mask is compared with the ground truth count.
5 Conclusion
In this paper, we introduce a more extended view of looking into anomaly localiza-
tion to make this information more useful for real-life rescue resource deployments.
a localization mask is initially generated and features of source/starting point, crowd
Anomaly Localization at High-Density Crowd … 903
Fig. 10 Crowd density map and density count inside divergence mask. 1st row: UCF; 2nd row:
PETS2009; 3rd row: PILGRIM left-divergence; 4th row: PILGRIM right divergence; 5th row: two
divergences at CONCERT
904 M. U. Farooq et al.
Fig. 11 Divergence (localized) direction and angle estimation for UCF dataset
density estimation, and direction/angle of the localized region are computed. Exper-
imentation results show that the proposed method outperforms the existing state-of-
the-art method for source/starting point detection at the high-density crowd and CDE
and direction/angle estimations are close to the ground truth.
Anomaly Localization at High-Density Crowd … 905
Fig. 12 Direction and angle identification of localized region for PETS (1st row), PILGRIM (2nd
row), and MELAN (3rd row)
Acknowledgements The Massmotion crowd simulation software is supported by the Center for
Intelligent Signal for Imaging Research (CISIR) under PO number 3920089787/30.10.2017.
References
1. Chen CY, Shao Y (2015) Crowd escape behavior detection and localization based on divergent
centers. IEEE Sens J 15(4):2431–2439. https://doi.org/10.1109/JSEN.2014.2381260
906 M. U. Farooq et al.
2. Ongun C, Temizel A, Temizel TT (2014) Local anomaly detection in crowded scenes using
finite-time lyapunov exponent based clustering. 11th IEEE international conference advance
video signal-based surveillance, AVSS, pp 331–336. https://doi.org/10.1109/AVSS.2014.691
8690.
3. Ren WY, Li GH, Chen J, Liang HZ (2012) Abnormal crowd behavior detection using behavior
entropy model. In International conference on wavelet analysis and pattern recognition, pp
212–221. https://doi.org/10.1109/ICWAPR.2012.6294781
4. Wu S, Moore BE, Shah M (2010) Chaotic invariants of lagrangian particle trajectories for
anomaly detection in crowded scenes. In: Proceedings of the IEEE computer society conference
on computer vision and pattern recognition, pp 2054–2060
5. Wu S, Wong HS, Yu Z (2014) A bayesian model for crowd escape behavior detection. IEEE
Trans Circ Syst Video Technol 24(1):85–98. https://doi.org/10.1109/TCSVT.2013.2276151
6. Chaker R, Al Aghbari Z, Junejo IN (2017) Social network model for crowd anomaly detection
and localization. Pattern Recognit 61:266–281. https://doi.org/10.1016/j.patcog.2016.06.016
7. Mehran R, Moore BE, Shah M (2010) A streakline representation of flow in crowded scenes.
Lect Notes Comput Sci (including Subser Lect Notes Artif Intell Lect Notes Bioinformatics),
vol 6313 LNCS, no PART 3, pp 439–452. https://doi.org/10.1007/978-3-642-15558-1_32
8. Muhammad Farooq U (2020) Divergence behavior detection and localization using Motion
Shape Image (MSI). Universiti Teknologi PETRONAS
9. Krausz B, Bauckhage C (2012) Loveparade 2010: automatic video analysis of a crowd disaster.
Comput Vis Image Underst 116(3):307–319. https://doi.org/10.1016/j.cviu.2011.08.006
10. Li Y, Zhang X, Chen D (2018) CSRNet: dilated convolutional neural networks for under-
standing the highly congested scenes. Proc IEEE Comput Soc Conf Comput Vis Pattern
Recognit February: 1091–1100. https://doi.org/10.1109/CVPR.2018.00120
11. https://github.com/ZhengPeng7/CSRNet-Keras
12. Khan SD, Bandini S, Basalamah S, Vizzari G (2016) Analyzing crowd behavior in naturalistic
conditions: Identifying sources and sinks and characterizing main flows. Neurocomputing
177:543–563. https://doi.org/10.1016/j.neucom.2015.11.049
13. Berens P, Velasco MJ (2009) Journal of statistical software. Wiley Interdiscip. Rev. Comput.
Stat. 1(1):128–129. https://doi.org/10.1002/wics.10
14. Ahmed F, Tarlow D, Batra D (2015) Optimizing expected intersection-over-union with
candidate-constrained CRFs. Proc IEEE Int Conf Comput Vis Inter 1850–1858. https://doi.
org/10.1109/ICCV.2015.215
EEG Based Brain Controlled RC Car
with Attention Level
1 Introduction
Spinal Cord Injuries (SCI) are increasing rapidly. According to the study of the
National Spinal Cord Injury Statistical Center (NSCISC) at UAB. The most promi-
nent cause of SCI is vehicular accidents, which are approx. 39.3%, then by falling
31.8% and so on. One of the US studies shows that around 1.25 million died in road
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 907
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_85
908 M. Talha Ejaz et al.
accidents in the world [1]. Quadriplegia is generally caused by road accidents, which
becomes a global catastrophe and leading cause of death by injury. Quadriplegia is
one of the spinal cord injuries in which the lower body becomes paralysis. It is caused
by damage to the spinal cord at a level of C1–C7 or sometimes brain injury.
Quadriplegia caused several complications, including pain, blood clots, pressure
sores, related injuries, spastic muscles, respiratory problems, automatic dysreflexia,
and loss of bladder and bowel control. The most important complication is the loss of
the motor sensing control of limbs and sometimes hands also. These complications
make a person handicapped. Different types of wheelchairs, including manual hand-
controlled, hand gesture-controlled, voice recognition control, eye motion, finger
movement control, and remote control, are available in the market. Still, some of
them are not feasible for Quadriplegia patients, and some are expensive.
Brain-controlled wheelchairs are designed for those patients who are unable to
use the powered wheelchairs for mobility. After the successful advancements in
robotics, artificial intelligence, and sensor technology, more advanced methods come
into existence for designing a smart wheelchair. EEG based controlled wheelchairs
are very popular nowadays as they are cost-effective and easy to operate. EEG or
electroencephalography is a process to record and detect brain activity in terms of
signals in real-time. The human brain comprises 100 billion neurons, and signals
are transmitted from one neuron to another through electric impulses. Alpha, Beta,
Theta, Delta, and Gamma are five different waveforms found in the human brain, and
each one is different from another in terms of frequency and amplitude. Electrodes
are attached on a scalp, and conductive gel is used to get better results to record or
monitor an EEG signal. It takes around 30 min for the setup, which is not suitable
to drive a wheelchair. To consider this issue, Neurosky is being used to detect the
brain signals, which does not need any gel before using it, and this method is fast and
cost-effective. However, this method is not accurate as detecting the signals using
electrodes with gel.
Brain-Computer Interface (BCI) is an approach where a physical device is operates
using brain activity. Attention level and eye blink are the main methods used in the
BCIs application. In [2–4], attention level was used to analyze human activities such
as sleep patterns and recognized the degree of human attention. When a person blinks
an eye, a deflection in a waveform is detected. By capturing that deflection, several
applications were proposed in [5–8]. In [9, 10], methods were discussed for the
wheelchair for quadriplegia patients using the head movement and retina movement
utilizing the image processing and accelerometer technique. These methods were not
suitable for a person, as it causes fatigue while moving the head. A neuro handset is
used to record the attention level for driving a wheelchair that mitigates the issues.
It makes a system faster and allows the patient to control the wheelchair by focusing
the directions.
In this paper, we proposed a conceptional idea where RC car is drive using the
attention level. This method helps the handicapped especially quadriplegia patients
to move from one place to another on their own. We also incorporate IOT in the
proposed method, where signals are transmitted to an android application.
EEG Based Brain Controlled RC Car with Attention Level 909
The rest of the paper is organized as follows. Section 2, covers the background
knowledge of different types of EEG signals and recording methods. The proposed
methodology and results are discussed in Sects. 3, 4 respectively. Section 5 covered
the conclusion.
2 Background
3 Methodology
The novelty of the proposed method is that Neurosky was used as a sensor. It is
an EEG headset for acquiring the EEG signal. This setup not reduced the cost of
the overall system but increase efficiency. Neurosky is a one channel band with
a sampling rate of 512 Hz. It is the only sensor used to drive an RC car in four
directions: right, left, forward, and backward. The specification of Neurosky (Mind
wave) is mentioned in Table 2.
Neurosky is placed on a frontal side of a head precisely at the FP1 location. It has
a dry electrode, which senses the EEG and gives the value during attention and eye
blink. The Neurosky and the site where it is placed is shown in Fig. 2.
We dealt with the attention level of a person. Attention was calculated by acquiring
a signal from a device and passed to the signal processing unit. EEG is recorded as a
sampling frequency of 512 Hz with a 16-bit quantized level. Fast Fourier Transform
(FFT) is being used to convert time signal into a frequency domain with a low pass
filter with a cutoff frequency of 50 Hz.
(a) (b)
Fig. 2 a shows the Neurosky headset and b depicts the position where Neurosky attached to the
scalp
In this section, we are discussing the hardware and software flow of the proposed
method. (Appendix shows the hardware and software flow of a system). The flow
starts by calibrating the headset, as all the movement depends upon the data acquired
from the headset. After obtaining the data, it was propagated to the micro-controller
to perform the actions. L298N motor driver was also attached to the microcontroller
that received the command and then powered the motors for the actions. The actions
were also monitored in the mobile application, which was designed in an android
studio for this particular application. For communication between the RC car and
application was done by the Wi-Fi module (ESP8266). In the proposed method,
four basic movements were considered: forward, backward, left, and right. One
more additional command, stop, was considered for the safety purpose. The RC car
moved when the attention level was greater than 10. The actions were classified as
mentioned in Table 3.
4 Experimental Results
(a) (b)
Fig. 6 a shows the Andriod application interface and b depicts the action selected by the subject
5 Conclusion
In this article, a conceptual idea is presented that can be used for Quadriplegic
patients to control the wheelchair using the EEG signals. The signals have extracted
using the Neurosky headset placed at the FP1 position of the user’s head. Actions
are classified based on attention levels such as forward, backward, left, right, and
stop. The proposed method was tested on five different subjects and results shown
that all the subjects were able to classify the actions according to the attention level.
An IoT-based approach was also used in the proposed method that transmits and
receives the data wirelessly with the help of NodeMCU. An android app was also
developed that shows the actions, and again, a user can control the car manually if
malfunctions happen. In the future, the idea will be implemented in the wheelchair
and also incorporate the GPS for tracking.
Appendix
References
12. Awais MA, Yusoff MZ, Yahya N, Ahmed SZ, Qamar MU (2020) Brain controlled wheelchair:
a smart prototype. J Phys: Conf Seri 1529(4):042075
13. Khan D, Kamel N, Muzaimi M, Hill T (2020) Effective connectivity for default mode network
analysis of alcoholism. Brain Connectivity
14. Guo X, Wu X, Gong X, Zhang L (2013) Envelope detection based on online ICA algorithm
and its application to motor imagery classification. In Proceeding of the 6th international
IEEE/EMBS conference on neural engineering (NER), San Diego, CA, USA, 6–8 November
2013
Automated Pill and Syringe Dispenser
Abstract This paper presents the Automated Fabrication of Pill and Syringe
Dispenser. As it is necessary to provide medication for the elderly people on time
a medication remainder is already developed. So, in addition to help the diabetic
patient, automatic pill and syringe dispenser is recommended. It is designed in such
a way that, based on the real time clock it dispenses the pill and syringe for the patients
on required time and it acts as a reminder. In addition to this, automatic syringe
dispenser is developed, which loads the insugen automatically. The major compo-
nents of the combination of pill and syringe dispenser are DS3231 RTC, Arduino
UNO, Servo motor, LCD, LED, Buzzer, and in addition to this, the software called
BLYNK, is used to pass the information through Bluetooth and the medication is
dispensed on the required time for the elderly ones. And the components used for
automatic syringe dispenser are Arduino UNO, lead screws, L293d, DVD stepper
motor board, servo motor, piezo buzzer.
R. Paviya (B)
N.G.P Institute of Technology, Kalapatti road, Coimbatore, India
S. Prabakar
Sona College of Technology, Salem, India
e-mail: [email protected]
K. Porkumaran
Sri Sairam Engineering College, Chennai, India
A. B. S. Saman
Universiti Teknologi PETRONAS, Seri Iskandar, Perak, Malaysia
e-mail: [email protected]
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 919
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_86
920 R. Paviya et al.
1 Introduction
As there are many elderly individuals who need constant help, to take their medi-
cations on time and to prevent any illness timing is must. So Pill Dispenser has
been developed. Also, many elders face problems like Alzheimer’s, poor vision and
Diabetics etc.., and most of the death occur due to the intake of wrong medications.
So, in addition to help the diabetic patient, Syringe dispenser is combined with the
Pill dispenser. In this combination of pill and syringe dispenser the patient themselves
has to load the insugen. So, the extension of Syringe dispenser is made. Where the
syringe is automatically loaded to the insugen and reminds the people to take their
drug on time.
These are items which release syringe at specified times. And automatic Syringe
dispenser is developed. This mainly helps in loading the insugen for the required
amount and it dispenses the syringe at the predefined time. This is done with the help
of stepper motor, servo motor, lead screws, L293d IC, 3D print software for holding
the syringe.
• Their purpose is to help the Diabetic patients those who are insulin dependent, to
take their medications on time.
• It dispenses Syringe at a set interval of time, especially for the Diabetic patient.
2 Literature Survey
(i) “Smart drugs: Improving Healthcare using Smart pill box for Medicine remainder
and monitoring system” Diaa Salama Abdul Minaam, Mohamed Abd-elfattah [1].
Automated Pill and Syringe Dispenser 921
The paper “Smart drugs: Improving Healthcare using Smart pill box for Medicine
remainder and monitoring system” published in the journal Science Direct in 2018.
This has the concept, design and creation of pill box prototype which has nine separate
sub boxes to solve the deficiency in medical area. It enables the care takers to take
the determined amount of pill and timing to take pill on every day. It requires the
client to stack the box consistently in order to refill it with medicines.
(ii) “Smart Medication Dispenser” Suraj shinde, Nitin bange, Monika kumbhar,
Snehal patil [2].
The paper “Smart Medication Dispenser” published in the International Journal
of advanced research in Electronics and Communication Engineering (IJARECE)
in 2017. This uses 4 × 4 matrix keypad, GSM module, LCD display and real time
clock (RTC) and this helps to provide information automatically to patients to take
their right dosages according to the prescription at appropriate time.
(iii) “Construction of a Smart Medication Dispenser with High Degree of
Scalability and Remote Manageability” Jugeon Pak, Keehyun Park [3].
The paper “Construction of a Smart Medication Dispenser with High Degree of
Scalability and Remote Manageability” published in the Journal of Biomedicine and
Biotechnology in 2012. In this each medication have a medication tray dispenser.
The time is set in RTC and when the pre-determined time reaches the time, it gives
some kind of notification and then the user has to push a button and the medication
is dispensed from the medication tray (MDT).
3.1 Introduction
For the combination of pill and syringe dispenser same methodology is used and this
dispenses the syringe and pill and it reminds the patient to take their drug on time
and in this, the patient has to load the insulin into the needle.
Construction: For both the pill and syringe dispenser, same methodology is used.
The Arduino UNO has two-way input, one from the power supply and another from
the ‘BLYNK’ app via Bluetooth. The time is set in BLYNK as well as in RTC and it
is compared. When the RTC (DS3231) reaches the BLYNK time, it starts dispensing
the medications on the predefined time. Thus both the pill and syringe is released for
the patients use, for the required time (Fig. 1).
Working: Once the power supply is fed into Arduino UNO and the information
from BLYNK is received, and it sends the signal to DS3231 and it again receives as
922 R. Paviya et al.
Bluetooth
LCD Display
DS3231
Arduino UNO
an input to the Arduino UNO and then the Arduino processes the program to run a
servo in order to dispense the pill and syringe on the required time. An Alarm gives
off with the buzzing action which is also controlled by the Arduino and then the LED
glows.
4.1 Introduction
In this, the extension of syringe dispenser is made as, the Syringe holder is made
by using the 3D printer and the syringe moves forward and backward as the syringe
holder is placed on the CD driver setup. This setup consists of the lead screws with
stepper motor which helps in the linear motion and IC L293d is used, it mainly helps
to run 2 DC motors with same IC. It can control both the speed and direction.
Construction: Arduino UNO is connected to the IC L293d and also connected to the
stepper motors of CD drive. The syringe holder is made by using the 3D printer and
the servo setup moves front and back with the help of lead screws in CD driver. And
the syringe gets dispensed and it starts to buzzer. This process continues (Fig. 2).
Working: When the power supply is fed into Arduino UNO and then it initiates
the IC L293d and the IC starts moving the CD driver motor forward and backward
with the help of lead screws and the syringe moves forward into the insugen bottle
and then the back holder of syringe moves backward, so that the insugen is sucked
inside the syringe and then both the setup moves backward and then the micro servo
Automated Pill and Syringe Dispenser 923
NO
Timer
If t=y?
YES
pushes the syringe out and it starts buzzing off and notifies the patient to take the
syringe.
The overall performance of this device is good. First the combination of both pill and
syringe dispenser (Fig. 3) has been developed. This helps to remind the patients or the
care taker to take the drug on time. It dispenses the Pill followed by the Syringe. In
this the syringe is dispensed without suction of insugen. So an extension of syringe
dispenser is done (Fig. 4). In this extension of syringe dispenser, it automatically
sucks the insugen with the help of the CD driver setup and IC. And the syringe gets
dispensed and alarm goes on and notifies the patient. This mainly helps the patient
in not missing their medications and it will not lead to any medication error.
There are some improvements in this device to be made in future. It includes auto-
matic refilling of syringe after usage, sending message to the caretaker if the medi-
cation is not taken this can be done using the GPS, fixing the insugen bottle with the
needle, and also an input panel will be set so that the amount of insugen to be sucked
924 R. Paviya et al.
by the needle can be entered easily and time can be set easily by the care taker. This
would be very useful for the patients and also it reduces stress for the care takers.
7 Conclusion
There are many pill boxes available which are bit complicated and most of them are
costlier. But the combination of pill and syringe dispensers is not available, so it was
developed. In this fabrication of pill and syringe dispenser, with the help of BLYNK
app the time is set and when the time reaches the pre-determined time the pill gets
dispensed followed by the syringe. In this the syringe is not loaded with insugen.
Automated Pill and Syringe Dispenser 925
So, the extension of this process has been developed with the help of CD drivers
and IC L293d the syringe automatically loads the insugen and it is released with the
help of micro servo. This is user friendly and it helps the patients in not missing
the medications. It is less weight and portable, it reduces the medication errors. This
can be used in hospitals, homecare, and other old age homes. The main goal of the
system is to provide healthy, tension free life to those who are taking pills regularly
and to provide it at an affordable cost. This product is easy to design. So, thus it
requires less maintenance.
References
1. Abdul Minaam DS, Abd-Elfattah M (2018) Smartdrugs: improving healthcare using smart pill
box for medicine reminder and monitoring system. Sci Dir 29 November (2018)
2. Shinde S, Bange N, Kumbhar M, Patil S (2017) Smart medication dispenser of a proceedings
paper: IJARECE
3. Pak J, Park K (2012) Construction of a smart medication dispenser with high degree of scalability
and remote manageability. J Biomed Biotech
Multi-classification of Brain Tumor
Images Using Transfer Learning Based
Deep Neural Network
1 Introduction
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 927
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_87
928 P. Dutta et al.
images. For instance, Sajjad et al. [2] employed an unsupervised learning method
called the convolutional neural network (CNN) algorithm for the classification of the
brain tumor images in different classes. The accuracy was found to be almost 94.58%
for the classification of the multiple categories of tumors. Moreover, Amin et al. [3]
employed a fusion process using the discrete wavelet transform (DWT) method
to extract a more informative tumor region. Then the noise removal process was
applied based on a partial differential diffusion filter (PDDF) before segmentation.
After that, the CNN model was utilized for classifying the tumors as cancerous and
non-cancerous. In another study, Sultan et al. [4] developed an approach for the
multi-classification of brain tumors. The classification of the brain tumors images
in multiple classes was based on the CNN algorithm. The accuracy was found to be
almost 96.13% for the classification of the multiple categories of tumors. Abiwinanda
et al. [5] proposed a classification model based on CNN for the multi-classification
of brain MRI images. The architecture of the CNN model consisted of different
layers i.e., convolution, max-pooling, flattening, and fully connected one hidden
layer. The classification method based on CNN provided a classification accuracy of
approximately 94.68%.
The major contribution of this paper is to develop an efficient classification method
using CNN with the aid of the inception-v3 transfer learning method. Moreover, a
suitable learning rate and batch size are employed to make the designed model
robust and also faster and smooth the training process. In addition, the effectiveness
of the designed method is analyzed by comparing it with some existing classification
methods in terms of classification accuracy.
The rest of the paper is organized as follows: Sect. 2 provides the methodology
of the proposed system with a detailed explanation of each of the steps. Section 3
represents the results with analysis and the comparative study is also conducted in
this Section. Finally, Sect. 4 shows concluding remarks.
The process of the proposed multi-classification method is divided into three steps as
illustrated in Fig. 1. In the first step, the image preprocessing is performed using the
image augmentation technique to increase the amount of total image dataset. After
the preprocessing step, the image features are automatically generated by employing
the Inception-v3 transfer learning method. Finally, the extracted features are feed
into the modified DNN to classify the three brain tumor types.
The dataset used for this model is obtained from 233 patients with three categories of
brain tumor images at different slices (a) 994 axial images, (b) 1045 coronal images,
and (c) 1025 sagittal images. This T1-weighted contrast-enhanced image dataset is
provided by Cheng [6]. It comprised of 3064 brain tumor MRI images. The datasets
are formulated with 1426 meningioma images, 708 glioma images, and 930 pituitary
images. Each of the images has a size of 512 × 512 in pixels. For decreasing the
computational time and dimensionality the images are resized to 150 × 150 pixels.
Then the total image datasets are splitting into training (80% of the total dataset)
and validation (20% of the total dataset) dataset. The validation dataset is employed
to estimate the proposed classifier model. Moreover, the details explanation of each
step of the classifier model are described as following.
Before feeding the classifier model, the image augmentation process is performed
because of the lower number of the image dataset. For this case, various opera-
tion including rotation, zoom, horizontal flip, width shift, height shift, and shear is
performed to enhance the diversity of the brain tumor dataset.
After image augmentation, the features are automatically extracted by using the
Inception-v3 based transfer learning model. The model utilizes two or three layers
of a small convolutional layer based on factorized convolution operation instead of a
large convolution layer that reduces the parameter without reducing the efficiency of
the model. The factorization process is represented by the block as shown in Fig. 2a.
This model also employed a grid size reduction technique for mapping the features
930 P. Dutta et al.
Inception-v3
(a)
(b) (c)
Fig. 2 a The architecture of proposed classification model, b The detail of block, and c The detail
of block_1
as shown in Fig. 2b. For this factorization process, the 42 layers Inception-v3 model
with fewer parameters is much more efficient than the VGGNet [7]. Moreover, the
default weights are used for this particular model. This model contains many filters to
detect simple feature that is very effective for classification problem. For this reason,
the images are convoluted to extract the desired features.
In order to make the classifier more effective, the output of the Inception-v3 model
is coordinated with the customized DNN that results in fine tuning of the transfer
learning model. Figure 3 shows the network architecture of DNN with four layers
where the first layer is used to attend the output of the inception model and the second
layer consisting of 1024 neurons and the next layer is utilized to dropout. Finally,
the output layer is designed to generate the output with 3 neurons that represents the
brain tumor types of glioma, meningioma, and pituitary respectively.
For training the network, the Adam is employed as an optimizer and categorical
cross-entropy is utilized as a loss function with the learning rate of 0.00003 and
batch size of 32. The training process of the designed model is performed in the
Multi-classification of Brain Tumor Images … 931
Google Colab environment using the Graphics Processing Unit (GPU). The training
process takes about 6 min for 19 epochs (callback stop) with 20 s per epoch. During
the training, the model occupied 3.06 GB RAM and 2.53 GB GPU in the colab
environment.
3 Results Analysis
After completing the training process, the loss and accuracy of the designed classifier
model are evaluated at different epochs as shown in Fig. 4. The training and validation
loss are found 0.0614 and 0.1468 that ensure optimum performance of the classifier
model since the validation loss shows a decreasing outline with the increasing of
the number of epochs. Moreover, the accuracy is obtained 97.80% and 96.25% for
training and validation sets that ensure a good fit of the classifier model because both
sets show an increase of accuracy value for every epoch.
Table 1 Precision and recall of the proposed classifier set at different epochs
Epoch At epoch At epoch At epoch At epoch At epoch At epoch At epoch
1 4 7 10 13 16 19
Training 0.7859 0.9166 0.9490 0.9524 0.9644 0.9726 0.9816
precision
Validation 0.8094 0.9373 0.9522 0.9343 0.9161 0.9195 0.9639
precision
Training 0.6663 0.9013 0.9412 0.9466 0.9621 0.9718 0.9755
recall
Validation 0.7276 0.9266 0.9429 0.9282 0.9086 0.9135 0.9592
recall
For evaluating the performance of the proposed model, various metrics including
precision, recall are obtained at different epochs shown in table 1. The precision
and recall of the training set are found 98.16% and 96.39% and for validation set
are found 97.55% and 95.92% respectively. Moreover, the F1 Score for training
and validation set is calculated 97.85% and 96.15% respectively depending on the
following equation.
precision * Recall
F1 Score = 2 × × 100% (1)
Pricision + Recall
Table 2 shows the comparison of the proposed work with some existing research
works based on the classification method and performance parameter. From this table,
it is shown that the accuracy is quite improved by using the transfer learning-based
CNN classifier proposed in this work. Therefore, the proposed method is proficient to
significantly improve the multi-classification accuracy comparing with other models
presented in the table.
4 Conclusion
References
1. Siegel RL, Miller KD, Jemal A (2020) Cancer statistics. CA A Cancer J Clin 70:7–30 https://
doi.org/10.3322/caac.21590
2. Sajjad M, Khan S, Muhammad K, Wu W, Ullah A, Baik SW (2019) Multi-grade brain tumor
classification using deep CNN with extensive data augmentation. Journal of computational
science 1(30):174–182
3. Amin J, Sharif M, Gul N, Yasmin M, Shad SA (2020) Brain tumor classification based on DWT
fusion of MRI sequences using convolutional neural network. Pattern Recogn Lett 1(129):115–
122
4. Sultan HH, Salem NM, Al-Atabany W (2019) Multi-classification of brain tumor images using
deep neural network. IEEE Access. 27(7):69215–69225
5. Abiwinanda N, Hanif M, Tafwida Hesaputra S, Handayani A, Mengko TR (2018) Brain tumor
classification using convolutional neural network. In: World congress on medical physics and
biomedical engineering 2018. Springer, Singapore, pp 183–189
6. Cheng J (2017) Brain tumor dataset (version 5) 2017. https://doi.org/10.6084/m9.figshare.151
2427.v5.
7. Tsang S-H (2018) Review: inception-v3 — 1st Runner Up (Image Classification) in
ILSVRC 2015. https://sh-tsang.medium.com/review-inception-v3-1st-runner-up-image-classi
fication-in-ilsvrc-2015-17915421f77c. Accessed 10 September 2018
8. Talo M, Yildirim O, Baloglu UB, Aydin G, Acharya UR (2019) Convolutional neural networks
for multi-class brain disease detection using MRI images. Comput Med Imaging Graph. https://
doi.org/10.1016/j.compmedimag.2019.101673
9. Shahzadi I, Tang TB, Meriadeau F, Quyyum A (2018) CNN-LSTM: cascaded framework for
brain Tumour classification. In: 2018 IEEE-EMBS conference on biomedical engineering and
sciences (IECBES) 2018 Dec 3. IEEE, pp 633–637
A Meta Model Based Particle Swarm
Optimization for Enhanced Global
Search
Rasel Ahmed , Shuhaimi Mahadzir ,
and Nor Erniza Mohammad Rozali
1 Introduction
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 935
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_88
936 R. Ahmed et al.
ways: tuning of parameters and parameter control [5]. Traditionally, researchers used
to do fine tuning of algorithm parameters manually to get better performance of the
algorithm. Shi and Eberhart conducted studies to manually improve the performance
of algorithms [7, 8]. Some researchers analyzed particle swarm optimization (PSO)
and showed how the parameters affect the convergence of particles to the global
optimum [9, 10]. However, these approaches are sometimes not systematic and have
limited scope. Another widely used path followed by researchers is to develop new
variants of algorithms [11, 15]. Two common ways to this are: hybridize one algo-
rithm with another for better performance and controlling the particles contraction in
the swarm [11–13]. This article presents a meta-optimization of PSO and its evalua-
tion on 8 unimodal and 6 multi-modal benchmark functions (Table 1) to test both of
its exploration (global search) and exploitation (local search) ability. The proposed
Input parameters
n =1
If n < Niteration
j =1
No Yes
J < Nsuper_swarm
n=n+1
c1, c2, w, ns
J=J+1
Update links
Evaluation
(a) (b)
(a) (b)
Fig. 3 Convergence curve for Dixon Price and Zakharov benchmark functions
It agreed with the no free launch theorem, that if we want to emphasize any specific
property of any algorithm, we need to sacrifice some other properties as a tradeoff.
4 Conclusion
The main objective of this article is to analyze the performance of meta model based
PSO algorithm in case of optimization benchmark functions. It outperforms the other
algorithms in terms of efficiency, convergence and robustness, it also overcomes
the limitations of manually selecting optimal behavioral parameters. The proposed
A Meta Model Based Particle Swarm … 941
(a) (b)
(a) (b)
Fig. 5 Convergence curve for Branin and Modified Schaffer3 benchmark functions
meta optimization increases the exploration and exploitation ability of the algorithms
by continuously updating the parameter values and by maintaining a good balance
between exploration and exploitation, that makes the meta model based PSO algo-
rithm to outperform other renowned algorithms and keep its performance consistent
with increasing modality and dimensionality.
942 R. Ahmed et al.
(a) (b)
(a) (b)
(a) (b)
References
13. Krink T, Løvbjerg M (2002) The lifecycle model: combining particle swarm optimisation,
genetic algorithms and hillclimbers. In: International conference on parallel problem solving
from nature. Springer, Berlin, Heidelberg, pp 621–630. https://doi.org/10.1007/3-540-45712-
7_60
14. Atashnezhad A, Wood DA, Fereidounpour A, Khosravanian R (2014) Designing and opti-
mizing deviated wellbore trajectories using novel particle swarm algorithms. J Nat Gas Sci
Eng 21:1184–1204. https://doi.org/10.1016/j.jngse.2014.05.029
15. Xu X, Tang Y, Li J, Hua C, Guan X (2015) Dynamic multi-swarm particle swarm optimizer
with cooperative learning strategy. Appl Soft Comput 29:169–183. https://doi.org/10.1016/j.
asoc.2014.12.026
HCShare: Blockchain Technology
in Healthcare System
Abstract On this generous growing world, the medical healthcare system is expe-
diting by driving new modernizations in the healthcare system by providing contem-
poraneous varieties for an inmate’s management. The enhancement of the medical
system is done by collecting enormous preceding data, accommodating those records,
scrutinizing of the data, and interchanging them with the different medical system
layers. In the headway to enhance the medical system with further preservation
of the health records, the Blockchain Technology takes the part in the system. The
prospective system likewise supports to transfer or interchange the data of the inmates
between different healthcare system, which is in the guaranteed way. HCShare is a
confidential, protected, reliable and translucent data interchanging scheme between
the different healthcare organizations in a scattered decentralized mechanism. In
which some of the cryptographic algorithms are used to ensure the protection of the
inmate’s confidential records stored in the ledger and to enhance the data integrity
of the records.
1 Introduction
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 945
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_89
946 R. Sangeetha and M. Krishnamoorthi
development of the healthcare system has delivered by the most looming sciences
like Blockchain Technology, Internet of Things, and Edge Computing techniques.
These are the major technologies drives the healthcare system in the smartest way
with better security and integration among the other healthcare organization’s data
[4]. HCShare Healthcare system is a confidential, protected, reliable and translucent
data interchanging scheme between the different healthcare organizations in a scat-
tered decentralized mechanism [5]. HCShare system helps consistently to distribute
and digitization [6] of the records between multiple medical organizations [6]. The
transaction of the records of the inmate are transferred from one of the hospitals to
other hospitals can be done by the blockchain network. As shown in the Fig. 1, every
data that are stored in the medical organizations are shared in the scattered distributed
blockchain network, in which all the data are stored in the secured mechanism by
encrypting the data of the inmates [2].
The blockchain is referred to as the amalgam of couple of decrepit automation
equally like cryptography [6] and peer-peer communication mechanisms [6]. It is a
kind of scattered storage of records in a distributed-ledger-technology [6] in which
the ledger can be increased generously within a massive collection of data directory.
In the HCShare System model, the blockchain technology helps in the transaction of
the medical records of the patients over the different medical institutions or over the
different clinics depends on the patient’s need about the records. For example, if the
patient has been treated from hospital 1 for over the one years of duration, then in
the critical situation, they have to claim another hospice for a quick treatment. In that
situation the patient can’t able to get all their previous medical reports again by doing
all the test cases of the hospital to ensure his problem, or they can’t go the hospital A
for records of their medical reports. This HCShare system model helps the patients
to get their medical data reports whenever they need through the blockchain system.
HCShare: Blockchain Technology in Healthcare System 947
2 Related Works
In the HS-BC [7] for the electronic medical record [8], that provides the permission
to approach the patients records in a secured manner, by implementing the KUNodes
algorithm [7]. Similarly, in the [9], for securing the data of the inmates the algorithms
likely SHA-256 [9] which is to calculate the hash key value, ECC (Elliptic Curve
Cryptography) algorithm [9] for creating irregularity hash key values and for the
encryption of the records of the patients, the AES (Advanced Encryption Standard)
Algorithm is used. In [9] there have been some of the inadequacy such as absence
of integrity of the records, and the system can’t be able to perform massive requests
simultaneously [2]. When the system uses cloud transaction via a blockchain tech-
nology, it is found that there is no security for the confidential data of the inmate’s
records. In which there is lack of the data integrity such as problems in the correct-
ness of the records, lack of accuracy in the data, lack of consistency of the records.
In some of the data sharing system, also used cloud environment, in which the data
can be presented in the different layers, along with the accessibility to the data [2]. In
such cases, the third party can filch the data, and can misuse the confidential data of
the inmate’s records. The SHealth works on the different layers such as government,
user, and the IoT layers [2]. In the ChainSDI [3], the home-based medical service
sharing programming introduced to manage the confidential data of the inmates [3].
In the Sshealth [4] system, massive records are analyzed also the difficulties in the
management to provide the secured storage for the inmate data preservation [4]. In
the [5], the objections in the blockchain scalability are well focused and found the
issues like size and number of transactions of a block, and the solutions to sort it out
was classified into storage optimization and redesigning blockchain [5].
3 Proposed Model
The proposed research work on the healthcare system is to yield the security of
the inmate data, from the third-party attacks. Figure 2 provides the structure of the
blocks [6] hash values in the blockchain system [10], in which the hash values are
ordered in the Merkle tree [6]. In the HCShare the physician depot the medical reports
in the corresponding servers of the specific hospice, in which the only authorized
administrator who were allowed creates and depot the private passcode to ensure
the safety of every dossier, in the permissioned [8] blockchain system to accomplish
secured scattered ledger of hospice records. So that, every data within the ledger can
be available for the hospice, physician and for the inmate of the hospice can access
them in anytime from anywhere.
948 R. Sangeetha and M. Krishnamoorthi
3.1 Blocks
In the HCShare System model, blocks in the blockchain transaction consists of the
couple of things, which are called as block header [11] and block body [11], in which
it contains the version numbers represented in the blocks, and hash content from the
Merkle tree as shown in Fig. 2, timestamp value and the parental hash key values.
These are shown in Fig. 3.
In the Merkle tree, the hash key generation starts from the nodes in the leaf level,
in which the parental node hash value is calculated from its combination of leaf
nodes, in which manner all the node’s hash value is calculated. For example, assume
that the leaf node of the Merkle tree as H0 , H1 , H2 , H3 , and H4 , in which the parental
node hash value is calculated as follows,
HV01 = Hash_VALUE [HV0 + HV1 ].
HV23 = Hash_VALUE [HV2 + HV3 ].
HV0123 = Hash_VALUE [HV01 + HV23 ].
In which the each and individual hash values of the leaf node contains the hash
value of the transactions. As it contains every transaction details of the blocks, the
complete Merkle tree contains the overall information of all transactions.
For the better secured data transaction among the multiple blockchain system in the
healthcare system, every blocks of the transaction contains the confidential private
key along with the hash values and the transaction details of the blocks. In the
transaction details the inmate’s records are stored according the specific hospice. For
the encryption of the inmate’s records, algorithms such as SHA-256, AES, KUNodes
algorithms are used to ensure the safety and security of the inmate records from the
malicious attacks.
In the encryption process of the HCShare System, pair of different keys has been
introduced to improve the security and safety of the ledger, they are as Priority
hash address key values and Secondary updated hash value. Where, with the help of
priority hash key, one can only view their records, which were updated in the system
by the reputed authorized hospice. So that the inmate and their family can view their
data, when it is needed in the emergency situations. Another secondary updated hash
value will help the hospice authorized admin or the physician to upload the records
the inmates, where only with the help of the secondary hash key value, one can access
completely of the ledger of particular inmate’s data.
In the HCShare system model, we propose SHA-256 for generating the priority
hash address key values, and to generate the secondary key values, KUNodes (Fig. 4),
AES algorithm is used to ensure the data integrity of the patient’s records.
Every block in the blockchain system has its own unique key, in which each node of
the blockchain system is connected to the prior blocks of the system via a individual
hash key data [12]. Ethereum [12] is most widely used blockchain technology to give
a better transaction over the blockchain system. In which the smart contracts [12]
are used to secure the transaction details of the inmates. Smart contracts are used
to execute set of programs in which the program is executed automatically when
it is needed. These smart contracts manage the transaction of the ledger and the
cryptocurrencies [12] of the blockchain system. These programs can be encountered
without depending any other data, which performs and control overall transaction
950 R. Sangeetha and M. Krishnamoorthi
In Fig. 6, the expected authentication delay of the HCShare system will be lesser
than the prior existing system, as the number of transaction increases over the commu-
nication of the inmates with the hospice, the response time is efficient as compared
to other systems.
HCShare: Blockchain Technology in Healthcare System 951
5 Conclusion
In this enormous world, the blockchain became most noteworthy technology in the
internet security [13], where the it gives all kinds of facility in the smart technological
world like in the fields of smart houses, smart grid [13] applications, as well as in
the food safety application [13], and so on. In this proposed HCShare system model,
the system gives better accessibility of the records in the blocks, by attribute-based
encryption [14] technology, in which the people without having the confidential keys
cannot have an access to the data of the inmates. So that the records of the inmates
will be stored in secured manner in the blockchain technology. We propose HCShare
system by using the Ethereum technology for the transaction of the medical reports
over the different medical institutions. Our research work plans to get the better data
integrity of the medical reports of the inmates in which all the records of the inmates
should be transferred from and to different hospice, with the complete records in
a secured manner. This HCShare is to give the better truncation efficiency over a
different block to ensure the performance of the data transaction. The lack of data
integrity is needs to be solved in the HCShare system, in which to overcome these
kinds of issues the data are secured using pair of different addresses to ensure the
safety and integrity of the data.
References
1. Zarour M et al (2020) Evaluating the impact of blockchain models for secure and trustworthy
electronic healthcare records. IEEE Access 8:157959–157973. https://doi.org/10.1109/ACC
ESS.2020.3019829
2. Zghaibeh M, Farooq U, Hasan NU, Baig I (2020) SHealth: a blockchain-based health system
with smart contracts capabilities. IEEE Access 8:70030–70043. https://doi.org/10.1109/ACC
ESS.2020.2986789
952 R. Sangeetha and M. Krishnamoorthi
Abstract BLDC motors possess high torque than the same rated Induction motor.
Due to electronic commutation and the absence of mechanic commutation and
brushes, BLDC motors are widely used in many industrial applications. The main
disadvantage of BLDC motor is high torque ripple. Commutation Ripple reduc-
tion of brushless direct current (BLDC) motor torque using the SAZZ converter is
proposed in this work; torque ripple is proposed to carry out under both conduction
and commutation intervals in this proposed work. The fuzzy logic speed controller
is used for generating reference pulse in the conduction interval by processing the
speed error. The performance of torque and speed based on the proposed control
method was analyzed. MATLAB/Simulink platform is used for implementing as
well as showing effective implementation of the proposed technique. PI and Fuzzy
logic control are used for evaluating the performance. SAZZ (Snubber Assisted Zero
Voltage and Zero Current Transition) topology was proposed to reduce switching
losses. The main switch snubber circuit reduces the voltage during turn off, and the
auxiliary circuit discharges the snubber before the next switching cycle.
1 Introduction
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 953
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_90
954 A. Senthilnathan et al.
cogging, commutation, and switching power devices [1, 2]. Many research works are
progressing to reduce current and torque ripples in PMBLDC motor. In the six-step
inverter conduction topology, at any time, only two phases are in conduction [3–
5]. Generally, the rectangular phase current is considered to be ideal for two-phase
conduction mode. Due to commutation, the incoming phase current’s slew rate time
is not equal to the outgoing phase current, which results in current pulsation [6].
This current pulsation produced torque ripple in the PMBLDC motor. Fuzzy logic
controlled SAZZ converter fed BLDC motor will create fewer torque ripples than
PI- Controller.
SAZZ (Snubber Assisted Zero Voltage and Zero Current Transition) converter is
used to reduce switching losses. The main switch snubber circuit is used to reduce
the voltage during turn off and the auxiliary circuit discharges the snubber before the
next switching cycle. Figure 2 Shows the block diagram of the SAZZ converter fed
SAZZ Converter Fed Fuzzy Logic Speed Controlled BLDC Motor Drive 955
SAZZ
3 Phase BLDC
Battery Con-
Inverter motor
verter
Power
Supply PI/Fuzzy logic controller
BLDC motor. The battery is used as the input source supply and the SAZZ converter
is used to step-up DC voltage and fed the supply to the inverter. Six step inverter is
used to energize the BLDC motor. At any point, only two phases are energized as
shown in Fig. 1. The Hall sensor is used to sense the rotor position and the PI and
fuzzy controller is used to control the MOSFET driver circuit. The error and change
in error are taken as two inputs to the controller and the output of the controller is
used to fire switches in SAZZ converters. The block diagram of proposed work is
shown in Fig. 2.
The reference voltage of the inverter is produced based on the speed received from
the brushless DC motor. The inverter’s output voltage is compared with the refer-
ence voltage, and an error voltage is generated. This error is used to control the PI
controller.
The PI controller has a proportional gain and an integral gain. The proportional
gain (KP = 0.8) supports the analog error function and an integrator (KI = 1.1)
to reduce the constant level error value to zero. Based on the fault voltage, the PI
controller produces pulses that regulates the MOSFET. Although the PMDC motor
current follows the change in torque and change in voltage, the armature current from
Fig. 3a has different deformations, and Fig. 3b and c back EMF and torque are more
distorted and take longer to resolve.
956 A. Senthilnathan et al.
Time in sec
b) Back EMF
Ea (V)
Time in sec
c) Torque
Te(N-m)
Time in sec
Fig. 3 Performance of PI controller based SAZZ converter fed BLDC motor
3 FLC Operation
Fuzzy logic is the usually available control approach for the above two decades. Fuzzy
logic consists of fuzzification, inference engine, and defuzzification. The Mamdani
fuzzy control system is used here. The input will be given to a fuzzifier [7–9]. A
fuzzy rule base contains the fuzzy rule set which we have to design. This input is
given to a fuzzy interface engine that compares the input with the fuzzifier inputs
to give the required output to the defuzzifier. The defuzzifier gives the output after
defuzzification and gives the respective output to the next system. The same input is
obtained at the output but after fuzzification i.e. the character of the output does not
SAZZ Converter Fed Fuzzy Logic Speed Controlled BLDC Motor Drive 957
change. This output is given to the converter to boost up the voltage and obtain the
required output. The Fuzzy Logic Control is designed by MATLAB R2014a.
Where,
• N B- negative Big
• N M- negative Medium
• N S- negative Small
• Z E- zero
• P S-positive Small
• P M- positive Medium
• P B- positive Big
Figure 4 shows the Fuzzy partition. Each input and output set is assigned with
seven linguistic variables and forty-nine rules are enclosed in FLC. Mamdani algo-
rithm is used to design the set of rules for the FLC. The fuzzy partition is shown in
Fig. 4 which includes all the 7 sets of variables that use the IF–THEN logic to design
the 49 sets of fuzzy rules.
From Fig.5a it is very clear that stator current is almost rectangular in nature.
From Fig. 5c it is very clear that percentage average speed error is less than 1% and
the speed follow the command as fast as possible. Moreover the step change from
200V to 150 V at t = 1 sec produce a change in speed within 0.25 s.
4 Conclusion
A Fuzzy Logic Controller based SAZZ converter for BLDC motor has been designed
in this paper. This technique has been implemented as it proves to be more efficient
than any other method. By employing the Fuzzy Logic Controller (FLC), the varying
outputs can be brought in control to get a constant output. Here two sets of input
blocks have been used where several voltage variations have been obtained on the
input side in an FLC. Thus the simulation and its outputs have shown that even
though voltage variations occur to a large extent, the output tends to remain the same
throughout the whole process.
958 A. Senthilnathan et al.
Time in sec
b) Back EMF
Ea
Time in sec
c) Speed
Speed (rpm)
Time in Sec
Fig. 5 Performance of Fuzzy logic controller based SAZZ Converter fed BLDC motor
References
5. Vasallo Vázquez MJ, Andújar Márquez JM, Segura Manzano F (2008) A methodology for
optimizing stand-alone PV-system size using parallel-connected DC/DC converters. IEEE Trans
Ind Electron 55(7)
6. Senthilnathan, Palanivel P (2020) A new approach for commutation torque ripple reduction of
FPGA based brushless DC motor with outgoing phase current control. Microprocess Microsyst
75
7. Veerachary M, Senjyu T, Uezato K (2003) Neural-network-based maximum-power-point
tracking of coupled-inductor interleaved-boost-converter-supplied PV system using fuzzy
controller. IEEE Trans Ind Electron 50(4)
8. Muyeen SM, Al-Durra A (2013) Modeling and control strategies of fuzzy logic controlled
inverter system for grid interconnected variable speed wind generator. IEEE Syst J 7, No. 4,
Dec. 2013.
9. Mutoh N, Ohno M, Inoue T (2006) A method for MPPT control while searching for parameters
corresponding to weather conditions for PV generation systems. IEEE Trans Ind Electron 53(4)
Pedestrian Re-identification in Video
Surveillance System with Improved
Feature Extraction
1 Introduction
Research in the area of full body pedestrian Re-identification (re-id) has been thriving
in recent years. Pedestrian re-id aims at identifying full body shots of people in
non-overlapping camera views [1–7]. There are a number of challenges involved in
identifying people in a new location from a different angle, such as drastic appearance,
lighting and image quality changes. These types of variations can make the task of
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 961
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_91
962 S. Salehian et al.
person re-id seem quite challenging. The main goal of any re-id system is to identify
a subject based on a few sample images that it has seen of that subject before, these
images could be from a different camera at a different location. Pedestrians re-id is
of great importance for applications such as surveillance, security systems, activity
analysis and tracking. Some of the other non-obvious challenges are: pose, occlusion,
activity, angle and background variations in two images from two different cameras
[8–12].
In this work, we are investigating the video-based re-id, by implementing the state-
of-the-art image-based re-id techniques. Image based re-id is a more challenging task
compare to video-based re-id due to the lack of information in the very few images
available (in some cases only a single image) for each identity and also the lack of
spatial and temporal information. However, an image-based re-id model would be
more practical in a scenario with a small database of each ID. Moreover, the public
datasets that are typically used by the image-based re-id models are much more
diverse in terms of pose variation, lighting condition and angle variation, as compare
to the datasets used by the video-based re-id models. Thus, alleviating biases in our
model. Hence, our motivation for applying image-based re-id to perform re-id on a
video [1–3, 13, 14].
The first step in re-id is to create descriptors that characterize the appearance of a
person based on color and textural information of the person’s clothes. The descriptor
has to have high discriminative power in order to distinguish different people, but at
the same time be able to handle variations in pose, angle, lighting and etc. [15].
Typically, the images have large variation of poses and are low quality; conse-
quently, it has been proven by both, machine learning methods and also the conven-
tional image processing methods that color and texture are the two most important
features for the task of person re-id [15] and [16].
Recent work in the area of pedestrian re-id has been mainly on either creating
discriminative features that are view, lighting, pose and background in variance [4, 6,
8, 11] or producing a distance metric between two positive or negative pairs of images
[5, 9, 10]. In some cases, both of these methods are being implemented together, to
achieve a better performance [1–3, 7, 12, 17]. Distance metric learning has proven
to outperform discriminative feature learning in most cases. Hence, the focus of this
work is on model utilizing both, distance metric and discriminative feature learning.
Using a single dataset that is bias to a particular pose, feature, background or
appearance would not generalize well to real world scenarios. Thus, it is important
to use a well-mixed batch of training data, in order to avoid bias towards a particular
category. For example, most of the pedestrian re-id datasets suffer from this problem
because either they are captured in an environment with a particular group of people
such as CUHK1 [18] and CUHK3 [19] in which most the subjects are university
students that are carrying backpacks or contains constant background information
like crosswalks in PRID [20] dataset. VIPeR [21] dataset is a good example of cases
where there is significant difference in resolution between the images. One solution
is to use a combination of these datasets to create a more diverse dataset which in
turn would help the model learn more robust features [22] the other solution is to use
better and deeper models.
Pedestrian Re-identification in Video Surveillance … 963
2 Related Work
3 Methodology
cameras the subject is then detected and tracked throughout the field of view of that
camera. The cropped images generated by the detector are extracted, labeled and
stored in a database.
The pedestrian detector estimates the ROI (region of interest), which would be
a bounding box around the area with a person in a given frame. The tracker keeps
track of the identity of each box in consecutive frames in a video [32]. There are now
multiple images from subject in our database.
At this stage the database contains images of the subject, these images can now
be used to identify that subject in different cameras and locations. The new images
are passed to the re-id network to be compared against the database to be identified.
As shown in Fig. 1, the two images are resized appropriately to the input size of
the network, tabulated in Table 1. Each image is then fed to a CNN, which would
result in a feature map. The two feature maps are then stacked on top of one another
to form a cross-neighborhood difference layer based on the method described in
the original paper [1]. The result is then fed into a fully connected layer to make a
final classification. A positive match indicates that the two input images are from a
same individual and a negative match indicates that they are not depicting the same
person. The final stage is a SoftMax classification which would result in a probability
between 0–1 for each class.
In this paper we are going to focus on the re-id block which consists of 2 parts.
• Feature extraction stage
• Difference learning stage
We have implemented the difference learning stage similar to the work done
by [1], by computing cross-input neighborhood differences from the output of the
feature extraction stage, which is then fed to a patch summary feature layer which
summarizes the neighborhood difference maps by producing a holistic representation
of the differences in each block. It is then followed by another convolutional layer
in order to incorporate the spatial information into the model. This output is then
connected to a fully connected layer, to perform the final classification of whether
or not the 2 images belong to the same ID.
In this work we are going to mainly focus on the feature extraction stage, which
in the case of our baseline is just 2 convolutional layers. We have experimented with
3 different convolutional neural networks (CNNs). Some of the main parameters
of these networks are sown in Table 1. With the goal of achieving a much higher
accuracy of re-id by learning a deeper feature map from the input.
A pair of images are fed through the network, they each go through the same CNN
network and a feature map is extracted from each image. These 2 feature maps are
then passed to the difference learning stage. We have replaced the CNN used in the
baseline with AlexNet, DenseNet121 and InceptionV3. The networks have to first
be truncated because the difference learning stage expects a feature map and all of
the networks mentioned above are designed as classifiers with fully connected layers
at the end. In all 3 cases the output was taken from the last layer before the fully
connected layers (no top). In the case of AlexNet the original network has a max
pooling layer before the fully connected layers in order to reduce the dimensions of
the convolutional feature map so for AlexNet the output was taken from the max
pooling layer. The network with no top is then connected to the rest of the summary
network. The dimension of the feature map is different for each of the networks refer
to Table 1 for input, feature map dimensions.
The database of new IDs that are saved during registration contains a 50 image for
each subject. When a new image is being evaluated it has to be checked against all the
images in the database. Every comparison result in a probability score which is then
used to calculate an average for the 50 images that belong to each ID. The ID with
the highest average probability score is chosen as the final prediction. The accuracy
score is calculated based on the rate of correctly identified images in a batch.
In CUHK03, the dataset that is used for training, there are 13,164 images of
1,360 pedestrians. Therefore, the number of negative pairs is much greater than the
positive pairs. If the model is trained on this dataset, it will just learn to identify every
image as a mismatch which is an example of a data imbalance and overfitting. We
performed a 2D translation on the dataset to produce 5 images from every image,
based on [1]. However, the dataset is still imbalance and the negative pairs have to be
down sampled to just twice the number of positive pairs. We randomly divide 1360
identities into non-overlapping train (1160), test (100), and validation (100) sets.
966 S. Salehian et al.
4 Results
The performance of these models was evaluated on the test data from the training
dataset and also our own dataset that we created. Our dataset is divided into 2 Batches
as follows: In batch number 1, all the videos have the same background with two IDs
moving from side to side, recorded from different distances of 4, 6 and 8 m. There
are also different conditions such as single subjects moving from left to right and
vice versa, complete occlusion of either subjects in the middle of the frame by the
other subject and subjects following each other. For batch number 1 all the videos
are recorded 1280 × 720 pixels resolution. In batch number 2, there are multiple
backgrounds and for each background there are multiple videos of two subjects
(different from batch number 1) walking from side to side at different distance (3,
5 or 7 m). However, there are no special conditions. All the videos in this batch are
recorded with a resolution of 320 × 240 pixels. Figure 2 depicts some sample images
from our dataset.
In order to select the best Re-id dataset for the purpose of training, the baseline
model was trained and on 3 different datasets and tested on our own dataset. The
result from Table 2 shows that Market1501 performed well on ID0-2 but it had a
very bad performance on ID3. CUHK03 had an overall higher accuracy as compare
to CUHK01 hence, it was chosen as the primary training dataset. Choosing a diverse
dataset is an important factor in order to reduce unwanted biases, which could lead
to an underperforming re-id system.
Figure 3 also shows the accuracy for each ID which is a measure of how many
times the model correctly identified that ID for the given number of images.
Finding a perfect dataset with no bias is nearly impossible. Looking at the test
result from Fig. 3 we can observe that selecting a better model can alleviate some
of these biases. Ideally, we expect that the accuracy of the re-id be closer to a single
value. However, we can see that in the case of the baseline and DenseNet121 model
there is a 12% difference between the highest and lowest values. Whereas in AlexNet
this difference is only 6%.
The second important criteria in evaluating an identification system is space
between each user. In short, the difference in prediction probability of the model for a
match and non-match should be maximized. Table 3. Shows the average probability
of each ID compared to all the IDs in the database.
For example, for ID 0 the average probability of all those images when they
pass through the baseline model is 96.50% meanwhile, the second highest average
probability for the same images is for ID_1 at 39.15% which means that the network
is very confident in distinguishing between images of ID 0 and all the other 3 IDs
in the database. However, the same cannot be said about ID2 where the average
probability of ID_2 and ID_3 are very close 98.54% and 95.87% respectively. This
is mainly caused by the bias in CUHK03 dataset that different subjects wear clothes
with different colors and ID 2 and 3 have similar color clothes which increases the
chance of misclassification. By optimizing the models, we can see that in all 3 cases
AlexNet, DenseNet121 and InceptionV3 the differentiation between the IDs is more
pronounced.
The probability column shows the average probability of those images belonging
to each of the 4 IDs. Ideally the correct ID should have a probability of 100% and the
rest be 0%. Therefore, a larger difference between these probabilities would ensure
a lower chance of misclassification.
Typically, the datsets are recorded either in a university, a shopping mall or in the
street and the subjects are wearing various color clothes and they appear in different
angles and distance. This is good for training but in order to investigate what are
the shortcomings of the system an engineered dataset like the one used in this paper
helps with analyzing each parameter separately.
Pedestrian Re-identification in Video Surveillance … 969
Baseline AlexNet
DenseNet121 IncepƟonV3
All the networks are trained on CUHK03 and the test accuracies are listed in Table
7. DenseNet121 performed the best with an accuracy of 92.09 as compare to the 88.7
which is the highest accuracy reported in the literature.
Table 6. AlexNet has the highest accuracy with InceptionV3 and baseline coming
at second and third place respectively and DenseNet121 had the lowest accuracy on
our dataset.
Figure 4 shows the accuracy of re-id of different models at different distances. Our
dataset is recorded at different distances which enables us to investigate the effects of
distance on performance of re-id networks. The database of existing IDs that every
new image is compared against is consist of images from ID0 at 8 m and right side
towards the camera, ID1 at 8 m and left side towards the camera, ID2 at 5 m and
right side towards the camera and ID3 at 5 m and right side towards the camera
As evident by Fig. 4 and Table 4 for ID 2 and 3 the accuracy is higher at 5 m
as compare to 7 m. This is mainly due to the lower quality of the images which
causes the far objects to be a lot blurrier than the images at 5 m. However, for ID2
the difference between the average probabilities are much closer at 5 m than they are
at 7 m. This is true for all of the models which means the models are struggling to
distinguish between images of ID2 and ID3 at 5 m. The main reason for this unusual
behavior is the fact that the images of ID 2 and 3 at 5 m have a lot of more similar
features which results in a higher probability on both IDs.
The accuracy of the re-id for different point of view where in one case the camera
sees the right side of the person and another that sees the left are shown in Fig. 5.
As mentioned earlier the database contains only images from one side therefore this
test can demonstrate the importance of having images from different angle in the
database.
Looking at Fig. 5 we can see a pattern where the accuracy is either higher or similar
when evaluating images with the same view. We can also observe that InceptionV3
970 S. Salehian et al.
Baseline AlexNet
DenseNet121 IncepƟonV3
has a more uniform characteristic which results a similar accuracy for each ID,
regardless of the side.
Table 5 shows that the difference between the average probabilities are above
10% for AlexNet, DenseNet121 and InceptionV3 which is an improvement over the
7% of the baseline model. The models are capable of handling variations in viewing
direction which is evident by both the accuracy result and the difference in average
probability values (Tables 6 and 7).
As we mentioned earlier, our dataset is consisting of 2 batches that are quite
different in a few ways. Table 6 shows that all the 4 networks perform relatively
good on batch number 1 with 3 of them having an accuracy of above 97% and only
DenseNet121 is lower at around 95.6%. This could be either because batch 1 uses a
similar background for all the images or because the 2 IDs are wearing distinctive
clothing or just the fact that batch 1 has images with higher quality. As for batch 2,
AlexNet performed the best with an accuracy of almost 95% which is only 3% lower
than its batch 1 accuracy. InceptioV3 came second, with an accuracy around 91% in
this case 8% lower than its batch 1 accuracy. The baseline and DenseNet121 had an
accuracy around 90, with their accuracies dropping by 7 and 6% compare to batch 1
respectively. Because AlexNet had a high performance on both batches it has a much
higher overall accuracy than the other networks.
In terms of difference in average probability all three networks outperformed the
baseline based on values in Table 3.
As shown in Table 7 all three networks AlexNet, DenseNet121 and Incep-
tionV3 have outperform previous networks in the literature. This is mainly because
these designs are much deeper and more computationally expensive than the other
networks. Which is a tradeoff between speed and performance. DenseNet + SumNet
achieved the highest accuracy with AlexNet + SumNet in the second place and
InceptionV3 + SumNet in third.
Looking at Figs. 3, 4 and 5 we can see a more uniform pattern in results from
AlexNet and InceptionV3 where the results from different IDs are closer to one
another.
972 S. Salehian et al.
Table 7 Overall
Method Rank-1 (%)
accuracy comparison with
other re-id methods IDLA [1] 54.74
CAN [3] 77.6
MLS [4] 87.50
PAR [5] 85.4
MSDL [6] 76.87
DCAF [8] 74.21
PersonNet [10] 64.80
LDNS [12] 62.55
Gated Siamese [14] 68.1
PDC [33] 88.7
DPFL [34] 86.70
DeepAlign [35] 85.40
Spindle [36] 88.50
JLML [37] 83.20
AlexNet + SumNet 91.89
DenseNet121 + SumNet 92.09
InceptionV3 + SumNet 90.18
The networks were trained using the following parameters, softmax cross-entropy
loss function, Momentum Optimizer, learning rate 0.001 with a batch size of 12
for DenseNet121 and Inception, learning rate of 0.01 with a batch size of 150 for
AlexNet. The networks were all pretrained on ImageNet which helps them converge
faster. AlexNet was trained end to end for 100,000 iteration but for DenseNet121
974 S. Salehian et al.
and Inception only the top of the network was trained for the first 100,000 iterations
by freezing the weights in the CNN portion of the network, it was then trained end
to end for an additional 100,000 iterations.
5 Conclusion
References
1. Ahmed E, Jones M, Marks T (2015) An improved deep learning architecture for person re-
identification. In: 2015 IEEE conference on computer vision and pattern recognition (CVPR),
Available: https://doi.org/10.1109/cvpr.2015.7299016
2. Lukezic A, Vojir T, Zajc LC, Matas J, Kristan M (2017) Discriminative correlation filter
with channel and spatial reliability. In: 2017 IEEE conference on computer vision and pattern
recognition (CVPR)
3. Liu H, Feng J, Qi M, Jiang J, Yan S (Jul. 2017) End-to-end comparative attention networks for
person re-identification. IEEE Trans Image Process 26(7):3492–3506
4. Guo Y, Cheung N-M (2018) Efficient and deep person re-identification using multi-level
similarity. 2018 IEEE/CVF conference on computer vision and pattern recognition
5. Zhao L, Li X, Zhuang Y, Wang J (201) Deeply-learned part-aligned representations for person
re-identification. In: 2017 IEEE international conference on computer vision (ICCV)
6. Qian X, Fu Y, Jiang Y-G, Xiang T, Xue X (2017) Multi- scale deep learning architectures for
person re-identification. In: The IEEE international conference on computer vision (ICCV)
7. Chen W, Chen X, Zhang J, Huang K (2017) Beyond triplet loss: a deep quadruplet network for
person re-identification. In: The IEEE conference on computer vision and pattern recognition
(CVPR
Pedestrian Re-identification in Video Surveillance … 975
8. Li D, Chen X, Zhang Z, Huang K (2017) Learning deep context-aware features over body and
latent parts for person re-identification. In: The IEEE conference on computer Vi- sion and
pattern recognition (CVPR)
9. Chen W, Chen X, Zhang J, Huang K (2017) A multi-task deep network for person
re-identification. In: AAAI conference on artificial intelligence (AAAI)
10. Wu L, Shen C, van den Hengel A (2016) PersonNet: person re-identification with deep
convolutional neural networks. Comput Vis Pattern Recogn
11. Matsukawa T, Okabe T, Suzuki E, Sato Y (2016) Hierarchical gaussian descriptor for person
re-identification. In: Proceedings of the IEEE conference on computer vision and pattern
recognition, pp 1363–1372
12. Zhang L, Xiang T, Gong S (2016) Learning a discriminative null space for person re-
identification. CVPR
13. Xiao T, Li H, Ouyang W, Wang X (2016) Learning deep feature representations with domain
guided dropout for person re-identification. In CVPR
14. Varior RR, Haloi M, Wang G (2016) Gated siamese convolutional neural network architecture
for human re- identification. In: European conference on computer vision (ECCV)
15. Xiong F, Gou M, Camps O, Sznaier M (2014) Person re-identification using kernel-based
metric learning methods. Comput Vis ECCV 1–16
16. Liao S, Hu Y, Zhu X, Li S (2015) Person re-identification by local maximal occurrence represen-
tation and metric learning. In: 2015 IEEE conference on computer vision and pattern recognition
(CVPR)
17. Li Z, Chang S, Liang F, Huang T, Cao L, Smith J (2013) Learning locally-adaptive decision
functions for person verification. In: 2013 IEEE conference on computer vision and pattern
recognition
18. Guillaumin M, Verbeek J, Schmid C (2009) Is that you? Metric learning approaches for face
identification. In: 2009 IEEE 12th international conference on computer vision
19. Chen D, Yuan Z, Hua G, Zheng N, Wang J (2015) Similarity learning on an explicit polynomial
kernel feature map for person re-identification. In: 2015 IEEE conference on computer vision
and pattern recognition (CVPR)
20. Chen J, Zhang Z, Wang Y (2014) Relevance metric learning for person re-identification by
exploiting global similarities. In: 22nd international conference on pattern recognition
21. Li W, Zhao R, Xiao T, Wang X (2014) DeepReID: deep filter pairing neural network for person
re-identification. In: CVPR
22. Lukežič A, Vojíř T, Čehovin Zajc L, Matas J, Kristan M (2018) Discriminative correlation filter
tracker with channel and spatial reliability. Int J Comput Vis 126(7):671–688
23. Xiong F, Gou M, Camps O, Sznaier M (2014) Person re-identification using kernel-based metric
learning methods. Comput Vis ECCV, pp 1–16, 2014. Available: https://doi.org/10.1007/978-
3-319-10584-0_1
24. Liao S, Hu Y, Zhu X, Li SZ (2015) Person re-identification by Local Maximal Occurrence
representation and metric learning. In: 2015 IEEE conference on computer vision and pattern
recognition (CVPR), Boston, MA, pp 2197–2206. https://doi.org/10.1109/CVPR.2015.729
8832
25. Matsukawa T, Okabe T, Suzuki E, Sato Y (2020) Hierarchical gaussian descriptors with applica-
tion to person re-identification. IEEE Trans Pattern Anal Mach Intell 42(9):2179–2194. https://
doi.org/10.1109/TPAMI.2019.2914686
26. Li Z, Chang S, Liang F, Huang TS, Cao L, Smith JR (2013)Learning locally-adaptive decision
functions for person verification. In: 2013 IEEE conference on computer vision and pattern
recognition, Portland, OR, pp 3610–3617. https://doi.org/10.1109/CVPR.2013.463
27. Guillaumin M, Verbeek J, Schmid C (2009)Is that you? metric learning approaches for face
identification. In: 2009 IEEE 12th international conference on computer vision, Kyoto, pp
498–505. https://doi.org/10.1109/ICCV.2009.5459197
28. Chen D, Yuan Z, Hua G, Zheng N, Wang J (2015) Similarity learning on an explicit polynomial
kernel feature map for person re-identification. In: 2015 IEEE conference on computer vision
and pattern recognition (CVPR), Boston, MA, pp 1565–1573. https://doi.org/10.1109/CVPR.
2015.7298764
976 S. Salehian et al.
29. Chen J, Zhang Z, Wang Y (Dec. 2015) Relevance metric learning for person re-identification
by exploiting listwise similarities. IEEE Trans Image Process 24(12):4741–4755. https://doi.
org/10.1109/TIP.2015.2466117
30. Shi H et al (2016)Embedding deep metric for person re-identication a study against large
variations. Available: https://arxiv.org/abs/1611.00137v1
31. Varior RR, Shuai B, Lu J, Xu D, Wang G (2016) A siamese long short-term memory architecture
for human re-identification. In: Leibe B, Matas J, Sebe N, Welling M (eds) Computer vision—
ECCV 2016. ECCV 2016. Lecture Notes in Computer Science, vol 9911. Springer, Cham.
https://doi.org/10.1007/978-3-319-46478-7_9
32. Salehian S, Sebastian P, Sayuti AB (2019)Framework for pedestrian detection, tracking and
re-identification in video surveillance system. In: 2019 IEEE international conference on signal
and image processing applications (ICSIPA), Kuala Lumpur, Malaysia, pp 192–197
33. Su C, Li J, Zhang S, Xing J, Gao W, Tian Q (2017) Pose-driven deep convolutional model for
person re-identification. In: 2017 IEEE international conference on computer vision (ICCV),
Venice, pp 3980–3989. https://doi.org/10.1109/ICCV.2017.427
34. Chen Y, Zhu X, Gong S (2017) Person re-identification by deep learning multi-scale repre-
sentations. In: 2017 IEEE international conference on computer vision workshops (ICCVW),
Venice, 2017, pp 2590–2600, https://doi.org/10.1109/ICCVW.2017.304
35. Zhao L, Li X, Wang J, Zhuang Y (2017) Deeply-learned part-aligned representations for person
re-Identification. In: The IEEE international conference on computer vision (ICCV), pp.3219–
3228
36. Zhao H, et al (2017) Spindle net: person re-identification with human body region guided
feature decomposition and fusion. In: 2017 IEEE conference on computer vision and pattern
recognition (CVPR), Honolulu, HI, pp 907–915. https://doi.org/10.1109/CVPR.2017.103
37. Li W, Zhu X, Gong S (2017) Person re-identification by deep joint learning of multi-loss clas-
sification. In: Proceedings of the 26th international joint conference on artificial intelligence,
pp 2194–2200
Smart Children Management Using Data
Analytics, Machine Learning and IoT
Abstract One of the biggest challenges faced by parents is managing the difficult
or defiant behavior of children of the twenty-first century. Technology can help
to provide support in managing children’s activities and helping parents to be in
control. Recent advances in IoT have sparked the interest of researchers to develop
a smart management system for children. This paper presents a review of the smart
management of children using the internet of things (IoT), machine learning, and
Data Analytics. The objective is to identify potential approaches regarding smart
management for children based on IoT for future researchers. The reviews were
conducted from secondary sources, where conference proceedings and journals have
been analyzed from Scopus and other databases. The finding indicates that IoT-based
smart system has robust accuracy and performance. Thus, exploiting IoT for the
development of a smart management system for children will provide significant
benefits to the future generation.
1 Introduction
Children’s programs are now more important to parents as they are encouraged to
engage their children in as many activities as possible to help them plan for their
future [1, 2]. Many studies have described various methods used by families to handle
all the work they need to do around the overlapping duties of work, education, home,
and enrichment events [1–7]. Managing the challenging or rebellious actions of
Md J. H. Faruk
College of Computing and Software Engineering, Kennesaw State University, Marietta, USA
e-mail: [email protected]
M. H. M. Adnan (B)
Computing Department, Faculty of Art, Computing and Creative Industry, Universiti Pendidikan
Sultan Idris, Tanjung Malim, Malaysia
e-mail: [email protected]
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 977
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_92
978 Md J. H. Faruk and M. H. M. Adnan
twenty-first century children is one of the greatest challenges faced by parents [8].
As information and communication technologies (ICT) and smart devices can be an
integral part of self-management plans, technology can help support the management
of children’s activities and help parents to be in charge [1–9].
The concept of a smart system has been proposed by many scholars where the
generic term ‘smart system’ involves several classes of systems composed of sensors,
actuators, a network, and a signal processing unit that delivers definite signals and
fulfill subtasks or complete tasks [10]. Smart systems may visualize as an integration
of the ideas of smart people, smart environments, and smart living and it expands into
not only IoT but also data analytics and machine learning which is contextualized
with an open and comprehensive network of intelligent objects that can auto-organize,
share information, data and resources while reacting and acting in face of situations
and changes in the environment [4–13].
In this paper, we present potential approaches towards an ideal smart management
system for children that shall help future researchers to get a comprehensive knowl-
edge of plenty of existing techniques. Figure 1. presents some potential approaches
for the smart management of children.
The paper is organized as follows: In the second section, potential methods of
IoT based on the smart management system for children are presented followed by
presenting the discussions. The last section provides the conclusion.
Bed-Activities and Vital Signs Monitoring System: Monitoring sleep and physical
activities are vital because of their uncertain moves and emerging technologies offer
new opportunities for parents to monitor children regardless of their position and
pursuit. Besides, monitoring vital signs also significant as indicators of one’s health
condition and assurance of proper circulatory, respiratory, neural and endocrinal
functions towards keep track of the children’s health condition and it refers to the
parameters of the body’s basic functions [14]. A group of researchers from the Univer-
sity of Georgia proposed vibrations based on a real-time, contactless bed-mounted
monitoring system that can estimate vital signs, body movement, falls from bed,
posture changes and on/off bed detection using during sleep time [15, 16].
One of its primary components is Raspberry Pi 3 that connects with a seis-
mometer to get real-time measurement data, maxima statistics and an instanta-
neous property-based approach towards enhancing the quality of monitoring sleep
condition and posture. Envelop based approach is integrated to estimate BPMh for
measuring heart rate (HR) [16]. System evaluation indicates its robust accuracy of
both approaches to monitoring bed activities and detecting vital signs and body move-
ments. This approach can be one of the potential components of a smart management
system because of its robust filtering techniques to estimate the heart rate (HR) and
respiration rate (RR) as well as other sleep activities.
Health Care Monitoring System: According to the National Academy of Science
[17], In addition to supporting this viewpoint, wellness during childhood sets the
stage for adult health, but also provides an important legal, social and economic
obligation to ensure that all children are as healthy as they can be. Amrita University
researchers [18] present an IoT-based smart and secure health care monitor system
for children to monitor behavioral and health issues and predict child disorders.
C4.5 algorithm has been adopted to predict the disorder of children by analyzing
collected therapeutic games and bodily parameter data. Besides. an android game
has been presented that enables the device to evaluate the disorders of the child using
visuospatial and strop test, sustained in attention-al blindness, and ADD/ADHD test.
A case study is carried out among 500 children from schools and health centers. The
result indicates its capability in predicting disorders in the child with 93.4% accuracy.
Contact-free and privacy-preserving solutions for monitoring bed activities need to
be emphasized towards the smart management of children.
3 Discussion
A review of smart management of children based on IoT, Machine Learning and Data
Analytics has been presented where potential methods and techniques proposed by
researchers have been highlighted. The summary of the findings from the literature
review is illustrated in Table 1. Throughout the studies, we discovered the approaches
from different areas and a combination of suitable methods that may lead us towards
the development of a new smart management system for children.
Considering the challenges that the smart management applications for children
faced some common approaches have been identified. Eye-tracking data analysis,
emotion recognition, facial expression, infant activity monitoring and real-time inter-
active growth management are some of the areas that have huge potentials for the
development of a smart management system for children. Similarly, the vital signs
monitoring approach is an important technique that can be further investigated and
integrated into a smart management system to keep track of the children’s health
condition. The development of such kind of SMART systems for children may
initiate a huge impact on the people globally, especially among the people in Asia.
We conducted a preliminary survey among 22 parents aged 35–45 from Malaysia and
Bangladesh to estimate the guardian’s approaches to monitoring their children. A self-
administered questionnaire was utilized to collect the data from the parents and there
were asked to answer and complete the questionnaire online where 48.2% are female.
The study indicates that most respondents have medium knowledge of technologies,
and they are more familiar and dependent on conventional techniques. A 14% of
respondents utilize the existing monitoring approaches, for instance, CCTV cameras,
and tracking systems. However, such a tracking system was not efficient since each
component needs to monitor separately which leads the users dispassionate.
982 Md J. H. Faruk and M. H. M. Adnan
4 Conclusion
management system may be proposed. As similar to smart homes and smart cities,
not only sensors-based IoT but also data analytics and machine learning shall be the
revolutionary technique towards developing a smart management system. However,
it is expected to conduct future research to design and develop a fully functional smart
management system by interpreting the methods and approaches, and hundreds of
thousands of people shall be benefited using the futuristic system.
References
1. Davidoff S, Lee MK, Yiu C, Zimmerman J, Dey AK (2006) Principles of smart home control.
Lect Notes Comput Sci (including Subser Lect Notes Artif Intell Lect Notes Bioinformatics).
4206 LNCS:19–34
2. Lee MK, Davidoff S, Zimmerman J, Dey AK (2007) Smart bag: managing home and raising
children. Proc 2007 Conf Des Pleasurable Prod Interfaces, DPPI’07. 434–437.
3. Lee MK, Davidof S, Zimmerman J, Dey A (2006) Smart homes, families, and controls. 2006;
Available from: https://pdfs.semanticscholar.org/b751/6530ebffe5b7202f0ac201aefc7850b87
a52.pdf?_ga=2.123838666.1183646419.1597302163-2076110809.1586498474
4. Yang H, Lee W, Lee H. IoT smart home adoption: the importance of proper level automation.
J Sens (2018)
5. Taylor AS, Swan L (2005) Artful systems in the home. CHI 2005 Technol Safety, Community
Conf Proc - Conf Hum Factors Comput Syst 641–650
6. Tolmie P, Pycock J, Diggins T, MacLean A, Karsenty A (2002) Unremarkable computing. Conf
Hum Factors Comput Syst - Proc 4(1):399–406
7. Darrah CN, English-Lueck JA, Freeman JM (2000) Living in the eye of the storm: controlling
the maelstrom in silicon valley, 1–10
8. Childmind. Managing problem behavior at home: a guide to more confident, consistent and
effective parenting. Child Mind Inst Inst [Internet]. Available from: https://childmind.org/art
icle/managing-problem-behavior-at-home/
9. Sendra S, Parra L, Lloret J, Tomás J (2018) Smart system for children’s chronic illness
monitoring. Inf Fusion. 40:76–86
10. Kanoun O, Universit T, View E, Kanoun O (2014) Smart systems and devices : innovative key
modules for engineering applications
11. Lanting C, Lionetto A (2015) Smart systems and cyber physical systems paradigms in an IoT
and industrie/y4.0 context. S5002
12. Madakam S, Ramaswamy R, Tripathi S (2015) Internet of things (IoT): a literature review. J
Comput Commun 03(05):164–173
13. Apanaviciene R, Vanagas A, Fokaides PA (2020) Smart building integration into a smart city
(SBISC): development of a new evaluation framework. Energies 13(9)
14. Teixeira CC, Boaventura RP, Souza ACS, Paranaguá TT de B, Bezerra ALQ, Bachion MM et al
(2015) Aferição de sinais vitais: Um indicador do cuidado seguro em idosos. Texto e Context
Enferm 24(4):1071–1078
15. Clemente J, Li F, Valero M, Song WZ (2019) Demo: contactless device for monitoring on-bed
activities and vital signs. Proc - 2019 IEEE Int Conf Smart Comput SMARTCOMP 2019.
472–474.
984 Md J. H. Faruk and M. H. M. Adnan
16. Clemente J, Valero M, Li F, Wang C, Song WZ (2020) Helena: real-time contact-free monitoring
of sleep activities and events around the bed. In: 18th annual IEEE international conference on
pervasive computing and communications
17. (2005) Children’s health, the nation’s wealth: assessing and improving child health. Choice
Rev Online 42(10):42–5896–42–5896
18. Binu PK, Akhil V, Mohan V (2017) Smart and secure IoT based child behaviour and health
monitoring system using hadoop. In: 2017 International conference on advances in computing,
communications and informatics (ICACCI) 2017. 2017-Janua:418–423
19. Boughorbel S, Breebaart J, Bruekers F, Flinsenberg I, Ten Kate W (2011) Child-activity
recognition from multi-sensor data. ACM Int Conf Proceeding Ser
20. Sindhu P (2018) Toddler activity recognition using machine learning. Loyola eCom-
mons [Internet]. 2018; Available from: https://ecommons.luc.edu/cgi/viewcontent.cgi?article=
4755&context=luc_theses
21. Guo J, Li Y, Hou M, Han S, Ren J (2020) Recognition of daily activities of two residents in a
smart home based on time clustering. Sensors (Switzerland) 20(5)
22. Rawashdeh M, Al MG, Samarah S, Hossain MS, Muhammad G (2020) A knowledge-driven
approach for activity recognition in smart homes based on activity profiling. Futur Gener
Comput Syst 107:924–941
23. Dzedzickis A, Kaklauskas A, Bucinskas V (2020) Human emotion recognition: review of
sensors and methods. Sensors (Switzerland) 20(3)
24. Sivasangari A, Ajitha P, Rajkumar I, Poonguzhali S (2019) Emotion recognition system for
autism disordered people. J Ambient Intell Humaniz Comput
25. Hossain MS, Muhammad G (2018) Emotion-aware connected healthcare big data towards 5G.
IEEE Internet Things J 5(4):2399–2406
26. Hossain Faruk MJ, Valero M, Shahriar H (2021) An investigation on non-invasive brain-
computer interfaces: emotiv epoc+ neuroheadset and its effectiveness. In: 2021 IEEE 45th
Annual Computers, Software, and Applications Conference (COMPSAC), pp 580–589. https://
doi.org/10.1109/COMPSAC51774.2021.00086
Optical Wireless Communication Based
Wireless Body Area Network for Remote
Patient Monitoring
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 985
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_93
986 K. Masroor et al.
1 Introduction
2 System Description
An intensity modulation and direct detection (IM/DD) channel model has been
considered for VLC-based LED transmitters, given by Ghassemlooy et al. [12]:
y = Hγ x + n (1)
988 K. Masroor et al.
Table 1 Summary of
Parameters Values
simulation parameters
For transmitter
Center luminous intensity per LED chip 0.73 Cd
Lambert’s order 1
Number of LEDs per array 3600 (60 × 60)
For receiver
Field-of-View 85o
Physical detector area 1 cm2
Gain of an optical filter 1.0
Refractive index of lens 1.5
Height (from the floor) → Sitting 0.85 m
Height (from the floor) → Standing 1.2 m
where y is the received signal power, H is the channel static gain, γ is the receiver
responsivity, x is the transmitted signal, and n denotes the Additive White Gaussian
Noise (AWGN) with double-sided noise power spectral density N 0 .
The total received power (Pr-total ) due to LoS (H (0)) and first-order NLoS (Hnlos
k
)
reflections can be calculated as [12]:
∞
Pr −total = H (0) + k
Hnlos Pt (2)
k=1
In this study, simulations were carried out for a 5 × 5 × 3 m3 room to examine the
effects of physical parameters i.e. the reflectivity of walls and SAHP of LEDs on the
performance of an OBSN link. Illustrated in Fig. 2 is the RXPWR obtained in Watts
for LoS configuration at different angles of irradiance. It can be observed in Fig. 2a
that with single transmitter (Tx ), the RXPWR is mainly concentrated at the centre
when angle of irradiance is 10º while it is more evenly distributed when LED transmits
at 60º (see Fig. 2b). Although the RXPWR has decreased in the latter due to the
increasing distance of the PD from the Tx , it provides better coverage. Consequently,
to ensure maximum coverage, further analysis will be based on LED transmissions
at SAHP = 60º. Furthermore, Fig. 2c illustrates the effects of reflectivity of walls on
the RXPWR via NLoS link for two values i.e. for ρ min and ρ max corresponding to 0.1
and 0.8, respectively. It can be observed that the reflected power reduces considerably
with low reflectivity and vice versa indicating that reflectivity of surfaces contributes
Optical Wireless Communication Based ... 989
(a) (b)
(c) (d)
Fig. 2 Received Power for Single Transmitter with a SAHP 10º and b 60º, c NLoS distribution at
ρ = [0.1, 0.8] and d SAHP = 10º
(a) (b)
4 Conclusion
References
1 Introduction
One of the effective methods to bounce back from stroke effectively is by taking
prevention steps such as identifying and understanding the signs of a stroke, although
prompt care can save lives and significantly affect the chance of full recovery. Unfor-
tunately, in stroke survivors the risk of a second stroke experiencing increase. About
25% of stroke survivors will suffer a second stroke, according to The National Stroke
Association. The probability of a second stroke is around 40% higher within the first
5 years after the first stroke. Apparently, it is predicted that about 80% of all secondary
strokes are preventable with improvements in lifestyle and medical care.
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 993
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_94
994 H. F. B. Hawari and S. B. Abu
By 2019, Malaysia be the 113 in the world according to stroke patients. This type
of disease is very dangerous and silent killer among Malaysia citizen. Due to slowly
growth disease in human body this stroke disease almost unnoticeable until disease
attack at the paralyzed stage. he goal of long-term recovery is to help insure that the
victim of the stroke is as healthy as possible. Theoretically this is achieved in a way
that keeps the patient comfortable and inspires them to learn the new simple skills
like bathing, eating, sitting and walking.
Rehabilitation usually begins in hospital following a stroke. If your diagnosis is
stable, recovery will begin during two days of the stroke and proceed after discharge
from hospital. It can often take several weeks, months, or even years for anyone
after a stroke recovery period is dissimilar. Few people completely recover while
others have long-term or lifelong disabilities. Rehabilition may include consultation
with speech therapists, physiotherapists and occupational therapists. Speech therapy
supports individuals who have trouble generating expression or hearing expression.
Because of the stroke, physical therapy uses exercises to help you relearn balance
and agility skills you may have lost. The aim of occupational therapy is to change
everyday habits, such as feeding, drinking, dressing, bathing, reading and writing.
A stroke rehabilitation system can be divided to upper limb muscle and lower limb
muscle [1]. The top limb or upper limb is the limb of a vertebrate extending from
the deltoid area to and including the neck, including the head, axilla and shoulder
Although, in the general sense of the term, the human leg is the entire lower limb
for the human body, including the foot, groin, and even the hip or gluteal region.
Nonetheless, in human anatomy the term only applies to the part of the lower limb
extended from the knee to the ankle, also known as the crus. Legs are used for
balancing, and all types of locomotion, including leisure activities such as dancing,
are an important part of the mass of a individual. Figure 1, show that the people using
armband upper limb stroke rehabilitation system. In past several past years, stroke
rehabilitation system widely have been invent to detect the movement of upper limb
muscle. A stroke rehabilitation system is a device that can detect lower limb muscle
expand and contract and in the same time, able to analyses the movement of lower
limb muscle.
Basically a stroke rehabilitation system consists of the 2 major parts. The first
part consists of the stroke rehabilitation equipment usually in a form of machine
exercise. Secondly, a wearable device that can detect the movement of lower limb
muscle during the stroke rehabs. In past several years, stroke rehabilitation system
widely been invent to detect the movement of upper limb muscle. From this concept,
it shows that there is a lot of room of possible application of stroke rehabilitation
system that can apply on lower limb of the body. In this research, we will develop an
stroke rehabilitation system that will be able to detect the movement of lower limb
muscle and able to provide useful information through the analysis of the stroke
rehabs to diagnose the movement of lower body muscle by using muscle sensor. By
looking at several criteria from lower limb muscle movement, such as how much the
muscle expand and contract.
Development of an IoT-Enabled Stroke Rehabilitation System 995
The severity of stroke complications and each person’s ability to recover vary
widely. Researchers have found that people who participate in a focused stroke reha-
bilitation program perform better than most people who don’t have stroke rehabili-
tation. For example, these skills can include coordinating leg movements in order to
walk or carrying out the steps involved in any complex activity [2]. According to Joel
Stein, MD, a physician under department of Rehabilitation Machine of Columbia
University’s College, repetitive exercising is the key to stroke rehabilitation. Since
there is no commercial device that is currently available to monitor real time stroke
rehabs at home, it is important to develop a system that a user is able to monitor their
health performance and in the same time, its data being monitor by medical expertise
to ensure the if abnormal sign happens in data.
In IoT, health information of the user can be stored in cloud and the data accessible
by third party like doctor and or medical officer to review the data and analyze the
condition of the user without having a appointment between them. This exchange
information of the user and the doctor will improve the medical check-up quality and
able to track the rehabs activity [3, 4]. With IoT, it is hope the application of stroke
rehabilitation system rehabs diagnostic will help society to boost the quality of life
and the same time, provide user as alternative device to monitor their health.
996 H. F. B. Hawari and S. B. Abu
NodeMCU is the controller which have an open source system for IOT devices and
application which make this controller special the device equipped with ESP8266
WiFi SoC make the device ready to be used for IOT loop from devices and to the
cloud based system [5]. This controller is improving version from Arduino type of
controller. The functioning and software for this controller are same with Arduino
and the only different is WIFI part inside the controller that makes this controller
ready to connected unlike Arduino need to buy Wifi Module and write a programmer
to activate the Wifi by pairing the IP Address.
The major process flow of whole project as per Fig. 2. Myoware muscle sensor
sense the muscle movement and the accelerometer give 3 axis data of the motion
and send the analog signal to the microcontroller which is NodeMCU powered by
battery. The data that been read by sensor will be connected to the ESP8266 which
a module Wi-Fi which sends the data to the mobile application and store inside
the cloud storage. Within this research, the most significant and one of the latest
technologies is a mobile application, based on extracting real-time information from
medical instruments attached to calculate the cardio details of patients [12].
As in Fig. 3, the wiring diagram show the wiring part and the wireless part is
being described. This project base an IoT technology show that the device portable
and connected to the smartphone and cloud database and smartphone via WiFi
technology.
998 H. F. B. Hawari and S. B. Abu
As in Fig. 4, the description of the development of the lower limb stroke rehabilitation
system is highlighted. The prototype of this project is attached with the e-box that
contain microcontroller, muscle sensor, accelerometer and battery. The dimension of
the system prototype will be 360 × 290 × 180 mm. The user is required to place their
leg on the rehabilitation platform where they are required to step down the pedal for
a number of tries. The output of accelerometer and muscle sensor will be displayed
real time in the Blynk mobile application and Thing Speak cloud.
The system prototype test is conducted at old folk’s home at Rumah Seri Kenangan,
Seri Iskandar Perak. The field testing is tested on random elderly at Rumah Seri
Kenangan. The testing is conducted using the whole system communicate together
with Thing Speak Cloud and main central system using wifi module. The data is
then observed through the Thing Speak Cloud. The objective of this field testing is
to study the effectiveness and the efficiency of the overall system.
Person X is a male, 42 years old person with related to asthma problem. Based
on Fig. 5, the muscle sensor attached to the leg is able to provide a good response
whenever the person tries to perform the steps using the system prototype where the
result is posted real time on the Thing Speak cloud. For the accelerometers, it also
good response on different axis, indicating good movement on the leg.
Figure 6 shows the Think Speak for Person Y. Person Y is a female, 46 years old
person with related to diabetic problem and osteoporosis history. The result for the
muscle sensor shows that the muscle movement is less compared to Person X. The
first response is almost similar to Person X, but as the Person Y continue to perform
more steps on the prototype, the movement is slightly stuttered possibly due to her
health condition. As a result, the muscle sensor did not show good response for the
remaining steps. Similarly, for the accelerometer the x axis indicate similarity for
person X, but for Y and Z axis, the results are not different due to stuttered movement
of the leg.
Person Z is an elderly male, 65 years old person with related to diabetic and gout
problem. Person Z has problem with his leg and has problem to perform the steps
movement on the prototype. Based on Fig. 7, the muscle sensor attached to the leg
show that person Z try to perform a step. However, due to his leg problem, he was
only to perform partial step and take about 10 min to perform it. From the Think
Speak cloud, the accelerometer response also shows difficulty in the leg movement
through indication of linear response for different x, y and z axis.
1000 H. F. B. Hawari and S. B. Abu
Fig. 5 Lower limb stroke system prototype think speak result for person ‘X’
Fig. 6 Lower limb stroke system prototype think speak result for person ‘Y’
1002 H. F. B. Hawari and S. B. Abu
Fig. 7 Lower limb stroke system prototype think speak result for person ‘Z’
4 Conclusion
Acknowledgements I would like to thank the personnel and elderly from the Rumah Seri
Kenangan, Seri Iskandar Perak who despite of being busy with their schedule, managed to take
time out to provide support on the testing.
References
Abstract Brain tumor is a deathly disease and it is indispensable to point out the
tumor immediately. Detection of brain tumor from MRI with higher accuracy has
become a major research region for the medical sector. In this paper, an automatic
brain tumor classification procedure applying fuzzy C-means and artificial neural
network is proposed which provides higher precision. For this proposed technique,
inputted MRI images are resized and then sharpening filter is used for preprocessing.
After that, fuzzy C-means cluster process is chosen for image segmentation. At
the next step, discreate wavelet transform is utilized for feature extraction and then
features quantity are reduced by principal component analysis. Furthermore, reduced
features are taken to artificial neural network for brain tumor classification. An
effective training function Levenberg–Marquardt is used for neural network. This
proposed method provides 99.8% accuracy, 100% sensitivity and 99.59% specificity
which is comparatively better than other existing detection techniques.
1 Introduction
Brain tumor is another disease which is playing a dominant role to cause death.
ASCO published that 18,020 adults die from brain tumor in a year [1]. Cancerous
brain tumors can snatch life within 5 years from the diagnosis time [1]. Brain tumors
are mainly abnormal growth of cells that fill up the necessary brain space [2]. 120
types of brain tumors exist but they said all tumor types are not cancerous [2]. Glial
cells are the point from where Glioma tumor arises. Pituitary tumor exists in pituitary
gland [3]. A patient needs to detect the tumor type quickly to initiate treatment. It is
already proved that Artificial Intelligence (AI) has enough skill to detect object more
perfectly than manual process [4]. Again, without trained-experienced radiologists,
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 1005
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_95
1006 A. Biswas and M. S. Islam
2 Methodology
The classification technique of proposed paper includes FCM and ANN to explore
the higher accuracy. The total process of classification can be divided into four steps:
pre-processing steps, image segmentation step, feature extraction and reduction and
finally, classification step with ANN. The workflow of technique is exhibited in
Fig. 1.
2.1 Pre-processing
This proposed technique applied resizing and sharpening filter for preprocessing:
MRI Brain Tumor Classification Technique … 1007
Fig. 1 Proposed
classification technique flow
chart
Image Resizing. Dataset consisted of different size of images. To work with all
images at a time, input images were resized by 200 × 200 pixels which means to
change the pixel information.
Sharpening Filter. Sharpening filter was used to strengthen the contrast shade of
the bright and dark area of the input image. This also enhanced the object’s edge.
Sharpening filter provides better accuracy in case of image processing.
1008 A. Biswas and M. S. Islam
In 1981, Bezdek is the person who introduced FCM algorithm at first and FCM is an
improved version than previous clustering algorithms [11] and FCM allows two or
more clusters in one pixel [12]. FCM is unsupervised clustering process where data
points and centers of cluster’s distance create clusters [13]. It performs better than
k-means (result section). FCM works based on optimizing Objective Function [11].
n
c
m 2
J (U, V ) = µi j xi − v j (1)
i=1 j=1
where, Objective Function = J (U, V), data points = n, fuzziness index = m, member-
ship = µi j , jth cluster center = v j , No. of
cluster center = c, Euclidean distance of ith
particular to jth mid-point of cluster = xi −v j , k = iteration step, di j = Euclidean
distance.
FCM Algorithm
Step 1: Cluster center ‘c’ is selected randomly
Step 2: Calculation of Fuzzy membership function ‘µi j ’
c
2 −1
µi j = 1 di j − dik ( m ) (2)
k=1
Fig. 2 a Glioma MRI. b Pituitary MRI. c Sharpening filtered image of Glioma MRI. d Sharpening
filtered image of Pituitary MRI. e FCM clustered image. f FCM clustered image
MRI Brain Tumor Classification Technique … 1009
Here, Fig. 2a is Glioma and Fig. 2b is Pituitary MRI images. Contrast was strength-
ened by sharpening filter, shown in Fig. 2c and d. Then in Fig. 2e and f, fuzzy C-means
clustering was applied after filtering to segment tumor from the image background
properly.
Two-dimensional DWT was used in this proposed method and that was three-level
decomposition wavelet. Discreate Wavelet Transform is employed to extract multiple
features from the filtered image. Fourier transformation has a serious disadvan-
tage as it doesn’t contain any information about time domain [14]. LL, LH, HH,
and HL are the four sub-bands for each scale [15]. Principal Component Anal-
ysis (PCA), an efficient way of feature reduction was used to reduce the quan-
tity of features after feature extraction to reduce computational time and memory.
Extracted multiple features were: Contrast, Energy, Kurtosis, Mean, Entropy, Corre-
lation, RMS, Variance, Homogeneity, Smoothness, Standard Deviation, Skewness,
IDM.
Input of this ANN was collection of 13 features, hidden neurons were 10 in hidden
layer and output layer had 2 neurons to indicate glioma and pituitary classes of
brain tumor. This neural network was constructed with the fastest training function
‘Levenberg–Marquardt’. ANN is dexterous tool for classifying brain tumors and
makes flawless decisions. ANN took 70% data for training firstly, 15% sample was
used for validation and then 15% was used for prediction in the testing stage.
3.1 Dataset
512 × 512 pixels dimension, T1-weighted data was obtained from [16]. Dataset
consisted of total 489 MRI images where 246 glioma and 243 brain tumors. Axial
and coronal plane images are used for this experiment and dataset was unbiased.
1010 A. Biswas and M. S. Islam
Fig. 3 a Glioma MRI. b K-means clustered of Glioma. c FCM clustered image of Glioma. d
Pituitary MRI. e K-means clustered of Pituitary. f FCM clustered image of Pituitary
The best result confusion matrix is shown in Fig. 4 where 99.8% accuracy was found
for 18 iterations. Figure 4 proves that the proposed methodology can classify glioma
and pituitary brain tumors flawlessly.
4 Conclusion
A unique and flawless technique is proposed based on FCM clustering and ANN
classifier to classify glioma tumors and pituitary tumors using MRI. This method the
removes previous limitation of low accuracy by using 489 patient’s data. MRI images
1012 A. Biswas and M. S. Islam
were pre-processed by sharpening filter to enhance the contrast. Then better clustering
algorithm ‘fuzzy C-means (FCM)’ was selected for segmentation. DWT was utilized
to extract features from clustered images and for feature reduction, PCA was applied.
13 features were collected from images and then used for artificial neural network
training and testing. The fastest ANN function ‘Levenberg–Marquardt’ was used
for proposed network training. Test result provided 99.8% accuracy at 18 iterations,
100% sensitivity, 99.59% specificity. This proposed method is able to provide 99.8%
accuracy which is much better than some previous classification techniques. This
result is enough to claim that proposed technique can classify glioma and pituitary
tumor types magnificently. This improved accuracy is beneficial for medical sector
to classify automatically.
References
Abstract A large variety of precious lives are lost because of road traffic accidents
each day that may occur due to the mistake of driver and delayed response from the
place of accident to the emergency services. The most important aspect of driver is
to own a good road accident detection and data communication system in place to
save injured persons. A system that sends information messages to close emergency
services regarding the accident location for timely response is completely essential.
The accident detection process uses raspberry pi, in detection of the data from the
vibration sensor and accelerometer during accident. The process of rescue system
uses GSM, in the process of sending message to the neighboring medical Centre or
relatives. GPS is used to track the location of the vehicle and used in the calculation
of the speed of the vehicle based on the position of (latitude, longitude) and time
difference. GSM in the system is used to send the location of accident through
message. Vehicle ad-hoc network (Wi-Fi) is used to transfer of the data from the
controlling unit to the application server in the later retrieval of data. Hyper Text
Transfer Protocol is used to transfer web page from remote server after internet
connection is established. The FireBase database is used in the storage of collected
data, and uploads over in the web-page and in the process informing the accident
location and in alerting.
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 1013
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_96
1014 K. Gayathri Devi et al.
1 Introduction
Nearly 1.25 million individuals die in road crashes annually, on average 3287 deaths
on a daily basis. An additional 20–50 million are injured or disabled. Over 1/2 all
road traffic deaths occur among young adults ages 15–44. There are 1012 deaths in
2019. This can be 74 deaths (+7.9%) over an equivalent amount in 2018. Within the
12 months complete October 2019 there have been 1209 road deaths. Throughout
calendar 2018 there have been 1135 road deaths, 7% not up to in 2017. A recent
World Health Organization (WHO) report showed that each year 1.35 million people
die and 50 million place get injured. Road accidents are ranked because the eighth
leading reason behinds deaths (up from ninth its previous report in 2015), with the
Association for Safe International Road Travel (ASIRT) predicting that it’s going to
rise to the fifth leading reason behind deaths within close to future, unless forceful
changes occur. Moreover, because the social damage caused by road traffic accidents,
there is a big price. ASIRT estimates that between one and two percent of the annual
budget of each country is spent on road accidents. Now-a-days, it has become terribly
troublesome to understand that an accident has occurred and to find the position
wherever it’s happened. There is no system of identification and intimation relating
to an accident in the past. For intimation purpose messaging services is used. GPS
makes use of identification of location and GSM is built for the usage of intimation.
The most intention of an accident detection is to locate the accident spot anywhere
and intimating the emergency through the GPS and GSM networks. The GPS based
vehicle accident identification module contains Mostly of different sensors such
as accelerometer, MEMS etc., GSM module and a GPS modem connected to the
microcontroller. Global System for Mobiles (GSM) technology is employed for a
proper cellular mobile communication. Position of the vehicle accident location is
traced down by GPS.
2 Literature Survey
The self-collision detection sensor is used to detect the accident and the information
is uploaded to cloud based database server collision vehicle accident recognition,
and a connected emergency notification is provided [1]. The detection of accidents
supported by monitored speed and send the accident location acquired from the GPS
together with the time by GSM and therefore the speed by utilizing the GPS network
[2], sent to a OpenGTS server, providing a time period OpenStreetMap visualization
of traffic scenario. MongoDB appraise the close mobile APPs with alert messages
[3]. Auto continuous collision accident collision mechanism, exploitation physical
phenomenon preview iterative algorithm, trace examination reconstruction algo-
rithm, serial collision contact position reconstruction localization algorithm [4], self-
tuning iterative hard thresholding (ST-IHT) algorithm for learning distributed spatio-
temporal features and a weighted extreme learning machine (W-ELM) [5], Sudden
An IoT Based Automatic Vehicle Accident Detection … 1015
Lane Departure Avoidance Module and Rear-End Collision Avoidance Module [6]
is used over in the process of detection of accident. The accident is detected by
the in-Vehicle sensor and updates in database, verifies and compares the severity of
accident supported to different accident data [7] and notifies the emergency services,
uses remote cut off mechanism [8] and using GPS and GSM intimates the emergency
services [9, 10] and provides over an crash path reconstruction using Kalman filter
and GPS [11]. Data continuously collected from the smartphones accelerometer and
analyzed using Dynamic Time Warping (DTW) [12], inputs from the accelerometer,
magnetometer, gyroscope and on vehicle crash detection [13] are analyzed and inti-
mated as e-Call with owner’s information. The local prediction is exchanged from
each sensor with neighboring vehicles aggregates the native predictions it receives
using a weighted majority aggregation rule to get a final prediction [14], crash signals
at different locations of a vehicle and their implications on crash analyzes the severity
of accident [15], the system also sends a warning message. The system uses CAD-
CVIS consists of varied sorts of accident types, atmospheric conditions and accident
location, a deep neural network model YOLO-CA supported CAD-CVIS and deep
learning algorithms to detect accidents and in alerting [16]. The system detects and
assesses the severity of accidents with the assistance of an on-board unit and the
TestBed analyses different accident scenarios and helps in rejecting false alert and
serially detects the rollover and force [17]. Besides over these many systems, they
are costly and not applicable in every cars and lacking over in the process of saving
location for later use.
3 Proposed Model
This works focuses on designing and developing an accident detection using vibration
sensors and wireless modules. The block diagram of the proposed system is shown in
Fig. 1. In order to enhance real time accident detection and monitoring, the system has
Vibration sensor SW-420, GPS and GSM module, which is connected to Raspberry
pi controller, which works with Wi-Fi in transfer of data. The Vibration sensor SW-
420 produces an digital output, on the detection of any collision or accident. GPS is
used to track over the location in which the vehicle travels. GSM is used to send the
location of accident as an URL link with the latitude and longitudinal information on
the place of accident. Raspberry pi controlling unit is used to sense the information
and to perform the entire controlling functions with database through Wi-Fi. For the
enhancement of real time accident monitoring, the system with sensors which works
on IoT basis [18, 19]. By using IoT, status on speed and the location of accident is
recorded and the database manages over the data and uploads over in the webpage,
in tracking of vehicle and its speed.
The connection and the flow diagram of the Vibration Sensor SW420 with Rasp-
berry pi is given in Figs. 2 and 3. The accident detection system operates with the
vibration sensor producing digital pulse output on the detection of any accident or
collision. It produces over an output based on the threshold which is been set over
1016 K. Gayathri Devi et al.
Power Supply
CLOUD
Vibration Database-Firebase
Sensor
Raspberry
GPS Pi 3
Tracks Location
and Speed
Reset GSM
Button
Update Location
Message
in the potentiometer. The sensor is tightly fitted over in any part of the car. The
variations in the output of the sensor are trapped and those are fed as input to the
micro-controller.
GPS is commonly used for both tracking and navigation in vehicles. Tracking
systems helps the base station to keep track of the vehicles in the way of travel without
the intervention of the driver whereas, the system also helps the driver to reach the
destination by navigation. The architecture of navigation or tracking system is more
or less similar. The connection and the flow diagram of the GPS with Raspberry pi
is given in Fig. 4. During the occurrence of accident in any place, tracking on the
position of the vehicle by GPS system and send the information by alerting person
through call or SMS by using GSM.
The connection and the flow diagram of the GSM with Raspberry pi is given in
Fig. 5. Now based on these variations, the micro-controller is programmed to take
decision which in turn.
• Sent as an message an through the GSM to relative or nearby emergency service.
• Data is updated over in the database through wireless transmission.
• Webpage provides over the alert message and update the location in the map.
Wireless communication system acts as a bridge for duplex communication for data
collection and control message delivery and is an essential part for IoT infrastruc-
ture. Various IoT applications can be applied using Wireless communication system,
which includes health care monitoring and home automation etc. and gains a lot of
attention in connecting devices to internet.
Internet cloud services like an application service provider is used to exchange data
and manage the traffic of messages that will be connected directly to IoT devices
in a device-to-cloud communication model. There are numerous advantages when
a connection is established between the device and the IP network through cloud
service when compared with the existing communications mechanisms like Ethernet
or Wi-Fi connections.
The Device-to-cloud communication model that is proposed in our work is shown
in Fig. 6 and maintained by more consumer IoT devices like the nest labs learning
devices and many SmartTv and Smart phones. The Nest Learning Thermostat uses the
devices to broadcast data to a cloud database where the data can be used to analyzed
and problems can be identified. Cloud connection may help or enable user to obtain
remote accessing of data through smartphones or web interface and supporting in
some software updates to device.
The SmartTV technology make use of the Internet connectivity for the process
of transmitting the information viewed by the user to do further analysis by the
service provider and this enables the voice recognition features. The user can expand
capability of device further than its original features to device-cloud module with
respect to certain cases. Interoperability problem exist when there is an amalgamation
of device detail from many diverse manufacturer it can be avoided if the authorized
data protocols are used between the device and the cloud services. The device owner
or user can be coupled to a precise cloud service in order to prevent them the usage
of alternative service providers. This is referred to as “vendor lock-in”, a term the
compasses other facts of the relationship with the provider such as ownership of
and access to the data. In mean time, users can have confidence that devices can be
integrated based on the design for the specific platform.
IoT has made the change in the evolution today’s world by connecting things
over the Internet, which makes very much easier remote access. IoT is the emerging
and trending technology that can be implemented over in the safety and security
monitoring systems and in rescuing.
The Raspberry pi controls the functionality of the above mentioned sensor and
modules. Programming for the controller is done using Python coding. After comple-
tion of the entire program, debugging is carried out. Finally, the program is dumped
to the controller. Then, the working mechanism is verified. Figure 7 gives the snap-
shot of the prototype developed with major components. IoT plays a major role in
collection of database, from the user. Each user can reach the server using their own
e-mail id. This is achieved by means of mobile-to-cloud communication. When the
user access the server, any variation in sensor detection is updated in form of database
with the date and time. Therefore, real time accident detection is achieved by the
implementation of IoT.
GSM SIM800A
Raspberry pi 3
Figure 8 gives the snapshots of the location of the accident that were transmitted
through GSM. In this prototype firebase help in cloud based data storage and helps
in easy retrieval of data. Firebase provides developers in the platform of mobile and
web app development and which provides a plenty of tools and services which help in
developing high-quality apps, grow their user base, and earn more profit without any
back end server. Figure 9 gives the snapshot of the updation of location and speed in
the database along with the information on accident occurrence through wi-fi using
HTTP protocol. Webpage provides over the view in which location the vehicle travels,
and provides over a pop-up information on the accident occurrence. Figure 10 gives
the snapshot of the webpage during accident occurrence which gathers information
from the database. The information based on the location helps in easy retrieval of
victim people.
The Comparison between the proposed and existing system is given in Table 1.
The existing system was designed with need to transmit the accident location using
IoT technology to cater the needs of people who are undergoing accident.
5 Conclusion
This proposed system with advanced features reduces manual work and allows
detecting the accident and reporting to the emergency services and updating data
in database. The IoT technology used in this system helps in treating the victim
people instantly in case of any emergencies and the route to the location of accident.
1022 K. Gayathri Devi et al.
References
1. Chang WJ, Chen LB, Su KY (2019) DeepCrash: a deep learning-based internet of vehicles
system for head-on and single-vehicle accident detection with emergency notification. IEEE
Access 7:148163–148175
2. Amin MS, Jalil J, Reaz MBI (2012) Accident detection and reporting system using GPS, GPRS
and GSM technology. In: IEEE international conference on informatics, electronics & vision
(ICIEV), pp 640–643
3. Celesti A, Galletta A, Carnevale L, Fazio M, Lay-Ekuakille A, Villari M (2017) An IoT cloud
system for traffic monitoring and vehicular accidents prevention based on mobile sensor data
processing. IEEE Sens J 18(12):4795–4802
4. Lang W, Biao G, Tao C (2013) Vehicle continuous collision accident reconstruction system
development. Procedia-Soc Behav Sci 96:1659–1669
5. Yu Y, Xu M, Gu J (2019) Vision-based traffic accident detection using sparse spatio-temporal
features and weighted extreme learning machine. IET Intel Transp Syst 13(9):1417–1428
6. Chang TH, Hsu CS, Wang C, Yang LK (2008) Onboard measurement and warning module for
irregular vehicle behavior. IEEE Trans Intell Transp Syst 9(3):501–513
7. Fogue M, Garrido P, Martinez FJ, Cano JC, Calafate CT, Manzoni P (2013) A system for auto-
matic notification and severity estimation of automotive accidents. IEEE Trans Mob Comput
13(5):948–963
8. Srabanti S, Asaduzzaman M, Mokter MKB, Anannya TT, Tumpa SN, Afroze L, Karim
MM (2018) A proposed system for automatic vehicle monitoring and accident detection in
Bangladesh. In: 2018 international conference on computer, communication, chemical, material
and electronic engineering (IC4ME2). IEEE, pp 1–5
9. Kumar A, Jaganivasan V, Sathish T, Mohanram S (2018) Accident detection and alerting system
using GPS & GSM. Int J Pure Appl Math 119(15):885–891
10. Taj FW, Masum AKM, Reza ST, Chy MKA, Mahbub I (2018) Automatic accident detection and
human rescue system: assistance through communication technologies. In: 2018 international
conference on innovations in science, engineering and technology (ICISET). IEEE, pp 496–500
11. Aldimirov M, Arnaudov R (2018) Method for automated reconstruction of a car’s path during
crash from GPS/INS data using a Kalman filter. Adv Eng Softw 115:386–390
12. Aloul F, Zualkernan I, Abu-Salma R, Al-Ali H, Al-Merri M (2015) iBump: smartphone
application to detect car accidents. Comput Electr Eng 43:66–75
13. Fernandes B, Alam M, Gomes V, Ferreria J, Oliveira A (2016) Automatic accident detection
with multi-modal alert system implementation for ITS. Veh Commun 3:1–11
14. Canzian L, Demiryurek U, van der Schaar M (2015) Collision detection by networked sensors.
IEEE Trans Signal Inf Proc Over Networks 2(1):1–15
15. Chan CY (2002) On the detection of vehicular crashes-system characteristics and architecture.
IEEE Trans Veh Technol 51(1):180–193
16. Tian D, Zhang C, Duan X, Wang X (2019) An automatic car accident detection method based
on cooperative vehicle infrastructure systems. IEEE Access 7:127453–127463
17. Khaliq KA, Raza SM, Chughtai O, Qayyum A, Pannek J (2018) Experimental validation of an
accident detection and management application in vehicular environment. Comput Electr Eng
71:137–150
18. Kokila J, Gayathri Devi K, Dhivya M, Haritha Jose C (2017) Design and implementation of
IoT based waste management system. Middle-East J Sci Res 25(5):995–1000
19. Hou Z, Chen Y (2017) A real time vehicle collision detecting and reporting system based on
internet of things technology. In: 2017 3rd IEEE international conference on computer and
communications (ICCC). IEEE, pp 1135–1139
20. Krishnan P (2018) Design of collision detection system for smart car using Li-Fi and ultrasonic
sensor. IEEE Trans Veh Technol 67(12):11420–11426
Automatic Firefighting System Using
Unmanned Aerial Vehicle
1 Background
The implementation of the systems with drone and various remote sensing tech-
nologies is rapidly increasing to provide a mobile, and low-cost alternative for the
previous systems such as ground-based systems, air controlled vehicles, and satellite-
systems [1]. The images captured by the satellites with low resolution and noise are
not adequate to provide effective result. It is also proves ineffective in populated areas
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 1023
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_97
1024 K. Gayathri Devi et al.
[2, 3] since there exist problem to classify the region of interest and there may be
misclassification. The constrained surveillance ranges affect measurement of ground
equipment [4]. The early stages of fire detection UAV system with image capturing
is now replaced by infrared cameras for detecting the radiation emitted by fires [5].
A direct view of the radiation is required by the infrared sensors for effective results
[3]. Visual cameras, a small size high resolution camera have been used for detection
of smoke produced by fire in daylight [6]. Visual cameras provide the details such as
flame height, angle, location and width. These cameras produce different contrast,
texture, and motion analysis [3]. Another mentioned is the usage of LIDAR devices
to identify smoke particles [7].
Holman et al. [4] proposed that UAV system can be used to capture images from
a camera fixed to systems in a real time environment. It was concluded that image
processing quality of topographic images which are in existence and the meta data
of the specific image is rectified and stored for later usage purposes. The paper
discusses the quality of resolution of images and how they are processed frame by
frame. He constructed a similar exiting prototype of unmanned aerial vehicle (UAV)
with a quadcopter setup to collect test images for specific enhancements of low cost
and flexibility. The autonomous functions in the UAV is capable of excellent ground
station. The accuracy of images from ground position could be found with accuracy.
Kumar et al. [8] had developed a optimization model for the detection of fire
and to extinguish the fire with fire suppressing fluid for a region where there are no
habitant with multiple UAV. Two step approach is considered, the first step represents
the fire by a two dimensional shape with a boundary, it requires all the UAV to work
in coordination within the boundary and there should be no collision. The second
step proposes a model as a function of fire and time. This approach aims at reducing
the Euclidean distance between the quadcopter and the fire at any instant of time.
There are few studies that considered fire suppressants other than water. They tested
this theory with master model with use of 10 drones via simulation. The suppressant
fluid is not practical. The undergraduate project designed as a quadcopter with a
ball-dropping mechanism that can drop a single unit of fire-extinguishing ball by
using compressed springs.
Wallar et al. [9] suggested a approach for the surveillance of risky areas by using
three quadcopters to evaluate the coverage area of the sensors on board, the quality
of the data collected with reference to various altitudes and risk involved with ground
level. The two stage approach is implemented in which the movement of the quad
copter to maximize the coverage area is considered in first stage and the second
age optimizes the risk involved. It was intended for the examination of regions to
diminish the detection risk and to ensure the data provided by the sensor is of high
quality. This approach is more costly as more number of quad copters are used.
Mois et al. [10] has focused on the data transmission, security and success rate of
huge amount of environmental parameter data with three the Internet of Things (IoT)
based protocol, the cloud computations, and cyber-physical systems. Three types of
IoT-based wireless sensors for monitoring the environment and the surroundings
were analyzed in this paper. The web based communication established with User
Datagram Protocol (UDP) was not reliable as the number of packets that was lost
Automatic Firefighting System Using Unmanned Aerial Vehicle 1025
in each minute and only the success of the transmission was 95.4% and it is mainly
due to the network and is not related with the hardware. The success rate of the
transmission was slightly increased, there exist a reduced battery life time when
the communication was established through Wi-Fi and Hypertext Transfer Protocol
(HTTP).The packets transmitted are received only when the devices is within a
certain range in the case of third one using Bluetooth Smart. This paper concludes
with various advantages and disadvantages of using the above three communication
protocol with respect to the factors such as power consumption and life time of the
devices, Internet facility, success and reliability of the transmission and revealed that
all three can be used for web based application. Wi-Fi and BLE are two technologies
suited for ZigBee protocol for monitoring applications that can successfully compete
with the concept established.
Zhou et al. [11] proposed a tracking framework and a systematic road detection
approach in video footage using UAV. Initially Graph cut algorithm is applied for the
extraction of the road regions from other non road regions. After the initialization,
for quick retrieval of the road region homography estimation is used. The proposed
framework has a precision of 97.4% and processing 35 frames per second for 1080
× 600 videos on average. Author has published a systematic approach for the identi-
fication of road by applying the relevant algorithm for different time instant in UAV
videos.
Lockheed Martin et al. conducted and demonstrated a collaborative system
comprising of drone and helicopter. The helicopter is designed to drop water exactly
on the hot spots identified by the UAV [12]. Companies such as Aerones, Nitrofirex,
and Singular Aircraft are focusing to design and implement drone systems that
make use of water to suppress fires [13–15]. Phan and Liu [16] developed a system
consisting of airship, UAV and unmanned ground vehicles (UGVs). The peak level
of the hierarchy the airship generates a mission plan by utilizing fire outbursts, UAV,
and UGV planned models [16].
The continuous research of developing UAV system for the detection and construc-
tion of wildfire implemented with drones, application program to program the hard-
ware and relevant algorithms for detection and tracking is in progress day by day
[4]. The main focus is different from the previous establishments or papers which
deliberately work in developing and resulting a remote-sensing capability which will
give an alarm not only for fires, but it also in detection of the fire head and flanks’
spreading. Thus remote sensing is designed to detect fires caused in buildings, fire-
fighters, and based on that will recognize the risk of the fire spreading to any other
area or a precise location parameters. The purview of this particular paper is designed
for this remote-sensing system.
Howden et al. developed a system with multiple that overlooks complex areas for
fires [17]. This system will split the area under surveillance into cells and the parame-
ters such as velocity, pheromone factor and the repulsion factor will be continuously
updated for every iteration. The simulation result for various parameters and giving
priority to the cell was also investigated in detail. Their algorithm proved successful
when implemented in real time environment as one of the UAV may have less fuel
and refilling is needed suddenly, one of them may crash and one may become failure.
1026 K. Gayathri Devi et al.
Pham et al. also developed a similar algorithm for multiple UAV system that iden-
tifies the fire that is spreading across the boundaries [18]. It ensures collisions are
avoided between UAVs. Yuan et al. implemented a forest fire detection system for
a forest fire when the fire is spreading in regions [19]. The video captured by an
aircraft in real time environment and the detection proves the systems reliable and
accurate. Similarly, a decentralized autonomous system of multiple UAVs to monitor
and govern the perimeter of the forest fire and containment was presented by Alexis
et al [20].
The detailed analysis of the implementation of various system for fire detection
and methods to extinguish the fire motivated to implement the drone of required
specification to extinguish the fire using fire balls attachable to drones. This system
is implemented on low power chip Raspberry Pi 3 and Navio 2. The Raspberry
Pi 3 is interfaced with Navio 2 module for better specific controls such as barom-
eter, accelerometer etc. The hardware modules interfaced with raspberry pi 3 are 4
BLDC motor, 4 ESC, Power distribution board, a transceiver, and a thermal sensor.
The system design is a combination of internet of things and dedicated server for
continuous storage and monitoring system using Amazon web services.
2 Proposed System
speed controller (ESC) for the stable movement. The purpose of using electronic
speed control or ESC is for varying the speed of Brushless DC electric motor (BLDC)
motor, its direction and also to act as a dynamic brake. The Raspberry Pi 3 is build
with powerful processor, ten times speeder than the first generation, wireless LAN
and with Bluetooth connectivity.
The drone frame used in the system is built from quality glass fiber and polyamide
nylon. The main setup of the UAV is connecting the motors with the power distri-
bution board where the power to the motors is controlled by the electronic speed
controller. The power distributed board is connected to the flight controller where
the input output to the ESC is connected to the flight controller. Then the LIPO
battery is connected to the power distribution board and tested for the function of
the motors. For wireless communications a transmitter is connected to the raspberry
pi. On bottom of the UAV the fire ball drop is fixed. The balls effectiveness and
robustness for building fires is maintained by conducted experiments. The fire extin-
guisher ball (AFO) which is approximately 0.7 kg, is used. The AFO was dropped
directly into fire, whereas they were located in bottom of the drone frame which can
be released the fire a few meters above.
The stage 2 focuses upon the software connection using dedicated server. Here
the raspberry pi is loaded with a pre-configured operating system and connected to
amazon web services (AWS) which is used to monitor and track the UAV movements.
Here the operating system image configures the flight controller, so the dumping of
operating system is done on a memory card. The installation takes place containing
1028 K. Gayathri Devi et al.
all services, kind of a meeting point for nodes. Hence, the nodes are discoverable and
communicated to each other Raspberry Pi. There are different set of drivers within
one node.
Considering fire is around two meter diameter, 0.7 kg AFO ball was dropped, the
possibilities of which if the ball exploded but was not able to extinguish the fire and the
fir outbursts continues or if it did not explode and the fire passed an approximate area
of three meters in diameter, the fire was going to be put out manually by the firefighters
These conditions were set for safety purposes. Figure 3 gives drone developed. The
drone is connected to the ground station and controlled via internet. Using the AWS
console a new account is created with the following services such as Lambda Service,
Simple Storage Service, Cloud watch, Cloud alarm, Simple Notification Services and
some additional services for external image processing. The servos are backed up by
the power to so a Bit Error Correction (BEC) should be present. BEC would serve as
back-up power supply to Navio + incase the main power source supply is intervened.
ESCs will heat each other if many ESC central wire connected to Navio 2.
3 Results
Drone configurations are usually connected to motors and move when the downward
thrust of UAV system is equal to the gravitational pull working against it. When the
rotors provide an upward force greater than gravity, they automatically descend, and
the UAV is directed in the desired direction using pitch, yaw, and roll. When the
rotors provide an upward force greater than gravity, they automatically descend, and
the UAV is directed in the desired direction using pitch, yaw, and roll. Drones are a
microcosm because they are dependent to provide two-way communications for the
Internet of Things. Testing of the drone must be done for safety and reliability for
minimum cost and requirements. They must have fail-proof software that is certified
for safety and effective in communication systems. They also should have efficient
Automatic Firefighting System Using Unmanned Aerial Vehicle 1029
4 Discussion
Consider the assumption that the UAVs could be re-loaded with a new method, it
could be possible to prevent fire outbursts from spreading in multiple directions, or
to extinguishing fires igniting the surrounding objects in large amount. The system
should be increased exponentially if there is relation between increases in swarm size
to the number of balls per UAV increase. The Table 1 depicts the idea of required
components.
Power will be 200 amps combination of continuous current and 400 amps peak
current. Motor power capability can pull up to the amps which is 25 amps continu-
ously without any disruptions. The thrust tests, each motor provides of 3 kg with 1867
carbon propellers, at 22.3 V, 24.1 amps. A ball-release mechanism was designed
and built as a mechatronic system attachable to the firefighting UAV. The system
consists of electronic components such as microcontroller, power supply, sender and
receiver, and motor as well as a mechanism to carry and release the ball and its
connections to the fire-fighting UAV. The ball-releasing mechanism was attached
with the mechatronic components.
1030 K. Gayathri Devi et al.
The project can be further improved by widening the obstacle detection range with
ultrasonic detectors. Further improvement can be made for this system. This system
can be improved with less latency and more accuracy. Deep learning algorithms and
Machine learning algorithms could be initiated. This proposed work with advanced
features reduces manual work and allows people to move to that place immedi-
ately and independently. The advancements in software side can reduce the latency
and produce high throughput. In the hardware side adding high resolution cameras
providing high sustainable circuit boards which have high durability. Therefore, real
time UAV is achieved by the implementation of web services in IOT.
These experiments were constructed and developed for a research project which
the main objective is to design a system of UAVs which can be automatic connected
to web server and monitored through internet and minimize the fire by extinguishing
the AFO balls above them. The idea can be simple if a single system of UAV could
drop them to optimal points, in optimal numbers, on time then there can be multiple
systems simultaneously working at same wavelength. Remote-sensing and image
capturing technology will be needed to determine points of attack and the web server
interfaced should be a dedicated one which cannot be used for other services. The
size of the fire-extinguishing balls depends upon the drones frame and the thrust it
can lift along with its own weight is a parameter that requires further investigation
for the usage of this system in real time applications. This is a promising result for
further studies that will be conducted for developing the proposed drone-assisted
fire-fighting system.
Automatic Firefighting System Using Unmanned Aerial Vehicle 1031
References
Zaira Hassan Amur, Yewkwang Hooi, Irum Naz Sodhar, Hina Bhanbhro,
and Kamran Dahri
1 Introduction
Natural language processing the semantic similarity put its tremendous impact on
question answers systems. Whereas short answers/texts observed to be most import
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 1033
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_98
1034 Z. H. Amur et al.
techniques. Which commonly recognized as STSS. These short texts include social
posts, question/answers, conversations, keywords and comments etc. In STSS short-
questions provide a useful understanding of specific concepts in a particular subject
domain. Short texts semantic similarity (STSS) helps to measure 10–20 words long.
Due to limited information these texts consider the most challengeable for number
of applications. These texts contains more than one meaning so they are generally
difficult to understand. Short texts contains lots of textual errors and produce result
ambiguity. These Texts generate semantic errors like grammatical and real world
errors. The main reason of these errors are user’s insufficient knowledge. Analyzing
and fixing these errors in short-texts improves the effectiveness of learning environ-
ment. Semantic calculation of short texts influenced the text words that causes the
poor accuracy [1]. Usually, understanding the hidden semantic meaning of word is
difficult task.
In natural language processing the text represents sentence. Each sentence/texts
has set of words, these words counted with different Parts of speech. Machines only
understand the language of symbols rather than text. To recognize the particular
and intended meaning of words or sentences it’s highly needed to convert text into
symbols. Part-of-Speech (POSs) determines that each word in the sentence contains a
specific sense of information. Generally machines don’t understand the dual meaning
of words/sentences. So therefore, natural language processing enables machines to
learn and seek the human natural language.
Text understanding plays a vital role in NLP. Most of the current methods are task
oriented such as information retrieval, relationship extraction and Q&A systems. The
main reason of using text in natural language processing make machines to know the
human languages. The aim of this paper is to summarize STSS techniques to identify
the trend of research.
The Paper is organized as follows: Section 2 is on literature study; Sect. 3 analyses the
various similarity approaches; Sect. 4 proposed the research framework and Sect. 5.
Gives the conclusion where section 6 provides the references of study.
2 Literature Review
Journal articles and proceeding papers from ISI and Scopus-indexed publications or
reputable institutions on STSS research in the past years have been carefully selected.
The papers are further filtered with focus on ML applications for A&Q systems. The
literature review identifies interesting development and trends of the study.
Refer to Table 1.1 in Attachment I. The table characterizes the STSS methods and
show the years and trends.
State-of-the-Art: Short Text Semantic Similarity (STSS) Techniques … 1035
or more words are collocated, there must be one word that works as the center word
and the other word as the collocation word. Parts of speech works well to construct
the center word.
However, Kim [15] suggest that related features can also help with disambigua-
tion. Furthermore, they worked to capture semantic relations between terms using
the novel approach LDA, this approach helps to improve the accuracy of short text
conceptualization by using context semantics into consideration. Whereas POSs like
verbs, adjectives, and other attributes, can also help to detect instance ambiguity.
Moreover, the study [16] introduces the framework for short texts that detect the errors
from text. More specifically, the work has divided into three subtasks to understand
the short text such as text segmentation, type detection, and concept labeling.
Wang [4] present an approach to solve the problem of semantic similarity in test
papers, with the help of density entropy they selected the test papers to screen the
various question papers from item bank and then applies the calculation of semantic
similarity to detect the intelligent test paper from the corpus.
Furthermore, Huang [17] developed the algorithm to improve the performance of
STSS with low time complexity. This algorithm incorporated the different WordNet-
based word measures to address the word pairs with specific POSs that helps to
enhance the evaluation of semantic similarity of unknown words in shorts texts.
Hesham [1, 13] represented the work, that focusing one Semantic Textual Similarity
(STS) of a question pair. In this study they finds if two questions have the same
answers, then they are semantically equivalent.
Otherwise, if the answers are different to one another, or partially different then
(the answer of the first question is included in the answer of the second question but it
considered the incomplete answers, or the questions share a portion of their answers
yet different otherwise) then these answers subsequently considered the semantically
nonequivalent.
To compute semantic similarity for short texts is important in many areas. Many
approaches have been proposed that use linguistic analysis. These methods determine
whether the words in two short texts look alike, in terms of largest common substring
[18]. These approaches usually work for trivial cases.
This kind of methods creates a major problems, because the complex features are
derived from other NLP tools and components which use for other purposes. Some
application depends on parse trees that are completely restricted to syntactically
well-formed texts. More ever, the error from these tools also creates the problems in
Text semantic similarity models that are very hard to solve individually.
Furthermore Younes [19] proposes a technique for text summarization they
focused on identifying significant parts of text to create summary. To solve the
problem they represented an optimized algorithm that helps to produce summary
with the lexical chains and WordNet dictionary.
However, the Study [20], proposed methodology that consists on rules, based on
NLP algorithms like sentence splitting, tokenization and POS tagging. These rules
were applicable on user provided text which he want to summarized. In Algorithm
each word is assigned a score based on its frequency, this algorithm will run sentence
by sentence.
State-of-the-Art: Short Text Semantic Similarity (STSS) Techniques … 1037
In this section, this study proposes the methods that have potential for further devel-
opment and the rationals. There are various approaches that used to measure the
semantic similarity. This section represents the proposed method for extracting simi-
larity in short-text questions in detail. These approaches needs to acquire similarity
from semantic and syntactic information that exist in the compared natural language
sentences. In a natural language processing (NLP) a sentence can be defines as a set
of words with Part-of-speeches (POSs) [22] (Fig. 1).
Fig. 1 Techniques for measuring the short text semantic similarity (STSS) [33]
1038 Z. H. Amur et al.
corpora. Furthermore, analyzing words using Eq. 1 would allow the massive corpus
helps to extract the words similarity accordingly. Corpus-based approaches also
known as statistical-based approach. Corpus/statistical based similarities can be
measured by two.
Different techniques, normal statistical analysis (LSA) and the other one is deep
learning. A large corpora analyzed statistically through counting the words within
the corpus and documents. Tf-idf is a vital statistical object for the corpus analysis
which is used to employ as a word co-efficient [23].
In this technique, words usually expressed as vectors. In LSA, the word matrix and
vectors constructed by analyzing the large text. The word matrix represents in the
form of rows and columns. In this corpus based analysis, usually cosine similarity
is highly used for calculating the word similarity text, tables and figures. Substantial
changes in content, e.g., new results, corrected values, title and authorship, are not
possible and can-not be processed.
This approach also known as word embedding technique which used to find the
words semantically. In the semantic space, a very big corpus is required to extract
the word representation. This technique is specifically used to guess the word from
the given surrounding words, usually known as bag of words model. This model also
helps to understand the vector representation. Deep learning also use the Skip gram
method to understand the context meaning of word. It also helps to understand the
word/vector semantically. For example if we perform some calculation on vectors
semantically,
The result of this vector shows the word “queen” which is generally very close
to the specific semantic calculation of word representation. Deep learning also use
web based similarity.
These methods works well to understand the meaning of word to process compu-
tationally. It actually use the WordNet for the lexical relationships. There are three
methods that helps WordNet to extract the similarity. These can be categorized as
node-based/Ic, edge-based and Glass-based measure.
This approach uses concepts to extract the similarity. This measure includes the
hierarchical structure. In this method every node uses IS-A taxonomy that are kept
in on set. There are also other measures that considers the node/Ic based measures
for example, RES, LIN and JCN [1]. Before implementing the node based model the
concepts can be divide into words(c) and classes defined as classes (w) [24].
Here in the whole hierarchy the class works as a subtree and the class (w) contains
all sets of possible words [24].
This is the direct approach for computing the similarity. This technique counts the
edges of those nodes that are correspond to each other. Distance based measures used
the hierarchal concept for WordNet where the PATH length can be used to measure
the similarity There are three other approaches based on Distance measures PATH,
WUP, LCH [1].
1040 Z. H. Amur et al.
These measures uses the glosses between the concepts of WordNet. Vector and LESK
are the type of this measure [1]. Table 1 Shows the Summary of various methods and
dataset used for the semantic similarity of short texts. This table presents the latest
work. Various datasets has been used by the different approaches the most common
is Microsoft paraphrase corpus. This dataset contain 5801 sentence pairs that can
extract from news, documents and various social network platforms.
This study proposed Research framework (Fig. 2) that has one independent variable
(Text assessment) whereas the other’s work as dependent variables. In this Frame-
work the text assessment will be done by the student’s essays. The evaluation will be
based on examiners or expert essays that will be available to check the relatability
among essays. These short texts divided into text extraction process that actually
extract the core keyword from text through the machine learning process. Further
the machine normalize the text through stemming and lemmatizing techniques to
remove the suffixes. In summarization the machine performs the text prepressing
that helps to reduce the noise from text to detect the stop words or punctuation
marks from Texts. Moreover the Intent recognition check the similarity among the
short texts to identify the relatable Keywords that has the highest similarity. This
technique helps to give the rank to the text (answers). Whereas (Table 2) Text assess-
ment, summarization, as well as similarity all these effects on the accuracy of short
texts. The measurable variables in framework are: Text Assessment, Text extraction,
intent recognition and accuracy.
In Q &A systems usually the Reponses works independently. So if the student’s
and examiner’s responses relate with each other than machine will understand the
context of word/sentence and indicates the predicate feature that has the higher
similarity. If the predicate feature has high similarity then the accuracy of STSS will
increase. If the predicate or candidate word failed to provide the accurate response
then the accuracy of answer will not match to the specific answer.
The main consideration of this study is to give brief overview on short texts. There are
various approaches like string similarity, corpus based or knowledge based. Usually
these approaches provide good impact on text representation. From the literature and
previous work we have identified and noticed that corpus based and knowledge based
measure are most suitable measures for text summarization. Corpus based measure
Table 1 Summary of various methods and datanet used for the Semantic similarity of short texts
Research Methods
WordNet LCS Word embedding Structure based Vector Based PMI-IR LSA Base method Dataset
Huang et al. [3], ✓ ✓ ✓ String Based MSRPar(Microsoft
2019 paraphrase corpus)
Islam et al. [25], ✓ ✓ ✓ Combined (String & MSRPar(Microsoft
2009 Corpus based) paraphrase corpus)
Zhao et al. [26], ✓ ✓ Corpus based Gigaword &
2019 DUC-2004
Suleiman et al. [27], ✓ ✓ Corpus based OSAC
2019
Pawar et al. [28], ✓ ✓ Knowledge based PILOT
2018
Hien et al. [29], 2019 ✓ ✓ Corpus based MRPC
P4PIN.STS2015
Cai et al. [30], 2020 ✓ ✓ ✓ String & knowledge M&C, R&G, WS-
State-of-the-Art: Short Text Semantic Similarity (STSS) Techniques …
based 353
Zhiguo et al. [30], ✓ Corpus based QASent, MSRP
2016
Avani et al. [28], ✓ ✓ Corpus based Kaggle
2019
1041
1042 Z. H. Amur et al.
provides the advantage of cosine similarity. These approaches helps to minimize the
risk of errors and improve the chances of accuracy between short texts. Morever, this
study proposed the research Framework that helps to find out the similarity among
short texts (answers). For the future work, there is a lot of work yet to be done on
unsupervised learning in semantic similarity.
References
1. Huang PS, Chiu PS, Chang JW, Huang YM, Lee MC (2019) A study of using syntactic cues
in short-text similarity measure. J Internet Technol 20(3):839–850
2. Jimenez S, Cucerzan SP, Gonzalez FA, Gelbukh A, Dueñas G (2018) BM25-CTF: improving TF
and IDF factors in BM25 by using collection term frequencies. J Intell Fuzzy Syst 34(5):2887–
2899
State-of-the-Art: Short Text Semantic Similarity (STSS) Techniques … 1043
3. Song X, Min YJ, Da-Xiong L, Feng WZ, Shu C (2019) Research on text error detection and
repair method based on online learning community. Procedia Comput Sci 154:13–19
4. Wang H, Yang W (2019) An intelligent test paper generation method to solve semantic similarity
problem. In: IOP conference series: earth and environmental science Vol 252, No 5. IOP
Publishing, p 052126
5. Gashaw I, Shashirekha HL (2020) Machine learning approaches for amharic parts-of-speech
tagging. arXiv preprint arXiv:2001.03324
6. Croft D, Coupland S, Shell J, Brown S (2013) A fast and efficient semantic short text similarity
metric. In: 2013 13th UK workshop on computational intelligence (UKCI) (pp. 221–227).
IEEE
7. Chen HH, Lin MS, Wei, YC (2006) Novel association measures using web search with double
checking. In: Proceedings of the 21st international conference on computational linguistics and
44th annual meeting of the association for computational linguistics, pp 1009–1016
8. Martinez-Gil J, Pichler M (2014) Analysis of word co-occurrence in human literature for
supporting semantic correspondence discovery. In: Proceedings of the 14th international
conference on knowledge technologies and data-driven business, pp 1–7
9. Wang H, Zeng M, Xiong Z, Yang F (2017) Finding main causes of elevator accidents via
multi-dimensional association rule in edge computing environment. China communications
14(11):39–47
10. Radford A, Narasimhan K, Salimans T, Sutskever I (2018) Improving language understanding
by generative pre-training
11. Bauer L, Wang Y, Bansal M (2018) Commonsense for generative multi-hop question answering
tasks. arXiv preprint arXiv:1809.06309
12. Nickel M, Murphy K, Tresp V, Gabrilovich E (2015) A review of relational machine learning
for knowledge graphs. Proc IEEE 104(1):11–33
13. Phan XH, Nguyen LM, Horiguchi S (2008) Learning to classify short and sparse text & web
with hidden topics from large-scale data collections. In: Proceedings of the 17th international
conference on World Wide Web, pp 91–100
14. Tachbelie MY, Abate ST, Besacier L (2011) Part-of-speech tagging for underresourced and
morphologically rich languages—the case of Amharic. HLTD 2011:50–55
15. Song Y, Wang H, Wang Z, Li H, Chen W (2011) Short text conceptualization using a prob-
abilistic knowledgebase. In: Proceedings of the twenty-second international joint conference
on artificial intelligence-volume volume three, pp 2330–2336
16. Kim D, Wang H, Oh A (2013) Context-dependent conceptualization. In: Proceedings of the
twenty-third international joint conference on artificial intelligence, ser. IJCAI’13. AAAI Press,
pp 2654– 2661. [Online]. Available: http://dl.acm.org/citation.cfm?id=2540128 2540511
17. Hua W, Wang Z, Wang H, Zheng K, Zhou X (2015) Short text understanding through lexical-
semantic analysis. In: 2015 IEEE 31st international conference on data engineering. IEEE, pp
495–506
18. Al-Bataineh H, Farhan W, Mustafa A, Seelawi H, Al-Natsheh HT (2019). Deep contextualized
pairwise semantic similarity for Arabic language questions. In: 2019 IEEE 31st international
conference on tools with artificial intelligence (ICTAI). IEEE, pp 1586–1591
19. Slam A, Inkpen D (2008) Semantic text similarity using corpus-based word similarity and
string similarity. ACM Trans Knowled Discov Data (TKDD) 2(2):1–25
20. Jaafar Y, Bouzoubaa K (2018) Towards a new hybrid approach for abstractive summarization.
Procedia comput sci 142:286–293
21. Vuong La CS (2019) Artificial intelligence versus natural stupidity: evaluating AI readiness
for the vietnamese medical information system. J clin med, 8(2):168
22. Otter DW, Medina JR, Kalita, JK (2020) A survey of the usages of deep learning for natural
language processing. IEEE Tran Neural Netw Learn Syst
23. Farouk M (2019) Measuring sentences similarity: a survey. arXiv preprint arXiv:1910.03940
24. Majumder G, Pakray P, Gelbukh A, Pinto D (2016) Semantic textual similarity methods, tools,
and applications: a survey. Computación y Sistemas 20(4):647–665
1044 Z. H. Amur et al.
25. Islam A, Inkpen D (2009) Semantic similarity of short texts. Recent Advances in Natural
Language Processing V 309:227–236
26. Zhao F, Quan B, Yang J, Chen J, Zhang Y, Wang X (2019) Document summarization using
word and part-of-speech based on attention mechanism. In: Journal of Physics: Conference
Series, Vol 1168, No 3. IOP Publishing, p 032008
27. Suleiman D, Awajan AA (2019) Using part of speech tagging for improving word2vec model.
In: 2019 2nd International Conference on new trends in computing sciences (ICTCS). IEEE,
pp 1–7
28. Wang Z, Mi H, Ittycheriah A (2016) Sentence similarity learning by lexical decomposition and
composition. arXiv preprint arXiv:1602.07019
29. Sakhapara A, Pawade D, Chaudhari B, Gada R, Mishra A, Bhanushali S (2019) Subjective
answer grader system based on machine learning. In: Soft computing and signal processing.
Springer, Singapore, pp 347–355
30. Cai Y, Pan S, Wang X, Chen H, Cai X, Zuo M (2020) Measuring distance-based semantic
similarity using meronymy and hyponymy relations. Neural Comput Appl 32(8):3521–3534
31. Tsatsaronis G, Varlamis I, Vazirgiannis M (2010) Text relatedness based on a word thesaurus.
J Artif Intell Res 37:1–39
32. Jin O, Liu NN, Zhao K, Yu Y, Yang, Q (2011) Transferring topical knowledge from auxiliary
long texts for short text clustering. In: Proceedings of the 20th ACM international conference
on information and knowledge management, pp 775–784
33. Han M, Zhang X, Yuan X, Jiang J, Yun W, Gao C (2021) A survey on the techniques, applica-
tions, and performance of short text semantic similarity. Concurrency Comput Pract Experience
33(5):e5971
Modeling of Wireless Power
Transmission for Pacemaker Battery
Charging in MATLAB Simulink
Environment
G. P. Kiruthiga (B)
Dr.N.G.P Institute of Technology, Kalapatti road, Coimbatore, India
M. B. Romlie
UniversitiTeknologi Petronas, Seri Iskandar, Perak, Malaysia
e-mail: [email protected]
K. Porkumaran
Sri Sairam Engineering College, Chennai, India
S. Prabakar
Sona College of Technology, Salem, India
e-mail: [email protected]
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 1045
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_99
1046 G. P. Kiruthiga et al.
1 Introduction
These days the world without electric force is troublesome thing to live. For instance,
the day-day items like electronic vehicle, versatile, PC, clothes washer and so on,
requires electric force. Like that force move through the little attachment containing
numerous wires may cause fire mishaps and demise [1]. Pacemaker is only it is a little
gadget that utilizes electrical motivation conveyed to the terminal by contracting heart
muscles after that they can direct the pulsating of the heart [5]. In 1958, the pacemaker
was embedded first with battery-powered battery made of nickel cadmium. In 1973,
lithium battery was utilized to control the pacemaker. A pacemaker is put in the chest
(just underneath the collarbone) to help manage moderate electrical issues with the
heart.
Pacemaker leads might be situated in the chamber (upper chamber) or ventricle
(lower chamber) or both, contingent upon the clinical condition [6]. Inherent coro-
nary illness in conceived babies needs to the implantation of pacemaker from their
introduction to the world to whole lifetime. And furthermore utilized in mature age
individuals to aid appropriate working of heart. Boundaries of pacemaker are; capac-
itor, timing circuit, terminal, beat generator, battery. In this paper we zeroed in on to
charge the battery.
In pacemaker battery generally they are utilizing Lithium batteries in light of longer
life expectancy and 5 V has been utilized, battery substitution is needed after 3.3 V.
Mercury batteries likewise utilized however 3–5 cells are utilized to make one
mercury battery, it makes costly. Natural force source battery is as yet in explo-
ration not yet emerged. Plutonium based batteries were utilized in times past, and
its life expectancy around 30 years however plutonium is atomic receptive and
undependable, harmful subsequently not utilized.
Remote Power Transfer (WPT) is a strategy for sending the electrical energy starting
with one point then onto the next in a remote medium (without actual association).
Two sorts in WPT,
• Near field it is non radiative and the force can be communicated to the short
distance and inductive coupling is utilized.
• Far field it is radiative and the force can be communicated to the significant
distance and capacitive coupling is utilized. For remote force transmission close
documented is utilized on the grounds that it is non radiative (Fig. 1).
Modeling of Wireless Power Transmission for Pacemaker Battery … 1047
From the above circuit when the power can be feed into the transmitter side then
by through the coupling devices the power can be goes to the receiver side and then
voltage is transferred to the battery. This is the basic wireless power transmission [7]
(Table 1).
Once the life span of battery is over, repeated changing of pacemaker would affect
the patients in these ways,
• Re-implantation of whole device cause defect to the surrounding tissues.
• Frequent surgeries for every 10 years interval for the new born babies with
congenital heart disease may lead to weakness of the patient body.
• As well as senior citizens who are above 60 years cannot opt surgery method
frequently [4].
To overcome the above issues, we adopted wireless power transmission to charge
the cardiac pacemaker can be the alternative for surgery. In this paper we designed
a simple model using MATLAB Simulink software for pacemaker wireless charger
that constitutes amplifier, oscillator, inductive power transfer coupler, rectifier, LC
filter and resistance load which represents pacemaker battery.
1048 G. P. Kiruthiga et al.
2 Proposed Work
In the wireless power transmission, the output circuit consists of a coil, a transmitter
which transmits the power and receiver receives the power and send it to the battery.
Whereas the transmitter side consist of an input voltage that is an AC voltage source,
amplifier and the oscillator. The receiver circuit consists of rectifier, monitoring
circuit and battery. All my simulation and output were generated using MATLAB
(Fig. 2).
2.1 Amplifier
From the transmitter side foremost component is class A amplifier with Darlington
pair. Moreover, in biomedical devices, the most commonly used type of amplifier is
the class A amplifier, it is the simplest form of power amplifier during one complete
cycle of input signal waveform producing minimum distortion and maximum ampli-
tude of the output signal. Amplifier is used to increase the power of the signal.
The current gain of the two-transistor individual is multiplied together along with
collector currents are compared to the single transistor circuit (Fig. 3).
From the above circuit the input power 5 V and the frequency is 50 Hz then the
supply goes to the capacitor C1 and value is 0.22 µF and it reaches the resistor
R1, R2 and value is 2000, 7250 k and then supply goes to the Darlington pair
of collectors and the process continues along the Darlington pair and the amplified
voltage is approximately 9 V (Fig. 4).
2.2 Oscillator
The acceptable range of frequency for a human body is 300 kHz and 3.55 MHz.
Hence the 5 V is given as the input voltage and obtaining the frequency in the range
300 kHz after oscillation (Fig. 5).
After we get the amplified voltage form the amplifier it reaches the oscillator. The
oscillator can oscillate the frequency. In the above circuit 555 timer is an 8 pin IC.
Pin 2 trigger is connected to the capacitor and the value is 0.1 nF, then the discharge
pin is said as pin 7 and pin 4 is said as reset pin, this is connected to the input DC
voltage source which is 5 V and control voltage pin is pin 5, which is grounded with
capacitor value 0.1 nF. We are using the Astable mode 555 timer as an oscillator and
1050 G. P. Kiruthiga et al.
it can generate the pulse. Generally, the Astable mode generates a square wave, in
order to obtain a Sine wave, we add L resonance of value 100 µH and C of value
2.5 nF along with the Astable mode 555 timer. The obtaining frequency range after
simulation is about 318 kHz (Fig. 6).
f = 1 ÷ 2π )Lc
Then substituting the values for above equation then the frequency is about, f =
318 kHz.
The output from the oscillator goes the capacitor which is connected parallel. The
AC voltage from the capacitor goes to the transmitter coil and the consists of mutual
inductance L1, L2 = 95.7 nH, by through the magnetic coupling signal goes to the
receiver coil. In the receiver side rectifier is present, here half rectifier is used, because
Modeling of Wireless Power Transmission for Pacemaker Battery … 1051
the secondary coil is placed inside the heart so component should not be complex.
Full wave rectifier is not used here because it is still in research. Rectifier is used
to convert alternating current into direct current. To the charge the battery, the input
will be in DC. The output from the rectifier is appeared as pulsated DC, for charging
of the battery pulsated DC is enough. By adding LC filter pure DC is formed. This
is more appropriate compared to pulsated DC (Fig. 7).
The above circuit mentioned is the overall circuit. When the input here is DC of
about 5 V then it goes to the amplifier circuit then it amplifies voltage about 9 V
then it goes to oscillator, and it converts DC voltage into AC voltage and it shapes
the frequency which is about 318 kHz and it goes to the transmitter coil through
the coupling device, it goes to the receiver coil, in that rectifier converts AC to DC
voltage. After rectification the output voltage is about 4 V, in spite of the loss occurred.
The simulation results show that the transferred efficiency is about 80.01% at the
distance 2 cm (Figs. 8 and 9).
4 Future Work
References
1. Gore VB, Gawali DH (2016) Wireless power transfer technology for medical applications
2. Peng K, Liu X, Huang P Study on the wireless energy supply system in the implantable cardiac
pacemaker
3. Campi T, Cruciani S, Palandrani F, De Santis V, Hirata A, Feliziani M Wireless power transfer
charging system for AIMDs and pacemakers
4. Lafrance A Who killed the rechargeable pacemaker? The Atlantic, 4 Feb 2014
5. Ali KK, Jarndal AH Remote-monitoring and wireless-powering system for pacemaker patients
6. Khan MU, Jafar A, Karimov KhS, Feroze S A proposed optimized solution for wireless power
transfer using magnetic resonance coupling
7. Hu Y, Zhang Y, Xu C, Lin L,Snyder RL, Wang ZL Self-powered system with wireless data
transmission
8. Vulfin V, Sayfan-Altman S, Ianconescu R Wireless power transfer for a pacemaker application
9. https://en.wikipedia.org/wiki/Wireless_power_transfer
10. Sung SH, Kim YS, Joe DJ, Mun BH, You BK, Keum DH, Hahn SK, Berggren M, Kim D,
Lee KJ Flexible wireless powered drug delivery system for targeted administration on cerebral
cortex
11. https://en.wikipedia.org/wiki/Artificial_cardiac_pacemaker
12. https://www.hopkinsmedicine.org/healthlibrary/test_procedures/cardiovascular/pacemaker_
insertion_92, p 07980
Recent t-way Test Generation Strategies
Based on Optimization Algorithms:
An Orchestrated Survey
Abstract In software testing, test case generation is the most challenging activities
and expensive tasks. Where has a considerable impact on the ability to produce
a desired or intended result (i.e., quality and efficacy) of software testing. As a
result, several researchers have developed a number of t-way test case generation
strategies (where t points the interaction strength between parameters) due to the
market demand to the various types of tests based on different approaches. This paper
presents an orchestrated survey of the latest test case generation strategies such as
Binary Black Hole (BBH), Sine Cosine Variable Strength (SCAVS), Combinatorial
Testing Based Jaya Algorithm (CTJ), deterministic genetic multi-parameter-order
(GAMIPOG) and Hybrid Artificial Bee Colony (HABC). This survey illustrates the
strengths and weaknesses of each strategy, and indicates potential research studies
in the field for future work.
1 Introduction
The researchers in the software testing have sophisticated a number of helpful tech-
niques in order to avoid bugs and promote bug detection the last few decades.
Although the sophisticated techniques are useful. However, the manufactured soft-
ware produced has no bug-free guarantee [1]. Thus, the only procedure that can
demonstrate the achieved quality and recognize the remaining problems, and risks
is software testing.
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 1055
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_100
1056 A. K. Alazzawi et al.
Every configuration system under test (CSUT) is consists of a different set of elements
called parameters (p), and each parameter is associated with different values (v).
Table 1 shows a simple software system to easily understand the interaction between
parameters and the associated values (i.e., t-way testing). Where this software system
consists of 3-parameters associated with 2-values each, 1-parameter associated with
3-values each and the interaction strength t = 2. This software system can be
expressed as covering array CA (N; 2, 23 31 ).
The exhaustive test suite size in this software system is 24 test cases (3 × 2×2
× 2 = 24) in order to test this system exhaustively as shown in Fig. 1. By executing
2-way testing, where interaction strength between parameter’s t = 2. The possible
combinations are 6 combinations between parameters (AB, AC, AD, BC, BD and
CD) respectively. The optimal test suite size achieved is 6 test cases (See Fig. 1) that
able to cover each interaction between parameters. Therefore, t-way testing achieved
a minimization from 24 test cases to 6 test case only (75% of total exhaustive testing).
The generation of optimal test suite size can be considered as an optimization problem
[12]. Test case generation strategies based meta-heuristics (i.e., optimization algo-
rithm) always outstanding in this regard. Numerous of the existing meta-heuristics
algorithms based strategies are using the one-test-at-a-time (OTAT) approach exclu-
sively. In general, all the test case generation strategies based meta-heuristics generate
non-deterministic solutions due to the randomization. In comparison, most studies
are limited to small interaction intensity values for pure computation-based equiva-
lents (e.g., 2 ≤ t ≤ 6). In addition, there was also inadequate investigation of support
for variable strength and constraints.
Binary Black Hole (BBH) is one of the most recently developed test case gener-
ation strategy based on meta-heuristic algorithms called Black Hole (BH) algorithm
[13, 14]. The inspiration of BBH strategy came by the interaction of the phenomenon
of the black hole and the star’s behaviour together. In case of the star become close
to the black hole, the black hole will swallow it. Where the algorithm will begins to
generate a new solution (i.e., new star) randomly and will be searched again by the
algorithm. BHH is an attractive strategy to solve the t-way testing problem because
it is a tuning-free. BHH does not need multiple parameters for balancing between
the global and local search like other generation strategies based on meta-heuristics.
These features make it simple and easy for implementation. Up until now, BBH
address small values of interaction strength (i.e., t ≤ 4). In fact, BBH do not address
variable strength interactions and do not provide the support for constraints.
1058 A. K. Alazzawi et al.
Sine Cosine Variable Strength (SCAVS) strategy [15] based on Sine Cosine Algo-
rithm (SCA) Algorithm for t-way variable test suite generation. SCA is meta-heuristic
based on exploit the mathematical properties of both sine and cosine trigonometric
functions. SCAVS is as any other algorithm starts first to produce random solutions.
SCAVS supports the interaction strength up to t = 6 and addressing variable strength
interaction. Although SCAVS addressing variable strength interaction, SCAVS does
not support the constraints.
Hybrid Artificial Bee Colony (HABC) strategy [16–19] is an extended version
based on Artificial Bee Colony (ABCVS) [10, 11] strategy in order to overcome the
limitations of the original Artificial Bee Colony (ABC) algorithm. HABC based on
hybridization of two meta-heuristic algorithms that can help the search capabilities of
each by compensating for the limitations of one algorithm with the strength of others.
HABC strategy based on merging the advantages of the original Artificial Bee Colony
(ABC) algorithm with the advantages of an original Particle Swarm Optimization
(PSO) algorithm. Unlike other strategies, HABC does support the high value of
interaction strength (i.e., t ≤ 6) and addresses the variable strength interactions.
HABC also addresses the support for constraints through its variant, called PhABC
[20, 21].
Unlike BBH, SCAVS and HABC that generate non-deterministic solutions,
GAMIPOG is deterministic strategy. Deterministic Genetic Multi-Parameter-Order
(GAMIPOG) strategy [22], based on Modified Input Parameter Order General
(MIPOG) for test case generation [23]. GAMIPOG overcomes many desired chal-
lenges by combined them together in one strategy in order to take the benefits of
each one such as meta-heuristics, one-parameter-at-a-time and one-test-at-a-time. In
addition, GAMIPOG comparing to other existing strategies have less complexity,
manageable test size and fast execution time. Where GAMIPOG works to find the
global solution by breaking the problem into sub-problems merging all solution’s
together iterative steps. Thus far, GAMIPOG addresses high values of interaction
strength (i.e., t ≤ 15) and addresses the variable strength interactions. Furthermore,
GAMIPOG does not provide the support for constraints.
Unlike all the above-mentioned strategies, CTJ is support input–output based
relation (IOR) for test case generation [24]. Combinatorial Testing based on Jaya
Algorithm (CTJ) strategy is one of the recently proposed strategies that support
input–output based relation (IOR) for test case generation. Where CTJ adopts the
Jaya algorithm as the main engine for generating the solutions. CTJ has two common
control parameters: (1) maximum iteration and (2) population size. This feature
makes CTJ performance to produce the solutions easily than other existing strategies.
Thus far, CTJ addresses small values of interaction strength (i.e., t ≤ 3) and addresses
the input–output based relation (IOR). Furthermore, CTJ does not provide the support
for constraints and variable strength interactions.
In conclusion, Table 2 gives a brief statement of the current t-way variable test
generation. Referring to Table 2, all the existing strategies are non-deterministic due
to the randomization (i.e., different test suite size in every implementation) except
GAMIPOG strategy. There was a lack of investigation of constraints support, input
output relations and as well as variable strength. GAMIPOG is the only strategy
Recent t-way Test Generation Strategies Based on Optimization … 1059
4 Conclusions
This paper summarizes the current existing t-way test case generation strategies based
on meta-heuristics and emphasises the strengths and weaknesses of each strategy. The
software market is massive, and therefore, the test is needed to ensure compliances
are varied; so new test ideas and test suite development are still needed. There are
also numerous other meta-heuristics algorithms that still do have good potential and
have not been adopted for t-way test suite generation.
References
1. Alsewari AA, Alazzawi AK, Rassem TH, Kabir MN, Homaid AAB, Alsariera YA, Tairan NM,
Zamli KZ (2017) ABC algorithm for combinatorial testing problem. J Telecommun Electron
Comput Eng (JTEC) 9:85–88
2. Alsariera YA, Zamli KZ (2015) A bat-inspired strategy for t-way interaction testing. Adv Sci
Lett 21:2281–2284
3. Alsariera YA, Majid MA, Zamli KZ (2015) Adopting the bat-inspired algorithm for interaction
testing. In: The 8th edition of annual conference for software testing, pp 14
4. Alsariera YA, Majid MA, Zamli KZ (2015) A bat-inspired strategy for pairwise testing. ARPN
J Eng Appl Sci 10:8500–8506
5. Esfandyari S, Rafe V (2018) A tuned version of genetic algorithm for efficient test suite
generation in interactive t-way testing strategy. Inf Softw Technol 94:165–185
6. Lei Y, Kacker R, Kuhn DR, Okun V, Lawrence J (2007) IPOG: a general strategy for t-way
software testing. In: Engineering of computer-based systems, 2007. ECBS’07. 14th annual
IEEE international conference and workshops on the, IEEE, pp 549–556
7. Lei Y, Kacker R, Kuhn DR, Okun V, Lawrence J (2008) IPOG/IPOG-D: efficient test generation
for multi-way combinatorial testing. Softw Testing Verif Reliabil 18:125–148
1060 A. K. Alazzawi et al.
8. Younis MI, Zamli KZ, Isa NM (2008) MIPOG-modification of the IPOG strategy for T-
Way software testing. In: Proceeding of the distributed frameworks and applications (DFmA),
Penang, Malaysia
9. Ahmed BS, Zamli KZ (2010) PSTG: a t-way strategy adopting particle swarm optimization. In:
The fourth asia international on mathematical/analytical modelling and computer simulation
(AMS), IEEE, pp 1–5
10. Alazzawi AK, Rais H, Basri S (2019) ABCVS: an artificial bee colony for generating variable
t-way test sets. Int J Adv Comput Sci Appl 10:259–274
11. Alazzawi AK, Rais HM, Basri S (2018) Artificial bee colony algorithm for t-way test suite
generation. In: 2018 4th international conference on computer and information sciences
(ICCOINS), IEEE, pp 1–6
12. Floudas C, Pardalos P, Adjiman C, Esposito W, Gumus Z, Harding S, Klepeis J, Meyer C,
Schweiger C (1999) Handbook of test problems in local and global optimization. Kluwer
Academic Publishers, Dordrecht
13. Nsaif HN, Jawawi DNA (2020) Binary black hole-based optimization for t-way testing. In:
IOP conference series: materials science and engineering, IOP Publishing, pp 012073
14. Al-Sammarraie HNN, Jawawi DN (2020) Multiple black hole inspired meta-heuristic searching
optimization for combinatorial testing. IEEE Access 8:33406–33418
15. Altmemi JM, Othman R, Ahmad R (2021) SCAVS: implement sine cosine algorithm for gener-
ating variable t-way test suite. In: IOP conference series: materials science and engineering,
IOP Publishing, pp 012011
16. Alazzawi AK, Rais HM, Basri S, Alsariera YA (2020) Pairwise test suite generation based on
hybrid artificial bee colony algorithm. In: Advances in electronics engineering, Springer, pp
137–145
17. Alazzawi AK, Rais HM, Basri S (2019) HABC: hybrid artificial bee colony for generating
variable T-Way test sets. J Eng Sci Technol 7:13
18. Alazzawi AK, Rais HM, Basri S (2019) Parameters tuning of hybrid artificial bee colony search
based strategy for t-way testing. Int J Innov Technol Explor Eng (IJITEE) 8
19. Alazzawi AK, Rais HM, Basri S (2019) Hybrid artificial bee colony algorithm for t-way
interaction test suite generation. In: Computer science on-line conference. Springer, pp 192–199
20. Alazzawi AK, Rais HM, Basri S, Alsariera YA, Balogun AO, Imam AA (2021) A hybrid
artificial bee colony strategy for t-way test set generation with constraints support. J Phys:
Conf Ser 042068. IOP Publishing
21. Alazzawi AK, Rais HM, Basri S, Alsariera YA (2019) PhABC: a hybrid artificial bee colony
strategy for pairwise test suite generation with constraints support. In: 2019 IEEE student
conference on research and development (SCOReD). IEEE, pp 106–111
22. Younis M (2020) Gamipog: a deterministic genetic multi-parameter-order strategy for the
generation of variable strength covering arrays. J Eng Sci Technol 15:3142–3161
23. Younis MI, Zamli KZ (2011) MIPOG-an efficient t-way minimization strategy for combinato-
rial testing. Int J Comput Theory Eng 3:388
24. Younis MI, Alsewari ARA, Khang NY, Zamli KZ (2020) CTJ: input-output based relation
combinatorial testing strategy using jaya algorithm. Baghdad Sci J 17
Research and Accomplishments
in Applications of Non-negative Matrix
Factorization
Phong Cao Nguyen, Cao Hong Nga, and Huong Hoang Luong
1 Introduction
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 1061
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_101
1062 P. C. Nguyen et al.
in the respective subjects in Sects. 3–5. Finally, it ends by describing how NMF is
being applied in even more fields, and how research to enhance NMF’s performance
in current applications is being carried out in Sect. 6, which is followed by a brief
conclusion.
Methods for summarizing and analyzing mass datasets are needed in order to effec-
tively utilize the huge amount of data available nowadays. For example, an online
movie platform should be able to recommend new movies to users based on their
past ratings, other users’ ratings, similarities between movies and so on. Consider
a simple case where users’ ratings (m users on n movies) are stored in the m × n
matrix X . The (i, j)-entry of X denotes the rating of user i towards movie j. Such
a matrix could easily be large; one million users and 10,000 movies constitute 10
billion entries. What NMF does is, for a relatively small integer k, finding an m × k
matrix W and a k × n matrix H such that every entry in W and H is non-negative
and X ≈ W H . The W and H matrices are usually called basis and weight matrices
respectively because each column in X can be viewed as a linear combination of the
rows in W with coefficients provided by columns in H [1] (see Fig. 1). The n movies
are also said to be divided into k clusters.
The NMF problem itself is NP-hard [3], which means algorithms for exact NMF
have exponential time complexity, thus too slow for real-world scenarios. Conse-
quently, most practical algorithms are heuristic with acceptable accuracy. One of the
most common algorithms is that presented by Lee and Seung in [1], on which many
variants have been based. The core of the standard algorithm lies in the iterative
updates of W and H [1]. The main steps of the algorithm are briefly listed below.
1. Define a function that measures the quality of the approximation, called a cost
function. A simple measure is the Euclidean distance E(A, B) between two m × n
matrices A and B:
n
m
Ai j
D (A||B) = Ai j log − Ai j + Bi j (2)
i=1 j=1
Bi j
(a) If the Euclidean distance was chosen, the (i, j)-entry of the matrices are updated
as follows.
X HT ij
Wi j ← Wi j (3)
W H HT ij
WT X
ij
Hi j ← Hi j T
(4)
W W H ij
(b) If the divergence was chosen, the (i, j)-entry of the matrices are updated as
follows.
H jk X ik
k (W H )ik
Wi j ← Wi j (5)
l H jl
Wki X k j
k (W H )k j
Hi j ← Hi j (6)
l Hli
Iteration continues until the value of the chosen function falls within the pre-specified
desired range.
The authors proved the convergence property of the algorithm, by which contin-
uous iterations would converge to a locally optimal factorization [1].
The algorithm is not without its flaws, however. Certain concerns regarding the
use of heuristic algorithms like this have been raised, including the non-uniqueness of
W and H (multiple satisfactory pairs may exist, in which case initialization must be
carefully determined) and the choice of rank k (an understanding of the underlying
structure of the objects may be required). Research into these issues is still on-
going, along with the development of variants of the standard algorithm like graph
regularized NMF [4] or hierachical NMF [5], just to name a few. Readers interested
in such research are referred to [1, 3, 6] and the mentioned works.
1064 P. C. Nguyen et al.
For each of PCA and NMF, training was conducted with dimensions of 50, 100
and 150 to reveal, if any, the relationship between dimensionality and the results of
face recognition. Initialization is effectively random. The experiments pointed out
that NMF performed better than PCA at recognizing smiles and was comparable
with FaceIt when it came to anger. However, scream proved too difficult for both
NMF and PCA. For the illumination set, NMF, when used in a high dimensional
space (150), even outperformed both FaceIt and Bayesian [8].
Under the effect of sunglasses, recognition rates for all approaches decreased
remarkably, implying the importance of eyes in recognizing faces [8]. NMF did best
when there was no light, but when lighting conditions affect the photos, it became
less capable. Similar results were observed for the scarf set, suggesting NMF was
good for a certain level of occlusions without another change in the scene [8].
NMF’s ability to recognize face could be enhanced by adding various sets of
constraints to the matrices W and H , enforcing these constraints by using different
cost functions, which in turn produce variants of pure NMF. A number of such
variants has been proposed, including local NMF (LNMF), Fisher NMF (FNMF)
or PCA NMF (PNMF). For example, LNMF adds three constraints to pure NMF,
which are (1) the number of bases should be minimized, which essentially means a
basis component should not be decomposed any further, (2) bases must be made as
orthogonal as possible, which is to minimize redundancy among bases and (3) only
1066 P. C. Nguyen et al.
bases with the most important information should be retained. LNMF consolidates
these constraints into its divergence function (specific limits for i and j have been
omitted for simplicity):
Xi j
D (X ||W H ) = X i j log − X i j + (W H )i j +α Ui j − β Q ii
i, j
(W H )i j i, j i
(7)
where α, β > 0 are constants, U = W T W and Q = H H T .
The reader is referred to [10] for a detailed description of these variants. Other
than the mentioned constraints, the core steps should remain more or less the same
to what was described in Sect. 2. Experiments with different databases demonstrated
the improved performance of FNMF and PNMF against NMF or LNMF in certain
circumstances due to the ability of these methods to better localize basis images to
the certain facial parts [10].
Network distances (including, but not limited to, round-trip time between hosts or
network latency) have a great effect on the operation of distributed applications. A
user who wants to download a file is able to choose the fastest server to fetch the file
from if information about network distances is available. However,
obtaining all pair-
wise distances for N hosts is expensive: it would take O N 2 measurements. Thus,
an issue of interest is estimating unknown distances from O(N ) measurements and
many studies have been dedicated to solving this problem. Early models are based on
Euclidean distance, in which hosts are represented as coordinates in a d-dimensional
systems.
The distance
between H i and H j represented by Hi 1 Hi 2 , . . . , Hi d and
,
H j1 , H j2 , . . . , H jd respectively is
d
2
Di j = Hik − H jk (8)
k=1
→ −→
−
Di j = Wi · H jT (9)
Under this model, distances do not have to observe the triangle inequality, and
Di j is not necessarily equal to D ji . Hence, it better predicts real-life distances com-
pared to earlier Euclidean distance-based models. In the paper, the authors proposed
two factorization algorithms—Single Value Decomposition (SVD) and NMF—and
found that NMF had a major advantage over SVD: it could “fill” missing values
in the distance matrix with a slight modification in the algorithm, while SVD had
to “skip” hosts with missing values [18]. The predictions were closer to the actual
values, but were still negatively affected by landmarks that often produced incorrect
values. Chen et al. [19] introduced weight to the system—landmarks that provided
correct values more often would play a larger part in calculating the prediction. They
demonstrated a significant decrease in relative error on predictions of certain datasets
when weight was added compared to the IDES model in [18, 19].
The NMF algorithms behind IDES and Phoenix both try to minimize the error
function E, which is the Euclidean error function presented in Sect. 2 [18, 19]:
N
N
2
E= Di j − D ji (10)
i=1 j=1
Chai et al. showed that higher accuracy in predicting network distances can be
achieved if the error function used is the divergence of D from R = W H , also
mentioned in Sect. 2 [20]:
N
N
Di j
K (D||R) = Di j log − Di j + Ri j (11)
i=1 j=1
Ri j
1068 P. C. Nguyen et al.
This class of NMF is usually called NMF using KL divergence (KL-NMF), and
it was also discussed in [1]. Apart from this difference, the steps in the algorithm
of [20] are similar to those in [18], making use of landmarks or reference hosts. Chai
et al. observed a 10–14.2% increase in prediction accuracy on four datasets when
compared with the version of NMF used in IDES [20]. The authors attributed the
improved performance to the ability of KL divergence to model network delay as
having a Poisson distribution rather than the normal distribution that the Euclidean
distance-based error function assumes [20].
Some genes in our body belong to the same group and contribute to an expression
or a disease. Clustering a large number of genes into a small number of metagenes
and analyzing the expression patterns of these metagenes are useful in discovering
hidden or unclear biological structure or relationships between groups of genes and
certain expressions. These clusters are comparable to facial parts in face recognition
applications. Brunet et al. applied NMF on Leukemia, Medulloblastoma and Cen-
tral Nervous System Tumors datasets to classify cancer subtypes [21]. This section
describes the experiment with the Leukemia set.
N genes and their expression levels in M samples (in the Leukemia dataset [24],
M = 38) are stored in the N × M matrix X . When an approximation X ≈ W H
is found (W is an N × k matrix), each column in W defines a metagene that is a
positive linear combination of the N genes, and each column in H represents an
expression pattern of each sample as a positive linear combination of the metagenes.
This factorization also divides M samples into k clusters; sample j belong to cluster
i if the (i, j)-entry of H is the largest in column j (see Fig. 4) [21].
Acute leukemia is generally divided into two types—acute myelogenous leukemia
(AML) and acute lymphoblastic leukemia (ALL), with the latter being further divided
into two subtypes—T and B cells. The authors compared NMF with two other meth-
ods, namely hierarchical clustering (HC) and self-organizing maps (SOM). HC is
essentially a process of clustering that involves merging the closest pair of clusters
at the previous step into a larger one until a strong cluster, which consists of all
objects, are obtained [22]. The result is a tree-like structure of clustering, hence the
name hierarchical. An SOM consists of nodes that are initially placed at random
into k-dimensional space, whose positions are then adjusted by moving the points
toward a random data point selected for that iteration. After the specified number
of iterations (about 20,000–50,000), the points that are near each other are likely to
form a related cluster [23]. Among the three, better methods would correctly classify
the samples into the 2 or 3 abovementioned clusters (depending on the rank used in
the algorithm) and provide more stable results (sample j belongs or does not belong
to cluster j in almost all of the runs of the algorithm with different initial conditions).
The findings were as follows [21].
HC
(1) HC split ALL-B samples into two subgroups (when they should have been
together).
SOM
(2) While SOM could divide the samples into two clusters, such classification was
not stable as the two groups could be either [AML] vs [ALL] or [AML + ALL-T] vs
[ALL-B].
(3) SOM with k = 3 classes did not correctly divide the samples into the three known
classes (ALL-T and ALL-B were mixed).
NMF
(4) NMF correctly split the samples into two clusters with remarkable stability regard-
less of initialization.
(5) When used with rank k = 3, NMF correctly distinguished ALL-T from ALL-B.
(6) Under higher ranks, NMF found a 4-cluster division quite robust while showing
considerable dispersion when k = 5, implying that many clusters were not likely.
The authors concluded NMF had multiple strong points when compared with HC
and SOM, and was better at discovering biological substructure [21]. sThis was con-
sistent with the observation that HC imposes strict hierachical structure to objects,
which may not always be the case, and at times needs subjective clustering definition,
while SOM, though imposes only a partial structure on the objects, which is more
appropriate, is unstable and may produce different clusterings with different initial-
izations [21]. Y. Gao and G. Church later demonstrated the enhanced performance
of sparse NMF, a variation, over classic NMF [25].
The non-uniqueness of W and H , however, may produce inconsistent results
between various runs of NMF. Yang and Seoighe studied some post-processing meth-
ods to improve the stability of the algorithm [26]. They standardized W and H using
1-norm, 2-norm, 3-norm and maximum norm in addition to quantiles and sample
standard deviation. Normalization maps the matrix W to a diagonal matrix D and
updates the solution as follows: W = W D −1 and H = D H . The authors also tried
applying filters to the resulting matrices so that irrelevant genes were removed before
running the algorithm again. Irrelevant genes were defined as those that had similar
1070 P. C. Nguyen et al.
values in the clusters, and thus it was not clear as to which metagenes they belonged
to [26]. When performed on the Leukemia dataset, maximum norm remarkably made
no errors even when there were two well-known outliers in the dataset [21]. In con-
trast, both outliers were incorrectly categorized when basic NMF was used. For most
of the other datasets in the experiments, NMF with maximum norm was also singled
out as the best classifier, and a filter was also proved to be able to boost performance
in certain cases [26].
In medical, NMF is also used to detect new subtypes of diseases. Chalise et al.
proposed combining NMF and similar network-based clustering (nNMF) across mul-
tiple data sources to recognize latent clustering structures in the data [31]. In this
method, each consensus matrix constructed from each dataset represents the simi-
larity between patients. These matrices are integrated into a robust consensus matrix
which represents the network of patient similarities and the nNMF clusters. Since
consensus matrices are calculated separately, they are not affected by the discrepan-
cies in the distributions and scales of the datasets. Consensus matrices are built on a
cross-validation approach that partially solves the overfitting and underfitting issues
of unsupervised methods. This method is applied on both real and simulated data
and proves the applicability of this method.
Applications of NMF are not limited to the above three fields. Texts of hundreds of
thousands of words are clustered into a few topics [27]. In astronomy, Ren et al. used
NMF to extract images of astrophysical objects out of direct imaging data [28] and
Berne et al. analyzed the emission of dust particles from spectro-imagery data [29].
New recommendation systems using NMF are being proposed [30]. Along with
improvements in the algorithm for NMF itself, new applications in existing and new
fields are certain to be discovered.
References
1. Lee DD, Seung HS (2001) Algorithms for non-negative matrix factorization. Adv Neu Inf Proc
Syst 14:556–562
2. Paatero P, Tapper U (1994) Positive matrix factorization: a non-negative factor model with
optimal utilization of error estimates of data values. Environmetrics 5(2):111–126. https://doi.
org/10.1002/env.3170050203
3. Gillis N (2014) The why and how of nonnegative matrix factorization. In: Suykens JAK, Sig-
noretto M, Argynou A (eds) Regularization, optimization, kernels, and support vector machines.
Chapman and Hall/CRC, New York, NY, USA, pp 257–291
4. Cai D, He X, Han J, Huang TS (2011) Graph regularized nonnegative matrix factorization for
data representation. IEEE Trans Patt Anal Mach Intell 33(8):1548–1560. https://doi.org/10.
1109/TPAMI.2010.231
5. Li Y et al (2013) Hierarchical non-negative matrix factorization (hNMF): a tissue pattern
differentiation method for glioblastoma multiforme diagnosis using MRSI. NMR in Biomed
26(3):307–319. https://doi.org/10.1002/nbm.2850
6. Berry MW, Browne M, Langville AN, Pauca VP, Plemmons RJ (2007) Algorithms and applica-
tions for approximate nonnegative matrix factorization. Comput Stat Data Anal 52(1):155–173.
https://doi.org/10.1016/j.csda.2006.11.006
7. Lee DD, Seung HS (1999) Learning the parts of objects by non-negative matrix factorization.
Nature 401(6755):788–791. https://doi.org/10.1038/44565
8. Guillamet D, Vitria J (2002) Non-negative matrix factorization for face recognition. Catalonian
Conf Artif Intell 336–344
9. Martínez AM, Benavente R (1998) The AR face database. Comput Vis Center Tech Rep
#24:1998
10. Wang Y, Jia Y, Hu C, Turk M (2005) Non-negative matrix factorization framework for
face recognition. Int J Patt Recogn Artif Intell 19(4):495–511. https://doi.org/10.1142/
S0218001405004198
11. Ng TSE, Zhang H (2002) Predicting Internet network distance with coordinates-based
approaches. In: Proceeding twenty-first annual joint conference of the IEEE computer and
communications societies, New York, NY, USA, pp 170–179 vol 1. https://doi.org/10.1109/
INFCOM.2002.1019258.
12. Dabek F, Cox R, Kaashoek F, Morris R (2004) Vivaldi: a decentralized network coordinate sys-
tem. ACM SIGCOMM Comput Commun Rev 34(4):15–26. https://doi.org/10.1145/1030194.
1015471
13. Paxson V (1997) End-to-end routing behavior in the Internet. IEEE/ACM Trans Netw 5(5):601–
615. https://doi.org/10.1109/90.649563
14. Lakshminarayanan K, Padmanabhan VN (2003) Some findings on the network performance
of broadband hosts. In: Proceeding 3rd ACM SIGCOMM Conference Internet Measurement,
pp 45–50. https://doi.org/10.1145/948205.948212
15. Zheng H, Lua EK, Pias M, Griffin TG (2005) Internet routing policies and round-trip-times.
In: International workshop on passive and active network measurement, pp 236–250. https://
doi.org/10.1007/978-3-540-31966-5_19
16. Lee S, Zhang ZL, Sahu S, Saha D (2006) On suitability of euclidean embedding of Inter-
net hosts. ACM SIGMETRICS Perform Eval Rev 34(1):157–168. https://doi.org/10.1145/
1140103.1140296
1072 P. C. Nguyen et al.
17. Wang G, Zhang B, Ng TE (2007) Towards network triangle inequality violation aware dis-
tributed systems. In: Proceeding 7th ACM SIGCOMM conference internet measurement, pp
175–188. https://doi.org/10.1145/1298306.1298331
18. Mao Y, Saul LK, Smith JM (2006) IDES: an internet distance estimation service for large
networks. IEEE J Select Areas Commun 24(12):2273–2284. https://doi.org/10.1109/JSAC.
2006.884026
19. Chen Y et al (2011) Phoenix: a weight-based network coordinate system using matrix factor-
ization. IEEE Trans Netw Serv Manage 8(4):334–347. https://doi.org/10.1109/TNSM.2011.
110911.100079
20. Chai L, Luo X, Zhao F, Li M, Liu S (2017) Network coordinate system using non-negative
matrix factorization based on KL divergence. In: 2017 19th international conference on
advanced communication technology (ICACT), Bongpyeong, pp 193–198. https://doi.org/10.
23919/ICACT.2017.7890082
21. Brunet JP, Tamayo P, Golub TR, Mesirov JP (2004) Metagenes and molecular pattern discovery
using matrix factorization. Proc. Nat Acad Sci 101(12):4164–4169. https://doi.org/10.1073/
pnas.0308531101
22. Johnson SC (1967) Hierarchical clustering schemes. Psychometrika 32(3):241–254. https://
doi.org/10.1007/BF02289588
23. Tamayo P et al (1999) Interpreting patterns of gene expression with self-organizing maps:
methods and application to hematopoietic differentiation. Proc Nat Acad Sci 96(6):2907–2912.
https://doi.org/10.1073/pnas.96.6.2907
24. Golub TR et al (1999) Molecular classification of cancer: class discovery and class prediction
by gene expression monitoring. Science 286(5439):531–537. https://doi.org/10.1126/science.
286.5439.531
25. Gao Y, Church G (2005) Improving molecular cancer class discovery through sparse
non-negative matrix factorization. Bioinformatics 21(21):3970–3975. https://doi.org/10.1093/
bioinformatics/bti653
26. Yang H, Seoighe C (2016) Impact of the choice of normalization method on molecular cancer
class discovery using nonnegative matrix factorization. In: PloS One 11(10):e0164880. https://
doi.org/10.1371/journal.pone.0164880
27. Hassani A, Iranmanesh A, Mansouri N (2019) Text mining using nonnegative matrix factor-
ization and latent semantic analysis. arXiv:1911.04705
28. Ren B, Pueyo L, Zhu GB, Debes J, Duchêne G (2018) Non-negative matrix factorization:
robust extraction of extended structures. Astrophys J 852(2). https://doi.org/10.3847/1538-
4357/aaa1f2
29. Berne O et al (2007) Analysis of the emission of very small dust particles from Spitzer spectro-
imagery data using blind signal separation methods. Astronomy Astrophys 469(2):575–586.
https://doi.org/10.1051/0004-6361:20066282
30. Benzi K, Kalofolias V, Bresson X, Vandergheynst P (2016) Song recommendation with non-
negative matrix factorization and graph total variation. In: IEEE international conference acous-
tics, speech and signal processing (ICASSP). Shanghai 2439–2443. https://doi.org/10.1109/
ICASSP.2016.7472115
31. Prabhakar C, Ni Y, Fridley BL (2020) Network-based integrative clustering of multiple types
of genomic data using non-negative matrix factorization. Comput Biol Med. https://doi.org/
10.1016/j.compbiomed.2020.103625
Feature Selection Using Correlation
Matrix on Metagenomic Data with
Pearson Enhancing Inflammatory Bowel
Disease Prediction
Huong Hoang Luong, Trong Thanh Tran, Ngoc Van Nguyen, An Duc Le,
Huyen Thi Thanh Nguyen, Khoi Dinh Nguyen, Nghi Cong Tran,
and Hai Thanh Nguyen
Abstract The Fourth Industrial Revolution has brought up a vast amount of new
innovative implementations. These can be used for numerous areas to make wealth
and to improve human ways of living. Our point of view is to consider medical
problems for enhancing prediction first. In this study, we like to rise a question about
whether if we could enhancing Inflammatory Bowel Disease (IBD) prediction for
early detect related sickness by feature selection on metagenomic data. Over the
last few years, the prediction has been a challenge. Because of rare information and
lacking data, the problem is not well considered enough. To bring back the subject, in
this work, we propose a new way of enhancing Inflammatory Bowel Disease (IBD)
prediction by using the Correlation Matrix with the Pearson on Metagenomic Data.
Our implications have the purpose of finding out whether we could do predictions
better using a specific amount of features selected by Pearson correlation coefficient.
The result with the proposed method is quite promising, when we address some high
correlation features out, the model can predict better comparing to randomly select
features.
H. T. Nguyen
College of Information and Communication Technologies (CICT),
Can Tho University, Can Tho, Vietnam
e-mail: [email protected]
H. H. Luong (B) · T. T. Tran · N. Van Nguyen · A. D. Le · H. T. T. Nguyen · K. D. Nguyen
Information Systems Department, FPT University, Can Tho, Vietnam
e-mail: [email protected]
N. C. Tran
National Central University, Taoyuan, Taiwan
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2022 1073
R. Ibrahim et al. (eds.), International Conference on Artificial Intelligence
for Smart Community, Lecture Notes in Electrical Engineering 758,
https://doi.org/10.1007/978-981-16-2183-3_102
1074 H. H. Luong et al.
1 Introduction
In recent years, the quantity of deaths related to the gastrointestinal tract is increas-
ing. Inflammatory Bowel Disease (IBD) is not a doom disease but worth of mention
because of its part to develop some deadly diseases. IBD is also one of the primary
causes of colorectal cancer (CRC) [1]. IBD [2] is a term for two main categories
including Crohn’s disease (CD) [3] and Ulcerative colitis (UC) [4], which is char-
acterized by chronic inflammation of the digestive tract. Digestive tract damages
could be caused by prolonged inflammation which is also a major risk factor for
the formation and development of malignancies. Generally, cancer [5–7] is a dis-
ease characterized by an uncontrolled nature division and the existence of abnormal
cells. There are abnormal growth occurs in the colon or rectum, it is called colorectal
cancer [8, 9].
There are many factors that result in the disease including lifestyle, genetics,
and environmental surroundings. According to WHO statistics in 2018 [10], CRC
is the second most deadly cancer and the third most commonly diagnosed cancer
in the world. There are 1.80 million new cases and 862 000 deaths. In the United
States, CRC is the second most common cancer diagnosis. The American Cancer
Society’s [11] estimate for 2020 is 147,950 new cases. It is expected to cause around
53,200 deaths by 2020. The dangers if the person has this disease: cancer cells only
form and grow in the colon area in an early stage. Therefore, treatment is not too
difficult. Moving into the second stage, the pathogenic cells attack other parts of
the colon. This is a milestone that marks the development of cancer in the colon.
If cancer cells are not controlled in time, they begin to work harder, spreading to
lymph nodes. More lymph nodes that are attacked, the more severe your condition
becomes, and the harder it is to control. Finally, when the disease enters a terminal
stage, almost every organ of the patient is affected by the cancer cells. Treatment
does not bring much hope for the patient. The risk of death is high. For prevention
and early detection of the disease, an improved approach is required to cut back the
number of individuals who are becoming this illness, further leading patients into a
brand new healthy lifestyle.
Later to improve patients’ experiences with the growth of personalized medicine.
We see that patients are diagnosed with the disease in less time, assess the risks,
evaluate the optimal treatment, and last but not least help them improve health while
reducing expenses. On the other side, medical manufacturers also are provided a
chance to develop agents that are targeted to patient groups that do not respond to
medications as intended. New value assessments for personalized medicine products,
together with return-on-investment models, will be required as these new strategies
for pharmaceutical and diagnostic products emerge. For the direct side, hospitals,
health care providers, and health plan sponsors, personalized medicine represent yet
one more challenge in uncertain times. Innovation in provider and benefits manage-
ment, together with clarity in regulatory and legal constructs are required, even as
new national insurance reforms begin to emerge. The applications of PM are diag-
nosis and intervention, drug development and usage, respiratory proteomics, cancer
Feature Selection Using Correlation Matrix on Metagenomic … 1075
genomics, and population screening. And through that use metagenomic to research
methods for personalized medicine. Metagenomics is the research of metagenomes
that aims to obtain genetic material directly from samples within the real living envi-
ronment. This field is included in environmental genetics, ecological genetics, or
simply genetics generally. Due to its important role within the discovery of microbial
diversity, metagenomics are often considered as a prism to assist us better understand
the planet of small organisms, contributing to humanity’s understanding of the entire
living world. In medicine, personalized medicine could be a type of using informa-
tion a couple of person’s genes or proteins, from the data above wont to diagnose,
prevent, or treat a disease that someone may have. For that individual field of study,
aimed by the expansion of machine learning technology. We will combine medi-
cal data and machine learning by determining the concentration of bacteria that are
characteristic of the chance of making tumors that cause sickness.
Some current applications of personalized medicine and machine learning tech-
nology are extremely helpful not only within the way of supporting doctors in early
diagnosis but also for the sake of all stakeholders [12]. Machine learning has created
an unlimited amount of applications in healthcare. These are a number of imple-
menting machine learning in healthcare like diagnosis and identifying diseases, drug
discovery, medical imaging diagnosis, etc. Personalized medicine is one in these
applications which is referenced during this article [13].
During this particular article, we illustrate a way approaches for using the correla-
tion matrix to boost Inflammatory Bowel disease prediction. Our goal is to represent a
brand new efficient set of methods thanks to the trustworthy results of the prediction.
Our paper consists of 4 main sections. In the next Sect. 2, we present some high-
lights of related work. Afterward, the methodology Sect. 3 where we introduced to
you the methods which we use on the topic later. The next Sect. 4 will be details about
the Experiments. Finally, in the final Sect. 5, we conclude our paper and summarize
the point of our experiments.
2 Related Work
Numerous research studies have applied Machine Learning to analyze and visual-
ize metagenomic datasets. In particular, authors have done a study on the method
to increase disease prediction based on Metagenomic using Unsupervised Binning
Approaches. Their research aims to present one-dimensional (1D) using a binning
and scaler approach. They also use Multi-layer Perceptron (MLP) and traditional
artificial neural networks. In the paper, their purpose is to divide the bin into differ-
ent which includes frequency, width, and proposed breaks. From there, the process of
learning with MLP is to classify diseases with multi-fold (K-fold). The Binning clas-
sification method was implemented in the test with 10 bins. The implemented MLPs
are the Rectified Linear Unit (ReLU) [14]. Additionally, the authors investigate many
methods for visualizing features of metagenomic datasets relate to colorectal cancer.
Their objective is to propose an efficient method to present data distribution visually
1076 H. H. Luong et al.
and use for classification. They proposed three methods including Image Generation
and Models for Evaluation, Visualizations Based on Dimensionality Reduction Algo-
rithms, and Visualization Based on Data Density. Image Generation and Models for
Evaluation use Species Bins and Quantile Transformation Bins to generate the image,
then CNN is used to evaluate the considered visualization and models. Visualiza-
tions Based on Dimensionality Reduction Algorithms apply dimensionality reduc-
tion algorithms such as t-SNE, Principal Component Analysis, Random Projection,
Spectral Embedding, Isomap, Locally Linear Embedding, Multidimensional scaling
and Non-negative matrix factorization. They are used to shaped high-dimensional
data in 2D images for improving prediction results. Visualizations based on the Data
Density illustrate the data based on the data density with Fill-up [17]. Besides, data
visualization is a challenge for many areas. With measurement data, it is character-
ized by the size of the data that is difficult for humans to interpret. In diseases that
use measurement data to predict, DL is often less effective than Classical Machine
Learning in predicting CRC. Authors present an approach using Manifold Learning
with t-SNE [15] and Spectral embedding to visualize digital data into images and
using DL to improve performance in predicting CRC. Results come from 5 different
areas used to combine visualization with Deep Learning to improve performance in
diagnostics.
In the last few years, Metagenomics is a study that includes many genomes at
the same time. This Metagenomics can be collected from a variety of environments.
For example, Bacteria in the human intestine are estimated to have between 500 and
more than 1000 species in the human intestine. However, investigating the disease
remains a major challenge due to inconsistent results in disease prediction and dis-
ease complexity. High-dimensional researchers have difficulty understanding data.
Many studies use 2D illustrations to interpret data and find patterns in the data. To
explain metagenomic data, many studies devise methods to visualize data. Krona
[16] allows to display data in order and use the zoom feature of multi-layered pie
charts. The survey study shows that: Manifold Learning-based approach shows the
abundance of bacteria that cause colon cancer through data samples, from which 2
out of algorithms are selected. The best is Spectral embedding (SE) and t-SNE [15]
for further investigation in preparation for the filling method and based on the mean
magnitude of the feature. Visualization based on the fill approach: arrange and fill
the values of the features into a matrix in order of right to left and top to bottom. The
author fills in using phylogenetic ordering and random sorting. Then, the proposed
visuals are loaded into a DL algorithm for prediction. In short, this article uses Man-
ifold Learning with 2 methods of embedding SE and t-SNE to put data into a Deep
Learning algorithm to predict and then Fill-up [17, 18]. Moreover, authors show
how to use the CNN network on data that has no original visual structure. The first
is to map the data and define it to a 1D or 2D image. Based on that description they
apply the CNN network to predict the disease. The collection of data in the medical
field has created a revolution in medical research, application and biotechnology.
These data are called omics data, they reflect aspects of the biology of the system.
More and more data sets are available and models are using this information to make
decisions. Today, the most successful techniques applied to this omics data set are
Feature Selection Using Correlation Matrix on Metagenomic … 1077
mainly Random Forest (RF) and sparse regression. In this paper, authors use DL
directly onto 6 measurement datasets to reflect the abundance of bacterial species,
their presence in the intestines of sick and healthy patients. This technology works
well in image classification we focus on using CNN. The purpose of the paper is to
propose an effective, concise presentation in images and demonstrate the DL tech-
niques as an effective tool for the metagenomics prediction tasks. Methods include
the Visualization of image-based representations. Its steps are 1 color is chosen and
applied according to the abundance of distribution. The division can be done on a lin-
ear or logarithmic scale. The features are displayed as images in 1 of 2 ways, namely
phylogenetic-sorting (PLG) or t-SNE. In short, this paper proposes the MET2IMG
method to predict the patient’s disease using measurement data. Authors use two
main methods to construct synthetic images which are Fill-up [17] and t-SNE [19].
The Fill-up [17] used small and simple images. The T-SNE [19] feature is embedded
in two-dimensional space using classic embedding in ML [20].
3 Methodology
In our study, we select features using the Correlation Matrix with Pearson [21]. It is
a widely used method to classify and visually. We use the Correlation Matrix with
Pearson which measures linear dependence between two variables [22]. Later we
will be able to pick out some of the features that have a high correlation coefficient.
The datasets as mention in 4.1 Datasets Description, consisting of 6 sets of data.
Each of the datasets includes the number of rows corresponding to be called, the
subjects. These next columns contain the code of the bacteria and their value Metage-
nomic Data corresponding to the subjects. After implementing the Correlation Matrix
with Pearson, we get a matrix n · n, with n is the number of features. Matrix values
are the Correlation Coefficient values range from −1.0 to +1.0 with −1.0 is a total
negative correlation, 0.0 is no correlation and +1.0 is a total positive correlation.
Regardless of the sign correlation, we only take absolute values. These absolute val-
ues will be considered to be greater than n to let the features to be selected. In the
study, we set n to 0.1, 0.2 and 0.3.
For the comparison purpose, we used randomly select features to compare the Feature
Selection using the Correlation Matrix with Pearson for the final result. We used the
IBD datasets mentioned in the previous Sect. 3.1 as well. With this method, we based
on the number of features that using Feature Selection using Correlation Matrix with
1078 H. H. Luong et al.
Pearson has filtered out to random. From there we will have the Feature Selection
randomly correspond to Feature Selection using Correlation Matrix with Pearson in
quantity terms. The later process is the same.
For both methods, randomly selection or feature selection using Correlation Matrix
with Pearson, we apply K-Fold cross-validation [23] after finding out Feature Selec-
tion using Correlation Matrix with Pearson or randomly which one predict better.
K-Fold cross-validation is a statistical technique to estimate efficiency of ML model
to independent datasets. The goal of this method is to test the model’s skill to eval-
uate how accurately the model’s prediction result can generate with an unknown
dataset with the purpose of avoiding overfitting [24] or selection bias [25]. Due to
the amount of data is not abundant, we assigned K with values 5 which means divides
datasets into 5 groups. In every group of data, the train data and the test data was
divided by the algorithm so that we can evaluate the effectiveness of the model, and
the remaining groups for training the model. Based on the metrics mention in 4.2.
4 The Experiments
In the previous Sect. 4.1, we have been explored 6 informative datasets. Further to the
problem, we divide these datasets into two groups: one group for train and another
group for tests. In the first group, we choose UCr for feature selection later on for the
train. In the other group, five others set later on for tests. For evaluating the results
we used three scoring metrics:
• Accuracy (ACC): ACC metrics results in range from 0.0 to 1.0. The value 1.0
indicates perfect prediction and value 0.0 means that prediction all false. Depend-
ing on the number of correct predictions n corr ect and total predictions n total the
equation accuracy be:
n corr ect
ACC =
n total
We describe three scenarios for selecting features using Correlation Matrix with
Pearson on the dataset UCr. When all features have the absolute correlation coefficient
greater than n, these features later will be used for the trained model and give out
prediction.
1080 H. H. Luong et al.
In the first scenario, we select all features which have n > 0.1 (159 features are
selected). That means a feature has a positive correlation or negative correlation
which is greater than 0.1 will be selected. The matrix is shown in Fig. 1
In the second scenario, we select all features which have n > 0.2 (30 features
are selected). That means a feature has a positive correlation or negative correlation
which is greater than 0.2 will be selected. The matrix is shown in Fig. 2.
In the final scenario, we select all features which have n > 0.3 (10 features are
selected). That means a feature has a positive correlation or negative correlation
which is greater than 0.3 will be selected. The matrix is shown in Fig. 3.
Following Sect. 3, we got three sets of selected features from three scenarios : cor-
relation >0.1 (159 features are selected), correlation >0.2 (30 features are selected)
and correlation >0.3 (10 features are selected). We also take three sets of random
features with the same size, correspondingly. Totally, we got 6 sets from UCr. These
Feature Selection Using Correlation Matrix on Metagenomic … 1081
sets will be used for the trained model, then test on the other datasets (iCDr, UCf,
iCDf, CDr and CDf).
After predict using K-fold cross-validation method we got results in three scoring
metrics. These metrics include Accuracy (ACC), Matthews correlation coefficient
(MCC) and Area Under Curve (AUC). The results shown in Tables 2 and 3.
5 Conclusion
As the prediction result has shown, we could state that using the Correlation Matrix on
Metagenomic Data with Pearson for Feature Selection has an extraordinary improve-
ment during predicting IBD patients compared to random features selection. This is
made evident by the number that same size random features are not well suited for
the train prediction model. The way of using the Correlation Matrix to take the high
relate mutually feature to train is better for enhancing the prediction. Further research
based on that premise could lead people to a healthy life with early detection of any
kind of illness not just only IBD.
1082 H. H. Luong et al.
References
1. Kim ER, Chang DK (2014) Colorectal cancer in inflammatory bowel disease: the risk, patho-
genesis, prevention and diagnosis. World J Gastroenterol 20(29):9872–9881. https://doi.org/
10.3748/wjg.v20.i29.9872
2. Centers for Disease Control and Prevention: Inflammatory bowel disease (IBD), from https://
www.cdc.gov/ibd/what-is-IBD.htm
3. NIH-U.S. National Library of Medicine: Crohn’s Disease, from https://medlineplus.gov/
crohnsdisease.html
4. National Institute of Diabetes and Digestive and Kidney Diseases (NIDDK): Ulcerative Colitis,
from https://www.niddk.nih.gov/health-information/digestive-diseases/ulcerative-colitis
5. NHS: Overview Cancer, from https://www.nhs.uk/conditions/cancer/
6. NIH-National Cancer Institute: Understanding Cancer, from https://www.cancer.gov/about-
cancer/understanding/what-is-cancer
7. American Cancer Society: What Is Cancer? from https://www.cancer.org/cancer/cancer-
basics/what-is-cancer.html
8. (ASCRS) American Society of Colon & Rectal Surgeons: The Colon: What it is, What it
Does and Why it is Important: Overview Cancer, from https://fascrs.org/patients/diseases-
and-conditions/a-z/the-colon-what-it-is,-what-it-does
9. Innerbody: Rectum, from https://www.innerbody.com/image_digeov/dige14-new3.html
10. World Health Organization: Cancer. Retrieved September 28, 2020, from https://www.who.
int/news-room/fact-sheets/detail/cancer
11. American Cancer Society: Key Statistics for Colorectal Cancer. Retrieved September 28, 2020,
from https://www.cancer.org/cancer/colon-rectal-cancer/about/key-statistics.html
12. Vogenberg F, Isaacson Barash C, Pursel M (2010) Personalized medicine: Part 1: Evolution
and development into theranostics. Retrieved September 27, 2020, from https://www.ncbi.nlm.
nih.gov/pmc/articles/PMC2957753/
13. Top 10 Applications of Machine Learning in Healthcare - FWS. (n.d.). Retrieved September
27, 2020, from https://www.flatworldsolutions.com/healthcare/articles/top-10-applications-
of-machine-learning-in-healthcare.php
14. Nguyen TH, Zucker J (2019) Enhancing metagenome-based disease prediction by unsupervised
binning approaches. In: 2019 11th international conference on knowledge and systems engi-
neering (KSE), da nang, Vietnam, 2019, pp 1–5. https://doi.org/10.1109/KSE.2019.8919295
1084 H. H. Luong et al.
15. Nguyen TH, Nguyen TN (2019) Disease prediction using metagenomic data visualizations
based on manifold learning and convolutional neural network. Lecture Notes in Computer
Science, vol 11814. Springer, Cham. https://doi.org/10.1007/978-3-030-35653-8_9
16. O ndov BD, Bergman NH, Phillippy AM (2011) Interactive metagenomic visualization in a web
browser. BMC Bioinform. 12:385. https://doi.org/10.1186/1471-2105-12-385. (ISSN:1471-
2105)
17. Nguyen TH et al (2018) Disease classification in metagenomics with 2D embeddings and deep
learning. In: The annual French conference in machine learning (CAp 2018). France: Rouen;
June 2018. arXiv: 1806.09046
18. Thanh-Hai N, Thai-Nghe N (2020) Diagnosis approaches for colorectal cancer using manifold
learning and deep learning. SN COMPUT. SCI. 1:281
19. Laurens van der Maaten GH (2008) Visualizing data using t-sne. J Mach Learn Res 9:8
20. Nguyen T, Chevaleyre Y, Prifti E, Sokolovska N, Zucker J (2017) Deep learning for metage-
nomic data: using 2D embeddings and convolutional neural networks. arXiv: 1712.00244
21. Benesty J, Chen J, Huang Y, Cohen I (2009) Pearson correlation coefficient. In: Noise reduction
in speech processing. Springer Topics in Signal Processing, vol 2. Springer, Berlin, Heidelberg.
from https://doi.org/10.1007/978-3-642-00296-0_5
22. Correlation Test Between Two Variables in R. (n.d.). Retrieved October 13, 2020, from http://
www.sthda.com/english/wiki/correlation-test-between-two-variables-in-r
23. The ‘K’ in K-fold cross-validation: davide anguita, Luca Ghelardoni, Alessandro Ghio, Luca
Oneto and Sandro Ridella https://www.elen.ucl.ac.be/Proceedings/esann/esannpdf/es2012-
62.pdf
24. Overfitting (2020) Retrieved 13 October 2020, from https://en.wikipedia.org/wiki/Overfitting
25. Selection bias (2020) Retrieved 13 October 2020, from https://en.wikipedia.org/wiki/
Selection_bias
26. Sokol H, Leducq V, Aschard H et al (2017) Gut 66:1039–1048
27. Fioravanti D et al (2018) Phylogenetic convolutional neural networks in metagenomics. BMC
Bioinformatics 19.S2 (2018): n. pag. Crossref. Web
28. Boughorbel S, Jarray F, El-Anbari M (2017) Optimal classifier for imbalanced data using
Matthews Correlation Coefficient metric. PLoS ONE 12(6): e0177678. https://doi.org/10.1371/
journal.pone.0177678
29. En.wikipedia.org (2020) Matthews correlation coefficient. Retrieved October 12, 2020, https://
en.wikipedia.org/wiki/Matthews_correlation_coefficient
30. Huang J, Ling CX (2005) Using AUC and accuracy in evaluating learning algorithms. IEEE
Trans Knowl Data Eng 17(3):299–310. https://doi.org/10.1109/TKDE.2005.50