Download Image Processing and Communications Techniques Algorithms and Applications Michał Choraś ebook All Chapters PDF
Download Image Processing and Communications Techniques Algorithms and Applications Michał Choraś ebook All Chapters PDF
com
https://textbookfull.com/product/image-processing-and-
communications-techniques-algorithms-and-applications-
michal-choras/
OR CLICK BUTTON
DOWNLOAD NOW
https://textbookfull.com/product/advances-and-applications-of-
optimised-algorithms-in-image-processing-1st-edition-diego-oliva/
textboxfull.com
https://textbookfull.com/product/image-processing-and-gis-for-remote-
sensing-techniques-and-applications-second-edition-liu/
textboxfull.com
https://textbookfull.com/product/nature-inspired-optimization-
techniques-for-image-processing-applications-jude-hemanth/
textboxfull.com
https://textbookfull.com/product/image-processing-and-communications-
challenges-7-1st-edition-ryszard-s-choras-eds/
textboxfull.com
Metaheuristic Algorithms for Image Segmentation Theory and
Applications Diego Oliva
https://textbookfull.com/product/metaheuristic-algorithms-for-image-
segmentation-theory-and-applications-diego-oliva/
textboxfull.com
https://textbookfull.com/product/image-color-feature-extraction-
techniques-fundamentals-and-applications-jyotismita-chaki/
textboxfull.com
https://textbookfull.com/product/fractals-applications-in-biological-
signalling-and-image-processing-1st-edition-aliahmad/
textboxfull.com
https://textbookfull.com/product/methods-and-techniques-for-fire-
detection-signal-image-and-video-processing-perspectives-1st-edition-
cetin/
textboxfull.com
https://textbookfull.com/product/signal-and-image-processing-
techniques-for-the-development-of-intelligent-healthcare-systems-e-
priya/
textboxfull.com
Advances in Intelligent Systems and Computing 1062
Michał Choraś
Ryszard S. Choraś Editors
Image
Processing and
Communications
Techniques, Algorithms and
Applications
Advances in Intelligent Systems and Computing
Volume 1062
Series Editor
Janusz Kacprzyk, Systems Research Institute, Polish Academy of Sciences,
Warsaw, Poland
Advisory Editors
Nikhil R. Pal, Indian Statistical Institute, Kolkata, India
Rafael Bello Perez, Faculty of Mathematics, Physics and Computing,
Universidad Central de Las Villas, Santa Clara, Cuba
Emilio S. Corchado, University of Salamanca, Salamanca, Spain
Hani Hagras, School of Computer Science and Electronic Engineering,
University of Essex, Colchester, UK
László T. Kóczy, Department of Automation, Széchenyi István University,
Gyor, Hungary
Vladik Kreinovich, Department of Computer Science, University of Texas
at El Paso, El Paso, TX, USA
Chin-Teng Lin, Department of Electrical Engineering, National Chiao
Tung University, Hsinchu, Taiwan
Jie Lu, Faculty of Engineering and Information Technology,
University of Technology Sydney, Sydney, NSW, Australia
Patricia Melin, Graduate Program of Computer Science, Tijuana Institute
of Technology, Tijuana, Mexico
Nadia Nedjah, Department of Electronics Engineering, University of Rio de Janeiro,
Rio de Janeiro, Brazil
Ngoc Thanh Nguyen , Faculty of Computer Science and Management,
Wrocław University of Technology, Wrocław, Poland
Jun Wang, Department of Mechanical and Automation Engineering,
The Chinese University of Hong Kong, Shatin, Hong Kong
The series “Advances in Intelligent Systems and Computing” contains publications
on theory, applications, and design methods of Intelligent Systems and Intelligent
Computing. Virtually all disciplines such as engineering, natural sciences, computer
and information science, ICT, economics, business, e-commerce, environment,
healthcare, life science are covered. The list of topics spans all the areas of modern
intelligent systems and computing such as: computational intelligence, soft comput-
ing including neural networks, fuzzy systems, evolutionary computing and the fusion
of these paradigms, social intelligence, ambient intelligence, computational neuro-
science, artificial life, virtual worlds and society, cognitive science and systems,
Perception and Vision, DNA and immune based systems, self-organizing and
adaptive systems, e-Learning and teaching, human-centered and human-centric
computing, recommender systems, intelligent control, robotics and mechatronics
including human-machine teaming, knowledge-based paradigms, learning para-
digms, machine ethics, intelligent data analysis, knowledge management, intelligent
agents, intelligent decision making and support, intelligent network security, trust
management, interactive entertainment, Web intelligence and multimedia.
The publications within “Advances in Intelligent Systems and Computing” are
primarily proceedings of important conferences, symposia and congresses. They
cover significant recent developments in the field, both of a foundational and
applicable character. An important characteristic feature of the series is the short
publication time and world-wide distribution. This permits a rapid and broad
dissemination of research results.
Editors
Image Processing
and Communications
Techniques, Algorithms and Applications
123
Editors
Michał Choraś Ryszard S. Choraś
Institute of Telecommunications Department of Telecommunications,
and Computer Science Computer Sciences
University of Science and Electrical Engineering
and Technology (UTP) University of Science
Bydgoszcz, Poland and Technology (UTP)
Bydgoszcz, Poland
This Springer imprint is published by the registered company Springer Nature Switzerland AG
The registered company address is: Gewerbestrasse 11, 6330 Cham, Switzerland
Preface
The monograph contains high-level papers which address all aspects of image
processing (from topics concerning low-level to high-level image processing),
pattern recognition, novel methods and algorithms as well as modern
communications.
We would like to thank all the authors and also the reviewers for the effort they
put into their submissions and evaluation.
We are grateful to Agata Gielczyk and Dr Karolina Skowron for their
management work, to Dr Adam Marchewka for hard work as Publication Chair,
and also to Springer for publishing this book in their Advances in Intelligent
Systems and Computing series.
Those papers have also been presented at IP&C 2019 Conference in Bydgoszcz.
Michał Choraś
Conference Chair
v
Organization
Organization Committee
Conference Chair
Michał Choraś, Poland
Honorary Chairs
Ryszard Tadeusiewicz, Poland
Ryszard S. Choraś, Poland
vii
viii Organization
Organizing Committee
Łukasz Apiecionek
Sławomir Bujnowski
Piotr Kiedrowski
Rafał Kozik
Damian Ledziński
Zbigniew Lutowski
Adam Marchewka (Publication Chair)
Beata Marciniak
Tomasz Marciniak
Ireneusz Olszewski
Karolina Skowron (Conference Secretary)
Mścisław Śrutek
Łukasz Zabłudowski
Contents
ix
x Contents
Boguslaw Cyganek(B)
1 Introduction
Contemporary sensors produce huge amounts of multi-dimensional signals. The
most popular are ubiquitous video recordings produced on mobiles, but also
signals arising in various branches of industry, such as surveillance cameras, pro-
cess control, finance, as well as in science, in such domains as particle physics,
astronomy, seismology, biology, variety of experimental simulations, to name a
few. With no much exaggeration we can say that we live in times of big data. Pro-
cessing of big data was recently underpinned with artificial intelligence methods,
such as deep learning and widely applied convolutional neural networks (CNN).
These also entail processing of huge amounts of data for which high compu-
tational power computers and GPU are employed [10,11,13,14]. All these can
be characterized as high streams of multi-dimensional data [3]. Hence, develop-
ment of methods for their efficient processing is one of the important research
topics. In this context the methods of signal change detection, as well as sig-
nal compression seem to be very useful. Tensor based methods offer a natural
tool for multi-dimensional signal processing [3–7]. Originally proposed by Tucker
[21], then adopted to the signal processing domain [6,12,15,16], tensor decom-
position methods play the key role. In this context the special stress is put
upon methods of signal clustering based on abrupt change detection of various
c Springer Nature Switzerland AG 2020
M. Choraś and R. S. Choraś (Eds.): IP&C 2019, AISC 1062, pp. 3–5, 2020.
https://doi.org/10.1007/978-3-030-31254-1_1
4 B. Cyganek
forms and duration [4,7,20]. The second actively investigated are methods for
multi-dimensional signal compression. Connection of the two domains offers new
possibilities as well [5]. Such hybrid methods can be used for compression of sig-
nal chunks. They can find broad applications in compression of video or CNN
weights [1,2,18,19], to name a few. However, tensor processing is not free from
computational problems such as curse of dimensionality, missing data, storage
limitations and computational complexity to name a few. In this keynote, an
overview of the above methods will be presented with special focus upon appli-
cations and further developments. The basic theory behind tensor analysis, with
underpinned few examples, will be presented. Then, an overview of the basic
tensor decompositions will be discussed, underlying those especially suited for
signal change detection and compression. The talk will conclude with examples
as well as ideas for future research in these areas.
References
1. Asghar, M.N., Hussain, F., Manton, R.: Video indexing: a survey. Int. J. Comput.
Inf. Technol. 03(01), 148–169 (2014)
2. de Avila, S.E.F., Lopes, A.P.B., da Luz Jr., A., Araújo, A.A.: VSUMM: a mecha-
nism designed to produce static video summaries and a novel evaluation method.
Pattern Recogn. Lett. 32, 56–68 (2011)
3. Cyganek, B.: Recognition of road signs with mixture of neural networks and arbi-
tration modules. In: Advances in Neural Networks, ISNN 2006. Lecture Notes in
Computer Science, vol. 3973, pp. 52–57. Springer (2006)
4. Cyganek, B., Woźniak, M.: Tensor-based shot boundary detection in video streams.
New Gener. Comput. 35(4), 311–340 (2017)
5. Cyganek, B., Woźniak, M.: A tensor framework for data stream clustering and com-
pression. In: International Conference on Image Analysis and Processing, ICIAP
2017, Part I. LNCS, vol. 10484, pp. 1–11 (2017)
6. Cyganek, B., Krawczyk, B., Woźniak, M.: Multidimensional data classification with
chordal distance based kernel and support vector machines. J. Eng. Appl. Artif.
Intell. 46, 10–22 (2015). Part A
7. Cyganek, B.: Change detection in multidimensional data streams with efficient
tensor subspace model. In: Hybrid Artificial Intelligent Systems: 13th International
Conference, HAIS 2018, Lecture Notes in Artificial Intelligence, LNAI, Oviedo,
Spain, 20–22 June, vol. 10870, pp. 694–705. Springer (2018)
8. Del Fabro, M., Böszörmenyi, L.: State-of-the-art and future challenges in video
scene detection: a survey. Multimedia Syst. 19(5), 427–454 (2013)
9. Fu, Y., Guo, Y., Zhu, Y., Liu, F., Song, C., Zhou, Z.-H.: Multi-view video summa-
rization. IEEE Trans. Multimedia 12(7), 717–729 (2010)
10. Gama, J.: Knowledge Discovery from Data Streams. CRC Press, Boca Raton
(2010)
11. Gama, J., Žliobaitė, I., Bifet, A., Pechenizkiy, M., Bouchachia, A.: A survey on
concept drift adaptation. ACM Comput. Surv. (CSUR) 46(4), 44:1–44:37 (2014)
Overview of Tensor Methods 5
12. Kolda, T.G., Bader, B.W.: Tensor decompositions and applications. SIAM Rev.
51, 455–500 (2008)
13. Krizhevsky, A., Sutskever, I., Hinton, G.E.: ImageNet classification with deep con-
volutional neural networks. In: Proceedings of the 25th International Conference
on Neural Information Processing Systems - Volume 1, NIPS 2012, pp. 1097–1105
(2012)
14. Ksieniewicz, P., Woźniak, M., Cyganek, B., Kasprzak, A., Walkowiak, K.: Data
stream classification using active learned neural networks. Neurocomputing 353,
74–82 (2019)
15. de Lathauwer, L.: Signal processing based on multilinear algebra. Ph.D. disserta-
tion. Katholieke Universiteit Leuven (1997)
16. de Lathauwer, L., de Moor, B., Vandewalle, J.: A multilinear singular value decom-
position. SIAM J. Matrix Anal. Appl. 21(4), 1253–1278 (2000)
17. Lee, H., Yu, J., Im, Y., Gil, J.-M., Park, D.: A unified scheme of shot boundary
detection and anchor shot detection in news video story parsing. Multimedia Tools
Appl. 51, 1127–1145 (2011)
18. Mahmoud, K.A., Ismail, M.A., Ghanem, N.M.: VSCAN: an enhanced video sum-
marization using density-based spatial clustering. In: Image Analysis and Process-
ing, ICIAP 2013. LNCS, vol. 1, pp. 733–742. Springer (2013)
19. Medentzidou, P., Kotropoulos, C.: Video summarization based on shot bound-
ary detection with penalized contrasts. In: IEEE 9th International Symposium on
Image and Signal Processing and Analysis (ISPA), pp. 199–203 (2015)
20. Sun, J., Tao, D., Faloutsos, C.: Incremental tensor analysis: theory and applica-
tions. ACM Trans. Knowl. Discov. Data 2(3), 11 (2008)
21. Tucker, L.R.: Some mathematical notes on three-mode factor analysis. Psychome-
trika 31, 279–311 (1966)
Head Motion – Based Robot’s Controlling
System Using Virtual Reality Glasses
Tomasz Hachaj(B)
1 Introduction
Remote wireless controlling is among basic functionalities of robotic platforms.
Handheld controllers are most popular and reliable type of controlling devices.
However there might be a need to operate a robot without using hands. It might
be necessary if a person controlling it wants to have free hands or if this person
has some hands disabilities. Additionally, if an operator cannot follow the robot
or observe it all the time it might be necessary to receive a broadcast from
cameras installed on the robot. Those two functionalists: hands-free controlling
and camera view displaying can be implemented with the help of virtual reality
(VR) glasses (goggles).
Virtual reality glasses contain optical system that enables displaying stereo-
graphic image and to monitor user’s head rotation using gyroscope. Cheap gog-
gles often utilize smartphones which commonly has accelerometers. The stere-
ographic image is obtained by displaying image on smartphone screen in split
mode and contains adjustable lenses which together with binocular disparity
(obtained from stereo vision camera) are used to generate three dimensional
image.
c Springer Nature Switzerland AG 2020
M. Choraś and R. S. Choraś (Eds.): IP&C 2019, AISC 1062, pp. 6–13, 2020.
https://doi.org/10.1007/978-3-030-31254-1_2
Head Motion – Based Robot’s Controlling System 7
In state of the art papers we did not find a solution that has all functionalities
we have mentioned, however separate elements are present in number of papers.
The virtual reality controlling systems are present in robotic from many years
[10,12]. Among possible visualization techniques it is known that virtual reality
glasses give sense of immersion and helps with remote navigation of mechani-
cal devices [3,7,8,13] in medical robotic applications [9,16], rehabilitation and
physical training [1,2].
This paper proposes head motion – based robot’s controlling system using
virtual reality glasses that can be implemented using various up-to date software
and hardware solutions. This article also describes a user - centered head rota-
tion system similar to yaw – pitch – roll that can be used to intuitively design
functions of user interface. All source codes that were made for system imple-
mentation: motor control program, Android VR Google Cardboard application
and R language implementation of method that recalculates quaternion rotation
into proposed rotation coordinates system can be downloaded and tested [6].
WiFI
camera
Motors
control
Wiring
WiFi
Commentary
Motors
For the purpose of creating head motion controlled system it might be more
convenient to recalculate head rotation coordinates from quaternions into three-
parameter angle-based system. To make this coordinates system user-centered
we use coordinates similar to yaw – pitch – roll triplet, however, with differ-
ent definition of axis and rotations directions which are more ‘intuitive’ for a
user. This type of description is more straightforward while defining behavior
of the human - computer interface. There are three rotation axes, which initial
position is dependent to user initial position (see Fig. 2). The vertical y axis
which governs horizontal rotation is defined as the normal vector of the top of
the user’s head (in fact it is normal vector of the side part of the smartphone).
The horizontal rotation is defined with left-handed screw order. When a person
is looking straight, the rotation horizontal rotation angle equals 0, when he or
she is looking left, rotation angle is negative, when he or she is looking right
rotation angle is positive. x axis which governs vertical rotation is perpendicular
to y axis and is defined by a vector which *more or less* links ears of a user (in
fact it is normal vector of the top part of the smartphone). The vertical rotation
is defined with right-handed screw order. When a person is looking straight, the
rotation vertical rotation angle equals 0, when he or she is looking up, rotation
angle is positive, when he or she is looking down rotation angle is negative. z axis
is a cross product of unit vectors of x and y. z governs sideways head bending.
The bending rotation is defined with left-handed screw order. When a person is
looking straight, the rotation vertical rotation angle equals 0, when he or she is
bending head right, rotation angle is positive, when he or she is bending head left
rotation angle is negative. The coordinates system is left-handed. To recalculate
quaternion-based rotation returned by a VR goggles the following calculations
have to be done. Let us assume that the output rotation angles are restricted to
Head Motion – Based Robot’s Controlling System 9
[−π, π] which is enough to define head rotation of stationary (for example sitting
or standing in one place) person.
y
Horizontal
Vertical
x
z
Bending
Fig. 2. This figure presents definition of user-centered coordinate system with ‘intu-
itive’ rotations definition.
⎡ ⎤
Q1 .W ∗ Q2 .X + Q1 .X ∗ Q2 .W + Q1 .Y ∗ Q2 .Z − Q1 .Z ∗ Q2 .Y
⎢Q1 .W ∗ Q2 .Y + Q1 .Y ∗ Q2 .W + Q1 .Z ∗ Q2 .X − Q1 .X ∗ Q2 .Z ⎥
⎢
Q1 · Q 2 = ⎣ ⎥ (1)
Q1 .W ∗ Q2 .Z + Q1 .Z ∗ Q2 .W + Q1 .X ∗ Q2 .Y − Q1 .Y ∗ Q2 .X ⎦
Q1 .W ∗ Q2 .W − Q1 .X ∗ Q2 .X − Q1 .Y ∗ Q2 .Y − Q1 .Z ∗ Q2 .Z
Let us assume that the initial head rotation is QH . In order to recalculate quater-
nion rotation Q to Vertical – Horizontal – Bending relatively to QH we apply
the following calculation:
π
V ertical ← acos(T (QH , [1, 0, 0]) ◦ T (Q, [0, 0, 1])) −
2
π
Horizontal ← π − acos(T (QH , [0, 0, 1]) ◦ T (Q, [0, 1, 0])) − (4)
2
π
Bending ← π − acos(T (QH , [1, 0, 0]) ◦ T (Q, [0, 1, 0])) −
2
10 T. Hachaj
Fig. 3. This figure presents final implementation of the robot (a) and screenshot from
VR glasses (b).
The communication between robot and VR glasses is done via Wi-fi using
TCP/IP protocol. Both smartphone with Android OS that is used by Google
Cardboard technology and NodeMCU micro controller can communicate via
HTTP however this transmission protocol is too slow for real-time motors control
with this hardware. Because of it TCP/IP socket-level communication protocol
has been applied. The Android application running on smartphone is monitor-
ing head motions. The applications determines initial rotation of the user’s head
(QH ) and for each following data acquisition Vertical – Horizontal – Bending
head rotation is calculated. A user can reinitialize QH by touching smartphone
screen and the current head rotation Q becomes QH . The Horizontal rotation
angle governs platform’s turning and Vertical rotation is responsible for mov-
ing forward or backward. If the head position changes above a threshold values
(those values are defined in Android application) application sends a TCP/IP
package to NodeMCU with appropriate command. Firmware on NodeMCU pro-
cesses those messages and changes voltages on output pins of motor shield. The
robot continues doing particular motion until a message to stop arrives from
Android application. The stop message is send when a head position of the user
is within the ‘neutral’ threshold value. Messages from smartphone are sent no
more often than a certain time span, which is also defined in application. This
prevents robot platform to be overflown by incoming commands. A user in VR
glasses see the view from stereo vision camera that is mounted on the front of
the robot.
Head Motion – Based Robot’s Controlling System 11
80
60
60
40
40
20
Angle [degress]
Angle [degress]
20
0
−20
0
−40
−20
−60
−40
0 2 4 6 8 10 0 2 4 6 8 10
40
20
0
0 2 4 6 8 10
Time [s]
Fig. 4. This figure presents plots of example head rotations of user in VR glasses
calculated using (4).
3 Results
After implementing the system (see Fig. 3(a)), tests have been performed on
its remote head-motion based controlling module. As can be seen in Fig. 3(b)
view in VR glasses is additional distorted by Google Cardboard API in order
to strengthen effect of stereo vision and immersive experience of a user. The
remote controlling systems worked smoothly allowing remote control of robot
with nearly close to real-time preview from stereo vision camera. Beside naviga-
tion commands mathematical model (4) was tested in order to check if it can be
used to produce intuitive motions descriptions and visualizations. B. In order to
do so a user operating VR glasses was asked to move head down and up, then
turn it left and right, and finally to bend it clockwise and return to initial head
position. Plots of obtained angles descriptions are presented in Fig. 4(a)–(c).
Additionally Fig. 4(d) presents vertical and horizontal rotation applied to unit
12 T. Hachaj
vector and its projection onto unit sphere. That last plot visualizes in 3D what
the head motion trajectory was.
As could be seen in previous section the proposed head motion – based robot’s
controlling system satisfies needs of remote controlling of robotic platform. The
mathematical model of head rotation calculation enables to generate intuitive
user – centered rotation description. In this description positive directions of axis
are right, up and towards front user’s head. The positive direction of rotations
is right, up, and clockwise in case of head side banding, opposite rotations have
negative angle coordinates. With the help of this description it is straightforward
to define how head motion should be interpreted and translated to remote com-
mands to robot. The communication protocol takes into account system latency
and prevents robot from being overflown by incoming commands.
Due to applying a stereo vision camera in proposed systems can be utilized as
the prototype in many scientific researches. After stereo calibration this system
can be used for measuring distances between robot and elements of environment
and three-dimensional points cloud generation. The system can be also applied
for developing and testing vision – based robotic odometry [14,15] or simulta-
neous localization and mapping (SLAM) algorithms [4,11]. There is also a large
potential in developing and testing methods for head motion analysis and recog-
nition. Many commands to system might be coded in head gestures that could
be classified by appropriate algorithm.
References
1. Borowska-Terka, A., Strumillo, P.: Algorithms for head movements’ recognition in
an electronic human computer interface. Przeglad Elektrotechniczny 93(8), 131–
134 (2017)
2. Brütsch, K., Koenig, A., Zimmerli, L., Mérillat-Koeneke, S., Riener, R., Jäncke, L.,
van Hedel, H.J., Meyer-Heim, A.: Virtual reality for enhancement of robot-assisted
gait training in children with neurological gait disorders. J. Rehabil. Med. 43(6),
493–499 (2011)
3. Dornberger, R., Korkut, S., Lutz, J., Berga, J., Jäger, J.: Prototype-based research
on immersive virtual reality and on self-replicating robots. In: Business Information
Systems and Technology 4.0 Studies in Systems, Decision and Control, pp. 257–274
(2018). https://doi.org/10.1007/978-3-319-74322-6 17
4. Engel, J., Stückler, J., Cremers, D.: Large-scale direct slam with stereo cam-
eras. In: 2015 IEEE/RSJ International Conference on Intelligent Robots and Sys-
tems (IROS), pp. 1935–1942, September 2015. https://doi.org/10.1109/IROS.2015.
7353631
5. Grygiel, R., Bieda, R., Wojciechowski, K.: Angles from gyroscope to complemen-
tary filter in IMU. Przeglad Elektrotechniczny 90(9), 217–224 (2014)
6. Hachaj, T.: GitHub repository of the project (2019). https://github.com/
browarsoftware/rpm rotation calculation. Accessed 22 Mar 2019
Head Motion – Based Robot’s Controlling System 13
7. Kato, Y.: A remote navigation system for a simple tele-presence robot with virtual
reality. In: 2015 IEEE/RSJ International Conference on Intelligent Robots and
Systems (IROS), pp. 4524–4529, September 2015. https://doi.org/10.1109/IROS.
2015.7354020
8. Kurup, P., Liu, K.: Telepresence robot with autonomous navigation and virtual
reality: demo abstract. In: SenSys (2016)
9. Lin, L., Shi, Y., Tan, A., Bogari, M., Zhu, M., Xin, Y., Xu, H., Zhang, Y., Xie, L.,
Chai, G.: Mandibular angle split osteotomy based on a novel augmented reality
navigation using specialized robot-assisted arms - a feasibility study. J. Cranio-
Maxillofac. Surg. 44(2), 215–223 (2016). https://doi.org/10.1016/j.jcms.2015.10.
024. http://www.sciencedirect.com/science/article/pii/S1010518215003674
10. Monferrer, A., Bonyuet, D.: Cooperative robot teleoperation through virtual reality
interfaces. In: Proceedings Sixth International Conference on Information Visuali-
sation, pp. 243–248, July 2002. https://doi.org/10.1109/IV.2002.1028783
11. Mur-Artal, R., Tardüs, J.D.: ORB-SLAM2: an open-source SLAM system for
monocular, stereo, and RGB-D cameras. IEEE Trans. Robot. 33(5), 1255–1262
(2017). https://doi.org/10.1109/TRO.2017.2705103
12. Nguyen, L., Bualat, M., Edwards, L., Flueckiger, L., Neveu, C., Schwehr, K., Wag-
ner, M., Zbinden, E.: Virtual reality interfaces for visualization and control of
remote vehicles. Auton. Robots 11(1), 59–68 (2001). https://doi.org/10.1023/a:
1011208212722
13. Regenbrecht, J., Tavakkoli, A., Loffredo, D.: A robust and intuitive 3D interface
for teleoperation of autonomous robotic agents through immersive virtual reality
environments. In: 2017 IEEE Symposium on 3D User Interfaces (3DUI), pp. 199–
200, March 2017. https://doi.org/10.1109/3DUI.2017.7893340
14. Usenko, V., Engel, J., Stückler, J., Cremers, D.: Direct visual-inertial odometry
with stereo cameras. In: 2016 IEEE International Conference on Robotics and
Automation (ICRA), p. 1885 (2016). https://doi.org/10.1109/ICRA.2016.7487335
15. Wang, R., Schworer, M., Cremers, D.: Stereo DSO: large-scale direct sparse visual
odometry with stereo cameras. In: IEEE International Conference on Computer
Vision (ICCV), October 2017
16. Zinchenko, K., Komarov, O., Song, K.: Virtual reality control of a robotic cam-
era holder for minimally invasive surgery. In: 2017 11th Asian Control Confer-
ence (ASCC), pp. 970–975, December 2017. https://doi.org/10.1109/ASCC.2017.
8287302
Robustness of Haar Feature-Based
Cascade Classifier for Face Detection
Under Presence of Image Distortions
1 Introduction
Face detection is a highly developed computer technology being used in variety
of applications and multiple sciences. Face detection can be used in the process
of identification of human faces, access control on mobile devices or for the
purpose of entertainment. Each face detection system, the same as any other
object-class detection method, tackles with a problem with the quality of image
upon which the system is operating. In the majority of detected images, quality
is a factor that considerably affects the effectiveness of system operation. In order
to achieve higher effectiveness of image detection, systems ought to be resistant
to low image quality [5].
This paper focuses on one of the most popular methods of face detection, that
is Cascade Classifier [1]. When the method was presented, the basic classifier was
c Springer Nature Switzerland AG 2020
M. Choraś and R. S. Choraś (Eds.): IP&C 2019, AISC 1062, pp. 14–21, 2020.
https://doi.org/10.1007/978-3-030-31254-1_3
Robustness of Haar Feature-Based Cascade Classifier 15
Haar, subsequently Ojala proposed a new LBP classifier [2], which was added
to Cascade Classifier. Herein, we intend only to concentrate on HAAR classifier
which operates by selection of squares containing dark and bright fragments. In
case of face detection, division of a square into dark and fair regions serves the
purpose of calculation of the sum of pixels’ value from under the selected square.
A dark square may represent fragments of eyes, mouth, and the fair square may
represent a nose, forehead or cheeks.
Each classifier consists of a few stages where multiple functions are performed.
In order for the classifier to consider an examined area to be the face under
search, the sum obtained from the tested areas must exceed a certain threshold
defined in a file containing Haar cascade model. Sometimes, even this type of
testing is not capable for errors elimination and the classifier is bound to accept
the rectangles that does not contains faces (false positive error).
Regardless the foregoing errors, scientists use Cascade Classifier in various
tasks and come up with brand new solutions based on this method. With a
view to improve results of face detection, another classifier was proposed with
features that better reflect the face structure [3,16,17]. Another extension on
the Cascade Classifier for the purpose of more accurate results is to apply a
few weaker models and to verify the results obtained after each stage [4]. Next
issue where Haar classifier has been tested is to detect the area around eyes
and to make use of the driver’s fatigue [6,7]. Cascade Classifier not only assists
the process of face or body composition detection but may also be applied in
medicine, e.g. for the purpose of recognition of leukocytes on ultrasound scans
[11] or detection of objects from the bottom of the sea [12].
Development of new technologies as well as neural networks accounts for the
fact that recognition of various objects or faces becomes more and more accurate
and faster, each of those technologies, however, depending significantly upon the
quality of images supplied for the purpose of detection [9,10].
In this paper have been prepared 4 test groups in which it has been checked
how the Cascade Classifier work for face detection under presence of image dis-
tortions. Each of test contains a specially prepared images with the appropriate
level of distortion. In the first test it was prepared a images with salt and pepper
noise. This noise can be find in images who taken in low light. In next test, it
was use low-quality images. To reduce the quality images was used the Gauss
filter and Median filter. This type of distortion can be obtained when taking
low-resolution images. The third test involves the modifying images for changes
in brightness and contrast values. This type of images can be obtained by man-
ually configuration of camera or when the lighting changes dynamically and the
camera has to adjust the settings. In the last test it was use images with “fish-
eye” effect, To get this effect can be used a specific lens or application witch will
added this effect to images. The “fisheye” effect can be found in art photography
and cameras with a wide of view.
The research presented in this article can help to creating a new solutions to
face detection or it will be helpful when creating a images database for learning
models in cascade classifier method.
16 P. Mazurek and T. Hachaj
(a) Salt and pepper - during “salt and pepper” noise simulation 11 tests have
been carried out with noise being inserted into the image within the range of
0–100, and a constant pace of 10. The noise level was the percentage of pixels
against all pixels in the image. Noise pixels have been inserted randomly into
the tested images.
(b) Blurring - blurring simulation has been conducted with the use of Gaussian
blur filter and median filter.
1 x2 +y 2
G(x, y) = 2
e 2σ2 (1)
2πσ
For each filter a value between 1–17 was applied, adjusting input data every 2,
thanks to which 9 tests for each filter have been completed. The use of Gaussian
and median filters of high calibre causes image blurring but not loss of edges or
important data in the image.
(c) Contrast and brightness - in brightness and contrast tests 11 tests have
been prepared with the value of pixels changed accordingly.
The value of x(i, j) corresponds to the pixels of the basic image, the value
of b is responsible for brightness, which change was carried out by adding an
appropriate value from the range between [−127:127] to pixel. In case of contrast
(variable a), the value of each pixel was multiplied by an adequate value ranging
between [0.0:2.0] with the constant pace of 0.2. In each test, after the pixel value
change, it was necessary to adjust it to appropriate range [0:255].
(d) Lenses - the “fisheye” effect can be obtained by applying special lenses or
adequate algorithms which help to modify and transform the traditional image
into the one with “fisheye” effect. During the test, an image was modified by
Robustness of Haar Feature-Based Cascade Classifier 17
means of setting up a new Q coordinate system with its centre being located
halfway through the height and width of the modified image. Subsequently,
Cartesian coordinates are replaced with polar coordinates using.
r = (x2 + y 2 ) (3)
θ = atan2(x, y) (4)
Next, image is mapped onto a spherical or elliptical object. Value of distortion
of an image is between a vector and a polar coordinate from a “virtual camera”
to the beginning of a coordinate system. Mathematically, the above phenomenon
can be presented in the following way:
3 Results
During test 1 measurements have been made to check how various noise levels
in an image affect effectiveness of detection rate. Based on the presented table
(Table 1) it can be concluded that images with noise level up to 20% are likely
to get high effectiveness of detection. An increase of the noise level accounts for
a decrease in effectiveness of image detection. In case of Haar classifier that was
used in the test, images containing noise above 60% are extremely difficult to
be detected or simply impossible. As it appears from the test, Haar classifier is
ultra-sensitive to noise, because salt and pepper adds to image black and white
pixel and this disrupts the effect of Haar wavelet. The Haar wavelet’s work
involves calculating the differences between square. High content of white and
18 P. Mazurek and T. Hachaj
Table 2. This table compares results of the tests with Median filter and Gaussian
filter.
Filter size 0 1 3 5 7 9 11 13 15 17
Median filter 95.6% 96.8% 96.0% 96.1% 95.9% 95.7% 95.3% 94.9% 94.4% 93.9%
Gaussian filter 95.6% 95.8% 95.8% 95.8% 95.0% 93.6% 92.9% 90.9% 88.7% 86.1%
black pixels in images results in incorrect calculation results. The most frequent
circumstances in which noise is generated is taking photos in poor illumination.
Another task was to apply the blurring effect in images and to check effec-
tiveness of the tested model. The effect in question can be seen in the photos or
films of low resolution. For that purpose median filter and Gaussian filter was
used. On the presented table (Table 2) it can be observed that the use of filters
of low values does not cause deterioration of effectiveness or even, as in case of
Gaussian filter, is responsible for an increase of detection rate (Table 2). Use of
filters of high value is bound to result in a decrease of image detection rate but
every filter responds to an increase of value in a different way. As far as median
filter is concerned (Table 2), application of value of 17 caused detection effective-
ness to drop to 86% whereas in case of Gaussian filter effectiveness decreased to
94%. Each filter removes certain minor information from images causing them
to blur but leaving, however, edges and data vital for HAAR classifier which
enables face detection.
Table 3. This table presents results of the contrast test for values from 0.0 to 2.0.
Filter size 0.0 0.2 0.4 0.6 0.8 1.0 1.2 1.4 1.6 1.8 2.0
Positive results 0% 44.5% 94.7% 94.8% 94.9% 95.6% 96.0% 95.7% 94.5% 92.0% 87.7%
Table 4. This table presents results of the brightness test for values from −127 to 127.
Page 31, " Ononix Natrix " a été remplacé par " Ononis "
Page 33, " — annua Wickstr. " a été remplacé par " Wikstr. "
Page 38, " Cerastium Atlantium " a été remplacé par " Atlanticum "
Page 62, " Kalbfussia Salzmanni Schulz. Bip. " a été remplacé par " Schultz. "
Page 69, " qui apparait dans " a été remplacé par " apparaît "
Page 89, " Clamydophora pubescens " a été remplacé par " Chlamydophora "
Page 90, " Dæmia cordata " a été remplacé par " Dœmia "
Page 101, " [Sinapis]— arvenis " a été remplacé par " arvensis "
Page 141, " Monocotydélones " a été remplacé par " Monocotylédones "
*** END OF THE PROJECT GUTENBERG EBOOK RAPPORT SUR
UN VOYAGE BOTANIQUE EN ALGÉRIE, DE PHILIPPEVILLE A
BISKRA ET DANS LES MONTS AURÈS, ENTREPRIS EN 1853
SOUS LE PATRONAGE DU MINISTÈRE DE LA GUERRE ***
1.D. The copyright laws of the place where you are located also
govern what you can do with this work. Copyright laws in most
countries are in a constant state of change. If you are outside
the United States, check the laws of your country in addition to
the terms of this agreement before downloading, copying,
displaying, performing, distributing or creating derivative works
based on this work or any other Project Gutenberg™ work. The
Foundation makes no representations concerning the copyright
status of any work in any country other than the United States.
1.E.6. You may convert to and distribute this work in any binary,
compressed, marked up, nonproprietary or proprietary form,
including any word processing or hypertext form. However, if
you provide access to or distribute copies of a Project
Gutenberg™ work in a format other than “Plain Vanilla ASCII” or
other format used in the official version posted on the official
Project Gutenberg™ website (www.gutenberg.org), you must, at
no additional cost, fee or expense to the user, provide a copy, a
means of exporting a copy, or a means of obtaining a copy upon
request, of the work in its original “Plain Vanilla ASCII” or other
form. Any alternate format must include the full Project
Gutenberg™ License as specified in paragraph 1.E.1.
• You pay a royalty fee of 20% of the gross profits you derive from
the use of Project Gutenberg™ works calculated using the
method you already use to calculate your applicable taxes. The
fee is owed to the owner of the Project Gutenberg™ trademark,
but he has agreed to donate royalties under this paragraph to
the Project Gutenberg Literary Archive Foundation. Royalty
payments must be paid within 60 days following each date on
which you prepare (or are legally required to prepare) your
periodic tax returns. Royalty payments should be clearly marked
as such and sent to the Project Gutenberg Literary Archive
Foundation at the address specified in Section 4, “Information
about donations to the Project Gutenberg Literary Archive
Foundation.”
• You comply with all other terms of this agreement for free
distribution of Project Gutenberg™ works.
1.F.
Most people start at our website which has the main PG search
facility: www.gutenberg.org.
Our website is not just a platform for buying books, but a bridge
connecting readers to the timeless values of culture and wisdom. With
an elegant, user-friendly interface and an intelligent search system,
we are committed to providing a quick and convenient shopping
experience. Additionally, our special promotions and home delivery
services ensure that you save time and fully enjoy the joy of reading.
textbookfull.com