Face Expression Detection Using CNN
Face Expression Detection Using CNN
Abstract:
The ability to recognize facial expressions using computer vision is a crucial task that has numerous potential
applications. Although deep neural networks have achieved high performance, their use in the recognition of facial
expressions is still challenging. This is because different facial expressions have varying degrees of similarities among
themselves, and numerous variations because diversity in the same facial images .We propose a novel divide-and-conquer-
based learning strategy to improve the performance of facial expression recognition (FER). Due to globalization and digital
divide facial expression detection has received the primary attention in order to identify the criminals and breaches. Facial
expressions are the changes occurring on the human face indicating a person's internal emotional states, intents or societal
communications. Depending on the expressions on the face, human face is the most principal mode of conveying and
deducing affective states of human ones. In real time, facial expression detection has become a prominent research area as it
plays an important role in Human Computer Interaction. The applications of the facial expression detection are computer
vision, biometric security, social interaction, emotional intelligence and social intelligence.
Keyword:
Convolutional Neural Networks, Emotion Recognition, Image Processing, Facial Feature Extraction, Facial
Expression Dataset, Machine Learning, Facial Feature Representation, Neural Network Architecture, Feature Extraction.
2. Pooling Layer:
This layer is the first layer that is used to extract the Usually, when all the features are connected to the FC
various features from the input images. In this layer, the layer, it can cause overfitting in the training dataset.
mathematical operation of convolution is performed Overfitting occurs when a particular model works so well
between the input image and a filter of a particular size on the training data causing a negative impact in the
model’s performance when used on a new data. To facial components. The final step is to use a Feature
overcome this problem, a dropout layer is utilised wherein Extraction (FE) classifier and produce the recognition
a few neurons are dropped from the neural network during results using the extracted features. Figure 1.1 shows the
training process resulting in reduced size of the model. On FER procedure for an input image where a face region and
passing a dropout of 0.3, 30% of the nodes are dropped out facial landmarks are detected. Facial landmarks are
randomly from the neural network. visually salient points such as the end of a nose, and the
ends of eyebrows and the mouth as shown in Figure 1.2.
The pairwise positions of two landmark points or the local
5. Activation Functions:
texture of a landmark are used as features. Table 1.1 gives
the definitions of 64 primary and secondary landmarks [8].
Finally, one of the most important parameters of the CNN The spatial and temporal features are extracted from the
model is the activation function. They are used to learn face and the expression is determined based on one of the
and approximate any kind of continuous and complex facial categories using pattern classifier.
relationship between variables of the network. In simple
words, it decides which information of the model should
fire in the forward direction and which ones should not at
the end of the network. It adds non-linearity to the
network. There are several commonly used activation
functions such as the ReLU, Softmax, tanH and the
Sigmoid functions. Each of these functions have a specific
usage. For a binary classification CNN model, sigmoid
and softmax functions are preferred and for a multi-class
classification, generally softmax us used. In simple terms,
activation functions in a CNN model determine whether a
neuron should be activated or not. It decides whether the
input to the work is important or not to predict using
mathematical operations. It drops neurons from the
neural networks during training.
Literature survey:
2. Longbiao Mao; Yan Yan; Jing-Hao Deep Multi-Task Multi- DMM-CNN jointly
Xue, “Deep Multi-Task Multi-Label Label CNN, facial optimizes two closely-
CNN for Effective Facial Attribute expression recognition related tasks (i.e., facial
Classification”, IEEE, Volume No:13, landmark detection and
2022 FAC) to improve the
performance of FAC by
taking advantage of multi-
task learning
3. Xiao Liu; Xiangyi Cheng; Kiju Lee, Genetic algorithm, This method employs less
“GA-SVM-Based Facial Emotion Convolutional neural complicated models and
Recognition Using Facial Geometric network. thus shows potential for
Features”, IEEE, Volume No:21, 2021 real-time machine vision
applications in automated
systems
4. Jun-Tong Liu; Fang-Yu Wu, “Domain Similarity preserving Competitive accuracy is
Adaption for Facial Expression generative adversarial reported when compared it
Recognition”, IEEE, Volume No:10, network (SPGAN with other state of the art
2020 works, which shows
promising results.