Facial Expression Recognition Using a Simplified Convolutional Neural Network Model | IEEE Conference Publication | IEEE Xplore

Facial Expression Recognition Using a Simplified Convolutional Neural Network Model


Abstract:

Facial Expression Recognition (FER) is one of the most important information channels by which Human-Computer Interaction (HCI) systems can recognize human emotions. The ...Show More

Abstract:

Facial Expression Recognition (FER) is one of the most important information channels by which Human-Computer Interaction (HCI) systems can recognize human emotions. The importance of FER is not limited to the direct interaction between the machine and humans but can be extended to security, virtual reality, education, and entertainment. In this paper, we propose two Convolutional Neural Network (CNN) models for FER. One of these models achieved 100% accuracy for the JAFFE and CK+ benchmark datasets with lower computational complexity. We applied image augmentation techniques and image enhancement techniques with the first model. The other CNN model is an extended version of the first model that h as been validated for t he more challenging FER2013 dataset and we obtained 69.32% for this dataset. By comparing to the recent state-of-the-art approaches to FER, we demonstrate the superior accuracy and efficiency of the proposed approaches.
Date of Conference: 16-18 March 2021
Date Added to IEEE Xplore: 02 April 2021
ISBN Information:
Conference Location: Sharjah, United Arab Emirates

I. Introduction

In the last decade, Facial Expression Recognition (FER) research gained a lot of attention due to the advancement achieved in related research areas such as face detection and recognition [1] [2]. These advancements encourage researchers to study facial expressions and to build real-time FER approaches. FER approaches can be divided into two main categories. The first category is the traditional approaches that extract handcrafted features using methods such as Gabor wavelets, Local Binary Patterns (LBP), and Principal Component Analysis (PCA) [3]. Subsequently, the features are categorized into the respective facial expression classes based on classification methods such as Support Vector Machine (SVM) and Nearest Neighbor (NN) [4]. The second category is deep learning-based approaches that rely on reducing the dependence on manual extraction of facial patterns and enable machines to learn directly from the input images [5]. The deep learning-based approaches are composed of three basic steps: pre-processing, feature learning, and feature extraction. The pre-processing step is employed before training to enhance the input images, such as face alignment, image cropping, and face normalization [6]. Feature learning and extraction are performed using Deep Neural Networks (DNNs), which use training data and artificial intelligence algorithms to learn the relations among the extracted features [4]. Convolutional Neural Network (CNN) and Recurrent Neural Network (RNN) are the most common deep learning approaches used in this era, which enable learning spatial data patterns and temporal data patterns respectively [5]. In the CNN model, the input image is convoluted through a combination of filters; each filter has a particular set of values to produce a specified feature map.

Contact IEEE to Subscribe

References

References is not available for this document.