Repository logo
 

Human Facial Emotion Recognition From Digital Images Using Deep Learning

Date

Supervisor

Yan, Wei Qi

Item type

Thesis

Degree name

Master of Computer and Information Sciences

Journal Title

Journal ISSN

Volume Title

Publisher

Auckland University of Technology

Abstract

In this thesis, our aim is to investigate the best-performing methods and algorithms of facial emotion recognition (FER) based on the seven classes of human facial expressions of emotion: Neutral, scared, angry, disgusted, sad, happy, and surprised. We classify human facial emotions from digital images. The existing methodology of FER has various limitations: low training and testing result, and difficulty in classifying certain emotions. Convolutional Neural Network (CNN), Xception, Visual Transformers (ViT), Simple Deep Neural Network (SDNN), and Graph Convolution Neural Network (GCN) have been proposed to the FER with two types of methods: Non-facial landmarking and facial landmarking. The first method that we proposed is the modified CNN with Haar Cascade algorithm for frontal face detection as an initial solution. While running our experiments in real-time, our accuracy for FER is up to 90.0%. Our first effort was with the same methodology and parameters as our initial method Mini Xception. With the Mini Xception model, we achieved stable result with the highest accuracy of 99%. In addition, with the popularity of the deep learning algorithm Transformers, we implemented Visual Transformer with the non-landmarking method. With the training and testing work conducted based on the proposed model, we achieved the highest training accuracy in fewer number of training epoch compared to the previous two models. Due to misclassification that occurred in the previous methods, we developed a new method of facial landmarking. The models we proposed are Simple Deep Neural Network (SDNN) and Graph Convolutional Network (GCN). Our SDNN model was employed as a baseline to test the proposed method. With this model, we achieved 96.0% accuracy in our real-time testing experiments. In future, we plan to train and test the GCN model with facial landmarking method to determine the best performing model. In addition, we will implement the ViT model in real-time using the non-landmarking method.

Description

Source

DOI

Publisher's version

Rights statement

Collections