An optimal hybrid deep learning-aided facial emotion detection and classification scheme to identify criminal activities

  • Abstract
  • Similar Papers
Abstract
Translate article icon Translate Article Star icon
Take notes icon Take Notes

In general, the most significant field of research presently is the identification & recognition of facial expressions or emotions. Moreover, recognition and categorization of face emotion are vital in several areas of research like criminal activities investigation, innovative card application, security, surveillance system and so on. Among these, criminal investigation plays a vibrant part. There exist several methods for facial emotion/expression recognition (FER) systems; however there were some drawbacks like low prediction rate, lower recognition rate, high error rate and so on. For rectifying these existing issues, a new enhanced optimal Deep learning (DL)-based model is presented in this manuscript. In this work, an input facial dataset is extracted and preprocessed using Weighted fuzzy Histogram Equalization (WF-HE). From this, the features are extracted using Deep CNN, followed by an Enhanced glowworm swarm optimization (EGSO)-based feature selection model, at which hyper-parameter tuning is carried out by attaining fitness function values. This, in turn, enhances the performance of the classifier. The categorization for the FER system is carried out using a Hybrid Deep Variational LSTM (DVLSTM) and DenseNet model. The results are estimated in terms of various performance measures, and the analysis is made on three input datasets such as JAFFE, Extended CK+ and FER2013 dataset.

Similar Papers
  • PDF Download Icon
  • Research Article
  • Cite Count Icon 11
  • 10.17485/ijst/v14i12.14
Facial expression recognition for low resolution images using convolutional neural networks and denoising techniques
  • Mar 27, 2021
  • Indian Journal of Science and Technology
  • Pavan Nageswar Reddy Bodavarapu + 1 more

Background/Objectives: There is only limited research work is going on in the field of facial expression recognition on low resolution images. Mostly, all the images in the real world will be in low resolution and might also contain noise, so this study is to design a novel convolutional neural network model (FERConvNet), which can perform better on low resolution images. Methods: We proposed a model and then compared with state-of-art models on FER2013 dataset. There is no publicly available dataset, which contains low resolution images for facial expression recognition (Anger, Sad, Disgust, Happy, Surprise, Neutral, Fear), so we created a Low Resolution Facial Expression (LRFE) dataset, which contains more than 6000 images of seven types of facial expressions. The existing FER2013 dataset and LRFE dataset were used. These datasets were divided in the ratio 80:20 for training and testing and validation purpose. A HDM is proposed, which is a combination of Gaussian Filter, Bilateral Filter and Non local means denoising Filter. This hybrid denoising method helps us to increase the performance of the convolutional neural network. The proposed model was then compared with VGG16 and VGG19 models. Findings: The experimental results show that the proposed FERConvNet_HDM approach is effective than VGG16 and VGG19 in facial expression recognition on both FER2013 and LRFE dataset. The proposed FERConvNet_HDM approach achieved 85% accuracy on Fer2013 dataset, outperforming the VGG16 and VGG19 models, whose accuracies are 60% and 53% on Fer2013 dataset respectively. The same FERConvNet_HDM approach when applied on LRFE dataset achieved 95% accuracy. After analyzing the results, our FERConvNet_HDM approach performs better than VGG16 and VGG19 on both Fer2013 and LRFE dataset. Novelty/Applications: HDM with convolutional neural networks, helps in increasing the performance of convolutional neural networks in Facial expression recognition. Keywords: Facial expression recognition; facial emotion; convolutional neural network; deep learning; computer vision

  • Preprint Article
  • 10.21203/rs.3.rs-6478988/v1
Enhanced Convolutional Neural Network for Robust Facial Expression Recognition on Fer2013 and Natural Image Datasets
  • Apr 22, 2025
  • Prof Prakash Sangle

This study presents an enhanced convolutional neural network (CNN) architecture tailored for accurate facial expression recognition. The model is trained on the FER2013 dataset and evaluated using both FER2013 and a custom dataset containing natural facial expressions. By incorporating multiple convolutional and pooling layers along with dropout regularization, the network effectively extracts and classifies emotion-related features. Experimental results demonstrate high recognition accuracy and strong generalization across controlled and real-world image scenarios. In order to study the application of convolutional neural networks in facial expression recognition, a 10-layer convolutional neural network model is designed to recognize facial expressions. The last layer uses the Softmax function to output the classification results of expressions. First, the convolution and pooling algorithms of convolutional neural networks were studied and the structure of the model was designed. Secondly, in order to more vividly display the features extracted by the convolutional layer, the extracted features are visualized and displayed in the form of feature maps. The convolutional neural network model in this work was tested on the Fer-2013 data set, and the experimental results demonstrated the superiority of the recognition rate. It is known that the Fer-2013 dataset contains data collected in an experimental environment, and in order to verify the generalization ability of model recognition, a self-made facial expression data set in natural state was created, and performed a series of preprocessing on the face images such as cropping, grayscale and pixel adjustment. The trained model, which was previously applied to the Fer-2013 dataset, was tested out on the new dataset. The experiment yielded promising results, one of which in the form of a recognition accuracy rate as high as 85.1%.

  • Research Article
  • Cite Count Icon 221
  • 10.1007/s00371-019-01630-9
Using CNN for facial expression recognition: a study of the effects of kernel size and number of filters on accuracy
  • Jan 23, 2019
  • The Visual Computer
  • Abhinav Agrawal + 1 more

Facial expression recognition is a challenging problem in image classification. Recently, the use of deep learning is gaining importance in image classification. This has led to increased efforts in solving the problem of facial expression recognition using convolutional neural networks (CNNs). A significant challenge in deep learning is to design a network architecture that is simple and effective. A simple architecture is fast to train and easy to implement. An effective architecture achieves good accuracy on the test data. CNN architectures are black boxes to us. VGGNet, AlexNet and Inception are well-known CNN architectures. These architectures have strongly influenced CNN model designs for new datasets. Almost all CNN models known to achieve high accuracy on facial expression recognition problem are influenced by these architectures. This work tries to overcome this limitation by using FER-2013 dataset as starting point to design new CNN models. In this work, the effect of CNN parameters namely kernel size and number of filters on the classification accuracy is investigated using FER-2013 dataset. Our major contribution is a thorough evaluation of different kernel sizes and number of filters to propose two novel CNN architectures which achieve a human-like accuracy of 65% (Goodfellow et al. in: Neural information processing, Springer, Berlin, pp 117–124, 2013) on FER-2013 dataset. These architectures can serve as a basis for standardization of the base model for the much inquired FER-2013 dataset.

  • Research Article
  • Cite Count Icon 46
  • 10.3389/fpsyg.2021.759485
Facial Expression Emotion Recognition Model Integrating Philosophy and Machine Learning Theory
  • Sep 27, 2021
  • Frontiers in Psychology
  • Zhenjie Song

Facial expression emotion recognition is an intuitive reflection of a person’s mental state, which contains rich emotional information, and is one of the most important forms of interpersonal communication. It can be used in various fields, including psychology. As a celebrity in ancient China, Zeng Guofan’s wisdom involves facial emotion recognition techniques. His book Bing Jian summarizes eight methods on how to identify people, especially how to choose the right one, which means “look at the eyes and nose for evil and righteousness, the lips for truth and falsehood; the temperament for success and fame, the spirit for wealth and fortune; the fingers and claws for ideas, the hamstrings for setback; if you want to know his consecution, you can focus on what he has said.” It is said that a person’s personality, mind, goodness, and badness can be showed by his face. However, due to the complexity and variability of human facial expression emotion features, traditional facial expression emotion recognition technology has the disadvantages of insufficient feature extraction and susceptibility to external environmental influences. Therefore, this article proposes a novel feature fusion dual-channel expression recognition algorithm based on machine learning theory and philosophical thinking. Specifically, the feature extracted using convolutional neural network (CNN) ignores the problem of subtle changes in facial expressions. The first path of the proposed algorithm takes the Gabor feature of the ROI area as input. In order to make full use of the detailed features of the active facial expression emotion area, first segment the active facial expression emotion area from the original face image, and use the Gabor transform to extract the emotion features of the area. Focus on the detailed description of the local area. The second path proposes an efficient channel attention network based on depth separable convolution to improve linear bottleneck structure, reduce network complexity, and prevent overfitting by designing an efficient attention module that combines the depth of the feature map with spatial information. It focuses more on extracting important features, improves emotion recognition accuracy, and outperforms the competition on the FER2013 dataset.

  • Research Article
  • 10.54254/2755-2721/22/ojs/20231157
NA
  • Dec 25, 2023
  • Applied and Computational Engineering
  • Na Na

NA

  • PDF Download Icon
  • Research Article
  • Cite Count Icon 2
  • 10.54254/2755-2721/22/20231157
Convolutional neural network combined with the attention mechanism for facial emotion recognition
  • Oct 23, 2023
  • Applied and Computational Engineering
  • Xiguo Luo

Facial Emotion Recognition (FER) holds great importance in the fields of computer vision and machine learning. In this study, the aim is to improve the accuracy of facial expression recognition by incorporating attention mechanisms into Convolutional Neural Networks (CNN) with FER2013 dataset, which consists of grayscale images categorized into seven expressions. The combination of proposed CNN architecture and attention mechanisms is thoroughly elucidated, emphasizing the operations and interactions of their components. Additionally, the effectiveness of the new model is evaluated through experiments, comparing its performance with existing approaches in terms of accuracy. Besides, the results demonstrate that the CNN architecture with attention mechanisms outperforms the original CNN by achieving an improved accuracy rate of 69.07%, which is higher than 68.04% accuracy rate of original CNN. Moreover, the study further discusses the confusion matrix analysis, revealing the challenges faced in recognizing specific emotions due to limited training data and vague facial features. In the future, this study suggests addressing these limitations through data augmentation and to reduce the gap between training and testing accuracy. Overall, this research highlights the potential of attention mechanisms in enhancing facial expression recognition systems, paving the way for advanced applications in various domains.

  • Conference Article
  • 10.1109/iciss50791.2020.9307567
Comparative Analysis of Deep Convolutional Neural Networks Architecture in Facial Expression Recognition: A Survey
  • Nov 19, 2020
  • Rizky Andrian + 1 more

Facial expression recognition (FER) has made much progress and is supported by many scientific studies conducted in the past decade. The technique and architecture model used in the FER are the aspects that get the most improvement from the researchers. One of the famously used techniques in conducting FER is Deep Convolutional Neural Network (DCNN). The development of DCNN architecture has a vital role in increasing the accuracy of facial expression recognition. The choice of architecture will also affect the total computational costs required to perform facial expression recognition activities. This paper compares some of the DCNN architectures in the past FER research during 2010-2020 using various FER dataset. This paper presents detailed information on several DCNN architectures in terms of the dataset preprocessing techniques and accuracy value for some accessible FER datasets such as Fer2013, CK+, AFEW, and other datasets. This research will explain other FER researchers who are new in FER research to determine the DCNN architecture used based on several FER datasets.

  • Conference Article
  • Cite Count Icon 97
  • 10.1109/icic50835.2020.9288560
The Facial Emotion Recognition (FER-2013) Dataset for Prediction System of Micro-Expressions Face Using the Convolutional Neural Network (CNN) Algorithm based Raspberry Pi
  • Nov 3, 2020
  • Lutfiah Zahara + 4 more

One of the ways humans communicate is by using facial expressions. Research on technology development in artificial intelligence uses deep learning methods in human and computer interactions as an effective system application process. One example, if someone does show and tries to recognize facial expressions when communicating. The prediction of the expression or emotion of some people who see it sometimes does not understand. In psychology, the detection of emotions or facial expressions requires analysis and assessment of decisions in predicting a person's emotions or group of people in communicating. This research proposes the design of a system that can predict and recognize the classification of facial emotions based on feature extraction using the Convolution Neural Network (CNN) algorithm in real-time with the OpenCV library, namely: TensorFlow and Keras. The research design implemented in the Raspberry Pi consists of three main processes, namely: face detection, facial feature extraction, and facial emotion classification. The prediction results of facial expressions in research with the Convolutional Neural Network (CNN) method using Facial Emotion Recognition (FER-2013) were 65.97% (sixty-five point ninety-seven percent).

  • Research Article
  • Cite Count Icon 1
  • 10.1142/s219688882330003x
Real-Time Facial Expression Recognition: Advances, Challenges, and Future Directions
  • Dec 22, 2023
  • Vietnam Journal of Computer Science
  • Christine Dewi + 3 more

Facial emotion recognition (FER) is the technology or process of identifying and interpreting human emotions based on the analysis of facial expressions. It involves using computer algorithms and machine learning techniques to detect and classify emotional states from images or videos of human faces. Further, FER plays a vital role in recognizing and understanding human emotions to better interpret someone’s feelings, intentions, and attitudes. In the present time, it is widely used in various fields such as healthcare, human–computer interaction, law enforcement, security, and beyond. FER has a wide range of practical applications across various industries including Emotion Monitoring, Adaptive Learning, and Virtual Assistants. This paper presents a comparative analysis of FER algorithms, focusing on deep learning approaches. The performance of different datasets, including FER2013, JAFFE, AffectNet, and Cohn–Kanade, is evaluated using convolutional neural networks (CNNs), deep face, attentional convolutional networks (ACNs), and deep belief networks (DBNs). Among the tested algorithms, DBNs outperformed other algorithms, reaching the highest accuracy of 98.82%. These results emphasize the effectiveness of deep learning techniques, particularly DBNs, in FER. Additionally, outlining the advantages and disadvantages of current research on facial emotion identification might direct future research efforts in the direction of the most profitable directions.

  • Research Article
  • Cite Count Icon 14
  • 10.3390/electronics12224608
Automated Facial Emotion Recognition Using the Pelican Optimization Algorithm with a Deep Convolutional Neural Network
  • Nov 11, 2023
  • Electronics
  • Mohammed Alonazi + 5 more

Facial emotion recognition (FER) stands as a pivotal artificial intelligence (AI)-driven technology that exploits the capabilities of computer-vision techniques for decoding and comprehending emotional expressions displayed on human faces. With the use of machine-learning (ML) models, specifically deep neural networks (DNN), FER empowers the automatic detection and classification of a broad spectrum of emotions, encompassing surprise, happiness, sadness, anger, and more. Challenges in FER include handling variations in lighting, poses, and facial expressions, as well as ensuring that the model generalizes well to various emotions and populations. This study introduces an automated facial emotion recognition using the pelican optimization algorithm with a deep convolutional neural network (AFER-POADCNN) model. The primary objective of the AFER-POADCNN model lies in the automatic recognition and classification of facial emotions. To accomplish this, the AFER-POADCNN model exploits the median-filtering (MF) approach to remove the noise present in it. Furthermore, the capsule-network (CapsNet) approach can be applied to the feature-extraction process, allowing the model to capture intricate facial expressions and nuances. To optimize the CapsNet model’s performance, hyperparameter tuning is undertaken with the aid of the pelican optimization algorithm (POA). This ensures that the model is finely tuned to detect a wide array of emotions and generalizes effectively across diverse populations and scenarios. Finally, the detection and classification of different kinds of facial emotions take place using a bidirectional long short-term memory (BiLSTM) network. The simulation analysis of the AFER-POADCNN system is tested on a benchmark FER dataset. The comparative result analysis showed the better performance of the AFER-POADCNN algorithm over existing models, with a maximum accuracy of 99.05%.

  • Research Article
  • Cite Count Icon 1
  • 10.54097/cx6rc461
Comparison of CNN and ResNet Neural Networks on the Performance of Facial Expression Recognition
  • Apr 26, 2024
  • Highlights in Science, Engineering and Technology
  • Zhiming Gao

Facial expression recognition is a crucial task in numerous applications, including human-computer interaction, mental health monitoring, and human behavior analysis. Previous studies have primarily focused on individual models or techniques for improving emotion classification accuracy. However, a comparative analysis of different neural network architectures' performance for facial expression recognition is lacking. The main objective of this study is to compare the performance of Convolutional Neural Network (CNN) and Residual Network (ResNet) on the Fer2013 dataset. The author aims to analyze their behavior during the initial training phases and identify the architectural advantages and challenges associated with each model. Both models are trained using the same experimental setup to ensure a fair comparison. The models are trained using the Fer2013 dataset. The author employs a standard protocol for data preprocessing and augmentation. Results show that CNN achieves an accuracy of around 0.5 in the initial stages of training, which is significantly higher than ResNet's accuracy of 0.25. However, as training progresses, ResNet may outperform CNN because it has a more complicated structure that can capture more complex patterns. CNN exhibits superior performance during the initial training stage of the Fer2013 dataset. This reason may lie behind the fact that CNN has a simpler structure which makes it more sensitive on the basic features of the data.

  • Conference Article
  • Cite Count Icon 2
  • 10.1109/fie.2006.322738
Active Learning of Introductory Machine Learning
  • Jan 1, 2006
  • Maja Pantic + 1 more

This paper describes a computer-based training program for active learning of agent technology, expert systems, neural networks and case-based reasoning by undergraduate students using a simple agent framework. While many machine learning (ML) and artificial intelligence (AI) courses teach ML and AI concepts by means of programming assignments, these assignments have usually no connection to how the student will apply the newly obtained knowledge to previously unseen, real-world problems. The pedagogy that we adopted here is computer-based active learning: teams of students are presented with well-defined assignments aimed at building intelligent agents for person identification and recognition of facial expressions and emotions from video recordings of their faces. Classroom experience indicates that the students found the specified programming assignments highly motivating. Objective evaluation studies suggest that students learn much more effectively when a contextualized, collaborative, constructive, and reflective approach is used than when an orthodox, objectivist approach to teaching ML and AI techniques is used alone

  • Research Article
  • Cite Count Icon 48
  • 10.1142/s0218001419400159
Facial Emotion Recognition Using an Ensemble of Multi-Level Convolutional Neural Networks
  • Oct 1, 2019
  • International Journal of Pattern Recognition and Artificial Intelligence
  • Hai-Duong Nguyen + 5 more

Emotion recognition plays an indispensable role in human–machine interaction system. The process includes finding interesting facial regions in images and classifying them into one of seven classes: angry, disgust, fear, happy, neutral, sad, and surprise. Although many breakthroughs have been made in image classification, especially in facial expression recognition, this research area is still challenging in terms of wild sampling environment. In this paper, we used multi-level features in a convolutional neural network for facial expression recognition. Based on our observations, we introduced various network connections to improve the classification task. By combining the proposed network connections, our method achieved competitive results compared to state-of-the-art methods on the FER2013 dataset.

  • Research Article
  • Cite Count Icon 17
  • 10.1016/j.bpsc.2021.03.015
Facial Emotions Are Accurately Encoded in the Neural Signal of Those With Autism Spectrum Disorder: A Deep Learning Approach
  • Apr 16, 2021
  • Biological Psychiatry: Cognitive Neuroscience and Neuroimaging
  • Juan Manuel Mayor Torres + 5 more

Facial Emotions Are Accurately Encoded in the Neural Signal of Those With Autism Spectrum Disorder: A Deep Learning Approach

  • Research Article
  • Cite Count Icon 51
  • 10.1017/s1355617714000939
Facial and bodily emotion recognition in multiple sclerosis: the role of alexithymia and other characteristics of the disease.
  • Nov 1, 2014
  • Journal of the International Neuropsychological Society
  • Cinzia Cecchetto + 6 more

Multiple sclerosis (MS) may be associated with impaired perception of facial emotions. However, emotion recognition mediated by bodily postures has never been examined in these patients. Moreover, several studies have suggested a relation between emotion recognition impairments and alexithymia. This is in line with the idea that the ability to recognize emotions requires the individuals to be able to understand their own emotions. Despite a deficit in emotion recognition has been observed in MS patients, the association between impaired emotion recognition and alexithymia has received little attention. The aim of this study was, first, to investigate MS patient's abilities to recognize emotions mediated by both facial and bodily expressions and, second, to examine whether any observed deficits in emotions recognition could be explained by the presence of alexithymia. Thirty patients with MS and 30 healthy matched controls performed experimental tasks assessing emotion discrimination and recognition of facial expressions and bodily postures. Moreover, they completed questionnaires evaluating alexithymia, depression, and fatigue. First, facial emotion recognition and, to a lesser extent, bodily emotion recognition can be impaired in MS patients. In particular, patients with higher disability showed an impairment in emotion recognition compared with patients with lower disability and controls. Second, their deficit in emotion recognition was not predicted by alexithymia. Instead, the disease's characteristics and the performance on some cognitive tasks significantly correlated with emotion recognition. Impaired facial emotion recognition is a cognitive signature of MS that is not dependent on alexithymia.

More from: Automatika
  • Research Article
  • 10.1080/00051144.2025.2567093
Practically finite-time H ∞ deployment of large-scale multi-agent systems with sequence-dependent switching topology: a fuzzy PDE-based approach
  • Oct 2, 2025
  • Automatika
  • Pengbin Wei + 3 more

  • Research Article
  • 10.1080/00051144.2025.2561424
Fusion of deep recurrent neural network models and fuzzy decision support system for tweet sentiment analysis and classification
  • Oct 2, 2025
  • Automatika
  • A Sherin + 3 more

  • Research Article
  • 10.1080/00051144.2025.2577008
A novel transformer-less eleven-level inverter with optimized Buck–Boost controller and minimal switch design for grid-connected PV systems
  • Oct 2, 2025
  • Automatika
  • N Gowtham + 3 more

  • Research Article
  • 10.1080/00051144.2025.2569114
A new non-commutative key exchange protocol on combinatorial semigroup
  • Oct 2, 2025
  • Automatika
  • Jing Zhang + 2 more

  • Research Article
  • 10.1080/00051144.2025.2569113
A cutting-edge approach to observer-driven finite-time adaptive fault-tolerant control for unmanned surface vessels leveraging advanced neural architectures
  • Oct 2, 2025
  • Automatika
  • Minghua Sun + 1 more

  • Research Article
  • 10.1080/00051144.2025.2575572
The backstepping double integral terminal sliding mode control of upper limb rehabilitation robot based on friction compensation
  • Oct 2, 2025
  • Automatika
  • Peng Wang + 1 more

  • Research Article
  • 10.1080/00051144.2025.2572148
FEDTWIN – trustworthy digital twin as a service for visually impaired
  • Oct 2, 2025
  • Automatika
  • Hema Priya Natarajan + 1 more

  • Research Article
  • 10.1080/00051144.2025.2569119
Fault-tolerant control for high-speed trains based on neural network embedded compensation control
  • Oct 2, 2025
  • Automatika
  • Zixu Hao + 4 more

  • Research Article
  • 10.1080/00051144.2025.2549163
Adaptive tracking control for a class of strict feedback systems with unknown dead-zone input
  • Oct 2, 2025
  • Automatika
  • Lujuan Shen + 2 more

  • Research Article
  • 10.1080/00051144.2025.2569115
Empowering integrity and confidentiality in smart healthcare systems through effective edge cryptographic strategies
  • Oct 2, 2025
  • Automatika
  • R Gowthamani + 2 more

Save Icon
Up Arrow
Open/Close
  • Ask R Discovery Star icon
  • Chat PDF Star icon

AI summaries and top papers from 250M+ research sources.

Search IconWhat is the difference between bacteria and viruses?
Open In New Tab Icon
Search IconWhat is the function of the immune system?
Open In New Tab Icon
Search IconCan diabetes be passed down from one generation to the next?
Open In New Tab Icon