Facial Emotion Recognition and Synthesis with Convolutional Neural Networks

Main Article Content

Karkuzhali S
Murugeshwari. R
Umadevi V

Abstract

A crucial component of human communication is conveying emotions, intentions, and social signals. In this era of artificial intelligence and computer vision, the development of automated systems for facial expression synthesis and recognition has attracted significant attention due to their wide range of applications, including human-computer interaction, virtual reality, emotional analysis, and healthcare. This research focuses on integrating deep convolutional neural networks (CNNs) to address challenges in both facial expression synthesis and recognition. On the synthesis front, a generative CNN architecture is proposed to generate realistic facial expressions, enabling the generation of various emotional states from neutral faces. The network learns to capture the intricate details of human expressions, including subtle muscle movements and spatial relationships among facial features. For facial expression recognition, a separate CNN-based model is developed to classify the synthesised expressions accurately. The recognition model is trained on a large dataset of annotated facial expressions and is designed to handle real-world variations in lighting, pose, and occlusions. The CNN leverages its ability to automatically learn relevant features from raw image data, eliminating the need for manual feature engineering. The experimental results demonstrate the effectiveness of the proposed approach. The synthesized expressions exhibit a high degree of realism and diversity, effectively capturing the nuances of human emotions. The recognition model achieves state-of-the art accuracy in classifying these synthesised expressions, surpassing traditional methods and demonstrating the power of deep learning in this domain. This research advances automatic facial expression synthesis and recognition, with potential applications in human-computer interaction, affective computing, and virtual environments. The deep CNN-based approach offers a promising avenue for enhancing our understanding of human expressions and enabling more emotionally aware and responsive AI systems. The significance of emotion classification in human-machine interactions has grown significantly. Over the past decade, businesses have become increasingly attuned to the insights that analysing a person's facial expressions in images or videos can provide into their emotional state. Various organizations are currently leveraging emotion recognition to gauge customer sentiments towards their products. The applications of this technology extend well beyond market research and digital advertising. Convolutional Neural Networks (CNNs) have emerged as a valuable tool for eliciting emotions from facial landmarks, as they can automatically extract relevant information. Challenges such as brightness variations, background changes, and other factors can be effectively mitigated by isolating the essential features using techniques such as face resizing and normalisation. However, it's important to note that neural networks depend on extensive datasets for optimal performance. In cases where data availability is limited, strategies like data augmentation through techniques such as rotation can be employed to compensate. Additionally, fine tuning the CNN's architecture can enhance its accuracy in predicting emotions. Consequently, this approach enables the real-time identification of seven distinct emotions – anger, sadness, happiness, disgust, neutrality, fear, and surprise – from facial expressions in images.

Downloads

Download data is not yet available.

Article Details

Section

Articles

How to Cite

Facial Emotion Recognition and Synthesis with Convolutional Neural Networks (Karkuzhali S, Murugeshwari. R, & Umadevi V , Trans.). (2026). International Journal of Emerging Science and Engineering (IJESE), 14(3), 31-42. https://doi.org/10.35940/ijese.F2559.14030226
Share |

References

Corneanu, C., Simn, M., Cohn, J. F., & Guerrero, S. (2016). Survey on RGB, 3D, thermal, and multimodal approaches for facial expression recognition: history, trends, and affect-related applications. IEEE Transactions on Pattern Analysis and Machine Intelligence, 38(8), 1548–1568. DOI: https://doi.org/10.1109/tpami.2016.2515606

Sariyanidi, E., Gunes, H., & Cavallaro, A. (2015). Automatic analysis of facial affect: a survey of registration, representation, and recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence, 37(6), 1113–1133.

DOI: https://doi.org/10.1109/TPAMI.2014.2366127, works remain significant, see the declaration

Soleymani, M., Asghari-Esfeden, S., Fu, Y., & Pantic, M. (2016). Analysis of EEG signals and facial expressions for continuous emotion detection. IEEE Transactions on Affective Computing, 7(1), 17–28. https://www.ibug.doc.ic.ac.uk/media/uploads/documents/journal_eegcontinuous.pdf

Zhang, Z., Ping, L., Chen, C., & Tang, X. (2016). From facial expression recognition to interpersonal relation prediction. International Journal of Computer Vision, 126(5), 550–569.DOI: https://doi.org/10.1007/s11263-017-1055-1

Xie, S., & Hu, H. (2019). Facial expression recognition using hierarchical features with deep comprehensive multipatches aggregation convolutional neural networks. IEEE Transactions on Multimedia, 21(1), 211–220. https://ieeexplore.ieee.org/iel7/6287639/10005208/10210543.pdf

Fan, Y., Li, V., & Lam, J. (2020). Facial Expression Recognition with Deeply-Supervised Attention Network. IEEE Transactions on Affective Computing, 1-16. doi.org/10.1109/taffc.2020.2988264

Iqbal, M., Abdullah-Al-Wadud, M., Ryu, B., Makhmudkhujaev, F., & Chae, O. (2020). Facial Expression Recognition with Neighbourhood-Aware Edge Directional Pattern (NEDP). IEEE Transactions on Affective Computing, 11(1), 125– 137. https://www.computer.org/csdl/journal/ta/2020/01/08350037/13rRUytnsVm

Kumawat, S., Verma, M., & Raman, S. (2019). LBVCNN: local binary volume convolutional neural network for facial expression recognition from image sequences. IEEE Transactions on Computer Vision and Pattern Recognition, 1904.07647. https://arxiv.org/abs/1904.07647

Tang, Y., Zhang, X., Hu, X., Wang, S., & Wang, H. (2021). Facial Expression Recognition Using Frequency Neural Network. IEEE Transactions on Image Processing, 30, 444–457. https://pure.bit.edu.cn/en/publications/

Bailly, K., & Dubuisson, S. (2019). Dynamic Pose-Robust Facial Expression Recognition by Multi-View Pairwise Conditional Random Forests. IEEE Transactions on Image Processing, 30, 167–181. https://arxiv.org/abs/1607.06250

Yan, Y., Huang, Y., Chen, S., Shen, C., & Wang, H. (2020). Joint Deep Learning of Facial Expression Synthesis and Recognition. IEEE Transactions on Multimedia, 22(11), 2792–2807. https://arxiv.org/abs/2002.02194

Agarwal, S., & Mukherjee, D. (2019). Synthesis of realistic facial expressions using an expression map. IEEE Transactions on Multimedia, 21(4), 902–914. DOI: https://doi.org/10.1109/TMM.2018.2871417

Huanga, W., Zhanga, S., Zhangc, P., Zhac, Y., Fangd, Y., & Zhangc, Y. (2021). Identity-aware Facial Expression Recognition via Deep Metric Learning based on Synthesized Images. IEEE Transactions on Multimedia, 1424-1445. https://dl.acm.org/doi/10.1109/TMM.2021.3096068

Wen, S., Zhang, Y., Li, K., & Qian, M. (2018). Deep Emotion Recognition With Enhanced CNN Features. IEEE Transactions on Affective Computing. https://ieeexplore.ieee.org/iel7/6287639/10005208/10234550.pdf

Prates, D. M. G., Penalva, E. M., & Giraldi, G. A. (2017). Facial Expression Recognition with Convolutional Neural Networks: Coping with Few Data and the Training Sample Order. Pattern Recognition Letters. https://www.lcad.inf.ufes.br/wiki/index.php/

Islam, A. M. J. S., Siddiquee, M. M., Shoyaib, A. M. A., & Alam, M. S. (2019). Deep Learning-Based Human Emotion Recognition from Facial Expression: A Review. Journal of Ambient Intelligence and Humanized Computing. DOI: http://doi.org/10.5120/ijca2024923159

Masi, I., Tran, A., Hassner, T., Leksut, J., & Medioni, G. (2016). Facial Expression Recognition in the Wild. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR). https://doi.org/10.3390/app9214678?urlappend=%3Futm_source

Wu, X., He, R., Sun, Z., & Tan, T. (2018). A Light CNN for Deep Face Representation with Noisy Labels. IEEE Transactions on Information Forensics and Security. https://ieeexplore.ieee.org/document/8353856

Kim, K. K., Park, W. T., & Kweon, I. S. (2018). DctNet: Face Recognition Using Discriminant Contextual Representation and Face Anti-Spoofing. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR). https://www.ijcaonline.org/archives/volume186/number11/facial-emotion-recognition-and-synthesis-with-convolutional-neural-networks/

Khorrami, M., Paine, T., & Huang, T. (2017). Do Deep Neural Networks

Learn Facial Action Units When Doing Expression Recognition? Proceedings of the IEEE International Conference on Computer Vision (ICCV). DOI: https://doi.org/10.48550/arXiv.1510.02969

Li, H., Lin, Z., Shen, X., Brandt, J., & Hua, G. (2015). A Convolutional Neural Network Cascade for Face Detection. Proceedings of The IEEE Conference on Computer Vision and Pattern Recognition (CVPR). https://openaccess.thecvf.com/content_cvpr_2015/papers/, works remain significant, see the declaration

Liu, M., Li, S., Shan, S., Wang, R., & Chen, X. (2015). Deep Learning Deformable Facial Action Parts Model for Dynamic Expression Analysis. Proceedings of the IEEE International Conference on Computer Vision (ICCV). https://www.jdl.link/doc/2011/201511618541847969, works remain significant, see the declaration

Song, Y., Li, M., Tao, D., & Sun, X. (2018). Facial Expression Recognition with Incomplete Data. IEEE Transactions on Image Processing. https://ieeexplore.ieee.org/stamp/stamp.jsp?arnumber=9367192

Yang, B., Cao, J., Ni, R., & Zhang, Y. (2018). Facial Expression Recognition Using Weighted Mixture Deep Neural Network Based on Double-Channel Facial Images. IEEE Access, 6, 4630-4640. https://ieeexplore.ieee.org/stamp/stamp.jsp?arnumber=8214102

The FER 2013 Dataset. [Online]. Available: https://www.kaggle.com/msambare/fer2013, works remain significant, see the declaration

Li, Y., Zeng, J., Shan, S., & Chen, X. (2019). Occlusion Aware Facial Expression Recognition Using CNN With Attention Mechanism. IEEE Transactions On Image Processing, 28(5), 2439-2450.DOI: https://doi.org/10.1109/tip.2018.2886767

Zhang, T., Zheng, W., Cui, Z., Zong, Y., Yan, J., & Yan, K. (2016). A Deep Neural Network-Driven Feature Learning Method for Multi-view Facial Expression Recognition. IEEE Transactions on Multimedia, 18(12), 2528-2536. https://ieeexplore.ieee.org/document/7530823/

Isola, P., Zhu, J., Zhou, T., & Efros, A. (2016). Image-to-image translation with conditional adversarial networks. arXiv preprint, pp. 1611.007004. DOI: https://doi.org/10.48550/arXiv.1611.07004

Gulrajani, I., Ahmed, F., Arjovsky, M., Dumoulin, V., & Courville, A. (2018). Improved training of Wasserstein GANs. arXiv preprint, pp. 1703-1711. DOI: https://doi.org/10.48550/arXiv.1704.00028

Lee, S. H., & Ro, Y. M. (2016). Partial Matching of Facial Expression Sequence Using Over-Complete Transition Dictionary for Emotion Recognition. IEEE Transactions On Affective Computing, 7(4), 387-408. https://pure.kaist.ac.kr/en/publications/partial-matching-of-facial-expression-sequence-using-over-complet/

Tanfous, A. B., Drira, H., & Amor, B. B. (2020). Sparse Coding of Shape Trajectories for Facial Expression and Action Recognition. IEEE Transactions On Pattern Analysis and Machine Intelligence, 42(10), 2594-2607. DOI: https://doi.org/10.1109/tpami.2019.2932979

Wen, L., Zhou, J., Huang, W., & Chen, F. (2022). A Survey of Facial Capture for Virtual Reality, pp. 6042-6052. https://ieeexplore.ieee.org/stamp/stamp.jsp?arnumber=9662378

Most read articles by the same author(s)

<< < 4 5 6 7 8 9 10 11 12 13 > >>