Method for combining CNN-based features with geometric facial descriptors in emotion recognition
DOI:
https://doi.org/10.20535/2786-8729.7.2025.333629Keywords:
emotion recognition, facial landmarks, convolutional neural networks, .NET framework, feature fusionAbstract
This study presents a method for combining CNN-based visual features with geometric facial descriptors to improve the accuracy of emotion recognition in static images. The method integrates deep convolutional embeddings extracted from a pre-trained ResNetV2_101 model within the ML.NET framework with handcrafted geometric features computed from facial landmarks. Open-source datasets containing labeled emotional categories were used for experiments. At the first stage, deep image embeddings were obtained through transfer learning. At the second stage, 68 facial landmarks were detected to calculate distances and proportional relationships such as interocular distance, mouth width, eyebrow height, and other geometry-based indicators. These visual and geometric representations were concatenated into a unified feature space and classified using a multiclass linear model. The hybrid method achieved approximately 4% higher accuracy than the baseline CNN model relying solely on pixel-level features (from about 63% to 67%), confirming that combining heterogeneous features enhances generalization and robustness. The results also highlight that geometric descriptors act as stabilizing factors, compensating for noise, occlusions, and lighting variations that degrade CNN-only models. The developed pipeline demonstrates the feasibility of integrating interpretable geometric cues with deep embeddings directly in C# using ML.NET. The research novelty lies in proposing an interpretable hybrid model for emotion recognition that improves reliability while maintaining compatibility with .NET-based applications. The approach offers an accessible solution for developers working within enterprise .NET ecosystems, enabling direct deployment without cross-language integration. Future research will focus on extending the model toward multimodal emotion analysis that incorporates speech, gesture, and physiological signals to enhance contextual understanding of affective states. Additionally, the hybrid model can serve as a diagnostic tool for studying emotion dynamics in psychological or behavioral research.
References
A. Mehrabian, Silent Messages. Belmont, CA, USA: Wadsworth, 1971, pp. 13–14.
T. K. Arora, “Optimal facial feature based emotional recognition using deep learning algorithm,” Computational Intelligence and Neuroscience, 2022, https://doi.org/10.1155/2022/8379202.
F. Ye, “Emotion recognition of online education learners by convolutional neural networks,” Computational Intelligence and Neuroscience, 2022, https://doi.org/10.1155/2022/4316812.
Y. Liu, W. Dai, C. Feng, W. Wang, G. Yin, J. Zeng, and S. Shan, “MAFW: A large-scale, multi-modal, compound affective database for dynamic facial expression recognition in the wild,” in Proc. 30th ACM Int. Conf. Multimedia (MM’22), Lisbon, Portugal, Oct. 2022, pp. 1–9, https://doi.org/10.1145/3503161.3548190.
Denissen, J. J. A., Butalid, L., Penke, L., & van Aken, M. A. G. “The effects of weather on daily mood: A multilevel approach”. Emotion, 8(5), 662–667, 2008, https://doi.org/10.1037/a0013497.
P. Ekman and W. V. Friesen, Facial Action Coding System (FACS). Palo Alto, CA, USA: Consulting Psychologists Press, 1978, https://doi.org/10.1037/t27734-000.
O. Gavrylenko and L. Zinchenko, “Detection of human emotions using machine learning tools,” in Global Learning: Problems, Causes, Solutions, and Theories, Eur. Conf. Proc., 2024. Online: https://eu-conf.com/wp-content/uploads/2024/10/GLOBAL-LEARNING-PROBLEMS-CAUSES-SOLUTIONS-AND-THEORIES.pdf.
Bo Xiao, Hai Zhu, Sujun Zhang, Zi OuYang, Tandong Wang and Saeed Sarvazizi, “Gray-Related Support Vector Machine Optimization Strategy and Its Implementation in Forecasting Photovoltaic Output Power”, International Journal of Photoenergy, 2022, https://doi.org/10.1155/2022/3625541.
C. Rahmad, R. A. Asmara, D. R. H. Putra, I. Dharma, H. Darmono and I. Muhiqqin, “Comparison of Viola–Jones Haar cascade classifier and histogram of oriented gradients (HOG) for face detection,” IOP Conference Series: Materials Science and Engineering, vol. 732, no. 1, 2020, https://doi.org/10.1088/1757-899X/732/1/012038.
R. Bridgelall, “Tutorial on support vector machines”, Preprints, 2022, https://doi.org/10.20944/preprints202201.0232.v1.
Purwono Purwono, Alfian Ma'arif, Wahyu Rahmaniar, Haris Imam Karim Fathurrahman, Aufaclav Zatu Kusuma Frisky and Qazi Mazhar ul Haq, “Understanding of convolutional neural network (CNN): A review,” International Journal of Robotics and Control Systems, vol. 2, no. 4, p.739-748, 2022, https://doi.org/10.31763/ijrcs.v2i4.888.
Nada Elgendy, Ahmed Elragal and Tero Päivärinta, “ ECAS: a modern data-driven decision theory for big data and analytics”, Journal of Decision Systems, vol. 31, no. 4, p. 337–373, 2022, https://doi.org/10.1080/12460125.2021.1894674.
H. Song, “Comparison of different depths of convolutional neural network: Deep and shallow CNN comparison based on FER-2013,” 8th International Conference on Computer-Aided Design, Manufacturing, Modeling and Simulation (CDMMS 2023), vol. 41, 2023, https://doi.org/10.54097/hset.v41i.6746.
Özge Cömert and Nurcan Yücel, “Review Mate: A cutting-edge model for analyzing the sentiment of online customer product reviews using ML.NET,” Journal of Soft Computing and Artificial Intelligence, vol. 5, no. 2, p.74-88, 2024, https://doi.org/10.55195/jscai.1595798.
Cheng Qian, João Alexandre Lobo Marques, Auzuir Ripardo de Alexandria and Simon James Fong, “Application of Multiple Deep Learning Architectures for Emotion Classification Based on Facial Expressions”, Sensors, vol. 25, no. 5, 2025, https://doi.org/10.3390/s25051478.
Marcos Rodrigo, Carlos Cuevas and Narciso García, “Comprehensive comparison between vision transformers and convolutional neural networks for face recognition tasks,” Scientific Reports, vol. 14, Article 21392, 2024, https://doi.org/10.1038/s41598-024-72254-w.
Downloads
Published
How to Cite
Issue
Section
License
Copyright (c) 2025 Information, Computing and Intelligent systems

This work is licensed under a Creative Commons Attribution 4.0 International License.