A Lightweight Method for Almost Real-Time Recognition of Emotions through Facial Geometry with Normalized Landmarks

Authors

DOI:

https://doi.org/10.5753/jbcs.2025.4456

Keywords:

Multimedia Processing, Affective Computing, Machine Learning, Multimodal Interaction, Facial Patterns

Abstract

Recognizing emotion is an intrinsic and fundamental part of human social interactions and is closely tied to behaviors that produce distinct facial patterns. Facial expressions indicate emotional context, so scientific, artistic, medical, and commercial interest in the field has increased, driving the development of computational techniques that can recognize emotion automatically. Although current methods provide satisfactory results, challenges persist, particularly challenges regarding the variable patterns of facial shapes and the response time achieved with low computational resources. For instance, some applications requiring instantaneous emotion recognition with high accuracy and low latency may be limited by the processing power, specially in the case of mobile devices. Here, we present a practical and simple method called REFEL (Recognizing Emotions through Facial Expression and Landmark Normalization), designed to identify facial expressions and human emotions in digital images. This method addresses sequential steps that reduce sample variability such as anatomical, scale factor and geometric variations and performs reductions of color, brightness and others in preprocessing tasks. REFEL normalizes facial fiducial points, commonly referred to as landmarks, and allows fine-tuning of informative aspects delineating facial patterns. Using landmark positions makes the process of recognizing emotions more reliable. REFEL also exploits classifiers explicitly tailored to identify facial emotions accurately. As in the case of related works, we employed Machine Learning algorithms, to achieve average accuracy higher than 90% for emotion recognition, when we applied REFEL before classification. We have experimented REFEL with various datasets including facial images that consider racial, age and gender factors as well as facial rotation. In this study, we also compared emotion classification without grouping emotions and with two emotion groups (Fear-Surprise and Anger-Disgust). Analysis of the ROC curves revealed that grouping emotions led to a slight improvement in the average performance of the REFEL method, with a 3% increase in accuracy. Our method represents an enhanced approach in terms of hit rate and response time, generates resilient outcomes, and relies less on the training set and classifier architecture. Furthermore, REFEL performs well, almost in real time, lowers the processing costs inherent to training, and is particularly suited to devices with limited processing capabilities, like cell phones. Emotion recognition methods usually have almost minimal real-time delay, which enables the system to react fast but not necessarily instantaneously. With REFEL, we hope to improve computational synthesis techniques and resources, and to help robust and motivating assistive technologies to advance. As future efforts, we intend to consider 3D images and videos.

Downloads

Download data is not yet available.

References

Agarwal, A. and Susan, S. (2023). Emotion recognition from masked faces using inception-v3. In 2023 5th International Conference on Recent Advances in Information Technology (RAIT), pages 1-6. DOI: 10.1109/RAIT57693.2023.10126777.

Alvarez, M., Luengo, D., and Lawrence, N. (2013). Linear latent force models using gaussian processes. IEEE Transactions on Pattern Analysis and Machine Intelligence, 35(11):2693-2705. DOI: 10.1109/TPAMI.2013.86.

Bohy, H., El Haddad, K., and Dutoit, T. (2022). A new perspective on smiling and laughter detection: Intensity levels matter. In 2022 10th International Conference on Affective Computing and Intelligent Interaction (ACII), pages 1-8. DOI: 10.1109/ACII55700.2022.9953896.

Chaugule, V., Abhishek, D., Vijayakumar, A., Ramteke, P. B., and Koolagudi, S. G. (2016). Product review based on optimized facial expression detection. In 2016 Ninth International Conference on Contemporary Computing (IC3), pages 1-6, Noida, India. IEEE. DOI: 10.1109/IC3.2016.7880213.

Cheng, Y. and Ling, S. (2008). 3d animated facial expression and autism in taiwan. In IEEE International Conference on Advanced Learning Technologies (ICALT 2008), pages 17-19, Los Alamitos, CA, USA. IEEE Computer Society. DOI: 10.1109/ICALT.2008.220.

Chollet, F. (2017). Deep Learning with Python. Manning Publications Co., Greenwich, CT, USA, 1st edition. Book.

Cohn, J. and Kanade, T. (2010). The extended cohn-kanade dataset (ck+): A complete dataset for action unit and emotion-specified expression. pages 94-101. IEEE Computer Society Conference on Computer Vision and Pattern Recognition. DOI: 10.1109/CVPRW.2010.5543262.

Cui, D., Huang, G.-B., and Liu, T. (2018). Elm based smile detection using distance vector. Pattern Recognition, 79:356-369. DOI: 10.1016/j.patcog.2018.02.019.

Dalal, D. and Triggs, B. (2005). Histograms of oriented gradients for human detection. In 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05), volume 1, pages 886-893 vol. 1, San Diego, CA, USA. IEEE. DOI: 10.1109/CVPR.2005.177.

Darwin, C. (2013). The Expression of the Emotions in Man and Animals. Cambridge Library Collection - Darwin, Evolution and Genetics. Cambridge University Press, England. DOI: 10.1017/CBO9781139833813.

Ekman, P. and Friesen, W. V. (1971). Constants across cultures in the face and emotion. Journal of Personality and Social Psychology, 17(2):124-129. DOI: 10.1037/h0030377.

Filho, Oge Marques; Vieira Neto, H. (1999). Processamento Digital de Imagens. Brasport, Brasil. Available online [link].

Garrido, G. and Joshi, P. (2018). OpenCV 3.X with Python By Example: Make the most of OpenCV and Python to build applications for object recognition and augmented reality. Packt Publishing, US, 2nd edition. Book.

Ghorbani, G; Targhi, A. T. and Dehshibi, M. (2015). Hog and lbp: Towards a robust face recognition system. In 2015 Tenth International Conference on Digital Information Management (ICDIM), pages 138-141, Jeju, South Korea. IEEE. DOI: 10.1109/ICDIM.2015.7381860.

Goeleven, E., Raedt, R. D., Leyman, L., and Verschuere, B. (2008). The karolinska directed emotional faces: A validation study. Cognition and Emotion, 22(6):1094-1118. DOI: 10.1080/02699930701626582.

Gonzales, R. C. and Woods, R. E. (2008). Digital Image Processing. Pearson, New Jersey, US, 3rd edition. Book.

Gupta, B. B., Gaurav, A., Chui, K. T., and Arya, V. (2024). Deep learning-based facial emotion detection in the metaverse. In 2024 IEEE International Conference on Consumer Electronics (ICCE), pages 1-6. DOI: 10.1109/ICCE59016.2024.10444217.

Guyon, I. and Elisseeff, A. (2003). An introduction to variable and feature selection. Journal of Machine Learning Research, 3:1157–1182. Available online [link].

Hassner, T., Harel, S.and Paz, E., and Enbar, R. (2015). Effective face frontalization in unconstrained images. In 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 4295-4304. DOI: 10.1109/CVPR.2015.7299058.

He, J., Wen, X., and Zhou, J. (2023). Advances and application of facial expression and learning emotion recognition in classroom. In Proceedings of the 2023 6th International Conference on Image and Graphics Processing, ICIGP '23, page 23–30, New York, NY, USA. Association for Computing Machinery. DOI: 10.1145/3582649.3582670.

Hess, U. (2001). The communication of emotion. In Emotions, Qualia and Consciousness, pages 397-409. Singapore. DOI: 10.1142/9789812810687_0031.

Hossain, M. A., Osman, M. H., Hamdan, A. A., Abdelhag, M. E., and Kechadi, M. T. (2023). Ferlp: Facial emotion recognition based on landmark points using artificial intelligence and machine learning. In 2023 14th International Conference on Computing Communication and Networking Technologies (ICCCNT), pages 1-6. DOI: 10.1109/ICCCNT56998.2023.10308392.

Ismail, N. and Sabri, M. I. M. (2009). Review of existing algorithms for face detection and recognition. In Proceedings of the 8th WSEAS International Conference on Computational Intelligence, Man-Machine Systems and Cybernetics, page 30–39, Stevens Point, Wisconsin, USA. World Scientific and Engineering Academy and Society (WSEAS). Available at[link].

Jia, S., Wang, S., Hu, C., Webster, P., and Li, X. (2021). Detection of genuine and posed facial expressions of emotion: Databases and methods. Front. Psychol. - Sec. Perception Science, 11. DOI: 10.3389/fpsyg.2020.580287.

Kim, W. (2022). Low-light image enhancement: A comparative review and prospects. IEEE Access, 10:84535-84557. DOI: 10.1109/ACCESS.2022.3197629.

Koohestani, F., Karimi, N., and Samavi, S. (2024). Revealing shadows: Low-light image enhancement using self-calibrated illumination. In 2024 32nd International Conference on Electrical Engineering (ICEE), pages 1-7. DOI: 10.1109/ICEE63041.2024.10667748.

Kwon, J., Oh, K. T., Kim, J., Kwon, O., Kang, H. C., and Yoo, S. K. (2023). Facial emotion recognition using landmark coordinate features. In 2023 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), pages 4916-4918. DOI: 10.1109/BIBM58861.2023.10385536.

Lahiri, U., Bekele, E., Dohrmann, E., Warren, Z., and Sarkar, N. (2011). Design of a virtual reality based adaptive response technology for children with autism spectrum disorder. In Affective Computing and Intelligent Interaction,Springer, pages 165-174. DOI: 10.1007/978-3-642-24600-5_20.

Langner, O., Dotsch, R., Bijlstra, G., Wigboldus, D. H. J., Hawk, S. T., and van Knippenberg, A. (2010). Presentation and validation of the radboud faces database. Cognition and Emotion, 24(8):1377-1388. DOI: 10.1080/02699930903485076.

Li, K., Xu, F., Wang, J., Dai, Q., and Liu, Y. (2012). A data-driven approach for facial expression synthesis in video. In 2012 IEEE Conference on Computer Vision and Pattern Recognition, pages 57-64, Providence, RI, US. IEEE. DOI: 10.1109/CVPR.2012.6247658.

Li, S. and Deng, W. (2018). Deep facial expression recognition: A survey. Computing Research Repository (CoRR), abs/1804.08348. DOI: 10.1109/TAFFC.2020.2981446.

Lyons, M., Kamachi, M., and Gyoba, J. (2017). Japanese Female Facial Expression (JAFFE) Database. DOI: 10.6084/m9.figshare.5245003.v2.

Macedo, A., Persona, L., and Meloni, F. (2024a). Recognition of emotions through facial geometry with normalized landmarks. In Proceedings of the 30th Brazilian Symposium on Multimedia and the Web, pages 257-266, Porto Alegre, RS, Brasil. SBC. DOI: 10.5753/webmedia.2024.243252.

Macedo, A. A., de Souza Gonçalves, V., Mandrá, P. P., Motti, V., Bulcão-Neto, R. F., and da Hora Rodrigues, K. R. (2024b). A mobile application and system architecture for online speech training in portuguese: design, development, and evaluation of sofiafala. DOI: https://doi.org/10.1007/s11042-024-19980-5.

Mehta, D., Siddiqui, M. F. H., and Javaid, A. Y. (2018). Facial emotion recognition: A survey and real-world user experiences in mixed reality. Sensors, 18(2). DOI: 10.3390/s18020416.

Mohan, K., Seal, A., Krejcar, O., and Yazidi, A. (2021). Facial expression recognition using local gravitational force descriptor-based deep convolution neural networks. IEEE Transactions on Instrumentation and Measurement, 70:1-12. DOI: 10.1109/TIM.2020.3031835.

Monzo, D; Albiol, A. and Mossi, M. J. (2010). A comparative study of facial landmark localization methods for face recognition using hog descriptors. In 2010 20th International Conference on Pattern Recognition, pages 1330-1333, Istanbul, Turkey. IEEE. DOI: 10.1109/ICPR.2010.1145.

Nhu, H. L., Dang, H. V., and Xuan, H. H. (2023). Facial emotion recognition by combining deep learning and averaged weight of face-regions. In 2023 15th International Conference on Knowledge and Systems Engineering (KSE), pages 1-4. DOI: 10.1109/KSE59128.2023.10299482.

Persona, L. (2022). Reconhecimento de emoções por meio da geometria facial com coordenadas normalizadas dos landmarks (em inglês, recognition of emotions through facial geometry using normalized landmark coordinates). Master thesis. DOI: 10.11606/D.59.2022.tde-21112023-114806.

Persona, L., Meloni, F., and Macedo, A. A. (2023). An accurate real-time method to detect the smile facial expression. In Proceedings of the 29th Brazilian Symposium on Multimedia and the Web, WebMedia '23, page 46–55, New York, NY, USA. Association for Computing Machinery. DOI: 10.1145/3617023.3617031.

Picard, R. W. (2016). Automating the recognition of stress and emotion: From lab to real-world impact. IEEE MultiMedia, 23(3):3-7. DOI: 10.1109/MMUL.2016.38.

Poulose, A., Kim, J. H., and Han, D. S. (2021). Feature vector extraction technique for facial emotion recognition using facial landmarks. In 2021 International Conference on Information and Communication Technology Convergence (ICTC), pages 1072-1076. DOI: 10.1109/ICTC52510.2021.9620798.

Revina, I. and Emmanuel, W. S. (2018). A survey on human face expression recognition techniques. Journal of King Saud University - Computer and Information Sciences. DOI: 10.1016/j.jksuci.2018.09.002.

Salman, F. Z., Madani, A., and Kissi, M. (2016). Facial expression recognition using decision trees. In 2016 13th International Conference on Computer Graphics, Imaging and Visualization (CGiV), pages 125-130, Beni Mellal, Morocco. IEEE. DOI: 10.1109/CGiV.2016.33.

Testa, R. L., Corrêa, C. G., Machado-Lima, A., and Nunes, F. L. S. (2019). Synthesis of facial expressions in photographs: Characteristics, approaches, and challenges. ACM Comput. Surv., 51(6):124:1-124:35. DOI: 10.1145/3292652.

Tottenham, N., Tanaka, J., Leon, A., Mccarry, T., Nurse, M., Hare, T., Marcus, D., Westerlund, A., Casey, B., and Nelson, C. (2009). The nimstim set of facial expressions: Judgments from untrained research participants. Psychiatry research, 168:242-9. DOI: 10.1016/j.psychres.2008.05.006.

Ullah, S. and Tian, W. (2021). A systematic literature review of recognition of compound facial expression of emotions. In Proceedings of the 2020 4th International Conference on Video and Image Processing, page 116–121. DOI: 10.1145/3447450.3447469.

Vaillancourt, J. (2010). Statistical methods for data mining and knowledge discovery. In Proceedings of the 8th International Conference on Formal Concept Analysis, ICFCA'10, pages 51-60, Berlin, Heidelberg. Springer-Verlag. DOI: 10.1007/978-3-642-11928-6_4.

Viola, P. and Jones, M. J. (2001). Robust real-time face detection. International Journal of Computer Vision, 57(2):137-154. DOI: 10.1023/B:VISI.0000013087.49260.fb.

Vonikakis, V. and Winkler, S. (2020). Identity-invariant facial landmark frontalization for facial expression analysis. In International Conference on Image Processing (ICIP), pages 2281-2285. DOI: 10.1109/ICIP40778.2020.9190989.

Winterle, P. (2014). Vetores e Geometria Analítica. MAKRON. Book.

Wu, Y. and Ji, Q. (2018). Facial landmark detection: A literature survey. International Journal of Computer Vision, 2:115-142. DOI: 10.1007/s11263-018-1097-z.

Xie, W.; Shen, L. and Jiang, J. (2017). A novel transient wrinkle detection algorithm and its application for expression synthesis. IEEE Transactions on Multimedia, 19(2):279-292. DOI: 10.1109/TMM.2016.2614429.

XIE, W., SHEB, L., YANG, M., and HOU, Q. (2015). Lighting difference based wrinkle mapping for expression synthesis. In 2015 8th International Congress on Image and Signal Processing (CISP), pages 636-641, Shenyang, China. IEEE. DOI: 10.1109/CISP.2015.7407956.

Zhang, Q., Wang, Z., Liu, Y., Qin, Z., Zhang, K., and Gedeon, T. (2023). Geometric-aware facial landmark emotion recognition. In 2023 6th International Conference on Software Engineering and Computer Science (CSECS), pages 1-6. DOI: 10.1109/CSECS60003.2023.10428424.

Downloads

Published

2025-05-19

How to Cite

Persona, L., Meloni, F., & Macedo, A. A. (2025). A Lightweight Method for Almost Real-Time Recognition of Emotions through Facial Geometry with Normalized Landmarks. Journal of the Brazilian Computer Society, 31(1), 325–337. https://doi.org/10.5753/jbcs.2025.4456

Issue

Section

Articles