Abstract
Emotional Expression Recognition (EER) and Facial Expression Recognition (FER) are active research areas in the affective computing field, which involves studying human emotion, recognition, and sentiment analysis. The main objective of this research is to develop algorithms that can accurately interpret and estimate human emotions from portrait images. The emotions depicted in a portrait can reflect various factors such as psychological and physiological states, the artist’s emotional responses, social and environmental aspects, and the period in which the painting was created. This task is challenging because (i) the portraits are often depicted in an artistic or stylized manner rather than realistically or naturally, (ii) the texture and color features obtained from natural faces and paintings differ, affecting the success rate of emotion recognition algorithms, and (iii) it is a new research area, where practically we do not have visual arts portrait facial emotion estimation models or datasets.
To address these challenges, we need a new class of tools and a database specifically tailored to analyze portrait images. This study aims to develop art portrait emotion recognition methods and create a new digital portrait dataset containing 927 images. The proposed model is based on (i) a 3-dimensional estimation of emotions learned by a deep neural network and (ii) a novel deep learning module (3DEmo) that could be easily integrated into existing FER models. To evaluate the effectiveness of the developed models, we also tested their robustness on a facial emotion recognition dataset. The extensive simulation results show that the presented approach outperforms established methods. We expect that this dataset and the developed new tools will encourage further research in recognizing emotions in portrait paintings and predicting artists’ emotions in the painting period based on their artwork.
- [1] . 2011. Human-computer interaction using emotion recognition from facial expression. In 5th European Symposium on Computer Modeling and Simulation (UKSim’11). 196–201.
DOI: Google ScholarDigital Library - [2] . 2021. ArtEmis: Affective language for visual art. In Conference on Computer Vision and Pattern Recognition (CVPR’21). 11569–11579.Google ScholarCross Ref
- [3] . 2014. Pleasure, arousal, dominance: Mehrabian and Russell revisited. Curr. Psychol. 33 (2014), 405–421.
DOI: Google ScholarCross Ref - [4] . 1999. The structure of current affect: Controversies and emerging consensus. Curr. Direct. Psychol. Sci. 8, 1 (1999), 10–14.
DOI: Google ScholarCross Ref - [5] . 2016. Training deep networks for facial expression recognition with crowd-sourced label distribution. In ACM International Conference on Multimodal Interaction (ICMI’16).Google ScholarDigital Library
- [6] . 2016. EmotioNet: An accurate, real-time algorithm for the automatic annotation of a million facial expressions in the wild. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR’16). 5562–5570.
DOI: Google ScholarCross Ref - [7] . 1984. FCM: The fuzzy c-means clustering algorithm. Comput. Geosci. 10, 2 (1984), 191–203.
DOI: Google ScholarCross Ref - [8] . 2017. A Deep Learning Perspective on the Origin of Facial Expressions.
DOI: Google ScholarCross Ref - [9] . 2023. Computational tracking of head pose through 500 years of fine-art portraiture. In Computer Vision and Analysis of Art. SPIE Electronic Imaging, San Francisco, CA.Google Scholar
- [10] . 2021. Deep learning-based facial emotion recognition for human–computer interaction applications. Neural Comput. Applic.
DOI: Google ScholarDigital Library - [11] . 2017. Learning spatial and temporal cues for multi-label facial action unit detection. 25–32.
DOI: Google ScholarDigital Library - [12] . 2018. Group emotion recognition strategies for entertainment robots. In IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS’18). 813–818.
DOI: Google ScholarDigital Library - [13] . 2020. EmotiW 2020: Driver Gaze, Group Emotion, Student Engagement and Physiological Signal Based Challenges. Association for Computing Machinery, New York, NY, 784–789. Retrieved from Google ScholarDigital Library
- [14] . 2014. Compound facial expressions of emotion. Proc. Nat. Acad. Sci. 111, 15 (2014), E1454–E1462.
DOI: Google ScholarCross Ref - [15] . 2015. Recurrent neural networks for emotion recognition in video. InACM International Conference on Multimodal Interaction (
ICMI’15 ). Association for Computing Machinery, New York, NY, 467–474.DOI: Google ScholarDigital Library - [16] . 1971. Constants Across Cultures in the Face and Emotion. 124–129 pages. Retrieved from Google ScholarCross Ref
- [17] . 2020. Facial expression recognition and ordinal intensity estimation: A multilabel learning approach. In Advances in Visual Computing, , , , , , , , , and (Eds.). Springer International Publishing, Cham, 581–592.Google Scholar
- [18] . 2021. Facial expression recognition: A review of trends and techniques. IEEE Access PP (
09 2021), 1–1.DOI: Google ScholarCross Ref - [19] . 2016. Video-based emotion recognition using CNN-RNN and C3D hybrid networks. In 18th ACM International Conference on Multimodal Interaction (ICMI’16). Association for Computing Machinery, New York, NY, 445–450.
DOI: Google ScholarDigital Library - [20] . 2002. Head-pose invariant facial expression recognition using convolutional neural networks. In 4th IEEE International Conference on Multimodal Interfaces. 529–534.
DOI: Google ScholarDigital Library - [21] . 2002. Robust face analysis using convolutional neural networks. In International Conference on Pattern Recognition, Vol. 2. 40–43 vol.2.
DOI: Google ScholarCross Ref - [22] . 2021. Virtual reality facial emotion recognition in social environments: An eye-tracking study. Internet Intervent. 25 (2021), 100432.
DOI: Google ScholarCross Ref - [23] . 2018. Children facial expression production: Influence of age, gender, emotion subtype, elicitation condition and culture. Front. Psychol. 9 (2018).
DOI: Google ScholarCross Ref - [24] . 2020. 3D-CNN for facial emotion recognition in videos. In International Symposium on Visual Computing (ISVC’20).Google Scholar
- [25] . 2012. A real time facial expression classification system using local binary patterns. In 4th International Conference on Intelligent Human Computer Interaction (IHCI’12). 1–5.
DOI: Google ScholarCross Ref - [26] . 2017. Facial expression recognition using enhanced deep 3D convolutional neural networks. In IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW’17). IEEE.
DOI: Google ScholarCross Ref - [27] . 2016. Deep residual learning for image recognition. In IEEE Conference on Computer Vision and Pattern Recognition. 770–778.Google ScholarCross Ref
- [28] . 2020. Multi angle optimal pattern-based deep learning for automatic facial expression recognition. Pattern Recog. Lett. 139 (2020), 157–165.
DOI: Google ScholarCross Ref - [29] . 2018. Driver’s facial expression recognition in real-time for safe driving. Sensors (Basel, Switz.) (
Dec. 2018), 12 4270.DOI: Google ScholarCross Ref - [30] . 2013. 3D convolutional neural networks for human action recognition. IEEE Trans. Pattern Anal. Mach. Intell. 35, 1 (2013), 221–231.
DOI: Google ScholarDigital Library - [31] . 2019. Multi-objective based spatio-temporal feature representation learning robust to expression intensity variations for facial expression recognition. IEEE Trans. Affect. Comput. 10 (2019), 223–236.Google ScholarCross Ref
- [32] . 2020. Facial expression monitoring system for predicting patient’s sudden movement during radiotherapy using deep learning. J. Appl. Clin. Med. Phys. 21, 191–199.
DOI: Google ScholarCross Ref - [33] . 2017. Adam: A Method for Stochastic Optimization.
arxiv:1412.6980 [cs.LG]Google Scholar - [34] . 2021. SEWA DB: A rich database for audio-visual emotion and sentiment research in the wild. IEEE Trans. Pattern Anal. Mach. Intell. 43, 3 (
Mar. 2021), 1022–1040.DOI: Google ScholarCross Ref - [35] . 2003. Real-time estimation of facial expression intensity. In IEEE International Conference on Robotics and Automation. 2567–2572.
DOI: Google ScholarCross Ref - [36] . 2018. How training data affect the accuracy and robustness of neural networks for image classification.Google Scholar
- [37] . 2020. Deep facial expression recognition: A survey. IEEE Trans. Affect. Comput. (2020), 1–1.
DOI: Google ScholarCross Ref - [38] . 2021. Novel features for art movement classification of portrait paintings. Image Vis. Comput. 108 (2021), 104121.
DOI: Google ScholarCross Ref - [39] . 1982. Least squares quantization in PCM. IEEE Trans. Inf. Theor. 28, 2 (1982), 129–137.
DOI: Google ScholarDigital Library - [40] . 2010. The extended Cohn-Kanade dataset (CK+): A complete dataset for action unit and emotion-specified expression. In IEEE Computer Society Conference on Computer Vision and Pattern Recognition—Workshops. 94–101.
DOI: Google ScholarCross Ref - [41] . 1998. Coding facial expressions with Gabor wavelets. In 3rd IEEE International Conference on Automatic Face and Gesture Recognition. 200–205.
DOI: Google ScholarCross Ref - [42] . 1998. The Japanese Female Facial Expression (JAFFE) Dataset.
DOI: Google ScholarCross Ref - [43] . 2018. Recognition of facial emotional expressions among Italian pre-adolescents, and their affective reactions. Front. Psychol. 9 (2018).
DOI: Google ScholarCross Ref - [44] . 2013. DISFA: A spontaneous facial action intensity database. IEEE Trans. Affect. Comput. 4, 2 (2013), 151–160.
DOI: Google ScholarDigital Library - [45] . 2018. Facial emotion recognition: A survey and real-world user experiences in mixed reality. Sensors (Basel, Switz.) (
Feb. 2018).DOI: Google ScholarCross Ref - [46] . 2021. A dataset and a convolutional model for iconography classification in paintings. J. Comput. Cult. Herit. 14, 4, Article
46 (July 2021), 18 pages.DOI: Google ScholarDigital Library - [47] . 2019. AffectNet: A database for facial expression, valence, and arousal computing in the wild. IEEE Trans. Affect. Comput. 10, 1 (
Jan. 2019), 18–31.DOI: Google ScholarDigital Library - [48] . 2017. Deep spatio-temporal features for multimodal emotion recognition. In IEEE Winter Conference on Applications of Computer Vision (WACV’17). 1215–1223.
DOI: Google ScholarCross Ref - [49] . 2016. Recognition and intensity estimation of facial expression using ensemble classifiers. In IEEE/ACIS 15th International Conference on Computer and Information Science (ICIS’16). 1–6.
DOI: Google ScholarCross Ref - [50] . 2005. Web-based database for facial expression analysis. In IEEE International Conference on Multimedia and Expo.
DOI: Google ScholarCross Ref - [51] . 2019. PyTorch: An imperative style, high-performance deep learning library. In Advances in Neural Information Processing Systems, Vol. 32, , , , , , and (Eds.). Curran Associates, Inc., 8024–8035. Retrieved from http://papers.neurips.cc/paper/9015-pytorch-an-imperative-style-high-performance-deep-learning-library.pdfGoogle Scholar
- [52] . 2020. Learn Distributed GAN with Temporary Discriminators.Google Scholar
- [53] . 2016. An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747 (2016).Google Scholar
- [54] . 2021. A stochastic view of varying styles in art paintings. Heritage 4, 1 (2021), 333–348.
DOI: Google ScholarCross Ref - [55] . 2012. Computer analysis of art. J. Comput. Cult. Herit. 5, 2, Article
7 (Aug. 2012), 11 pages.DOI: Google ScholarDigital Library - [56] . 2018. Blended emotion in-the-wild: Multi-label facial expression recognition using crowdsourced annotations and deep locality feature learning. Int. J. Comput. Vis. 127 (2018), 884–906.Google Scholar
- [57] . 2013. Levels of valence. Front. Psychol. 4 (2013).
DOI: Google ScholarCross Ref - [58] . 2020. Efficient Facial Feature Learning with Wide Ensemble-based Convolutional Neural Networks. Retrieved from https://www2.informatik.uni-hamburg.de/wtm/publications/2020/SMW20/SMW20.pdfGoogle Scholar
- [59] . 2015. Face expression detection on Kinect using active appearance model and fuzzy logic. Procedia Comput. Sci. 59 (2015), 268–274.
DOI: Google ScholarCross Ref - [60] . 2014. Real-time mobile facial expression recognition system—A case study. In IEEE Conference on Computer Vision and Pattern Recognition Workshops. 132–137.
DOI: Google ScholarDigital Library - [61] . 2015. Facial emotion recognition using depth data. In 8th International Conference on Human System Interaction (HSI’15). 271–277.
DOI: Google ScholarCross Ref - [62] . 2020. KaoKore: A Pre-modern Japanese Art Facial Expression Dataset.
arxiv:2002.08595 [cs.CV]Google Scholar - [63] . 2021. Estimation of continuous valence and arousal levels from faces in naturalistic conditions. Nat. Mach. Intell. 3 (2021), 42–50.
DOI: Google ScholarCross Ref - [64] . 2020. Automatic recognition of student emotions from facial expressions during a lecture. Comput. Educ. 148 (2020), 103797.
DOI: Google ScholarDigital Library - [65] . 2014. The MPI emotional body expressions database for narrative scenarios. PloS One 9, 12 (2014), e113647.
DOI: Google ScholarCross Ref - [66] . 2020. Suppressing Uncertainties for Large-Scale Facial Expression Recognition.
DOI: Google ScholarCross Ref - [67] . 2019. Towards facial expression analysis in a driver assistance system. In 14th IEEE International Conference on Automatic Face Gesture Recognition (FG’19). 1–4.
DOI: Google ScholarDigital Library - [68] . 2020. Facial expression distribution prediction based on surface electromyography. Expert Syst. Applic. 161 (2020), 113683.
DOI: Google ScholarCross Ref - [69] . 2022. On-road driver emotion recognition using facial expression. Appl. Sci. 12, 2 (2022).
DOI: Google ScholarCross Ref - [70] . 2020. Facial expression recognition with two-branch disentangled generative adversarial network. IEEE Trans. Circ. Syst. Vid. Technol. (2020).Google Scholar
- [71] . 2014. CASME II: An improved spontaneous micro-expression database and the baseline evaluation. PLoS One 9 (
01 2014), 1–8.DOI: Google ScholarCross Ref - [72] . 2020. A closer look at accuracy vs. robustness. In Advances in Neural Information Processing Systems, , , , , and (Eds.), Vol. 33. Curran Associates, Inc., 8588–8601. Retrieved from https://proceedings.neurips.cc/paper/2020/file/61d77652c97ef636343742fc3dcf3ba9-Paper.pdfGoogle Scholar
- [73] . 2019. The face of art: Landmark detection and geometric style in portraits. ACM Trans. Graph. 38, 4, Article
60 (July 2019), 15 pages.DOI: Google ScholarDigital Library - [74] . 2006. A 3D facial expression database for facial behavior research. (2006).Google Scholar
- [75] . 2014. BP4D-Spontaneous: A high-resolution spontaneous 3D dynamic facial expression database. Image Vis. Comput. 32, 10 (2014), 692–706.
DOI: Google ScholarCross Ref - [76] . 2015. Emotion distribution recognition from facial expressions. In 23rd ACM International Conference on Multimedia (MM’15). Association for Computing Machinery, New York, NY, 1247–1250.
DOI: Google ScholarDigital Library
Index Terms
- 3DEmo: For Portrait Emotion Recognition with New Dataset
Recommendations
A comparison of students' emotional self-reports with automated facial emotion recognition in a reading situation
TEEM'18: Proceedings of the Sixth International Conference on Technological Ecosystems for Enhancing MulticulturalityThis study investigated the measurement of students' emotional states during a common learning activity, digital reading of factual texts. The objective was to compare emotional self-reports with automated facial emotion recognition. The latter promises ...
FERMOUTH: Facial Emotion Recognition from the MOUTH Region
Image Analysis and Processing – ICIAP 2023AbstractPeople use various nonverbal communicative channels to convey emotions, among which facial expressions are considered the most important ones. Consequently, automatic Facial Expression Recognition (FER) is a crucial task for enhancing computers’ ...
Kriging Predictor for Facial Emotion Recognition Using Numerical Proximities of Human Emotions
Emotion recognition from facial expressions has gained much interest over the last few decades. In the literature, the common approach, used for facial emotion recognition (FER), consists of these steps: image pre-processing, face detection, facial ...
Comments