[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
research-article

3DEmo: For Portrait Emotion Recognition with New Dataset

Published: 23 February 2024 Publication History

Abstract

Emotional Expression Recognition (EER) and Facial Expression Recognition (FER) are active research areas in the affective computing field, which involves studying human emotion, recognition, and sentiment analysis. The main objective of this research is to develop algorithms that can accurately interpret and estimate human emotions from portrait images. The emotions depicted in a portrait can reflect various factors such as psychological and physiological states, the artist’s emotional responses, social and environmental aspects, and the period in which the painting was created. This task is challenging because (i) the portraits are often depicted in an artistic or stylized manner rather than realistically or naturally, (ii) the texture and color features obtained from natural faces and paintings differ, affecting the success rate of emotion recognition algorithms, and (iii) it is a new research area, where practically we do not have visual arts portrait facial emotion estimation models or datasets.
To address these challenges, we need a new class of tools and a database specifically tailored to analyze portrait images. This study aims to develop art portrait emotion recognition methods and create a new digital portrait dataset containing 927 images. The proposed model is based on (i) a 3-dimensional estimation of emotions learned by a deep neural network and (ii) a novel deep learning module (3DEmo) that could be easily integrated into existing FER models. To evaluate the effectiveness of the developed models, we also tested their robustness on a facial emotion recognition dataset. The extensive simulation results show that the presented approach outperforms established methods. We expect that this dataset and the developed new tools will encourage further research in recognizing emotions in portrait paintings and predicting artists’ emotions in the painting period based on their artwork.

References

[1]
Faiza Abdat, Choubeila Maaoui, and Alain Pruski. 2011. Human-computer interaction using emotion recognition from facial expression. In 5th European Symposium on Computer Modeling and Simulation (UKSim’11). 196–201. DOI:
[2]
Panos Achlioptas, Maks Ovsjanikov, Kilichbek Haydarov, Mohamed Elhoseiny, and Leonidas Guibas. 2021. ArtEmis: Affective language for visual art. In Conference on Computer Vision and Pattern Recognition (CVPR’21). 11569–11579.
[3]
Iris Bakker, Theo van der Voordt, Peter Vink, and Jan de Boon. 2014. Pleasure, arousal, dominance: Mehrabian and Russell revisited. Curr. Psychol. 33 (2014), 405–421. DOI:
[4]
Lisa Feldman Barrett and James A. Russell. 1999. The structure of current affect: Controversies and emerging consensus. Curr. Direct. Psychol. Sci. 8, 1 (1999), 10–14. DOI:
[5]
Emad Barsoum, Cha Zhang, Cristian Canton Ferrer, and Zhengyou Zhang. 2016. Training deep networks for facial expression recognition with crowd-sourced label distribution. In ACM International Conference on Multimodal Interaction (ICMI’16).
[6]
Fabian C. Benitez-Quiroz, Ramprakash Srinivasan, and Aleix M. Martinez. 2016. EmotioNet: An accurate, real-time algorithm for the automatic annotation of a million facial expressions in the wild. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR’16). 5562–5570. DOI:
[7]
James C. Bezdek, Robert Ehrlich, and William Full. 1984. FCM: The fuzzy c-means clustering algorithm. Comput. Geosci. 10, 2 (1984), 191–203. DOI:
[8]
Ran Breuer and Ron Kimmel. 2017. A Deep Learning Perspective on the Origin of Facial Expressions. DOI:
[9]
Jean-Peic Chou and David G. Stork. 2023. Computational tracking of head pose through 500 years of fine-art portraiture. In Computer Vision and Analysis of Art. SPIE Electronic Imaging, San Francisco, CA.
[10]
Tu N. Chowdary, M. Kalpana. Nguyen and D. Jude Hemanth. 2021. Deep learning-based facial emotion recognition for human–computer interaction applications. Neural Comput. Applic. DOI:
[11]
Wen-Sheng Chu, Fernando Torre, and Jeffrey Cohn. 2017. Learning spatial and temporal cues for multi-label facial action unit detection. 25–32. DOI:
[12]
Sarah Cosentino, Estelle I. S. Randria, Jia-Yeu Lin, Thomas Pellegrini, Salvatore Sessa, and Atsuo Takanishi. 2018. Group emotion recognition strategies for entertainment robots. In IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS’18). 813–818. DOI:
[13]
Abhinav Dhall, Garima Sharma, Roland Goecke, and Tom Gedeon. 2020. EmotiW 2020: Driver Gaze, Group Emotion, Student Engagement and Physiological Signal Based Challenges. Association for Computing Machinery, New York, NY, 784–789. Retrieved from
[14]
Shichuan Du, Yong Tao, and Aleix M. Martinez. 2014. Compound facial expressions of emotion. Proc. Nat. Acad. Sci. 111, 15 (2014), E1454–E1462. DOI:
[15]
Samira Ebrahimi Kahou, Vincent Michalski, Kishore Konda, Roland Memisevic, and Christopher Pal. 2015. Recurrent neural networks for emotion recognition in video. InACM International Conference on Multimodal Interaction (ICMI’15). Association for Computing Machinery, New York, NY, 467–474. DOI:
[16]
Paul. Ekman and Wallace V. Friesen. 1971. Constants Across Cultures in the Face and Emotion. 124–129 pages. Retrieved from
[17]
Olufisayo Ekundayo and Serestina Viriri. 2020. Facial expression recognition and ordinal intensity estimation: A multilabel learning approach. In Advances in Visual Computing, George Bebis, Zhaozheng Yin, Edward Kim, Jan Bender, Kartic Subr, Bum Chul Kwon, Jian Zhao, Denis Kalkofen, and George Baciu (Eds.). Springer International Publishing, Cham, 581–592.
[18]
Olufisayo Ekundayo and Serestina Viriri. 2021. Facial expression recognition: A review of trends and techniques. IEEE Access PP (092021), 1–1. DOI:
[19]
Yin Fan, Xiangju Lu, Dian Li, and Yuanliu Liu. 2016. Video-based emotion recognition using CNN-RNN and C3D hybrid networks. In 18th ACM International Conference on Multimodal Interaction (ICMI’16). Association for Computing Machinery, New York, NY, 445–450. DOI:
[20]
Beat Fasel. 2002. Head-pose invariant facial expression recognition using convolutional neural networks. In 4th IEEE International Conference on Multimodal Interfaces. 529–534. DOI:
[21]
Beat Fasel. 2002. Robust face analysis using convolutional neural networks. In International Conference on Pattern Recognition, Vol. 2. 40–43 vol.2. DOI:
[22]
Chris N. W. Geraets, Stéphanie Klein Tuente, Bart P. Lestestuiver, Marije van Beilen, Saskia A. Nijman, Jan-Bernard C. Marsman, and Wim Veling. 2021. Virtual reality facial emotion recognition in social environments: An eye-tracking study. Internet Intervent. 25 (2021), 100432. DOI:
[23]
Charline Grossard, Laurence Chaby, Stéphanie Hun, Hugues Pellerin, Jérémy Bourgeois, Arnaud Dapogny, Huaxiong Ding, Sylvie Serret, Pierre Foulon, Mohamed Chetouani, Liming Chen, Kevin Bailly, Ouriel Grynszpan, and David Cohen. 2018. Children facial expression production: Influence of age, gender, emotion subtype, elicitation condition and culture. Front. Psychol. 9 (2018). DOI:
[24]
Jad El Haddad, Olivier Lézoray, and Philippe Hamel. 2020. 3D-CNN for facial emotion recognition in videos. In International Symposium on Visual Computing (ISVC’20).
[25]
S. L. Happy, Anjith George, and Aurobinda Routray. 2012. A real time facial expression classification system using local binary patterns. In 4th International Conference on Intelligent Human Computer Interaction (IHCI’12). 1–5. DOI:
[26]
Behzad Hasani and Mohammad H. Mahoor. 2017. Facial expression recognition using enhanced deep 3D convolutional neural networks. In IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW’17). IEEE. DOI:
[27]
Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. 2016. Deep residual learning for image recognition. In IEEE Conference on Computer Vision and Pattern Recognition. 770–778.
[28]
Deepak Kumar Jain, Zhang Zhang, and Kaiqi Huang. 2020. Multi angle optimal pattern-based deep learning for automatic facial expression recognition. Pattern Recog. Lett. 139 (2020), 157–165. DOI:
[29]
Mira Jeong and Byoung Chul Ko. 2018. Driver’s facial expression recognition in real-time for safe driving. Sensors (Basel, Switz.) (Dec.2018), 12 4270. DOI:
[30]
Shuiwang Ji, Wei Xu, Ming Yang, and Kai Yu. 2013. 3D convolutional neural networks for human action recognition. IEEE Trans. Pattern Anal. Mach. Intell. 35, 1 (2013), 221–231. DOI:
[31]
Dae Hoe Kim, Wissam J. Baddar, Jinhyeok Jang, and Yong Man Ro. 2019. Multi-objective based spatio-temporal feature representation learning robust to expression intensity variations for facial expression recognition. IEEE Trans. Affect. Comput. 10 (2019), 223–236.
[32]
Kwang Hyeon Kim et al. 2020. Facial expression monitoring system for predicting patient’s sudden movement during radiotherapy using deep learning. J. Appl. Clin. Med. Phys. 21, 191–199. DOI:
[33]
Diederik P. Kingma and Jimmy Ba. 2017. Adam: A Method for Stochastic Optimization. arxiv:1412.6980 [cs.LG]
[34]
Jean Kossaifi, Robert Walecki, Yannis Panagakis, Jie Shen, Maximilian Schmitt, Fabien Ringeval, Jing Han, Vedhas Pandit, Antoine Toisoul, Bjorn Schuller, Kam Star, Elnar Hajiyev, and Maja Pantic. 2021. SEWA DB: A rich database for audio-visual emotion and sentiment research in the wild. IEEE Trans. Pattern Anal. Mach. Intell. 43, 3 (Mar.2021), 1022–1040. DOI:
[35]
Ka Keung Lee and Yangsheng Xu. 2003. Real-time estimation of facial expression intensity. In IEEE International Conference on Robotics and Automation. 2567–2572. DOI:
[36]
Suhua Lei, Huan Zhang, Ke Wang, and Zhendong Su. 2018. How training data affect the accuracy and robustness of neural networks for image classification.
[37]
Shan Li and Weihong Deng. 2020. Deep facial expression recognition: A survey. IEEE Trans. Affect. Comput. (2020), 1–1. DOI:
[38]
Shao Liu, Jiaqi Yang, Sos S. Agaian, and Changhe Yuan. 2021. Novel features for art movement classification of portrait paintings. Image Vis. Comput. 108 (2021), 104121. DOI:
[39]
Stuart P. Lloyd. 1982. Least squares quantization in PCM. IEEE Trans. Inf. Theor. 28, 2 (1982), 129–137. DOI:
[40]
Patrick Lucey, Jeffrey F. Cohn, Takeo Kanade, Jason Saragih, Zara Ambadar, and Iain Matthews. 2010. The extended Cohn-Kanade dataset (CK+): A complete dataset for action unit and emotion-specified expression. In IEEE Computer Society Conference on Computer Vision and Pattern Recognition—Workshops. 94–101. DOI:
[41]
Micheal Lyons, Shigeru Akamatsu, Miyuki Kamachi, and Jiro Gyoba. 1998. Coding facial expressions with Gabor wavelets. In 3rd IEEE International Conference on Automatic Face and Gesture Recognition. 200–205. DOI:
[42]
Michael Lyons, Miyuki Kamachi, and Jiro Gyoba. 1998. The Japanese Female Facial Expression (JAFFE) Dataset. DOI:
[43]
Giacomo Mancini, Roberta Biolcati, Sergio Agnoli, Federica Andrei, and Elena Trombini. 2018. Recognition of facial emotional expressions among Italian pre-adolescents, and their affective reactions. Front. Psychol. 9 (2018). DOI:
[44]
Mohammad S. Mavadati, Mohammad H. Mahoor, Kevin Bartlett, Philip Trinh, and Jeffrey F. Cohn. 2013. DISFA: A spontaneous facial action intensity database. IEEE Trans. Affect. Comput. 4, 2 (2013), 151–160. DOI:
[45]
Dhwani Mehta, Mohammad Faridul Haque Siddiqui, and Ahmad Y. Javaid. 2018. Facial emotion recognition: A survey and real-world user experiences in mixed reality. Sensors (Basel, Switz.) (Feb.2018). DOI:
[46]
Federico Milani and Piero Fraternali. 2021. A dataset and a convolutional model for iconography classification in paintings. J. Comput. Cult. Herit. 14, 4, Article 46 (July2021), 18 pages. DOI:
[47]
Ali Mollahosseini, Behzad Hasani, and Mohammad H. Mahoor. 2019. AffectNet: A database for facial expression, valence, and arousal computing in the wild. IEEE Trans. Affect. Comput. 10, 1 (Jan.2019), 18–31. DOI:
[48]
Dung Nguyen, Kien Nguyen, Sridha Sridharan, Afsane Ghasemi, David Dean, and Clinton Fookes. 2017. Deep spatio-temporal features for multimodal emotion recognition. In IEEE Winter Conference on Applications of Computer Vision (WACV’17). 1215–1223. DOI:
[49]
Hiroki Nomiya, Shota Sakaue, and Teruhisa Hochin. 2016. Recognition and intensity estimation of facial expression using ensemble classifiers. In IEEE/ACIS 15th International Conference on Computer and Information Science (ICIS’16). 1–6. DOI:
[50]
Maja Pantic, Michel Valstar, Ron Rademaker, and Ludo Maat. 2005. Web-based database for facial expression analysis. In IEEE International Conference on Multimedia and Expo. DOI:
[51]
Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, Alban Desmaison, Andreas Kopf, Edward Yang, Zachary DeVito, Martin Raison, Alykhan Tejani, Sasank Chilamkurthy, Benoit Steiner, Lu Fang, Junjie Bai, and Soumith Chintala. 2019. PyTorch: An imperative style, high-performance deep learning library. In Advances in Neural Information Processing Systems, Vol. 32, H. Wallach, H. Larochelle, A. Beygelzimer, F. d’Alché-Buc, E. Fox, and R. Garnett (Eds.). Curran Associates, Inc., 8024–8035. Retrieved from http://papers.neurips.cc/paper/9015-pytorch-an-imperative-style-high-performance-deep-learning-library.pdf
[52]
Hui Qu, Yikai Zhang, Qi Chang, Zhennan Yan, Chao Chen, and Dimitris Metaxas. 2020. Learn Distributed GAN with Temporary Discriminators.
[53]
Sebastian Ruder. 2016. An overview of gradient descent optimization algorithms. arXiv preprint arXiv:1609.04747 (2016).
[54]
G.-Fivos Sargentis, Panayiotis Dimitriadis, Theano Iliopoulou, and Demetris Koutsoyiannis. 2021. A stochastic view of varying styles in art paintings. Heritage 4, 1 (2021), 333–348. DOI:
[55]
Lior Shamir and Jane A. Tarakhovsky. 2012. Computer analysis of art. J. Comput. Cult. Herit. 5, 2, Article 7 (Aug.2012), 11 pages. DOI:
[56]
Li Shang and Weihong Deng. 2018. Blended emotion in-the-wild: Multi-label facial expression recognition using crowdsourced annotations and deep locality feature learning. Int. J. Comput. Vis. 127 (2018), 884–906.
[57]
Vera Shuman, David Sander, and Klaus Scherer. 2013. Levels of valence. Front. Psychol. 4 (2013). DOI:
[58]
Henrique Siqueira, Sven Magg, and Stefan Wermter. 2020. Efficient Facial Feature Learning with Wide Ensemble-based Convolutional Neural Networks. Retrieved from https://www2.informatik.uni-hamburg.de/wtm/publications/2020/SMW20/SMW20.pdf
[59]
Sujono and Alexander A. S. Gunawan. 2015. Face expression detection on Kinect using active appearance model and fuzzy logic. Procedia Comput. Sci. 59 (2015), 268–274. DOI:
[60]
Myunghoon Suk and Balakrishnan Prabhakaran. 2014. Real-time mobile facial expression recognition system—A case study. In IEEE Conference on Computer Vision and Pattern Recognition Workshops. 132–137. DOI:
[61]
Mariusz Szwoch and Pawel Pieniazek. 2015. Facial emotion recognition using depth data. In 8th International Conference on Human System Interaction (HSI’15). 271–277. DOI:
[62]
Yingtao Tian, Chikahiko Suzuki, Tarin Clanuwat, Mikel Bober-Irizar, Alex Lamb, and Asanobu Kitamoto. 2020. KaoKore: A Pre-modern Japanese Art Facial Expression Dataset. arxiv:2002.08595 [cs.CV]
[63]
Antoine Toisoul, Jean Kossaifi, Adrian Bulat, Georgios Tzimiropoulos, and Maja Pantic. 2021. Estimation of continuous valence and arousal levels from faces in naturalistic conditions. Nat. Mach. Intell. 3 (2021), 42–50. DOI:
[64]
Güray Tonguç and Betul Ozaydın Ozkara. 2020. Automatic recognition of student emotions from facial expressions during a lecture. Comput. Educ. 148 (2020), 103797. DOI:
[65]
Ekaterina Volkova, Stephan de la Rosa, Heinrich H. Bülthoff, and Betty Mohler. 2014. The MPI emotional body expressions database for narrative scenarios. PloS One 9, 12 (2014), e113647. DOI:
[66]
Kai Wang, Xiaojiang Peng, Jianfei Yang, Shijian Lu, and Yu Qiao. 2020. Suppressing Uncertainties for Large-Scale Facial Expression Recognition. DOI:
[67]
Torsten Wilhelm. 2019. Towards facial expression analysis in a driver assistance system. In 14th IEEE International Conference on Automatic Face Gesture Recognition (FG’19). 1–4. DOI:
[68]
Xugang Xi, Yan Zhang, Xian Hua, Seyed M. Miran, Yun-Bo Zhao, and Zhizeng Luo. 2020. Facial expression distribution prediction based on surface electromyography. Expert Syst. Applic. 161 (2020), 113683. DOI:
[69]
Huafei Xiao, Wenbo Li, Guanzhong Zeng, Yingzhang Wu, Jiyong Xue, Juncheng Zhang, Chengmou Li, and Gang Guo. 2022. On-road driver emotion recognition using facial expression. Appl. Sci. 12, 2 (2022). DOI:
[70]
Siyue Xie, Haifeng Hu, and Yizhen Chen. 2020. Facial expression recognition with two-branch disentangled generative adversarial network. IEEE Trans. Circ. Syst. Vid. Technol. (2020).
[71]
Wen-Jing Yan, Xiaobai Li, Su-Jing Wang, Guoying Zhao, Yong-Jin Liu, Yu-Hsin Chen, and Xiaolan Fu. 2014. CASME II: An improved spontaneous micro-expression database and the baseline evaluation. PLoS One 9 (012014), 1–8. DOI:
[72]
Yao-Yuan Yang, Cyrus Rashtchian, Hongyang Zhang, Russ R. Salakhutdinov, and Kamalika Chaudhuri. 2020. A closer look at accuracy vs. robustness. In Advances in Neural Information Processing Systems, H. Larochelle, M. Ranzato, R. Hadsell, M. F. Balcan, and H. Lin (Eds.), Vol. 33. Curran Associates, Inc., 8588–8601. Retrieved from https://proceedings.neurips.cc/paper/2020/file/61d77652c97ef636343742fc3dcf3ba9-Paper.pdf
[73]
Jordan Yaniv, Yael Newman, and Ariel Shamir. 2019. The face of art: Landmark detection and geometric style in portraits. ACM Trans. Graph. 38, 4, Article 60 (July2019), 15 pages. DOI:
[74]
Lijun Yin, Xiaozhou Wei, Yi Sun, Jun Wang, and Matthew J. Rosato. 2006. A 3D facial expression database for facial behavior research. (2006).
[75]
Xing Zhang, Lijun Yin, Jeffrey F. Cohn, Shaun Canavan, Michael Reale, Andy Horowitz, Peng Liu, and Jeffrey M. Girard. 2014. BP4D-Spontaneous: A high-resolution spontaneous 3D dynamic facial expression database. Image Vis. Comput. 32, 10 (2014), 692–706. DOI:
[76]
Ying Zhou, Hui Xue, and Xin Geng. 2015. Emotion distribution recognition from facial expressions. In 23rd ACM International Conference on Multimedia (MM’15). Association for Computing Machinery, New York, NY, 1247–1250. DOI:

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image Journal on Computing and Cultural Heritage
Journal on Computing and Cultural Heritage   Volume 17, Issue 2
June 2024
355 pages
EISSN:1556-4711
DOI:10.1145/3613557
Issue’s Table of Contents

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 23 February 2024
Online AM: 02 November 2023
Accepted: 28 August 2023
Revised: 10 July 2023
Received: 09 April 2023
Published in JOCCH Volume 17, Issue 2

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. Facial emotion recognition
  2. 3D emotion estimation

Qualifiers

  • Research-article

Funding Sources

  • Art Science Connect, Graduate Center, CUNY

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • 0
    Total Citations
  • 357
    Total Downloads
  • Downloads (Last 12 months)309
  • Downloads (Last 6 weeks)14
Reflects downloads up to 13 Dec 2024

Other Metrics

Citations

View Options

Login options

Full Access

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Full Text

View this article in Full Text.

Full Text

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media