Skip to main content

2024 | OriginalPaper | Buchkapitel

Bi-Model Emotional AI for Audio-Visual Human Emotion Detection Using Hybrid Deep Learning Model

verfasst von : Anandakumar Haldorai, Babitha Lincy R, Suriya Murugan, Minu Balakrishnan

Erschienen in: Artificial Intelligence for Sustainable Development

Verlag: Springer Nature Switzerland

Aktivieren Sie unsere intelligente Suche, um passende Fachinhalte oder Patente zu finden.

search-config
loading …

Abstract

The present attention of computer vision study is on AI emotion identification, which comprises the automatic acknowledgment of facial terminologies of feeling and the evaluation of sentiment in visual database. In order for artificially intelligent systems with visual capabilities to comprehend human interactions, the study of human–machine interaction is essential. Artificial emotional intelligence, sometimes referred to as affective computing and emotional AI, is a subfield of artificial intelligence that concentrates on the comprehension, examination, and replication of human emotions. Its goal is to advance the sincerity and organic nature of interactions between people and robots. Textual content, voice tone, facial expressions, and gestures are just a few of the cues that emotional AI uses to understand people’s emotions and alter its answers accordingly. Using computer vision technology, Visual Emotion AI analyzes facial expressions in photos and videos to determine a person’s emotional state. This study uses both audio and visual inputs to investigate the recognition of emotions using artificial intelligence.

Sie haben noch keine Lizenz? Dann Informieren Sie sich jetzt über unsere Produkte:

Springer Professional "Wirtschaft+Technik"

Online-Abonnement

Mit Springer Professional "Wirtschaft+Technik" erhalten Sie Zugriff auf:

  • über 102.000 Bücher
  • über 537 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Maschinenbau + Werkstoffe
  • Versicherung + Risiko

Jetzt Wissensvorsprung sichern!

Springer Professional "Technik"

Online-Abonnement

Mit Springer Professional "Technik" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 390 Zeitschriften

aus folgenden Fachgebieten:

  • Automobil + Motoren
  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Elektrotechnik + Elektronik
  • Energie + Nachhaltigkeit
  • Maschinenbau + Werkstoffe




 

Jetzt Wissensvorsprung sichern!

Springer Professional "Wirtschaft"

Online-Abonnement

Mit Springer Professional "Wirtschaft" erhalten Sie Zugriff auf:

  • über 67.000 Bücher
  • über 340 Zeitschriften

aus folgenden Fachgebieten:

  • Bauwesen + Immobilien
  • Business IT + Informatik
  • Finance + Banking
  • Management + Führung
  • Marketing + Vertrieb
  • Versicherung + Risiko




Jetzt Wissensvorsprung sichern!

Literatur
2.
Zurück zum Zitat H. Kim, J. Ben-othman, L. Mokdad and P. Bellavista, "A Virtual Emotion Detection Architecture With Two-Way Enabled Delay Bound toward Evolutional Emotion-Based IoT Services," in IEEE Transactions on Mobile Computing, vol. 21, no. 4, pp. 1172–1181, 1 April 2022, doi: https://doi.org/10.1109/TMC.2020.3024059. H. Kim, J. Ben-othman, L. Mokdad and P. Bellavista, "A Virtual Emotion Detection Architecture With Two-Way Enabled Delay Bound toward Evolutional Emotion-Based IoT Services," in IEEE Transactions on Mobile Computing, vol. 21, no. 4, pp. 1172–1181, 1 April 2022, doi: https://​doi.​org/​10.​1109/​TMC.​2020.​3024059.
3.
5.
Zurück zum Zitat C. Suman, R. Chaudhari, S. Saha, S. Kumar and P. Bhattacharyya, "Investigations in Emotion Aware Multimodal Gender Prediction Systems From Social Media Data," in IEEE Transactions on Computational Social Systems, vol. 10, no. 2, pp. 470-479, April 2023, doi: https://doi.org/10.1109/TCSS.2022.3158605. C. Suman, R. Chaudhari, S. Saha, S. Kumar and P. Bhattacharyya, "Investigations in Emotion Aware Multimodal Gender Prediction Systems From Social Media Data," in IEEE Transactions on Computational Social Systems, vol. 10, no. 2, pp. 470-479, April 2023, doi: https://​doi.​org/​10.​1109/​TCSS.​2022.​3158605.
10.
Zurück zum Zitat S. Lee, S. Lee and H. Kim, "Differential Security Barriers for Virtual Emotion Detection in Maritime Transportation Stations With Cooperative Mobile Robots and UAVs," in IEEE Transactions on Intelligent Transportation Systems, vol. 24, no. 2, pp. 2461–2471, Feb. 2023, doi: https://doi.org/10.1109/TITS.2022.3172668. S. Lee, S. Lee and H. Kim, "Differential Security Barriers for Virtual Emotion Detection in Maritime Transportation Stations With Cooperative Mobile Robots and UAVs," in IEEE Transactions on Intelligent Transportation Systems, vol. 24, no. 2, pp. 2461–2471, Feb. 2023, doi: https://​doi.​org/​10.​1109/​TITS.​2022.​3172668.
11.
Zurück zum Zitat L. Chen, M. Zhou, W. Su, M. Wu, J. She, and K. Hirota, “Softmax regression based deep sparse autoencoder network for facial emotion recognition in human-robot interaction,” Information Sciences, vol. 428, pp. 49–61, 2018.MathSciNetCrossRef L. Chen, M. Zhou, W. Su, M. Wu, J. She, and K. Hirota, “Softmax regression based deep sparse autoencoder network for facial emotion recognition in human-robot interaction,” Information Sciences, vol. 428, pp. 49–61, 2018.MathSciNetCrossRef
12.
Zurück zum Zitat A. Hassouneh, A. M. Mutawa, and M. Murugappan, “Development of a real-time emotion recognition system using facial expressions and EEG based on machine learning and deep neural network methods,” Informatics in Medicine Unlocked, vol. 20, Article ID 100372, 2020. A. Hassouneh, A. M. Mutawa, and M. Murugappan, “Development of a real-time emotion recognition system using facial expressions and EEG based on machine learning and deep neural network methods,” Informatics in Medicine Unlocked, vol. 20, Article ID 100372, 2020.
13.
Zurück zum Zitat O. Ekundayo and S. Viriri, “Multilabel convolution neural network for facial expression recognition and ordinal intensity estimation,” PeerJ Computer Science, vol. 7, p. e736, 2021. O. Ekundayo and S. Viriri, “Multilabel convolution neural network for facial expression recognition and ordinal intensity estimation,” PeerJ Computer Science, vol. 7, p. e736, 2021.
14.
Zurück zum Zitat K. Jayanthi and S. Mohan, “An integrated framework for emotion recognition using speech and static images with deep classifier fusion approach,” International Journal of Information Technology, pp. 1–11, 2022. K. Jayanthi and S. Mohan, “An integrated framework for emotion recognition using speech and static images with deep classifier fusion approach,” International Journal of Information Technology, pp. 1–11, 2022.
15.
Zurück zum Zitat P. Babajee, G. Suddul, S. Armoogum, and R. Foogooa, “Identifying human emotions from facial expressions with deep learning,” in Proceedings of the 2020 Zooming Innovation in Consumer Technologies Conference (ZINC), pp. 36–39, IEEE, Novi Sad, Serbia, May 2020. P. Babajee, G. Suddul, S. Armoogum, and R. Foogooa, “Identifying human emotions from facial expressions with deep learning,” in Proceedings of the 2020 Zooming Innovation in Consumer Technologies Conference (ZINC), pp. 36–39, IEEE, Novi Sad, Serbia, May 2020.
16.
Zurück zum Zitat C. Tan, M. Šarlija, and N. Kasabov, “NeuroSense: short-term emotion recognition and understanding based on spiking neural network modelling of spatio-temporal EEG patterns,” Neurocomputing, vol. 434, pp. 137–148, 2021.CrossRef C. Tan, M. Šarlija, and N. Kasabov, “NeuroSense: short-term emotion recognition and understanding based on spiking neural network modelling of spatio-temporal EEG patterns,” Neurocomputing, vol. 434, pp. 137–148, 2021.CrossRef
19.
Zurück zum Zitat B. Jackson, S. R, B. Balachander, A. Haldorai, V. S., and A. N. A. Sahaya, “Modified Design Structure Of A Metamaterial Microstrip Patch Array Antenna For Rf Energy Optimization,” Materiali in tehnologije, vol. 56, no. 2, Apr. 2022, doi: 10.17222/mit.2022.384 B. Jackson, S. R, B. Balachander, A. Haldorai, V. S., and A. N. A. Sahaya, “Modified Design Structure Of A Metamaterial Microstrip Patch Array Antenna For Rf Energy Optimization,” Materiali in tehnologije, vol. 56, no. 2, Apr. 2022, doi: 10.17222/mit.2022.384
21.
Zurück zum Zitat A. Kumar, K. Abhishek, X. Liu, and A. Haldorai, “An Efficient Privacy-Preserving ID Centric Authentication in IoT Based Cloud Servers for Sustainable Smart Cities,” Wireless Personal Communications, vol. 117, no. 4, pp. 3229–3253, Nov. 2020, doi: https://doi.org/10.1007/s11277-020-07979-8. A. Kumar, K. Abhishek, X. Liu, and A. Haldorai, “An Efficient Privacy-Preserving ID Centric Authentication in IoT Based Cloud Servers for Sustainable Smart Cities,” Wireless Personal Communications, vol. 117, no. 4, pp. 3229–3253, Nov. 2020, doi: https://​doi.​org/​10.​1007/​s11277-020-07979-8.
24.
Zurück zum Zitat M. Bejani, D. Gharavian, and N. M. Charkari, “Audiovisual EmotionRecognition using ANOVA Feature Selection Method and Multi-ClassifierNeural Networks,” Neural Computing and Applications, vol. 24, no. 2,pp. 399–412, 2014.CrossRef M. Bejani, D. Gharavian, and N. M. Charkari, “Audiovisual EmotionRecognition using ANOVA Feature Selection Method and Multi-ClassifierNeural Networks,” Neural Computing and Applications, vol. 24, no. 2,pp. 399–412, 2014.CrossRef
25.
Zurück zum Zitat N. E. D. Elmadany, Y. He, and L. Guan, “Multiview Learning via DeepDiscriminative Canonical Correlation Analysis,” in IEEE InternationalConference on Acoustics, Speech and Signal Processing. IEEE, 2016,pp. 2409–2413. N. E. D. Elmadany, Y. He, and L. Guan, “Multiview Learning via DeepDiscriminative Canonical Correlation Analysis,” in IEEE InternationalConference on Acoustics, Speech and Signal Processing. IEEE, 2016,pp. 2409–2413.
26.
Zurück zum Zitat S. Zhang, S. Zhang, T. Huang, and W. Gao, “Multimodal DeepConvolutional Neural Network for Audio-Visual Emotion Recognition,“in ACM on International Conference on Multimedia Retrieval. ACM,2016, pp. 281–284. S. Zhang, S. Zhang, T. Huang, and W. Gao, “Multimodal DeepConvolutional Neural Network for Audio-Visual Emotion Recognition,“in ACM on International Conference on Multimedia Retrieval. ACM,2016, pp. 281–284.
28.
Zurück zum Zitat R. Gross, S. Baker, I. Matthews, and T. Kanade, “Face Recognition Across Pose and Illumination,” Handbook of Face Recognition, pp. 193–216, doi: 10.1007/0-387-27257-7_10. R. Gross, S. Baker, I. Matthews, and T. Kanade, “Face Recognition Across Pose and Illumination,” Handbook of Face Recognition, pp. 193–216, doi: 10.1007/0-387-27257-7_10.
29.
Zurück zum Zitat A. Dhall, R. Goecke, S. Lucey, and T. Gedeon, “Static facial expression analysis in tough conditions: Data, evaluation protocol and benchmark,” 2011 IEEE International Conference on Computer Vision Workshops (ICCV Workshops), Nov. 2011, doi: https://doi.org/10.1109/iccvw.2011.6130508. A. Dhall, R. Goecke, S. Lucey, and T. Gedeon, “Static facial expression analysis in tough conditions: Data, evaluation protocol and benchmark,” 2011 IEEE International Conference on Computer Vision Workshops (ICCV Workshops), Nov. 2011, doi: https://​doi.​org/​10.​1109/​iccvw.​2011.​6130508.
30.
Zurück zum Zitat P. Lucey, J. F. Cohn, T. Kanade, J. Saragih, Z. Ambadar, and I. Matthews, “The Extended Cohn-Kanade Dataset (CK+): A complete dataset for action unit and emotion-specified expression,” 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition – Workshops, Jun. 2010, doi: https://doi.org/10.1109/cvprw.2010.5543262 P. Lucey, J. F. Cohn, T. Kanade, J. Saragih, Z. Ambadar, and I. Matthews, “The Extended Cohn-Kanade Dataset (CK+): A complete dataset for action unit and emotion-specified expression,” 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition – Workshops, Jun. 2010, doi: https://​doi.​org/​10.​1109/​cvprw.​2010.​5543262
33.
Zurück zum Zitat Lijun Yin, Xiaozhou Wei, Yi Sun, Jun Wang, and M. J. Rosato, “A 3D Facial Expression Database For Facial Behavior Research,” 7th International Conference on Automatic Face and Gesture Recognition (FGR06), doi: 10.1109/fgr.2006.6. Lijun Yin, Xiaozhou Wei, Yi Sun, Jun Wang, and M. J. Rosato, “A 3D Facial Expression Database For Facial Behavior Research,” 7th International Conference on Automatic Face and Gesture Recognition (FGR06), doi: 10.1109/fgr.2006.6.
37.
38.
39.
Zurück zum Zitat A. V. Savchenko, L. V. Savchenko and I. Makarov, "Classifying Emotions and Engagement in Online Learning Based on a Single Facial Expression Recognition Neural Network," in IEEE Transactions on Affective Computing, vol. 13, no. 4, pp. 2132–2143, 1 Oct.-Dec. 2022, doi: https://doi.org/10.1109/TAFFC.2022.3188390. A. V. Savchenko, L. V. Savchenko and I. Makarov, "Classifying Emotions and Engagement in Online Learning Based on a Single Facial Expression Recognition Neural Network," in IEEE Transactions on Affective Computing, vol. 13, no. 4, pp. 2132–2143, 1 Oct.-Dec. 2022, doi: https://​doi.​org/​10.​1109/​TAFFC.​2022.​3188390.
41.
42.
Zurück zum Zitat B. García-Martínez, A. Fernández-Caballero, R. Alcaraz and A. Martínez-Rodrigo, "Application of Dispersion Entropy for the Detection of Emotions With Electroencephalographic Signals," in IEEE Transactions on Cognitive and Developmental Systems, vol. 14, no. 3, pp. 1179–1187, Sept. 2022, doi: https://doi.org/10.1109/TCDS.2021.3099344. B. García-Martínez, A. Fernández-Caballero, R. Alcaraz and A. Martínez-Rodrigo, "Application of Dispersion Entropy for the Detection of Emotions With Electroencephalographic Signals," in IEEE Transactions on Cognitive and Developmental Systems, vol. 14, no. 3, pp. 1179–1187, Sept. 2022, doi: https://​doi.​org/​10.​1109/​TCDS.​2021.​3099344.
Metadaten
Titel
Bi-Model Emotional AI for Audio-Visual Human Emotion Detection Using Hybrid Deep Learning Model
verfasst von
Anandakumar Haldorai
Babitha Lincy R
Suriya Murugan
Minu Balakrishnan
Copyright-Jahr
2024
DOI
https://doi.org/10.1007/978-3-031-53972-5_15

Neuer Inhalt