Automated emotion recognition: Current trends and future perspectives

M. Maithri, U. Raghavendra, Anjan Gudigar*, Jyothi Samanth, Datta Barua Prabal Datta Barua, Murugappan Murugappan, Yashas Chakole, U. Rajendra Acharya

*Corresponding author for this work

Research output: Contribution to journalReview articlepeer-review

122 Citations (Scopus)

Abstract

Background: Human emotions greatly affect the actions of a person. The automated emotion recognition has applications in multiple domains such as health care, e-learning, surveillance, etc. The development of computer-aided diagnosis (CAD) tools has led to the automated recognition of human emotions. Objective: This review paper provides an insight into various methods employed using electroencephalogram (EEG), facial, and speech signals coupled with multi-modal emotion recognition techniques. In this work, we have reviewed most of the state-of-the-art papers published on this topic. Method: This study was carried out by considering the various emotion recognition (ER) models proposed between 2016 and 2021. The papers were analysed based on methods employed, classifier used and performance obtained. Results: There is a significant rise in the application of deep learning techniques for ER. They have been widely applied for EEG, speech, facial expression, and multimodal features to develop an accurate ER model. Conclusion: Our study reveals that most of the proposed machine and deep learning-based systems have yielded good performances for automated ER in a controlled environment. However, there is a need to obtain high performance for ER even in an uncontrolled environment.

Original languageEnglish
Article number106646
JournalComputer Methods and Programs in Biomedicine
Volume215
DOIs
Publication statusPublished - 03-2022

All Science Journal Classification (ASJC) codes

  • Software
  • Computer Science Applications
  • Health Informatics

Fingerprint

Dive into the research topics of 'Automated emotion recognition: Current trends and future perspectives'. Together they form a unique fingerprint.

Cite this