A voice-based real-time emotion detection technique using recurrent neural network empowered feature modelling

dc.contributor.authorChamishka, S
dc.contributor.authorMadhavi, I
dc.contributor.authorNawaratne, R
dc.contributor.authorAlahakoon, D
dc.contributor.authorDe Silva, D
dc.contributor.authorChilamkurti, N
dc.contributor.authorNanayakkara, V
dc.date.accessioned2023-06-21T08:02:58Z
dc.date.available2023-06-21T08:02:58Z
dc.date.issued2022
dc.description.abstractThe advancements of the Internet of Things (IoT) and voice-based multimedia applications have resulted in the generation of big data consisting of patterns, trends and associations capturing and representing many features of human behaviour. The latent representations of many aspects and the basis of human behaviour is naturally embedded within the expression of emotions found in human speech. This signifies the importance of mining audio data collected from human conversations for extracting human emotion. Ability to capture and represent human emotions will be an important feature in next-generation artificial intelligence, with the expectation of closer interaction with humans. Although the textual representations of human conversations have shown promising results for the extraction of emotions, the acoustic feature-based emotion detection from audio still lags behind in terms of accuracy. This paper proposes a novel approach for feature extraction consisting of Bag-of-Audio-Words (BoAW) based feature embeddings for conversational audio data. A Recurrent Neural Network (RNN) based state-of-the-art emotion detection model is proposed that captures the conversation-context and individual party states when making real-time categorical emotion predictions. The performance of the proposed approach and the model is evaluated using two benchmark datasets along with an empirical evaluation on real-time prediction capability. The proposed approach reported 60.87% weighted accuracy and 60.97% unweighted accuracy for six basic emotions for IEMOCAP dataset, significantly outperforming current state-of-the-art models.en_US
dc.identifier.citationChamishka, S., Madhavi, I., Nawaratne, R., Alahakoon, D., De Silva, D., Chilamkurti, N., & Nanayakkara, V. (2022). A voice-based real-time emotion detection technique using recurrent neural network empowered feature modelling. Multimedia Tools and Applications, 81(24), 35173–35194. https://doi.org/10.1007/s11042-022-13363-4en_US
dc.identifier.databaseSpringer Linken_US
dc.identifier.doi10.1007/s11042-022-13363-4en_US
dc.identifier.issn1573-7721en_US
dc.identifier.journalMultimedia Tools and Applicationsen_US
dc.identifier.pgnos35173–35194en_US
dc.identifier.urihttp://dl.lib.uom.lk/handle/123/21137
dc.identifier.volume81en_US
dc.identifier.year2022en_US
dc.language.isoen_USen_US
dc.publisherSpringer Netherlandsen_US
dc.subjectBag-of-audio-wordsen_US
dc.subjectMachine learningen_US
dc.subjectArtificial intelligenceen_US
dc.subjectBig data . Emotion analysisen_US
dc.titleA voice-based real-time emotion detection technique using recurrent neural network empowered feature modellingen_US
dc.typeArticle-Full-texten_US

Files