Hybrid fusion of face and speech information for bimodal emotion estimation

Estimation of human emotions during a conversation is difficult using a computer. In this study, facial expressions and speech are used in order to estimate emotions (angry, sad, happy, boredom, disgust and surprise). A proposed hybrid system through facial expressions and speech is used to estimate...

Full description

Saved in:
Bibliographic Details
Main Authors: Mohan Kudiri, K., Md. Said, A., Nayan, M.Y.
Format: Article
Published: Penerbit UTM Press 2016
Online Access:https://www.scopus.com/inward/record.uri?eid=2-s2.0-84988353604&doi=10.11113%2fjt.v78.9538&partnerID=40&md5=79e645383e5248b7358a33cf8bc1bc66
http://eprints.utp.edu.my/25867/
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Universiti Teknologi Petronas
Description
Summary:Estimation of human emotions during a conversation is difficult using a computer. In this study, facial expressions and speech are used in order to estimate emotions (angry, sad, happy, boredom, disgust and surprise). A proposed hybrid system through facial expressions and speech is used to estimate emotions of a person when he is engaged in a conversational session. Relative Bin Frequency Coefficients and Relative Sub-Image-Based features are used for acoustic and visual modalities respectively. Support Vector Machine is used for classification. This study shows that the proposed feature extraction through acoustic and visual data is the most prominent aspect affecting the emotion detection system, along with the proposed fusion technique. Although some other aspects are considered to be affecting the system, the effect is relatively minor. It was observed that the performance of the bimodal system was lower than the unimodal system through deliberate facial expressions. In order to deal with the problem, a suitable database is used. The results indicate that the proposed system showed better performance, with respect to basic emotional classes than the rest. © 2016 Penerbit UTM Press. All rights reserved.