Multimodal emotion recognition system for spontaneous vocal and facial signals: SMERFS
Human computer interaction is moving towards giving computers the ability to adapt and give feedback in accordance to a user's emotion. Initial researches on multimodal emotion recognition shows that combining both vocal and facial signals performed better compared to using physiological signal...
Saved in:
Main Authors: | , , , |
---|---|
Format: | text |
Language: | English |
Published: |
Animo Repository
2010
|
Subjects: | |
Online Access: | https://animorepository.dlsu.edu.ph/etd_bachelors/14653 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | De La Salle University |
Language: | English |
id |
oai:animorepository.dlsu.edu.ph:etd_bachelors-15295 |
---|---|
record_format |
eprints |
spelling |
oai:animorepository.dlsu.edu.ph:etd_bachelors-152952021-11-13T05:40:27Z Multimodal emotion recognition system for spontaneous vocal and facial signals: SMERFS Dy, Marc Lanze Ivan C. Espinoza, Ivan Vener L. Go, Paul Patrick V. Mendez, Charles Martin M. Human computer interaction is moving towards giving computers the ability to adapt and give feedback in accordance to a user's emotion. Initial researches on multimodal emotion recognition shows that combining both vocal and facial signals performed better compared to using physiological signals. In addition, majority of the emotion corpus used on both unimodal and multimodal systems were modeled based on acted data using actors that tend to exaggerate emotions. This study improves the accuracy of single modality systems by developing a multimodal emotion recognition system through vocal and facial expressions using a spontaneous emotion corpus. FilMED2, which contains spontaneous television clips from reality television shows, is the corpus used in this study. The clips contain discrete emotion labels where they are only labeled as happiness, sadness, anger, fear and neutral. The system makes use of the facial feature points and prosodic features which include pitch and energy that will undergo machine learning for classification. SVM is the machine learning technique used for classification and was first tested on each modality for both acted and spontaneous corpus. The acted corpus yielded higher results as compared to when using the spontaneous corpus for both modalities. Both modalities were then combined using decision-level fusion. Using solely the face gave 60% accuracy while using solely the voice gave 32% accuracy. Combining both results with a weight-distribution of 75% face and 25% voice gave an accuracy rate of 80%. 2010-01-01T08:00:00Z text https://animorepository.dlsu.edu.ph/etd_bachelors/14653 Bachelor's Theses English Animo Repository Human-computer interaction Pattern recognition systems Computer vision Artificial intelligence |
institution |
De La Salle University |
building |
De La Salle University Library |
continent |
Asia |
country |
Philippines Philippines |
content_provider |
De La Salle University Library |
collection |
DLSU Institutional Repository |
language |
English |
topic |
Human-computer interaction Pattern recognition systems Computer vision Artificial intelligence |
spellingShingle |
Human-computer interaction Pattern recognition systems Computer vision Artificial intelligence Dy, Marc Lanze Ivan C. Espinoza, Ivan Vener L. Go, Paul Patrick V. Mendez, Charles Martin M. Multimodal emotion recognition system for spontaneous vocal and facial signals: SMERFS |
description |
Human computer interaction is moving towards giving computers the ability to adapt and give feedback in accordance to a user's emotion. Initial researches on multimodal emotion recognition shows that combining both vocal and facial signals performed better compared to using physiological signals. In addition, majority of the emotion corpus used on both unimodal and multimodal systems were modeled based on acted data using actors that tend to exaggerate emotions. This study improves the accuracy of single modality systems by developing a multimodal emotion recognition system through vocal and facial expressions using a spontaneous emotion corpus. FilMED2, which contains spontaneous television clips from reality television shows, is the corpus used in this study. The clips contain discrete emotion labels where they are only labeled as happiness, sadness, anger, fear and neutral. The system makes use of the facial feature points and prosodic features which include pitch and energy that will undergo machine learning for classification. SVM is the machine learning technique used for classification and was first tested on each modality for both acted and spontaneous corpus. The acted corpus yielded higher results as compared to when using the spontaneous corpus for both modalities. Both modalities were then combined using decision-level fusion. Using solely the face gave 60% accuracy while using solely the voice gave 32% accuracy. Combining both results with a weight-distribution of 75% face and 25% voice gave an accuracy rate of 80%. |
format |
text |
author |
Dy, Marc Lanze Ivan C. Espinoza, Ivan Vener L. Go, Paul Patrick V. Mendez, Charles Martin M. |
author_facet |
Dy, Marc Lanze Ivan C. Espinoza, Ivan Vener L. Go, Paul Patrick V. Mendez, Charles Martin M. |
author_sort |
Dy, Marc Lanze Ivan C. |
title |
Multimodal emotion recognition system for spontaneous vocal and facial signals: SMERFS |
title_short |
Multimodal emotion recognition system for spontaneous vocal and facial signals: SMERFS |
title_full |
Multimodal emotion recognition system for spontaneous vocal and facial signals: SMERFS |
title_fullStr |
Multimodal emotion recognition system for spontaneous vocal and facial signals: SMERFS |
title_full_unstemmed |
Multimodal emotion recognition system for spontaneous vocal and facial signals: SMERFS |
title_sort |
multimodal emotion recognition system for spontaneous vocal and facial signals: smerfs |
publisher |
Animo Repository |
publishDate |
2010 |
url |
https://animorepository.dlsu.edu.ph/etd_bachelors/14653 |
_version_ |
1718382643566870528 |