Mood recognition through facial expression and speech cues

Emotion interpretation is important to allow one to comprehend his or her environment. Knowing the emotional state of the surroundings can influence the decision-making process. Previous work has shown that both Facial Expressions and Vocal Data can be used to determine the emotions of a person. The...

Full description

Saved in:
Bibliographic Details
Main Author: Tan, Jonathan Tian-Ci
Other Authors: Seet Gim Lee, Gerald
Format: Final Year Project
Language:English
Published: 2014
Subjects:
Online Access:http://hdl.handle.net/10356/61007
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-61007
record_format dspace
spelling sg-ntu-dr.10356-610072023-03-04T18:40:39Z Mood recognition through facial expression and speech cues Tan, Jonathan Tian-Ci Seet Gim Lee, Gerald School of Mechanical and Aerospace Engineering DRNTU::Engineering::Mechanical engineering Emotion interpretation is important to allow one to comprehend his or her environment. Knowing the emotional state of the surroundings can influence the decision-making process. Previous work has shown that both Facial Expressions and Vocal Data can be used to determine the emotions of a person. The systems relied on a unimodal approach, which only focused on one aspect. However, the accuracy for emotion recognition was low, and there was a vital reliance on specific hardware. This project presents a multimodal approach for the recognition of five different emotions (Anger, Fear, Neutral, Happy, and Sad) that integrates the information from facial expressions and speech. Various studies showed that a combination of modes would improve the overall accuracy in classifying emotions, than compared to individual modes. Audio-Visual data was collected from six participants, in a controlled environment. The participants were given a fixed script to portray their emotions, and hence the data collected was of a “posed” emotion rather than a spontaneous one. The features (Facial Expressions, Voice Features) were manually extracted to facilitate Supervised Machine Learning. Various methods such as Principal Component Analysis and Support Vector Machine were used to classify the emotions. The project was coded using both C++ and MATLAB, and a working real time MATLAB program was implemented on both the Windows OS and Ubuntu OS. Through training and optimization, the speech features produced a result of 94.3% whereas the facial expressions produced a perfect score of 100% accuracy. The combined data produced an emotion classifier with 100% accuracy. The emotion classifiers were Speaker-Dependant, where the training and evaluation data collected was from the same group of speakers. The findings for the project suggest that a multimodal system does not necessarily provide results that are more accurate. Alternatively, more channels of feedback within the same mode would provide a more accurate model to distinguish emotions. Bachelor of Engineering (Mechanical Engineering) 2014-06-04T02:12:34Z 2014-06-04T02:12:34Z 2014 2014 Final Year Project (FYP) http://hdl.handle.net/10356/61007 en Nanyang Technological University 52 p. application/pdf
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic DRNTU::Engineering::Mechanical engineering
spellingShingle DRNTU::Engineering::Mechanical engineering
Tan, Jonathan Tian-Ci
Mood recognition through facial expression and speech cues
description Emotion interpretation is important to allow one to comprehend his or her environment. Knowing the emotional state of the surroundings can influence the decision-making process. Previous work has shown that both Facial Expressions and Vocal Data can be used to determine the emotions of a person. The systems relied on a unimodal approach, which only focused on one aspect. However, the accuracy for emotion recognition was low, and there was a vital reliance on specific hardware. This project presents a multimodal approach for the recognition of five different emotions (Anger, Fear, Neutral, Happy, and Sad) that integrates the information from facial expressions and speech. Various studies showed that a combination of modes would improve the overall accuracy in classifying emotions, than compared to individual modes. Audio-Visual data was collected from six participants, in a controlled environment. The participants were given a fixed script to portray their emotions, and hence the data collected was of a “posed” emotion rather than a spontaneous one. The features (Facial Expressions, Voice Features) were manually extracted to facilitate Supervised Machine Learning. Various methods such as Principal Component Analysis and Support Vector Machine were used to classify the emotions. The project was coded using both C++ and MATLAB, and a working real time MATLAB program was implemented on both the Windows OS and Ubuntu OS. Through training and optimization, the speech features produced a result of 94.3% whereas the facial expressions produced a perfect score of 100% accuracy. The combined data produced an emotion classifier with 100% accuracy. The emotion classifiers were Speaker-Dependant, where the training and evaluation data collected was from the same group of speakers. The findings for the project suggest that a multimodal system does not necessarily provide results that are more accurate. Alternatively, more channels of feedback within the same mode would provide a more accurate model to distinguish emotions.
author2 Seet Gim Lee, Gerald
author_facet Seet Gim Lee, Gerald
Tan, Jonathan Tian-Ci
format Final Year Project
author Tan, Jonathan Tian-Ci
author_sort Tan, Jonathan Tian-Ci
title Mood recognition through facial expression and speech cues
title_short Mood recognition through facial expression and speech cues
title_full Mood recognition through facial expression and speech cues
title_fullStr Mood recognition through facial expression and speech cues
title_full_unstemmed Mood recognition through facial expression and speech cues
title_sort mood recognition through facial expression and speech cues
publishDate 2014
url http://hdl.handle.net/10356/61007
_version_ 1759856655584985088