Mood recognition through vocal prosody recognition

Mood recognition through vocal prosody recognition is designed to predict human‟s mood through speeches profiles. There are existing applications in vocal prosody such as Microsoft‟s “Speech to Text”, IOS‟s “Siri” and Andriod‟s “S Voice” that they are executing actions which are ordered by users but...

Full description

Saved in:
Bibliographic Details
Main Author: Wong, Yi Ben.
Other Authors: Seet Gim Lee, Gerald
Format: Final Year Project
Language:English
Published: 2013
Subjects:
Online Access:http://hdl.handle.net/10356/53290
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-53290
record_format dspace
spelling sg-ntu-dr.10356-532902023-03-04T19:02:58Z Mood recognition through vocal prosody recognition Wong, Yi Ben. Seet Gim Lee, Gerald School of Mechanical and Aerospace Engineering Robotics Research Centre DRNTU::Engineering::Mechanical engineering::Mechatronics Mood recognition through vocal prosody recognition is designed to predict human‟s mood through speeches profiles. There are existing applications in vocal prosody such as Microsoft‟s “Speech to Text”, IOS‟s “Siri” and Andriod‟s “S Voice” that they are executing actions which are ordered by users but they are not relate to mood recognition function. Therefore, this project seeks to develop a software package in mood recognition through human‟s speeches. Speaker-Dependent and Speaker- Independent mode were investigated to develop Real-Time Emotion Recognition System. Speeches database were collected and studied to improve the emotion recognition system since speeches database is one of the factors to define the quality of emotion recognition model. Besides, the process of handling speeches database was proposed to improve the accuracy and several experiments were completed for the improvement. Speeches database was reviewed by other users to prove the quality of speeches in term of expressing moods. Experimental results had shown that the Speaker-Dependent mode provides higher accuracy than Speaker-Independent mode as similar researches are found to support the findings. Besides, the number of emotion used in emotion recognition system does affect the accuracy in recognizing mood through speeches. Emotion- basis data division was found to have better accuracy instead of using Speaker- basis data division during the process of handling speeches database to train emotion recognition model. The human recognition on speeches database had shown it's less accurate to predict others‟ emotions under cross cultural background. Bachelor of Engineering (Mechanical Engineering) 2013-05-31T04:01:49Z 2013-05-31T04:01:49Z 2013 2013 Final Year Project (FYP) http://hdl.handle.net/10356/53290 en Nanyang Technological University 51 p. application/pdf
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic DRNTU::Engineering::Mechanical engineering::Mechatronics
spellingShingle DRNTU::Engineering::Mechanical engineering::Mechatronics
Wong, Yi Ben.
Mood recognition through vocal prosody recognition
description Mood recognition through vocal prosody recognition is designed to predict human‟s mood through speeches profiles. There are existing applications in vocal prosody such as Microsoft‟s “Speech to Text”, IOS‟s “Siri” and Andriod‟s “S Voice” that they are executing actions which are ordered by users but they are not relate to mood recognition function. Therefore, this project seeks to develop a software package in mood recognition through human‟s speeches. Speaker-Dependent and Speaker- Independent mode were investigated to develop Real-Time Emotion Recognition System. Speeches database were collected and studied to improve the emotion recognition system since speeches database is one of the factors to define the quality of emotion recognition model. Besides, the process of handling speeches database was proposed to improve the accuracy and several experiments were completed for the improvement. Speeches database was reviewed by other users to prove the quality of speeches in term of expressing moods. Experimental results had shown that the Speaker-Dependent mode provides higher accuracy than Speaker-Independent mode as similar researches are found to support the findings. Besides, the number of emotion used in emotion recognition system does affect the accuracy in recognizing mood through speeches. Emotion- basis data division was found to have better accuracy instead of using Speaker- basis data division during the process of handling speeches database to train emotion recognition model. The human recognition on speeches database had shown it's less accurate to predict others‟ emotions under cross cultural background.
author2 Seet Gim Lee, Gerald
author_facet Seet Gim Lee, Gerald
Wong, Yi Ben.
format Final Year Project
author Wong, Yi Ben.
author_sort Wong, Yi Ben.
title Mood recognition through vocal prosody recognition
title_short Mood recognition through vocal prosody recognition
title_full Mood recognition through vocal prosody recognition
title_fullStr Mood recognition through vocal prosody recognition
title_full_unstemmed Mood recognition through vocal prosody recognition
title_sort mood recognition through vocal prosody recognition
publishDate 2013
url http://hdl.handle.net/10356/53290
_version_ 1759855543030120448