Mood recognition through vocal prosody recognition
Mood recognition through vocal prosody recognition is designed to predict human‟s mood through speeches profiles. There are existing applications in vocal prosody such as Microsoft‟s “Speech to Text”, IOS‟s “Siri” and Andriod‟s “S Voice” that they are executing actions which are ordered by users but...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Final Year Project |
Language: | English |
Published: |
2013
|
Subjects: | |
Online Access: | http://hdl.handle.net/10356/53290 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-53290 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-532902023-03-04T19:02:58Z Mood recognition through vocal prosody recognition Wong, Yi Ben. Seet Gim Lee, Gerald School of Mechanical and Aerospace Engineering Robotics Research Centre DRNTU::Engineering::Mechanical engineering::Mechatronics Mood recognition through vocal prosody recognition is designed to predict human‟s mood through speeches profiles. There are existing applications in vocal prosody such as Microsoft‟s “Speech to Text”, IOS‟s “Siri” and Andriod‟s “S Voice” that they are executing actions which are ordered by users but they are not relate to mood recognition function. Therefore, this project seeks to develop a software package in mood recognition through human‟s speeches. Speaker-Dependent and Speaker- Independent mode were investigated to develop Real-Time Emotion Recognition System. Speeches database were collected and studied to improve the emotion recognition system since speeches database is one of the factors to define the quality of emotion recognition model. Besides, the process of handling speeches database was proposed to improve the accuracy and several experiments were completed for the improvement. Speeches database was reviewed by other users to prove the quality of speeches in term of expressing moods. Experimental results had shown that the Speaker-Dependent mode provides higher accuracy than Speaker-Independent mode as similar researches are found to support the findings. Besides, the number of emotion used in emotion recognition system does affect the accuracy in recognizing mood through speeches. Emotion- basis data division was found to have better accuracy instead of using Speaker- basis data division during the process of handling speeches database to train emotion recognition model. The human recognition on speeches database had shown it's less accurate to predict others‟ emotions under cross cultural background. Bachelor of Engineering (Mechanical Engineering) 2013-05-31T04:01:49Z 2013-05-31T04:01:49Z 2013 2013 Final Year Project (FYP) http://hdl.handle.net/10356/53290 en Nanyang Technological University 51 p. application/pdf |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
DRNTU::Engineering::Mechanical engineering::Mechatronics |
spellingShingle |
DRNTU::Engineering::Mechanical engineering::Mechatronics Wong, Yi Ben. Mood recognition through vocal prosody recognition |
description |
Mood recognition through vocal prosody recognition is designed to predict human‟s mood through speeches profiles. There are existing applications in vocal prosody such as Microsoft‟s “Speech to Text”, IOS‟s “Siri” and Andriod‟s “S Voice” that they are executing actions which are ordered by users but they are not relate to mood recognition function. Therefore, this project seeks to develop a software package in mood recognition through human‟s speeches. Speaker-Dependent and Speaker- Independent mode were investigated to develop Real-Time Emotion Recognition System. Speeches database were collected and studied to improve the emotion recognition system since speeches database is one of the factors to define the quality of emotion recognition model. Besides, the process of handling speeches database was proposed to improve the accuracy and several experiments were completed for the improvement. Speeches database was reviewed by other users to prove the quality of speeches in term of expressing moods.
Experimental results had shown that the Speaker-Dependent mode provides higher accuracy than Speaker-Independent mode as similar researches are found to support the findings. Besides, the number of emotion used in emotion recognition system does affect the accuracy in recognizing mood through speeches. Emotion- basis data division was found to have better accuracy instead of using Speaker- basis data division during the process of handling speeches database to train emotion recognition model. The human recognition on speeches database had shown it's less accurate to predict others‟ emotions under cross cultural background. |
author2 |
Seet Gim Lee, Gerald |
author_facet |
Seet Gim Lee, Gerald Wong, Yi Ben. |
format |
Final Year Project |
author |
Wong, Yi Ben. |
author_sort |
Wong, Yi Ben. |
title |
Mood recognition through vocal prosody recognition |
title_short |
Mood recognition through vocal prosody recognition |
title_full |
Mood recognition through vocal prosody recognition |
title_fullStr |
Mood recognition through vocal prosody recognition |
title_full_unstemmed |
Mood recognition through vocal prosody recognition |
title_sort |
mood recognition through vocal prosody recognition |
publishDate |
2013 |
url |
http://hdl.handle.net/10356/53290 |
_version_ |
1759855543030120448 |