Learning affective representations based on magnitude and dynamic relative phase information for speech emotion recognition
The complete acoustic features include magnitude and phase information. However, traditional speech emotion recognition methods only focus on the magnitude information and ignore the phase data, and will inevitably miss some information. This study explores the accurate extraction and effective use...
Saved in:
Main Authors: | , , , , |
---|---|
Other Authors: | |
Format: | Article |
Language: | English |
Published: |
2022
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/162646 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-162646 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1626462022-11-02T01:23:25Z Learning affective representations based on magnitude and dynamic relative phase information for speech emotion recognition Guo, Lili Wang, Longbiao Dang, Jianwu Chng, Eng Siong Nakagawa, Seiichi School of Computer Science and Engineering Engineering::Computer science and engineering Speech Emotion Recognition Magnitude Spectrogram The complete acoustic features include magnitude and phase information. However, traditional speech emotion recognition methods only focus on the magnitude information and ignore the phase data, and will inevitably miss some information. This study explores the accurate extraction and effective use of phase features for speech emotion recognition. First, the reflection of speech emotion in the phase spectrum is analyzed, and a quantitative analysis shows that phase data contain information that can be used to distinguish emotions. A dynamic relative phase (DRP) feature extraction method is then proposed to solve the problem in which the original relative phase (RP) has difficulty determining the base frequency and further alleviating the dependence of the phase on the clipping position of the frame. Finally, a single-channel model (SCM) and a multi-channel model with an attention mechanism (MCMA) are constructed to effectively integrate the phase and magnitude information. By introducing phase information, more complete acoustic features are captured, which enriches the emotional representations. The experiments were conducted using the Emo-DB and IEMOCAP databases. Experimental results demonstrate the effectiveness of the proposed DRP for speech emotion recognition, as well as the complementarity between the phase and magnitude information in speech emotion recognition. This work was supported by the National Key R&D Program of China (Grant NO. 2018YFB1305200), by the National Natural Science Foundation of China (Grant NO. 61771333), and by the Tianjin Municipal Science and Technology Project, China (Grant NO.18ZXZNGX00330). Additionally, we would like to acknowledge the financial support provided by the China Scholarship Council (NO. 201906250176) during a visit of Lili Guo to Nanyang Technological University. 2022-11-02T01:23:24Z 2022-11-02T01:23:24Z 2022 Journal Article Guo, L., Wang, L., Dang, J., Chng, E. S. & Nakagawa, S. (2022). Learning affective representations based on magnitude and dynamic relative phase information for speech emotion recognition. Speech Communication, 136, 118-127. https://dx.doi.org/10.1016/j.specom.2021.11.005 0167-6393 https://hdl.handle.net/10356/162646 10.1016/j.specom.2021.11.005 2-s2.0-85121962474 136 118 127 en Speech Communication © 2021 Elsevier B.V. All rights reserved. |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering::Computer science and engineering Speech Emotion Recognition Magnitude Spectrogram |
spellingShingle |
Engineering::Computer science and engineering Speech Emotion Recognition Magnitude Spectrogram Guo, Lili Wang, Longbiao Dang, Jianwu Chng, Eng Siong Nakagawa, Seiichi Learning affective representations based on magnitude and dynamic relative phase information for speech emotion recognition |
description |
The complete acoustic features include magnitude and phase information. However, traditional speech emotion recognition methods only focus on the magnitude information and ignore the phase data, and will inevitably miss some information. This study explores the accurate extraction and effective use of phase features for speech emotion recognition. First, the reflection of speech emotion in the phase spectrum is analyzed, and a quantitative analysis shows that phase data contain information that can be used to distinguish emotions. A dynamic relative phase (DRP) feature extraction method is then proposed to solve the problem in which the original relative phase (RP) has difficulty determining the base frequency and further alleviating the dependence of the phase on the clipping position of the frame. Finally, a single-channel model (SCM) and a multi-channel model with an attention mechanism (MCMA) are constructed to effectively integrate the phase and magnitude information. By introducing phase information, more complete acoustic features are captured, which enriches the emotional representations. The experiments were conducted using the Emo-DB and IEMOCAP databases. Experimental results demonstrate the effectiveness of the proposed DRP for speech emotion recognition, as well as the complementarity between the phase and magnitude information in speech emotion recognition. |
author2 |
School of Computer Science and Engineering |
author_facet |
School of Computer Science and Engineering Guo, Lili Wang, Longbiao Dang, Jianwu Chng, Eng Siong Nakagawa, Seiichi |
format |
Article |
author |
Guo, Lili Wang, Longbiao Dang, Jianwu Chng, Eng Siong Nakagawa, Seiichi |
author_sort |
Guo, Lili |
title |
Learning affective representations based on magnitude and dynamic relative phase information for speech emotion recognition |
title_short |
Learning affective representations based on magnitude and dynamic relative phase information for speech emotion recognition |
title_full |
Learning affective representations based on magnitude and dynamic relative phase information for speech emotion recognition |
title_fullStr |
Learning affective representations based on magnitude and dynamic relative phase information for speech emotion recognition |
title_full_unstemmed |
Learning affective representations based on magnitude and dynamic relative phase information for speech emotion recognition |
title_sort |
learning affective representations based on magnitude and dynamic relative phase information for speech emotion recognition |
publishDate |
2022 |
url |
https://hdl.handle.net/10356/162646 |
_version_ |
1749179218203246592 |