Generating music with emotions
We focus on the music generation conditional on human emotions, specifically the positive and negative emotions. There is no existing large-scale music datasets with the annotation of human emotion labels. It is thus not intuitive how to generate music conditioned on emotion labels. In this paper, w...
Saved in:
Main Authors: | , |
---|---|
Format: | text |
Language: | English |
Published: |
Institutional Knowledge at Singapore Management University
2022
|
Subjects: | |
Online Access: | https://ink.library.smu.edu.sg/sis_research/7557 https://ink.library.smu.edu.sg/context/sis_research/article/8560/viewcontent/TMM_Generating_Music_Final.pdf |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Singapore Management University |
Language: | English |
id |
sg-smu-ink.sis_research-8560 |
---|---|
record_format |
dspace |
spelling |
sg-smu-ink.sis_research-85602022-11-29T07:02:45Z Generating music with emotions BAO, Chunhui SUN, Qianru We focus on the music generation conditional on human emotions, specifically the positive and negative emotions. There is no existing large-scale music datasets with the annotation of human emotion labels. It is thus not intuitive how to generate music conditioned on emotion labels. In this paper, we propose an annotation-free method to build a new dataset where each sample is a triplet of lyric, melody and emotion label (without requiring any labours). Specifically, we first train the automated emotion recognition model using the BERT (pre-trained on GoEmotions dataset) on Edmonds Dance dataset. We use it to automatically ‘`label’' the music with the emotion labels recognized from the lyrics. We then train the encoder-decoder based model to generate emotional music on that dataset, and call our overall method as Emotional Lyric and Melody Generator (ELMG). The framework of ELMG is consisted of three modules: 1) an encoder-decoder model trained end-to-end to generate lyric and melody; 2) a music emotion classifier trained on labeled data (our proposed dataset); and 3) a modified beam search algorithm that guides the music generation process by incorporating the music emotion classifier. We conduct objective and subjective evaluations on the generated music pieces, and our results show that ELMG is capable of generating tuneful lyric and melody with specified human emotions. 2022-03-01T08:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/7557 info:doi/10.1109/TMM.2022.3163543 https://ink.library.smu.edu.sg/context/sis_research/article/8560/viewcontent/TMM_Generating_Music_Final.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Conditional Music Generation Seq2Seq Beam Search Transformer Databases and Information Systems Music |
institution |
Singapore Management University |
building |
SMU Libraries |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
SMU Libraries |
collection |
InK@SMU |
language |
English |
topic |
Conditional Music Generation Seq2Seq Beam Search Transformer Databases and Information Systems Music |
spellingShingle |
Conditional Music Generation Seq2Seq Beam Search Transformer Databases and Information Systems Music BAO, Chunhui SUN, Qianru Generating music with emotions |
description |
We focus on the music generation conditional on human emotions, specifically the positive and negative emotions. There is no existing large-scale music datasets with the annotation of human emotion labels. It is thus not intuitive how to generate music conditioned on emotion labels. In this paper, we propose an annotation-free method to build a new dataset where each sample is a triplet of lyric, melody and emotion label (without requiring any labours). Specifically, we first train the automated emotion recognition model using the BERT (pre-trained on GoEmotions dataset) on Edmonds Dance dataset. We use it to automatically ‘`label’' the music with the emotion labels recognized from the lyrics. We then train the encoder-decoder based model to generate emotional music on that dataset, and call our overall method as Emotional Lyric and Melody Generator (ELMG). The framework of ELMG is consisted of three modules: 1) an encoder-decoder model trained end-to-end to generate lyric and melody; 2) a music emotion classifier trained on labeled data (our proposed dataset); and 3) a modified beam search algorithm that guides the music generation process by incorporating the music emotion classifier. We conduct objective and subjective evaluations on the generated music pieces, and our results show that ELMG is capable of generating tuneful lyric and melody with specified human emotions. |
format |
text |
author |
BAO, Chunhui SUN, Qianru |
author_facet |
BAO, Chunhui SUN, Qianru |
author_sort |
BAO, Chunhui |
title |
Generating music with emotions |
title_short |
Generating music with emotions |
title_full |
Generating music with emotions |
title_fullStr |
Generating music with emotions |
title_full_unstemmed |
Generating music with emotions |
title_sort |
generating music with emotions |
publisher |
Institutional Knowledge at Singapore Management University |
publishDate |
2022 |
url |
https://ink.library.smu.edu.sg/sis_research/7557 https://ink.library.smu.edu.sg/context/sis_research/article/8560/viewcontent/TMM_Generating_Music_Final.pdf |
_version_ |
1770576372184383488 |