Generating music with emotions

We focus on the music generation conditional on human emotions, specifically the positive and negative emotions. There is no existing large-scale music datasets with the annotation of human emotion labels. It is thus not intuitive how to generate music conditioned on emotion labels. In this paper, w...

Full description

Saved in:
Bibliographic Details
Main Authors: BAO, Chunhui, SUN, Qianru
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2022
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/7557
https://ink.library.smu.edu.sg/context/sis_research/article/8560/viewcontent/TMM_Generating_Music_Final.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-8560
record_format dspace
spelling sg-smu-ink.sis_research-85602022-11-29T07:02:45Z Generating music with emotions BAO, Chunhui SUN, Qianru We focus on the music generation conditional on human emotions, specifically the positive and negative emotions. There is no existing large-scale music datasets with the annotation of human emotion labels. It is thus not intuitive how to generate music conditioned on emotion labels. In this paper, we propose an annotation-free method to build a new dataset where each sample is a triplet of lyric, melody and emotion label (without requiring any labours). Specifically, we first train the automated emotion recognition model using the BERT (pre-trained on GoEmotions dataset) on Edmonds Dance dataset. We use it to automatically ‘`label’' the music with the emotion labels recognized from the lyrics. We then train the encoder-decoder based model to generate emotional music on that dataset, and call our overall method as Emotional Lyric and Melody Generator (ELMG). The framework of ELMG is consisted of three modules: 1) an encoder-decoder model trained end-to-end to generate lyric and melody; 2) a music emotion classifier trained on labeled data (our proposed dataset); and 3) a modified beam search algorithm that guides the music generation process by incorporating the music emotion classifier. We conduct objective and subjective evaluations on the generated music pieces, and our results show that ELMG is capable of generating tuneful lyric and melody with specified human emotions. 2022-03-01T08:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/7557 info:doi/10.1109/TMM.2022.3163543 https://ink.library.smu.edu.sg/context/sis_research/article/8560/viewcontent/TMM_Generating_Music_Final.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Conditional Music Generation Seq2Seq Beam Search Transformer Databases and Information Systems Music
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Conditional Music Generation
Seq2Seq
Beam Search
Transformer
Databases and Information Systems
Music
spellingShingle Conditional Music Generation
Seq2Seq
Beam Search
Transformer
Databases and Information Systems
Music
BAO, Chunhui
SUN, Qianru
Generating music with emotions
description We focus on the music generation conditional on human emotions, specifically the positive and negative emotions. There is no existing large-scale music datasets with the annotation of human emotion labels. It is thus not intuitive how to generate music conditioned on emotion labels. In this paper, we propose an annotation-free method to build a new dataset where each sample is a triplet of lyric, melody and emotion label (without requiring any labours). Specifically, we first train the automated emotion recognition model using the BERT (pre-trained on GoEmotions dataset) on Edmonds Dance dataset. We use it to automatically ‘`label’' the music with the emotion labels recognized from the lyrics. We then train the encoder-decoder based model to generate emotional music on that dataset, and call our overall method as Emotional Lyric and Melody Generator (ELMG). The framework of ELMG is consisted of three modules: 1) an encoder-decoder model trained end-to-end to generate lyric and melody; 2) a music emotion classifier trained on labeled data (our proposed dataset); and 3) a modified beam search algorithm that guides the music generation process by incorporating the music emotion classifier. We conduct objective and subjective evaluations on the generated music pieces, and our results show that ELMG is capable of generating tuneful lyric and melody with specified human emotions.
format text
author BAO, Chunhui
SUN, Qianru
author_facet BAO, Chunhui
SUN, Qianru
author_sort BAO, Chunhui
title Generating music with emotions
title_short Generating music with emotions
title_full Generating music with emotions
title_fullStr Generating music with emotions
title_full_unstemmed Generating music with emotions
title_sort generating music with emotions
publisher Institutional Knowledge at Singapore Management University
publishDate 2022
url https://ink.library.smu.edu.sg/sis_research/7557
https://ink.library.smu.edu.sg/context/sis_research/article/8560/viewcontent/TMM_Generating_Music_Final.pdf
_version_ 1770576372184383488