Music generation with deep learning techniques
This report demonstrated the use of a deep convolutional generative adversarial network (DCGAN) in generating expressive music with dynamics. The existing deep learning models for music generation were reviewed. However, most research focused on musical composition and removed expressive attributes...
Saved in:
Main Author: | |
---|---|
Other Authors: | |
Format: | Final Year Project |
Language: | English |
Published: |
Nanyang Technological University
2021
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/148097 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-148097 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1480972021-04-23T13:24:50Z Music generation with deep learning techniques Toh, Raymond Kwan How Alexei Sourin School of Computer Science and Engineering assourin@ntu.edu.sg Engineering::Computer science and engineering::Computing methodologies::Artificial intelligence Visual arts and music::Music::Compositions This report demonstrated the use of a deep convolutional generative adversarial network (DCGAN) in generating expressive music with dynamics. The existing deep learning models for music generation were reviewed. However, most research focused on musical composition and removed expressive attributes during data preprocessing, which resulted in mechanical-sounding, generated music. To address the issue, music elements such as pitch, time, velocity were extracted from MIDI files and encoded with piano roll data representation. With the piano roll data representation, DCGAN learned the data distribution from the given dataset and generated new data derived from the same distribution. The generated music was evaluated based on its incorporation of music dynamics and a user study. The evaluation results verified that DCGAN was capable of generating expressive music comprising of music dynamics and syncopated rhythm. Bachelor of Engineering (Computer Engineering) 2021-04-23T13:24:49Z 2021-04-23T13:24:49Z 2021 Final Year Project (FYP) Toh, R. K. H. (2021). Music generation with deep learning techniques. Final Year Project (FYP), Nanyang Technological University, Singapore. https://hdl.handle.net/10356/148097 https://hdl.handle.net/10356/148097 en SCSE20-0007 application/pdf Nanyang Technological University |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering::Computer science and engineering::Computing methodologies::Artificial intelligence Visual arts and music::Music::Compositions |
spellingShingle |
Engineering::Computer science and engineering::Computing methodologies::Artificial intelligence Visual arts and music::Music::Compositions Toh, Raymond Kwan How Music generation with deep learning techniques |
description |
This report demonstrated the use of a deep convolutional generative adversarial network (DCGAN) in generating expressive music with dynamics. The existing deep learning models for music generation were reviewed. However, most research focused on musical composition and removed expressive attributes during data preprocessing, which resulted in mechanical-sounding, generated music. To address the issue, music elements such as pitch, time, velocity were extracted from MIDI files and encoded with piano roll data representation. With the piano roll data representation, DCGAN learned the data distribution from the given dataset and generated new data derived from the same distribution. The generated music was evaluated based on its incorporation of music dynamics and a user study. The evaluation results verified that DCGAN was capable of generating expressive music comprising of music dynamics and syncopated rhythm. |
author2 |
Alexei Sourin |
author_facet |
Alexei Sourin Toh, Raymond Kwan How |
format |
Final Year Project |
author |
Toh, Raymond Kwan How |
author_sort |
Toh, Raymond Kwan How |
title |
Music generation with deep learning techniques |
title_short |
Music generation with deep learning techniques |
title_full |
Music generation with deep learning techniques |
title_fullStr |
Music generation with deep learning techniques |
title_full_unstemmed |
Music generation with deep learning techniques |
title_sort |
music generation with deep learning techniques |
publisher |
Nanyang Technological University |
publishDate |
2021 |
url |
https://hdl.handle.net/10356/148097 |
_version_ |
1698713741429833728 |