Music generation with deep learning techniques

This report demonstrated the use of a deep convolutional generative adversarial network (DCGAN) in generating expressive music with dynamics. The existing deep learning models for music generation were reviewed. However, most research focused on musical composition and removed expressive attributes...

Full description

Saved in:
Bibliographic Details
Main Author: Toh, Raymond Kwan How
Other Authors: Alexei Sourin
Format: Final Year Project
Language:English
Published: Nanyang Technological University 2021
Subjects:
Online Access:https://hdl.handle.net/10356/148097
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
Description
Summary:This report demonstrated the use of a deep convolutional generative adversarial network (DCGAN) in generating expressive music with dynamics. The existing deep learning models for music generation were reviewed. However, most research focused on musical composition and removed expressive attributes during data preprocessing, which resulted in mechanical-sounding, generated music. To address the issue, music elements such as pitch, time, velocity were extracted from MIDI files and encoded with piano roll data representation. With the piano roll data representation, DCGAN learned the data distribution from the given dataset and generated new data derived from the same distribution. The generated music was evaluated based on its incorporation of music dynamics and a user study. The evaluation results verified that DCGAN was capable of generating expressive music comprising of music dynamics and syncopated rhythm.