z-logo
open-access-imgOpen Access
Music Generation Based on Convolution-LSTM
Author(s) -
Yongjie Huang,
Huang Xiao-feng,
Qiakai Cai
Publication year - 2018
Publication title -
computer and information science
Language(s) - English
Resource type - Journals
eISSN - 1913-8997
pISSN - 1913-8989
DOI - 10.5539/cis.v11n3p50
Subject(s) - computer science , convolution (computer science) , midi , speech recognition , convolutional neural network , artificial intelligence , feature (linguistics) , overlap–add method , domain (mathematical analysis) , musical notation , pattern recognition (psychology) , matrix (chemical analysis) , musical , algorithm , artificial neural network , fourier transform , mathematics , mathematical analysis , linguistics , philosophy , operating system , art , fourier analysis , materials science , composite material , fractional fourier transform , visual arts
In this paper, we propose a model that combines Convolutional Neural Network (CNN) and Long Short-Term Memory (LSTM) for music generation. We first convert MIDI-format music file into a musical score matrix, and then establish convolution layers to extract feature of the musical score matrix. Finally, the output of the convolution layers is split in the direction of the time axis and input into the LSTM, so as to achieve the purpose of music generation. The result of the model was verified by comparison of accuracy, time-domain analysis, frequency-domain analysis and human-auditory evaluation. The results show that Convolution-LSTM performs better in music genertaion than LSTM, with more pronounced undulations and clearer melody.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here
Accelerating Research

Address

John Eccles House
Robert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom