Paper ID | E-1-2.3 |
Paper Title |
BEAT AND DOWNBEAT TRACKING OF SYMBOLIC MUSIC DATA USING DEEP RECURRENT NEURAL NETWORKS |
Authors |
Yi-Chin Chuang, National Chung Hsing University, Taiwan; Li Su, Academia Sinica, Taiwan |
Session |
E-1-2: Music Information Processing 1, Audio Scene Classification |
Time | Tuesday, 08 December, 15:30 - 17:00 |
Presentation Time: | Tuesday, 08 December, 16:00 - 16:15 Check your Time Zone |
|
All times are in New Zealand Time (UTC +13) |
Topic |
Speech, Language, and Audio (SLA): |
Abstract |
Musical beat tracking is one of the most investigated tasks in music information retrieval (MIR). Research endeavors on this task have mostly been focused on the modeling of audio data representations. In contrast, beat tracking of symbolic music contents (e.g., MIDI, score sheets) has been relatively overlooked in the past years. In this paper, we revisit the task of symbolic music beat tracking, and propose to solve this task with modern deep learning approaches. The symbolic beat tracking framework performs joint beat and downbeat tracking in a multi-task learning (MTL) manner, and we investigate various types of networks which are based on the recurrent neural networks (RNN), such as bidirectional long short-term memory (BLSTM) network, hierarchical multi-scale (HM) LSTM, and BLSTM with the attention mechanism. In the experiments, a systematic comparison of these networks and state-of-art audio beat tracking methods are performed on the MusicNet dataset. Experiment results show that the BLSTM model trained specifically on symbolic data outperforms the state-of-the-art beat tracking methods utilized on synthesized audio. Such a comparison of performance also indicates the technical challenges of symbolic music beat tracking. |