Technical Program

Paper Detail

Paper IDE-1-2.3
Paper Title BEAT AND DOWNBEAT TRACKING OF SYMBOLIC MUSIC DATA USING DEEP RECURRENT NEURAL NETWORKS
Authors Yi-Chin Chuang, National Chung Hsing University, Taiwan; Li Su, Academia Sinica, Taiwan
Session E-1-2: Music Information Processing 1, Audio Scene Classification
TimeTuesday, 08 December, 15:30 - 17:00
Presentation Time:Tuesday, 08 December, 16:00 - 16:15 Check your Time Zone
All times are in New Zealand Time (UTC +13)
Topic Speech, Language, and Audio (SLA):
Abstract Musical beat tracking is one of the most investigated tasks in music information retrieval (MIR). Research endeavors on this task have mostly been focused on the modeling of audio data representations. In contrast, beat tracking of symbolic music contents (e.g., MIDI, score sheets) has been relatively overlooked in the past years. In this paper, we revisit the task of symbolic music beat tracking, and propose to solve this task with modern deep learning approaches. The symbolic beat tracking framework performs joint beat and downbeat tracking in a multi-task learning (MTL) manner, and we investigate various types of networks which are based on the recurrent neural networks (RNN), such as bidirectional long short-term memory (BLSTM) network, hierarchical multi-scale (HM) LSTM, and BLSTM with the attention mechanism. In the experiments, a systematic comparison of these networks and state-of-art audio beat tracking methods are performed on the MusicNet dataset. Experiment results show that the BLSTM model trained specifically on symbolic data outperforms the state-of-the-art beat tracking methods utilized on synthesized audio. Such a comparison of performance also indicates the technical challenges of symbolic music beat tracking.