Y. Shiu et al.: Robust On-line Beat Tracking with Kalman Filtering and Probabilistic Data Association (KF-PDA)
Manuscript received July 11, 2008 0098 3063/08/$20.00 © 2008 IEEE
1369
Robust On-line Beat Tracking with Kalman Filtering
and Probabilistic Data Association (KF-PDA)
Yu Shiu, Student Member, IEEE, Namgook Cho, Student Member, IEEE, Pei-Chen Chang
and C.-C. Jay Kuo, Fellow, IEEE
Abstract — A Kalman filtering (KF) approach to on-line
musical beat tracking with probabilistic data association
(PDA) is investigated in this work. We first formulate the beat
tracking process as a linear dynamic system of beat
progression, and then apply the Kalman filtering algorithm to
the dynamic system in estimating the time-varying tempo and
beat locations. Musical beat tracking using traditional
Kalman filtering is however not reliable in the presence of
tempo fluctuations and expressive timing deviations. To
address this problem, we adopt data association techniques to
assign probability masses to all possible beat interpretations,
and then locate the true beat according to the weighting. Two
methods are proposed. The first one (PDA-I) weighs the
distance between the candidate observation and the predicted
beat location while the second method (PDA-II) considers not
only the distance but also the onset intensity in weight
selection. Superior performance of the proposed beat tracking
algorithm is demonstrated with simulation results on the
Music Information Retrieval Evaluation Exchange (MIREX)
2006 beat tracking competition practice dataset and the
Billboard Top-10 database
1
.
Index Terms — Musical signal processing, on-line beat
tracking, Kalman filter, probabilistic data association, music
information retrieval.
I. INTRODUCTION
When listening to music, most people even without musical
education can grasp the speed of music and follow it by foot-
tapping or hand-clapping along with beats. However, the same
is not true for electronic devices. Automatic beat tracking has
been an active area of research for more than twenty years.
The beat is a fundamental unit of the temporal structure of
music, especially to Western music, and beat tracking is an
essential task in many musical applications such as musical
analysis, synchronization, editing of musical sounds, and
human-computer improvisation. This work presents an on-line
(or causal) musical beat tracking system, where beat
estimation at a given time depends only on past and present
data.
Beat tracking is defined by estimating the possibly time-
varying tempo and the time location of each beat, where the
beat is referred to as the foot tapping and tempo as the beat
rate [1]. Our research goal is to estimate the set of beat
1
Part of this work was presented at ICCE2008, Las Vegas, NV USA.
The authors are with Department of Electrical Engineering and Signal and
Image Processing Institute, University of Southern California, Los Angeles,
CA 90089-2564 USA (e-mails: atoultaro@gmail.com
, namgookc@usc.edu,
peichenc@usc.edu
, and cckuo@sipi.usc.edu).
locations from musical audio signals sequentially. Ideally,
when beat pulses are strong and the duration between adjacent
beats is perceptually clear, automatic beat tracking can be
done easily. Its performance nevertheless degrades
significantly in practice due to several reasons. The first one
comes from rest notes and missed-beat syncopation. The rest
notes hide beat tracking cues, whereas syncopation does not
have an onset pulse on expected beat location but with a small
shift. The second one is due to variability in human
performance. Even if a performer attempts to keep the
duration between two consecutive beats constant throughout
the whole music piece, the actual duration tends to vary along
time. The last one is that some music pieces have time-varying
tempo and, consequently, a time-varying beat period. The
performance of beat tracking algorithms is often less robust
when dealing with classical music, as compared with that
containing drum sounds [1], [2].
Early work on automatic beat tracking was done by
researchers in the fields of music perception and computer
science [3]. More recently, Brown [4] used the autocorrelation
function to examine the pulses in musical scores. Scheirer [5]
applied a bank of comb filters to a musical signal at different
fixed frequencies and searched for the filter that gives the
strongest response for tempo estimation. Afterwards, the beat
location was calculated by examining the phase of the filtered
signal. Goto [2] developed an on-line beat tracking system
that can process music with or without drum sounds. The
system recognizes the hierarchical beat structure using three
kinds of musical knowledge: onset times, chord changes, and
drum patterns. A probabilistic generative model for tempo
tracking was examined by Cemgil et al. [6],[7]. A Kalman
filtering process was used to track beats in [6], which was
followed by using the tempogram representation to assign
probability masses to all possible beat candidates, while
Monte Carlo methods were exploited to infer a hidden tempo
variable in [7]. Hainsworth and Macleod [8] used particle
filters to associate onsets from an audio signal to a time-
varying tempo process so as to determine the beat locations.
Most of earlier work for beat tracking used symbolic or
musical instrument digital interface (MIDI) data, e.g.,
[4],[6],[7]. Audio signals have been examined more recently,
e.g., [2],[5],[8]. In addition, most previous beat tracking
systems adopt a non-causal method that allows the use of
future data and backward decoding, which is not suitable for
real-time implementation in consumer electronic applications.
In this work, we present a method that extracts beat
locations from acoustic musical signals, not limited to any
particular music type, including both classical music and
Authorized licensed use limited to: BEIJING INSTITUTE OF TECHNOLOGY. Downloaded on April 21,2010 at 13:44:39 UTC from IEEE Xplore. Restrictions apply.