没有合适的资源?快使用搜索试试~ 我知道了~
Connectionist Temporal Classification Labelling Unsegmented.pdf
0 下载量 47 浏览量
2024-11-19
16:06:18
上传
评论
收藏 306KB PDF 举报
温馨提示
优质文献资料分享,希望可以帮助到你~
资源推荐
资源详情
资源评论
Connectionist Temporal Classification: Labelling Unsegmented
Sequence Data with Recurrent Neural Networks
Alex Graves
1
alex@idsia.ch
Santiago Fern´andez
1
santiago@idsia.ch
Faustino Gomez
1
tino@idsia.ch
J¨urgen Schmidhuber
1,2
juergen@idsia.ch
1
Istituto Dalle Molle di Studi sull’Intelligenza Artificiale (IDSIA), Galleria 2, 6928 Manno-Lugano, Switzerland
2
Technische Universit¨at M¨unchen (TUM), Boltzmannstr. 3, 85748 Garching, Munich, Germany
Abstract
Many real-world sequence learning tasks re-
quire the prediction of sequences of lab e ls
from noisy, unsegmented input data. In
speech recognition, for example, an acoustic
signal is transcribed into words or sub-word
units. Recurrent neural networks (RNNs) are
powerful sequence learners that would seem
well suited to such tasks. However, because
they require pre-segmented training data,
and post-processing to transform their out-
puts into label sequences, their applicability
has so far been limited. This paper presents a
novel method for training RNNs to label un-
segmented sequences directly, thereby solv-
ing both problems. An experiment on the
TIMIT speech corpus demonstrates its ad-
vantages over both a baseline HMM and a
hybrid HMM-RNN.
1. Introduction
Labelling unsegmented sequence data is a ubiquitous
problem in real-world sequence learning. It is partic-
ularly common in perceptual tasks (e.g. handwriting
recognition, speech recognition, gesture recognition)
where noisy, real-valued input streams are annotated
with strings of discrete labels, such as letters or words.
Currently, graphical models such as hidden Markov
Models (HMMs; Rabiner, 1989), conditional random
fields (CRFs; Lafferty et al., 2001) and their vari-
ants, are the predominant framework for sequence la-
App earing in Proceedings of the 23
rd
International Con-
ference on Machine Learning, Pittsburgh, PA, 2006. Copy-
right 2006 by the author(s)/owner(s).
belling. While these approaches have proved success-
ful for many problems, they have several drawbacks:
(1) they usually require a significant amount of task
specific knowledge, e.g. to design the state models for
HMMs, or choose the input features for CRFs; (2)
they require explicit (and often questionable) depen-
dency assumptions to make inference tractable, e.g.
the assumption that observations are independent for
HMMs; (3) for standard HMMs, training is generative,
even though sequence labelling is discriminative.
Recurrent neural networks (RNNs), on the other hand,
require no prior knowledge of the data, beyond the
choice of input and output representation. They can
be trained discriminatively, and their internal state
provides a powerful, general mechanism for modelling
time series. In addition, they tend to be robust to
temporal and spatial noise.
So far, however, it has not been possible to apply
RNNs directly to sequence labelling. The problem is
that the standard neural network objective functions
are defined separately for each p oint in the training se -
quence; in other words, RNNs can only be trained to
make a series of independent label classifications. This
means that the training data must be pre-segmented,
and that the network outputs must be post-processed
to give the final label sequence.
At present, the most effective use of RNNs for se-
quence labelling is to combine them with HMMs in the
so-called hybrid approach (Bourlard & Morgan, 1994;
Bengio., 1999). Hybrid systems use HMMs to model
the long-range sequential structure of the data, and
neural nets to provide localised classifications. The
HMM component is able to automatically segment
the sequence during training, and to transform the
network classifications into label sequences. However,
as well as inheriting the aforeme ntioned drawbacks of
资源评论
不脱发的程序猿
- 粉丝: 26w+
- 资源: 5872
上传资源 快速赚钱
- 我的内容管理 展开
- 我的资源 快来上传第一个资源
- 我的收益 登录查看自己的收益
- 我的积分 登录查看自己的积分
- 我的C币 登录后查看C币余额
- 我的收藏
- 我的下载
- 下载帮助
最新资源
资源上传下载、课程学习等过程中有任何疑问或建议,欢迎提出宝贵意见哦~我们会及时处理!
点击此处反馈
安全验证
文档复制为VIP权益,开通VIP直接复制
信息提交成功