Speech Recognition with a Generative Factor Analyzed Hidden Markov Model

No Thumbnail Available
File version
Author(s)
Yao, K
Paliwal, KK
Lee, TW
Griffith University Author(s)
Primary Supervisor
Other Supervisors
Editor(s)

Andrzej Drygajlo (Technical Program Chair)

Date
2003
Size
File type(s)
Location

Geneva, Switzerland

License
Abstract

We present a generative factor analyzed hidden Markov model (GFA-HMM) for automatic speech recognition. In a traditional HMM, the observation vectors are represented by mixture of Gaussians (MoG) that are dependent on discrete-valued hidden state sequence. The GFA-HMM introduces a hierarchy of continuous-valued latent representation of observation vectors, where latent vectors in one level are acoustic-unit dependent and the latent vectors in a higher level are acoustic-unit independent. An expectation maximization (EM) algorithm is derived for maximum likelihood parameter estimation of the model. The GFA-HMM can achieve a much more compact representation of the intra-frame statistics of observation vectors than traditional HMM. We conducted an experiment to show that the GFA-HMM can achieve better performances over traditional HMM with the same amount of training data but much smaller number of model parameters.

Journal Title
Conference Title

EUROSPEECH 2003 - 8th European Conference on Speech Communication and Technology

Book Title
Edition
Volume
Issue
Thesis Type
Degree Program
School
DOI
Patent number
Funder(s)
Grant identifier(s)
Rights Statement
Rights Statement
Item Access Status
Note
Access the data
Related item(s)
Subject
Persistent link to this record
Citation