ISCA Archive ICSLP 1998
ISCA Archive ICSLP 1998

Training of context-dependent subspace distribution clustering hidden Markov model

Brian Mak, Enrico Bocchieri

Training of continuous density hidden Markov models(CDHMMs) is usually time-consuming and tedious due to the large number of model parameters involved. Recently we proposed a new derivative of CDHMM, the subspace distribution clustering hidden Markov model(SDCHMM) which tie CDHMMs at the finer level of subspace distributions, resulting in many fewer model parameters. An SDCHMM training algorithm is also devised to train SDCHMMs directly from speech data without intermediate CDHMMs. On the ATIS task, speaker-independent context-independent(CI) SDCHMMs can be trained with as little as 8 minutes of speech with no loss in recognition accuracy --- a 25-fold reduction when compared with their CDHMM counterparts. In this paper, we extend our novel SDCHMM training to context-dependent(CD) modeling with the assumption of various prior knowledge. Despite the 30-fold increase of model parameters in the CD ATIS CDHMMs, their equivalent CD SDCHMMs can still be estimated with a few minutes of ATIS data.


doi: 10.21437/ICSLP.1998-185

Cite as: Mak, B., Bocchieri, E. (1998) Training of context-dependent subspace distribution clustering hidden Markov model. Proc. 5th International Conference on Spoken Language Processing (ICSLP 1998), paper 0699, doi: 10.21437/ICSLP.1998-185

@inproceedings{mak98_icslp,
  author={Brian Mak and Enrico Bocchieri},
  title={{Training of context-dependent subspace distribution clustering hidden Markov model}},
  year=1998,
  booktitle={Proc. 5th International Conference on Spoken Language Processing (ICSLP 1998)},
  pages={paper 0699},
  doi={10.21437/ICSLP.1998-185}
}