TY - GEN
T1 - Subvector-quantized high-density discrete hidden Markov model and its re-estimation
AU - Ye, Guoli
AU - Mak, Brian
PY - 2010
Y1 - 2010
N2 - We investigated two methods to improve the performance of high-density discrete hidden Markov model (HDDHMM). HDDHMM employs discrete densities with a very large codebook consisting of thousands to tens of thousands of vector quantization (VQ) codewords which are constructed as the product of per-dimension scalar quantization (SQ) codewords. Although the subsequent HDDHMM is fast in decoding, it is not accurate enough. In this paper, making use of the fact that, for a fixed number of bits, VQ is more efficient than SQ, subvector quantization (SVQ) was investigated to improve the quantization efficiency while keeping the (time and space) complexity of the quantizer sufficiently low. Model parameters of the resulting SVQ-HDDHMM were further re-estimated. For the Wall Street Journal 5K-vocabulary task, it is found that the proposed SVQ-HDDHMM could be a better model (both in terms of recognition time and error rate) than conventional continuous-density HMM for practical deployment.
AB - We investigated two methods to improve the performance of high-density discrete hidden Markov model (HDDHMM). HDDHMM employs discrete densities with a very large codebook consisting of thousands to tens of thousands of vector quantization (VQ) codewords which are constructed as the product of per-dimension scalar quantization (SQ) codewords. Although the subsequent HDDHMM is fast in decoding, it is not accurate enough. In this paper, making use of the fact that, for a fixed number of bits, VQ is more efficient than SQ, subvector quantization (SVQ) was investigated to improve the quantization efficiency while keeping the (time and space) complexity of the quantizer sufficiently low. Model parameters of the resulting SVQ-HDDHMM were further re-estimated. For the Wall Street Journal 5K-vocabulary task, it is found that the proposed SVQ-HDDHMM could be a better model (both in terms of recognition time and error rate) than conventional continuous-density HMM for practical deployment.
UR - https://openalex.org/W2023921487
UR - https://www.scopus.com/pages/publications/79851492314
U2 - 10.1109/ISCSLP.2010.5684838
DO - 10.1109/ISCSLP.2010.5684838
M3 - Conference Paper published in a book
SN - 9781424462469
T3 - 2010 7th International Symposium on Chinese Spoken Language Processing, ISCSLP 2010 - Proceedings
SP - 109
EP - 113
BT - 2010 7th International Symposium on Chinese Spoken Language Processing, ISCSLP 2010 - Proceedings
T2 - 2010 7th International Symposium on Chinese Spoken Language Processing, ISCSLP 2010
Y2 - 29 November 2010 through 3 December 2010
ER -