Soft GPD for minimum classification error rate training

Bertram E. Shi, Kaisheng Yao, Zhigang Cao

Research output: Chapter in Book/Conference Proceeding/ReportConference Paper published in a bookpeer-review

Abstract

Minimum classification error (MCE) rate training is a discriminative training method which seeks to minimize an empirical estimate of the error probability derived over a training set. The segmental generalized probabilistic descent (GPD) algorithm for MCE uses the log likelihood of the best path as a discriminant function to estimate the error probability. This paper shows that by using a discriminant function similar to the auxiliary function used in EM, we can obtain a «soft» version of GPD in the sense that information about all possible paths is retained. Complexity is similar to segmental GPD. For certain parameter values, the algorithm is equivalent to segmental GPD. By modifying the misclassification measure usually used, we can obtain an algorithm for embedded MCE training for continuous speech which does not require a separate N-best search to determine competing classes. Experimental results show error rate reduction of 20% compared with maximum likelihood training.

Original languageEnglish
Title of host publicationSpeech Processing II
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages1253-1256
Number of pages4
ISBN (Electronic)0780362934
DOIs
Publication statusPublished - 2000
Event25th IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2000 - Istanbul, Turkey
Duration: 5 Jun 20009 Jun 2000

Publication series

NameICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings
Volume3
ISSN (Print)1520-6149

Conference

Conference25th IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2000
Country/TerritoryTurkey
CityIstanbul
Period5/06/009/06/00

Bibliographical note

Publisher Copyright:
© 2000 IEEE.

Fingerprint

Dive into the research topics of 'Soft GPD for minimum classification error rate training'. Together they form a unique fingerprint.

Cite this