TY - GEN
T1 - Latent topic modeling of word vicinity information for speech recognition
AU - Chen, Kuan Yu
AU - Chiu, Hsuan Sheng
AU - Chen, Berlin
PY - 2010
Y1 - 2010
N2 - Topic language models, mostly revolving around the discovery of "word-document" co-occurrence dependence, have attracted significant attention and shown good performance in a wide variety of speech recognition tasks over the years. In this paper, a new topic language model, named word vicinity model (WVM), is proposed to explore the co-occurrence relationship between words, as well as the long-span latent topical information for language model adaptation. A search history is modeled as a composite WVM model for predicting a decoded word. The underlying characteristics and different kinds of model structures are extensively investigated, while the performance of WVM is thoroughly analyzed and verified by comparison with a few existing topic language models. Moreover, we also present a new modeling approach to our recently proposed word topic model (WTM), and design an efficient way to simultaneously extract "word-document" and "word-word" co-occurrence characteristics through the sharing of the same set of latent topics. Experiments on broadcast news transcription seem to demonstrate the utility of the presented models.
AB - Topic language models, mostly revolving around the discovery of "word-document" co-occurrence dependence, have attracted significant attention and shown good performance in a wide variety of speech recognition tasks over the years. In this paper, a new topic language model, named word vicinity model (WVM), is proposed to explore the co-occurrence relationship between words, as well as the long-span latent topical information for language model adaptation. A search history is modeled as a composite WVM model for predicting a decoded word. The underlying characteristics and different kinds of model structures are extensively investigated, while the performance of WVM is thoroughly analyzed and verified by comparison with a few existing topic language models. Moreover, we also present a new modeling approach to our recently proposed word topic model (WTM), and design an efficient way to simultaneously extract "word-document" and "word-word" co-occurrence characteristics through the sharing of the same set of latent topics. Experiments on broadcast news transcription seem to demonstrate the utility of the presented models.
KW - Broadcast news transcription
KW - Speech recognition
KW - Topic language model
KW - Word vicinity model
UR - http://www.scopus.com/inward/record.url?scp=78049354892&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=78049354892&partnerID=8YFLogxK
U2 - 10.1109/ICASSP.2010.5494942
DO - 10.1109/ICASSP.2010.5494942
M3 - Conference contribution
AN - SCOPUS:78049354892
SN - 9781424442966
T3 - ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings
SP - 5394
EP - 5397
BT - 2010 IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2010 - Proceedings
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2010 IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2010
Y2 - 14 March 2010 through 19 March 2010
ER -