TY - GEN
T1 - Position information for language modeling in speech recognition
AU - Chiu, Hsuan Sheng
AU - Chen, Guan Yu
AU - Lee, Chun Jen
AU - Chen, Berlin
PY - 2008
Y1 - 2008
N2 - This paper considers word position information for language modeling. For organized documents, such as technical papers or news reports, the composition and the word usage of articles of the same style are usually similar. Therefore, the documents can be separated into partitions consisting of identical rhetoric or topic styles by the literary structures, e.g., introductory remarks, related studies or events, elucidations of methodology or affairs, conclusions of the articles, and references, or footnotes of reporters. In this paper, we explore word position information and then propose two position-dependent language models for speech recognition. The structures and characteristics of these position-dependent language models were extensively investigated, while its performance was analyzed and verified by comparing it with the existing n-gram, mixture- and topic-based language models. The large vocabulary continuous speech recognition (LVCSR) experiments were conducted on the broadcast news transcription task. The preliminary results seem to indicate that the proposed position-dependent models are comparable to the mixture- and topic-based models.
AB - This paper considers word position information for language modeling. For organized documents, such as technical papers or news reports, the composition and the word usage of articles of the same style are usually similar. Therefore, the documents can be separated into partitions consisting of identical rhetoric or topic styles by the literary structures, e.g., introductory remarks, related studies or events, elucidations of methodology or affairs, conclusions of the articles, and references, or footnotes of reporters. In this paper, we explore word position information and then propose two position-dependent language models for speech recognition. The structures and characteristics of these position-dependent language models were extensively investigated, while its performance was analyzed and verified by comparing it with the existing n-gram, mixture- and topic-based language models. The large vocabulary continuous speech recognition (LVCSR) experiments were conducted on the broadcast news transcription task. The preliminary results seem to indicate that the proposed position-dependent models are comparable to the mixture- and topic-based models.
KW - Language model
KW - Language model adaptation
KW - Position information
KW - Speech recognition
KW - Topic information
UR - http://www.scopus.com/inward/record.url?scp=60849103094&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=60849103094&partnerID=8YFLogxK
U2 - 10.1109/CHINSL.2008.ECP.37
DO - 10.1109/CHINSL.2008.ECP.37
M3 - Conference contribution
AN - SCOPUS:60849103094
SN - 9781424429431
T3 - Proceedings - 2008 6th International Symposium on Chinese Spoken Language Processing, ISCSLP 2008
SP - 101
EP - 104
BT - Proceedings - 2008 6th International Symposium on Chinese Spoken Language Processing, ISCSLP 2008
T2 - 2008 6th International Symposium on Chinese Spoken Language Processing, ISCSLP 2008
Y2 - 16 December 2008 through 19 December 2008
ER -