TY - JOUR
T1 - Leveraging relevance cues for language modeling in speech recognition
AU - Chen, Berlin
AU - Chen, Kuan Yu
N1 - Funding Information:
This work was sponsored in part by “Aim for the Top University Plan” of National Taiwan Normal University and Ministry of Education, Taiwan, and the National Science Council, Taiwan, under Grants NSC 101-2221-E-003-024-MY3, NSC 101-2511-S-003-057-MY3, NSC 101-2511-S-003-047-MY3, NSC 99-2221-E-003-017-MY3, and NSC 98-2221-E-003-011-MY3.
PY - 2013
Y1 - 2013
N2 - Language modeling (LM), providing a principled mechanism to associate quantitative scores to sequences of words or tokens, has long been an interesting yet challenging problem in the field of speech and language processing. The n-gram model is still the predominant method, while a number of disparate LM methods, exploring either lexical co-occurrence or topic cues, have been developed to complement the n-gram model with some success. In this paper, we explore a novel language modeling framework built on top of the notion of relevance for speech recognition, where the relationship between a search history and the word being predicted is discovered through different granularities of semantic context for relevance modeling. Empirical experiments on a large vocabulary continuous speech recognition (LVCSR) task seem to demonstrate that the various language models deduced from our framework are very comparable to existing language models both in terms of perplexity and recognition error rate reductions.
AB - Language modeling (LM), providing a principled mechanism to associate quantitative scores to sequences of words or tokens, has long been an interesting yet challenging problem in the field of speech and language processing. The n-gram model is still the predominant method, while a number of disparate LM methods, exploring either lexical co-occurrence or topic cues, have been developed to complement the n-gram model with some success. In this paper, we explore a novel language modeling framework built on top of the notion of relevance for speech recognition, where the relationship between a search history and the word being predicted is discovered through different granularities of semantic context for relevance modeling. Empirical experiments on a large vocabulary continuous speech recognition (LVCSR) task seem to demonstrate that the various language models deduced from our framework are very comparable to existing language models both in terms of perplexity and recognition error rate reductions.
KW - Information retrieval
KW - Language model
KW - Relevance model
KW - Speech recognition
KW - Topic model
UR - http://www.scopus.com/inward/record.url?scp=84874836936&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84874836936&partnerID=8YFLogxK
U2 - 10.1016/j.ipm.2013.01.005
DO - 10.1016/j.ipm.2013.01.005
M3 - Article
AN - SCOPUS:84874836936
SN - 0306-4573
VL - 49
SP - 807
EP - 816
JO - Information Processing and Management
JF - Information Processing and Management
IS - 4
ER -