TY - GEN
T1 - Effective pseudo-relevance feedback for language modeling in speech recognition
AU - Chen, Berlin
AU - Chen, Yi Wen
AU - Chen, Kuan Yu
AU - Jan, Ea Ee
PY - 2013
Y1 - 2013
N2 - A part and parcel of any automatic speech recognition (ASR) system is language modeling (LM), which helps to constrain the acoustic analysis, guide the search through multiple candidate word strings, and quantify the acceptability of the final output hypothesis given an input utterance. Despite the fact that the n-gram model remains the predominant one, a number of novel and ingenious LM methods have been developed to complement or be used in place of the n-gram model. A more recent line of research is to leverage information cues gleaned from pseudo-relevance feedback (PRF) to derive an utterance-regularized language model for complementing the n-gram model. This paper presents a continuation of this general line of research and its main contribution is two-fold. First, we explore an alternative and more efficient formulation to construct such an utterance-regularized language model for ASR. Second, the utilities of various utterance-regularized language models are analyzed and compared extensively. Empirical experiments on a large vocabulary continuous speech recognition (LVCSR) task demonstrate that our proposed language models can offer substantial improvements over the baseline n-gram system, and achieve performance competitive to, or better than, some state-of-the-art language models.
AB - A part and parcel of any automatic speech recognition (ASR) system is language modeling (LM), which helps to constrain the acoustic analysis, guide the search through multiple candidate word strings, and quantify the acceptability of the final output hypothesis given an input utterance. Despite the fact that the n-gram model remains the predominant one, a number of novel and ingenious LM methods have been developed to complement or be used in place of the n-gram model. A more recent line of research is to leverage information cues gleaned from pseudo-relevance feedback (PRF) to derive an utterance-regularized language model for complementing the n-gram model. This paper presents a continuation of this general line of research and its main contribution is two-fold. First, we explore an alternative and more efficient formulation to construct such an utterance-regularized language model for ASR. Second, the utilities of various utterance-regularized language models are analyzed and compared extensively. Empirical experiments on a large vocabulary continuous speech recognition (LVCSR) task demonstrate that our proposed language models can offer substantial improvements over the baseline n-gram system, and achieve performance competitive to, or better than, some state-of-the-art language models.
KW - Speech recognition
KW - information retrieval
KW - language modeling
KW - pseudo-relevance feedback
KW - relevance
UR - http://www.scopus.com/inward/record.url?scp=84893690712&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84893690712&partnerID=8YFLogxK
U2 - 10.1109/ASRU.2013.6707698
DO - 10.1109/ASRU.2013.6707698
M3 - Conference contribution
AN - SCOPUS:84893690712
SN - 9781479927562
T3 - 2013 IEEE Workshop on Automatic Speech Recognition and Understanding, ASRU 2013 - Proceedings
SP - 13
EP - 18
BT - 2013 IEEE Workshop on Automatic Speech Recognition and Understanding, ASRU 2013 - Proceedings
T2 - 2013 IEEE Workshop on Automatic Speech Recognition and Understanding, ASRU 2013
Y2 - 8 December 2013 through 13 December 2013
ER -