TY - JOUR
T1 - A probabilistic generative framework for extractive broadcast news speech summarization
AU - Chen, Yi Ting
AU - Chen, Berlin
AU - Wang, Hsin Min
N1 - Funding Information:
Manuscript received November 15, 2007; revised July 20, 2008. Current version published December 11, 2008. This work was supported in part by the National Science Council of Taiwan under Grants NSC95-2221-E-003-014-MY3 and NSC95-2422-H-001-031. The associate editor coordinating the review of this manuscript and approving it for publication was Dr. Ruhi Sarikaya.
PY - 2009/1
Y1 - 2009/1
N2 - In this paper, we consider extractive summarization of broadcast news speech and propose a unified probabilistic generative framework that combines the sentence generative probability and the sentence prior probability for sentence ranking. Each sentence of a spoken document to be summarized is treated as a probabilistic generative model for predicting the document. Two matching strategies, namely literal term matching and concept matching, are thoroughly investigated. We explore the use of the language model (LM) and the relevance model (RM) for literal term matching, while the sentence topical mixture model (STMM) and the word topical mixture model (WTMM) are used for concept matching. In addition, the lexical and prosodic features, as well as the relevance information of spoken sentences, are properly incorporated for the estimation of the sentence prior probability. An elegant feature of our proposed framework is that both the sentence generative probability and the sentence prior probability can be estimated in an unsupervised manner, without the need for handcrafted document-summary pairs. The experiments were performed on Chinese broadcast news collected in Taiwan, and very encouraging results were obtained.
AB - In this paper, we consider extractive summarization of broadcast news speech and propose a unified probabilistic generative framework that combines the sentence generative probability and the sentence prior probability for sentence ranking. Each sentence of a spoken document to be summarized is treated as a probabilistic generative model for predicting the document. Two matching strategies, namely literal term matching and concept matching, are thoroughly investigated. We explore the use of the language model (LM) and the relevance model (RM) for literal term matching, while the sentence topical mixture model (STMM) and the word topical mixture model (WTMM) are used for concept matching. In addition, the lexical and prosodic features, as well as the relevance information of spoken sentences, are properly incorporated for the estimation of the sentence prior probability. An elegant feature of our proposed framework is that both the sentence generative probability and the sentence prior probability can be estimated in an unsupervised manner, without the need for handcrafted document-summary pairs. The experiments were performed on Chinese broadcast news collected in Taiwan, and very encouraging results were obtained.
KW - Extractive spoken document summarization
KW - Language model (LM)
KW - Probabilistic generative framework
KW - Relevance model (RM)
KW - Topical mixture model
UR - http://www.scopus.com/inward/record.url?scp=67149133555&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=67149133555&partnerID=8YFLogxK
U2 - 10.1109/TASL.2008.2005031
DO - 10.1109/TASL.2008.2005031
M3 - Article
AN - SCOPUS:67149133555
SN - 1558-7916
VL - 17
SP - 95
EP - 106
JO - IEEE Transactions on Audio, Speech and Language Processing
JF - IEEE Transactions on Audio, Speech and Language Processing
IS - 1
M1 - 4717223
ER -