TY - JOUR
T1 - Combining Relevance Language Modeling and Clarity Measure for Extractive Speech Summarization
AU - Liu, Shih Hung
AU - Chen, Kuan Yu
AU - Chen, Berlin
AU - Wang, Hsin Min
AU - Yen, Hsu Chun
AU - Hsu, Wen Lian
N1 - Publisher Copyright:
© 2014 IEEE.
PY - 2015/6/1
Y1 - 2015/6/1
N2 - Extractive speech summarization, which purports to select an indicative set of sentences from a spoken document so as to succinctly represent the most important aspects of the document, has garnered much research over the years. In this paper, we cast extractive speech summarization as an ad-hoc information retrieval (IR) problem and investigate various language modeling (LM) methods for important sentence selection. The main contributions of this paper are four-fold. First, we explore a novel sentence modeling paradigm built on top of the notion of relevance, where the relationship between a candidate summary sentence and a spoken document to be summarized is discovered through different granularities of context for relevance modeling. Second, not only lexical but also topical cues inherent in the spoken document are exploited for sentence modeling. Third, we propose a novel clarity measure for use in important sentence selection, which can help quantify the thematic specificity of each individual sentence that is deemed to be a crucial indicator orthogonal to the relevance measure provided by the LM-based methods. Fourth, in an attempt to lessen summarization performance degradation caused by imperfect speech recognition, we investigate making use of different levels of index features for LM-based sentence modeling, including words, subword-level units, and their combination. Experiments on broadcast news summarization seem to demonstrate the performance merits of our methods when compared to several existing well-developed and/or state-of-the-art methods.
AB - Extractive speech summarization, which purports to select an indicative set of sentences from a spoken document so as to succinctly represent the most important aspects of the document, has garnered much research over the years. In this paper, we cast extractive speech summarization as an ad-hoc information retrieval (IR) problem and investigate various language modeling (LM) methods for important sentence selection. The main contributions of this paper are four-fold. First, we explore a novel sentence modeling paradigm built on top of the notion of relevance, where the relationship between a candidate summary sentence and a spoken document to be summarized is discovered through different granularities of context for relevance modeling. Second, not only lexical but also topical cues inherent in the spoken document are exploited for sentence modeling. Third, we propose a novel clarity measure for use in important sentence selection, which can help quantify the thematic specificity of each individual sentence that is deemed to be a crucial indicator orthogonal to the relevance measure provided by the LM-based methods. Fourth, in an attempt to lessen summarization performance degradation caused by imperfect speech recognition, we investigate making use of different levels of index features for LM-based sentence modeling, including words, subword-level units, and their combination. Experiments on broadcast news summarization seem to demonstrate the performance merits of our methods when compared to several existing well-developed and/or state-of-the-art methods.
KW - Clarity measure
KW - KL divergence
KW - language modeling
KW - relevance modeling
KW - speech summarization
UR - http://www.scopus.com/inward/record.url?scp=84928124743&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84928124743&partnerID=8YFLogxK
U2 - 10.1109/TASLP.2015.2414820
DO - 10.1109/TASLP.2015.2414820
M3 - Article
AN - SCOPUS:84928124743
SN - 1558-7916
VL - 23
SP - 957
EP - 969
JO - IEEE Transactions on Audio, Speech and Language Processing
JF - IEEE Transactions on Audio, Speech and Language Processing
IS - 6
M1 - 7063924
ER -