TY - GEN
T1 - Exploiting graph regularized nonnegative matrix factorization for extractive speech summarization
AU - Liu, Shih Hung
AU - Chen, Kuan Yu
AU - Hsieh, Yu Lun
AU - Chen, Berlin
AU - Wang, Hsin Min
AU - Yen, Hsu Chun
AU - Hsu, Wen Lian
N1 - Publisher Copyright:
© 2016 Asia Pacific Signal and Information Processing Association.
PY - 2017/1/17
Y1 - 2017/1/17
N2 - Extractive summarization systems attempt to automatically pick out representative sentences from a source text or spoken document and concatenate them into a concise summary so as to help people grasp salient information effectively and efficiently. Recent advances in applying nonnegative matrix factorization (NMF) on various tasks including summarization motivate us to extend this line of research and provide the following contributions. First, we propose to employ graph-regularized nonnegative matrix factorization (GNMF), in which an affinity graph with its similarity measure tailored to the evaluation metric of summarization is constructed and in turn serves as a neighborhood preserving constraint of NMF, so as to better represent the semantic space of sentences in the document to be summarized. Second, we further consider sparsity and orthogonality constraints on NMF and GNMF for better selection of representative sentences to form a summary. Extensive experiments conducted on a Mandarin broadcast news speech dataset demonstrate the effectiveness of the proposed unsupervised summarization models, in relation to several widely-used state-of-the-art methods compared in the paper.
AB - Extractive summarization systems attempt to automatically pick out representative sentences from a source text or spoken document and concatenate them into a concise summary so as to help people grasp salient information effectively and efficiently. Recent advances in applying nonnegative matrix factorization (NMF) on various tasks including summarization motivate us to extend this line of research and provide the following contributions. First, we propose to employ graph-regularized nonnegative matrix factorization (GNMF), in which an affinity graph with its similarity measure tailored to the evaluation metric of summarization is constructed and in turn serves as a neighborhood preserving constraint of NMF, so as to better represent the semantic space of sentences in the document to be summarized. Second, we further consider sparsity and orthogonality constraints on NMF and GNMF for better selection of representative sentences to form a summary. Extensive experiments conducted on a Mandarin broadcast news speech dataset demonstrate the effectiveness of the proposed unsupervised summarization models, in relation to several widely-used state-of-the-art methods compared in the paper.
UR - http://www.scopus.com/inward/record.url?scp=85013825275&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85013825275&partnerID=8YFLogxK
U2 - 10.1109/APSIPA.2016.7820883
DO - 10.1109/APSIPA.2016.7820883
M3 - Conference contribution
AN - SCOPUS:85013825275
T3 - 2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference, APSIPA 2016
BT - 2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference, APSIPA 2016
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference, APSIPA 2016
Y2 - 13 December 2016 through 16 December 2016
ER -