A position-aware language modeling framework for Extractive broadcast news speech summarization

Shih Hung Liu, Kuan Yu Chen, Yu Lun Hsieh, Berlin Chen, Hsin Min Wang, Hsu Chun Yen, Wen Lian Hsu

研究成果: 雜誌貢獻文章同行評審

2 引文 斯高帕斯(Scopus)

摘要

Extractive summarization, a process that automatically picks exemplary sentences from a text (or spoken) document with the goal of concisely conveying key information therein, has seen a surge of attention from scholars and practitioners recently. Using a language modeling (LM) approach for sentence selection has been proven effective for performing unsupervised extractive summarization. However, one of the major difficulties facing the LM approach is to model sentences and estimate their parameters more accurately for each text (or spoken) document. We extend this line of research and make the following contributions in this work. First, we propose a position-aware language modeling framework using various granularities of position-specific information to better estimate the sentence models involved in the summarization process. Second, we explore disparate ways to integrate the positional cues into relevance models through a pseudo-relevance feedback procedure. Third, we extensively evaluate various models originated from our proposed framework and several well-established unsupervised methods. Empirical evaluation conducted on a broadcast news summarization task further demonstrates performance merits of the proposed summarization methods.

原文英語
文章編號27
期刊ACM Transactions on Asian and Low-Resource Language Information Processing
16
發行號4
DOIs
出版狀態已發佈 - 2017 八月

ASJC Scopus subject areas

  • Computer Science(all)

指紋 深入研究「A position-aware language modeling framework for Extractive broadcast news speech summarization」主題。共同形成了獨特的指紋。

引用此