A position-aware language modeling framework for Extractive broadcast news speech summarization

Shih Hung Liu, Kuan Yu Chen, Yu Lun Hsieh, Berlin Chen, Hsin Min Wang, Hsu Chun Yen, Wen Lian Hsu

Research output: Contribution to journalArticle

1 Citation (Scopus)

Abstract

Extractive summarization, a process that automatically picks exemplary sentences from a text (or spoken) document with the goal of concisely conveying key information therein, has seen a surge of attention from scholars and practitioners recently. Using a language modeling (LM) approach for sentence selection has been proven effective for performing unsupervised extractive summarization. However, one of the major difficulties facing the LM approach is to model sentences and estimate their parameters more accurately for each text (or spoken) document. We extend this line of research and make the following contributions in this work. First, we propose a position-aware language modeling framework using various granularities of position-specific information to better estimate the sentence models involved in the summarization process. Second, we explore disparate ways to integrate the positional cues into relevance models through a pseudo-relevance feedback procedure. Third, we extensively evaluate various models originated from our proposed framework and several well-established unsupervised methods. Empirical evaluation conducted on a broadcast news summarization task further demonstrates performance merits of the proposed summarization methods.

Original languageEnglish
Article number27
JournalACM Transactions on Asian and Low-Resource Language Information Processing
Volume16
Issue number4
DOIs
Publication statusPublished - 2017 Aug

Fingerprint

Conveying
Feedback

Keywords

  • Extractive summarization
  • Positional language modeling
  • Relevance modeling
  • Speech information

ASJC Scopus subject areas

  • Computer Science(all)

Cite this

A position-aware language modeling framework for Extractive broadcast news speech summarization. / Liu, Shih Hung; Chen, Kuan Yu; Hsieh, Yu Lun; Chen, Berlin; Wang, Hsin Min; Yen, Hsu Chun; Hsu, Wen Lian.

In: ACM Transactions on Asian and Low-Resource Language Information Processing, Vol. 16, No. 4, 27, 08.2017.

Research output: Contribution to journalArticle

Liu, Shih Hung ; Chen, Kuan Yu ; Hsieh, Yu Lun ; Chen, Berlin ; Wang, Hsin Min ; Yen, Hsu Chun ; Hsu, Wen Lian. / A position-aware language modeling framework for Extractive broadcast news speech summarization. In: ACM Transactions on Asian and Low-Resource Language Information Processing. 2017 ; Vol. 16, No. 4.
@article{29a3766a1c83444c814832cb53495349,
title = "A position-aware language modeling framework for Extractive broadcast news speech summarization",
abstract = "Extractive summarization, a process that automatically picks exemplary sentences from a text (or spoken) document with the goal of concisely conveying key information therein, has seen a surge of attention from scholars and practitioners recently. Using a language modeling (LM) approach for sentence selection has been proven effective for performing unsupervised extractive summarization. However, one of the major difficulties facing the LM approach is to model sentences and estimate their parameters more accurately for each text (or spoken) document. We extend this line of research and make the following contributions in this work. First, we propose a position-aware language modeling framework using various granularities of position-specific information to better estimate the sentence models involved in the summarization process. Second, we explore disparate ways to integrate the positional cues into relevance models through a pseudo-relevance feedback procedure. Third, we extensively evaluate various models originated from our proposed framework and several well-established unsupervised methods. Empirical evaluation conducted on a broadcast news summarization task further demonstrates performance merits of the proposed summarization methods.",
keywords = "Extractive summarization, Positional language modeling, Relevance modeling, Speech information",
author = "Liu, {Shih Hung} and Chen, {Kuan Yu} and Hsieh, {Yu Lun} and Berlin Chen and Wang, {Hsin Min} and Yen, {Hsu Chun} and Hsu, {Wen Lian}",
year = "2017",
month = "8",
doi = "10.1145/3099472",
language = "English",
volume = "16",
journal = "ACM Transactions on Asian and Low-Resource Language Information Processing",
issn = "2375-4699",
publisher = "Association for Computing Machinery (ACM)",
number = "4",

}

TY - JOUR

T1 - A position-aware language modeling framework for Extractive broadcast news speech summarization

AU - Liu, Shih Hung

AU - Chen, Kuan Yu

AU - Hsieh, Yu Lun

AU - Chen, Berlin

AU - Wang, Hsin Min

AU - Yen, Hsu Chun

AU - Hsu, Wen Lian

PY - 2017/8

Y1 - 2017/8

N2 - Extractive summarization, a process that automatically picks exemplary sentences from a text (or spoken) document with the goal of concisely conveying key information therein, has seen a surge of attention from scholars and practitioners recently. Using a language modeling (LM) approach for sentence selection has been proven effective for performing unsupervised extractive summarization. However, one of the major difficulties facing the LM approach is to model sentences and estimate their parameters more accurately for each text (or spoken) document. We extend this line of research and make the following contributions in this work. First, we propose a position-aware language modeling framework using various granularities of position-specific information to better estimate the sentence models involved in the summarization process. Second, we explore disparate ways to integrate the positional cues into relevance models through a pseudo-relevance feedback procedure. Third, we extensively evaluate various models originated from our proposed framework and several well-established unsupervised methods. Empirical evaluation conducted on a broadcast news summarization task further demonstrates performance merits of the proposed summarization methods.

AB - Extractive summarization, a process that automatically picks exemplary sentences from a text (or spoken) document with the goal of concisely conveying key information therein, has seen a surge of attention from scholars and practitioners recently. Using a language modeling (LM) approach for sentence selection has been proven effective for performing unsupervised extractive summarization. However, one of the major difficulties facing the LM approach is to model sentences and estimate their parameters more accurately for each text (or spoken) document. We extend this line of research and make the following contributions in this work. First, we propose a position-aware language modeling framework using various granularities of position-specific information to better estimate the sentence models involved in the summarization process. Second, we explore disparate ways to integrate the positional cues into relevance models through a pseudo-relevance feedback procedure. Third, we extensively evaluate various models originated from our proposed framework and several well-established unsupervised methods. Empirical evaluation conducted on a broadcast news summarization task further demonstrates performance merits of the proposed summarization methods.

KW - Extractive summarization

KW - Positional language modeling

KW - Relevance modeling

KW - Speech information

UR - http://www.scopus.com/inward/record.url?scp=85028536651&partnerID=8YFLogxK

UR - http://www.scopus.com/inward/citedby.url?scp=85028536651&partnerID=8YFLogxK

U2 - 10.1145/3099472

DO - 10.1145/3099472

M3 - Article

AN - SCOPUS:85028536651

VL - 16

JO - ACM Transactions on Asian and Low-Resource Language Information Processing

JF - ACM Transactions on Asian and Low-Resource Language Information Processing

SN - 2375-4699

IS - 4

M1 - 27

ER -