Positional language modeling for extractive broadcast news speech summarization

Shih Hung Liu, Kuan Yu Chen, Berlin Chen, Hsin Min Wang, Hsu Chun Yen, Wen Lian Hsu

Research output: Contribution to journalConference article

7 Citations (Scopus)

Abstract

Extractive summarization, with the intention of automatically selecting a set of representative sentences from a text (or spoken) document so as to concisely express the most important theme of the document, has been an active area of experimentation and development. A recent trend of research is to employ the language modeling (LM) approach for important sentence selection, which has proven to be effective for performing extractive summarization in an unsupervised fashion. However, one of the major challenges facing the LM approach is how to formulate the sentence models and estimate their parameters more accurately for each text (or spoken) document to be summarized. This paper extends this line of research and its contributions are three-fold. First, we propose a positional language modeling framework using different granularities of position-specific information to better estimate the sentence models involved in summarization. Second, we also explore to integrate the positional cues into relevance modeling through a pseudo-relevance feedback procedure. Third, the utilities of the various methods originated from our proposed framework and several wellestablished unsupervised methods are analyzed and compared extensively. Empirical evaluations conducted on a broadcast news summarization task seem to demonstrate the performance merits of our summarization methods.

Original languageEnglish
Pages (from-to)2729-2733
Number of pages5
JournalProceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH
Volume2015-January
Publication statusPublished - 2015 Jan 1
Event16th Annual Conference of the International Speech Communication Association, INTERSPEECH 2015 - Dresden, Germany
Duration: 2015 Sep 62015 Sep 10

Fingerprint

Language Modeling
Summarization
Broadcast
Feedback
Pseudo-relevance Feedback
Threefolds
Granularity
Estimate
Experimentation
Express
Integrate
Speech
News Broadcasts
Speech Summarization
Line
Evaluation
Modeling
Model
Demonstrate

Keywords

  • Extractive broadcast news summarization
  • Positional language modeling
  • Relevance modeling

ASJC Scopus subject areas

  • Language and Linguistics
  • Human-Computer Interaction
  • Signal Processing
  • Software
  • Modelling and Simulation

Cite this

Positional language modeling for extractive broadcast news speech summarization. / Liu, Shih Hung; Chen, Kuan Yu; Chen, Berlin; Wang, Hsin Min; Yen, Hsu Chun; Hsu, Wen Lian.

In: Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH, Vol. 2015-January, 01.01.2015, p. 2729-2733.

Research output: Contribution to journalConference article

@article{cd61bd841a0b4af2a7a790c776927d20,
title = "Positional language modeling for extractive broadcast news speech summarization",
abstract = "Extractive summarization, with the intention of automatically selecting a set of representative sentences from a text (or spoken) document so as to concisely express the most important theme of the document, has been an active area of experimentation and development. A recent trend of research is to employ the language modeling (LM) approach for important sentence selection, which has proven to be effective for performing extractive summarization in an unsupervised fashion. However, one of the major challenges facing the LM approach is how to formulate the sentence models and estimate their parameters more accurately for each text (or spoken) document to be summarized. This paper extends this line of research and its contributions are three-fold. First, we propose a positional language modeling framework using different granularities of position-specific information to better estimate the sentence models involved in summarization. Second, we also explore to integrate the positional cues into relevance modeling through a pseudo-relevance feedback procedure. Third, the utilities of the various methods originated from our proposed framework and several wellestablished unsupervised methods are analyzed and compared extensively. Empirical evaluations conducted on a broadcast news summarization task seem to demonstrate the performance merits of our summarization methods.",
keywords = "Extractive broadcast news summarization, Positional language modeling, Relevance modeling",
author = "Liu, {Shih Hung} and Chen, {Kuan Yu} and Berlin Chen and Wang, {Hsin Min} and Yen, {Hsu Chun} and Hsu, {Wen Lian}",
year = "2015",
month = "1",
day = "1",
language = "English",
volume = "2015-January",
pages = "2729--2733",
journal = "Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH",
issn = "2308-457X",

}

TY - JOUR

T1 - Positional language modeling for extractive broadcast news speech summarization

AU - Liu, Shih Hung

AU - Chen, Kuan Yu

AU - Chen, Berlin

AU - Wang, Hsin Min

AU - Yen, Hsu Chun

AU - Hsu, Wen Lian

PY - 2015/1/1

Y1 - 2015/1/1

N2 - Extractive summarization, with the intention of automatically selecting a set of representative sentences from a text (or spoken) document so as to concisely express the most important theme of the document, has been an active area of experimentation and development. A recent trend of research is to employ the language modeling (LM) approach for important sentence selection, which has proven to be effective for performing extractive summarization in an unsupervised fashion. However, one of the major challenges facing the LM approach is how to formulate the sentence models and estimate their parameters more accurately for each text (or spoken) document to be summarized. This paper extends this line of research and its contributions are three-fold. First, we propose a positional language modeling framework using different granularities of position-specific information to better estimate the sentence models involved in summarization. Second, we also explore to integrate the positional cues into relevance modeling through a pseudo-relevance feedback procedure. Third, the utilities of the various methods originated from our proposed framework and several wellestablished unsupervised methods are analyzed and compared extensively. Empirical evaluations conducted on a broadcast news summarization task seem to demonstrate the performance merits of our summarization methods.

AB - Extractive summarization, with the intention of automatically selecting a set of representative sentences from a text (or spoken) document so as to concisely express the most important theme of the document, has been an active area of experimentation and development. A recent trend of research is to employ the language modeling (LM) approach for important sentence selection, which has proven to be effective for performing extractive summarization in an unsupervised fashion. However, one of the major challenges facing the LM approach is how to formulate the sentence models and estimate their parameters more accurately for each text (or spoken) document to be summarized. This paper extends this line of research and its contributions are three-fold. First, we propose a positional language modeling framework using different granularities of position-specific information to better estimate the sentence models involved in summarization. Second, we also explore to integrate the positional cues into relevance modeling through a pseudo-relevance feedback procedure. Third, the utilities of the various methods originated from our proposed framework and several wellestablished unsupervised methods are analyzed and compared extensively. Empirical evaluations conducted on a broadcast news summarization task seem to demonstrate the performance merits of our summarization methods.

KW - Extractive broadcast news summarization

KW - Positional language modeling

KW - Relevance modeling

UR - http://www.scopus.com/inward/record.url?scp=84959107792&partnerID=8YFLogxK

UR - http://www.scopus.com/inward/citedby.url?scp=84959107792&partnerID=8YFLogxK

M3 - Conference article

AN - SCOPUS:84959107792

VL - 2015-January

SP - 2729

EP - 2733

JO - Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH

JF - Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH

SN - 2308-457X

ER -