Leveraging kullbackLeibler divergence measures and information-rich cues for speech summarization

Shih Hsiang Lin, Yao Ming Yeh, Berlin Chen

Research output: Contribution to journalArticle

15 Citations (Scopus)

Abstract

Imperfect speech recognition often leads to degraded performance when exploiting conventional text-based methods for speech summarization. To alleviate this problem, this paper investigates various ways to robustly represent the recognition hypotheses of spoken documents beyond the top scoring ones. Moreover, a summarization framework, building on the KullbackLeibler (KL) divergence measure and exploring both the relevance and topical information cues of spoken documents and sentences, is presented to work with such robust representations. Experiments on broadcast news speech summarization tasks appear to demonstrate the utility of the presented approaches.

Original languageEnglish
Article number5549862
Pages (from-to)871-882
Number of pages12
JournalIEEE Transactions on Audio, Speech and Language Processing
Volume19
Issue number4
DOIs
Publication statusPublished - 2011 Apr 6

    Fingerprint

Keywords

  • KullbackLeibler (KL) -divergence
  • multiple recognition hypotheses
  • relevance information
  • speech summarization
  • topical information

ASJC Scopus subject areas

  • Acoustics and Ultrasonics
  • Electrical and Electronic Engineering

Cite this