Distribution-based feature normalization for robust speech recognition leveraging context and dynamics cues

Yu Chen Kao, Berlin Chen

Research output: Contribution to journalConference article

Abstract

Recently, histogram equalization (HEQ) of speech features has received considerable attention in the area of robust speech recognition because of its relative simplicity and good empirical performance. In this paper, we present a novel extension to the conventional HEQ approach in two significant aspects. First, polynomial regression of various orders is employed to efficiently perform feature normalization building up the notion of HEQ. Second, not only the contextual distributional statistics but also the dynamics of feature values are taken as the input to the presented regression functions for better normalization performance. By doing so, we can to some extent relax the dimension-independence and bag-offrames assumptions made by the conventional HEQ approach. All experiments were carried out on the Aurora-2 database and task and further verified on the Aurora-4 database and task. The corresponding results demonstrate that our proposed methods can achieve considerable word error rate reductions over the baseline systems and offer additional performance gains for the AFE-processed features.

Original languageEnglish
Pages (from-to)2958-2962
Number of pages5
JournalProceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH
Publication statusPublished - 2013 Jan 1
Event14th Annual Conference of the International Speech Communication Association, INTERSPEECH 2013 - Lyon, France
Duration: 2013 Aug 252013 Aug 29

Keywords

  • Contextual information
  • Histogram equalization
  • Noise robustness
  • Spatial-temporal
  • Speech recognition

ASJC Scopus subject areas

  • Language and Linguistics
  • Human-Computer Interaction
  • Signal Processing
  • Software
  • Modelling and Simulation

Fingerprint Dive into the research topics of 'Distribution-based feature normalization for robust speech recognition leveraging context and dynamics cues'. Together they form a unique fingerprint.

  • Cite this