HMM-Based Mask Estimation for a Speech Recognition Front-End Using Computational Auditory Scene Analysis

  • PARK Ji Hun
    Department of Information and Communications, Gwangju Institute of Science and Technology (GIST)
  • YOON Jae Sam
    Department of Information and Communications, Gwangju Institute of Science and Technology (GIST)
  • KIM Hong Kook
    Department of Information and Communications, Gwangju Institute of Science and Technology (GIST) The Institute of Electronics, Information and Communication Engineers

この論文をさがす

抄録

In this paper, we propose a new mask estimation method for the computational auditory scene analysis (CASA) of speech using two microphones. The proposed method is based on a hidden Markov model (HMM) in order to incorporate an observation that the mask information should be correlated over contiguous analysis frames. In other words, HMM is used to estimate the mask information represented as the interaural time difference (ITD) and the interaural level difference (ILD) of two channel signals, and the estimated mask information is finally employed in the separation of desired speech from noisy speech. To show the effectiveness of the proposed mask estimation, we then compare the performance of the proposed method with that of a Gaussian kernel-based estimation method in terms of the performance of speech recognition. As a result, the proposed HMM-based mask estimation method provided an average word error rate reduction of 61.4% when compared with the Gaussian kernel-based mask estimation method.

収録刊行物

参考文献 (9)*注記

もっと見る

詳細情報 詳細情報について

問題の指摘

ページトップへ