Cross-Corpus Speech Emotion Recognition Based on Deep Domain-Adaptive Convolutional Neural Network

  • LIU Jiateng
    Key Laboratory of Child Development and Learning Science of Ministry of Education, Southeast University
  • ZHENG Wenming
    Key Laboratory of Child Development and Learning Science of Ministry of Education, Southeast University
  • ZONG Yuan
    Key Laboratory of Child Development and Learning Science of Ministry of Education, Southeast University
  • LU Cheng
    School of Information Science and Engineering, Southeast University
  • TANG Chuangao
    Key Laboratory of Child Development and Learning Science of Ministry of Education, Southeast University

抄録

<p>In this letter, we propose a novel deep domain-adaptive convolutional neural network (DDACNN) model to handle the challenging cross-corpus speech emotion recognition (SER) problem. The framework of the DDACNN model consists of two components: a feature extraction model based on a deep convolutional neural network (DCNN) and a domain-adaptive (DA) layer added in the DCNN utilizing the maximum mean discrepancy (MMD) criterion. We use labeled spectrograms from source speech corpus combined with unlabeled spectrograms from target speech corpus as the input of two classic DCNNs to extract the emotional features of speech, and train the model with a special mixed loss combined with a cross-entrophy loss and an MMD loss. Compared to other classic cross-corpus SER methods, the major advantage of the DDACNN model is that it can extract robust speech features which are time-frequency related by spectrograms and narrow the discrepancies between feature distribution of source corpus and target corpus to get better cross-corpus performance. Through several cross-corpus SER experiments, our DDACNN achieved the state-of-the-art performance on three public emotion speech corpora and is proved to handle the cross-corpus SER problem efficiently.</p>

収録刊行物

被引用文献 (1)*注記

もっと見る

参考文献 (18)*注記

もっと見る

詳細情報 詳細情報について

問題の指摘

ページトップへ