Towards expressive musical robots: A cross-modal framework for emotional gesture, voice and music
抄録
It has been long speculated that expression of emotions from different modalities have the same underlying 'code', whether it be a dance step, musical phrase, or tone of voice. This is the first attempt to implement this theory across three modalities, inspired by the polyvalence and repeatability of robotics. We propose a unifying framework to generate emotions across voice, gesture, and music, by representing emotional states as a 4-parameter tuple of speed, intensity, regularity, and extent (SIRE). Our results show that a simple 4-tuple can capture four emotions recognizable at greater than chance across gesture and voice, and at least two emotions across all three modalities. An application for multi-modal, expressive music robots is discussed.
収録刊行物
-
- EURASIP Journal on Audio, Speech, and Music Processing
-
EURASIP Journal on Audio, Speech, and Music Processing 2012 2012-01-17
SpringerOpen
- Tweet
詳細情報 詳細情報について
-
- CRID
- 1050564285744979712
-
- NII論文ID
- 120005439173
-
- ISSN
- 16874714
-
- HANDLE
- 2433/187380
-
- 本文言語コード
- en
-
- 資料種別
- journal article
-
- データソース種別
-
- IRDB
- CiNii Articles