首页 | 本学科首页   官方微博 | 高级检索  
     


Psychoacoustic cues to emotion in speech prosody and music
Authors:Eduardo Coutinho  Nicola Dibben
Affiliation:1. Swiss Centre for Affective Sciences , University of Geneva , Geneva , Switzerland;2. School of Music , University of Liverpool , Liverpool , UK eduardo.coutinho@unige.ch;4. Music Department , University of Sheffield , Sheffield , UK
Abstract:There is strong evidence of shared acoustic profiles common to the expression of emotions in music and speech, yet relatively limited understanding of the specific psychoacoustic features involved. This study combined a controlled experiment and computational modelling to investigate the perceptual codes associated with the expression of emotion in the acoustic domain. The empirical stage of the study provided continuous human ratings of emotions perceived in excerpts of film music and natural speech samples. The computational stage created a computer model that retrieves the relevant information from the acoustic stimuli and makes predictions about the emotional expressiveness of speech and music close to the responses of human subjects. We show that a significant part of the listeners’ second-by-second reported emotions to music and speech prosody can be predicted from a set of seven psychoacoustic features: loudness, tempo/speech rate, melody/prosody contour, spectral centroid, spectral flux, sharpness, and roughness. The implications of these results are discussed in the context of cross-modal similarities in the communication of emotion in the acoustic domain.
Keywords:Emotion  Arousal and valence  Music  Speech prosody  Psychoacoustics  Neural networks
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号