Search for a command to run...
The last decade has seen a substantial body of literature on the recognition of emotion from speech.However, in comparison to related speech processing tasks such as Automatic Speech and Speaker Recognition, practically no standardised corpora and test-conditions exist to compare performances under exactly the same conditions.Instead a multiplicity of evaluation strategies employed -such as cross-validation or percentage splits without proper instance definition -prevents exact reproducibility.Further, in order to face more realistic scenarios, the community is in desperate need of more spontaneous and less prototypical data.This INTERSPEECH 2009 Emotion Challenge aims at bridging such gaps between excellent research on human emotion recognition from speech and low compatibility of results.The FAU Aibo Emotion Corpus [1] serves as basis with clearly defined test and training partitions incorporating speaker independence and different room acoustics as needed in most reallife settings.This paper introduces the challenge, the corpus, the features, and benchmark results of two popular approaches towards emotion recognition from speech.