Emotion classification: How does an automated system compare to Naive human coders?
- Sefik Emre Eskimez ,
- Kenneth Imade ,
- Na Yang ,
- Melissa Sturge-Apple ,
- Zhiyao Duan ,
- Wendi Heinzelman
2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) |
Published by IEEE
The fact that emotions play a vital role in social interactions, along with the demand for novel human-computer interaction applications, have led to the development of a number of automatic emotion classification systems. However, it is still debatable whether the performance of such systems can compare with human coders. To address this issue, in this study, we present a comprehensive comparison in a speech-based emotion classification task between 138 Amazon Mechanical Turk workers (Turkers) and a state-of-the-art automatic computer system. The comparison includes classifying speech utterances into six emotions (happy, neutral, sad, anger, disgust and fear), into three arousal classes (active, passive, and neutral), and into three valence classes (positive, negative, and neutral). The results show that the computer system outperforms the naive Turkers in almost all cases. Furthermore, the computer system can increase the classification accuracy by rejecting to classify utterances for which it is not confident, while the Turkers do not show a significantly higher classification accuracy on their confident utterances versus unconfident ones.