2005 Presentations (Communicative Events)
Using Context to Improve Emotion Detection in Spoken Dialog Systems
Most research that explores the emotional state of users of spoken dialog systems does not fully utilize the contextual nature that the dialog structure provides. This paper reports results of machine learning experiments designed to automatically classify the emotional state of user turns using a corpus of 5,690 dialogs collected with the "How May I Help You" spoken dialog system. We show that augmenting standard lexical and prosodic features with contextual features that exploit the structure of spoken dialog and track user state increases classification accuracy by 2.6%.
Subjects
Files
-
liscombe_al_05a.pdf application/pdf 50 KB Download File
More About This Work
- Academic Units
- Computer Science
- Publisher
- Proceedings of Eurospeech'05
- Published Here
- June 4, 2013