Speaker Name: | Jackson Liscombe |
Speaker Info: | Graduate Student, NLP Group; jaxin@cs.columbia.edu |
Date: | Thursday October 14th |
Time: | 11:30am-12:30pm |
Location: | CS Conference Room |
Abstract:
In this talk I will present work from my summer internship at AT&T
Shannon Labs. It is joint work with Giuseppe Riccardi and Dilek
Hakkani-Tur. The research entailed the exploration of the affective component of human-machine communication. In state-of-the-art spoken dialog systems this dimension is usually ignored though it plays a major role in engaging users in communicating with machines. Past research in the prediction of human emotion in spoken language has tended to exploit only isolated speech utterances (be they acted or spontaneous), using little or no contextual information. Over the summer we explored the dynamic nature of emotion as it is manifested in human-machine spoken communication. We augmented traditional emotion indicators, such as prosody and language, with contextual information. Contextual information consists of temporal variation of emotion indicators inter- and intra-utterance. We observed 1.2% relative improvement in the prediction of negative user state using dialog acts over lexical and prosodic information alone and 3.8% improvement using additional innovative contextual features. Overall performance accuracy using all features was 79%. |