Fusion of Smile, Valence and NGram Features for Automatic Affect Detection

This paper addresses the problem of feature fusion between smile, as a visual feature, and text, as a transcription result. The influence of smile over semantic data has been considered before, without investigating multiple approaches for the fusion. This problem is multi-modal, which makes it more...

Full description

Saved in:
Bibliographic Details
Published inInternational Conference on Affective Computing and Intelligent Interaction and workshops pp. 264 - 269
Main Authors Serban, Ovidiu, Castellano, Ginevra, Pauchet, Alexandre, Rogozan, Alexandrina, Pecuchet, Jean-Pierre
Format Conference Proceeding
LanguageEnglish
Published IEEE 01.09.2013
Subjects
Online AccessGet full text
ISSN2156-8103
DOI10.1109/ACII.2013.50

Cover

Loading…
More Information
Summary:This paper addresses the problem of feature fusion between smile, as a visual feature, and text, as a transcription result. The influence of smile over semantic data has been considered before, without investigating multiple approaches for the fusion. This problem is multi-modal, which makes it more difficult. The goal of this article is to investigate how this fusion could increase the current interactivity of a dialogue system by boosting the automatic detection rate of the sentiments expressed by a human user. There are two original propositions in our approach. The first lies in the use of a segmented detection for text data, rather than predicting a single label for every document (video). Second, this paper studies the importance of several features in the process of multi-modal fusion. Our approach uses basic features, such as NGrams, Smile Presence or Valence to find the best fusion approach. Moreover, we test a two level classification approach, using a SVM.
ISSN:2156-8103
DOI:10.1109/ACII.2013.50