King's College London

Research portal

Automatic Prediction of Impressions in Time and across Varying Context: Personality, Attractiveness and Likeability

Research output: Contribution to journalArticle

Oya Celiktutan, Hatice Gunes

Original languageEnglish
JournalIEEE Transactions on Affective Computing
Early online date30 Dec 2015
DOIs
Publication statusPublished - Jan 2017

Documents

King's Authors

Abstract

In this paper, we propose a novel multimodal framework for automatically predicting the impressions of extroversion, agreeableness, conscientiousness, neuroticism , openness, attractiveness and likeability continuously in time and across varying situational contexts. Differently from the existing works, we obtain visual-only and audio-only annotations continuously in time for the same set of subjects, for the first time in the literature, and compare them to their audio-visual annotations. We propose a time-continuous prediction approach that learns the temporal relationships rather than treating each time instant separately. Our experiments show that the best prediction results are obtained when regression models are learned from audio-visual annotations and visual cues, and from audio-visual annotations and visual cues combined with audio cues at the decision level. Continuously generated annotations have the potential to provide insight into better understanding which impressions can be formed and predicted more dynamically, varying with situational context, and which ones appear to be more static and stable over time.

Download statistics

No data available

View graph of relations

© 2018 King's College London | Strand | London WC2R 2LS | England | United Kingdom | Tel +44 (0)20 7836 5454