Learning to Relate Literal and Sentimental Descriptions of Visual Properties

  • Mark Yatsgar ,
  • Svitlana Volkova ,
  • Asli Celikyilmaz ,
  • ,
  • Luke Zettlemoyer

Proceedings of NAACL-HLT 2013 |

Published by Association for Computational Linguistics | Organized by North American Chapter of the Association for Computational Linguistics

Language can describe our visual world at many levels, including not only what is literally there but also the sentiment that it invokes. In this paper, we study visual language, both literal and sentimental, that describes the overall appearance and style of virtual characters. Sentimental properties, including labels such as “youthful” or “country western,” must be inferred from descriptions of the more literal properties, such as facial features and clothing selection. We present a new dataset, collected to describe Xbox avatars, as well as models for learning the relationships between these avatars and their literal and sentimental descriptions. In a series of experiments, we demonstrate that such learned models can be used for a range of tasks, including predicting sentimental words and using them to rank and build avatars. ogether, these results demonstrate that sentimental language provides a concise (though noisy) means of specifying low-level visual properties.

Publication Downloads

Xbox Avatars Descriptions

March 4, 2019

We introduce a new corpus of descriptions of Xbox avatars created by actual gamers. Each avatar is specified by 19 attributes, including clothing and body type, allowing for more than 10^20 possibilities. Using Amazon Mechanical Turk, we collected literal and sentimental descriptions of complete avatars and many of their component parts. In all, there are over 100K descriptions, including relative and multilingual descriptions, which will support different tasks, such as learning to automatically describe an avatar or generate one from a description.