Deutsch
 
Hilfe Datenschutzhinweis Impressum
  DetailsucheBrowse

Datensatz

DATENSATZ AKTIONENEXPORT

Freigegeben

Poster

Virtual Storytelling of Fairy Tales: Towards Simulation of Emotional Perception of Text

MPG-Autoren
/persons/resource/persons84285

Volkova,  E
Department Human Perception, Cognition and Action, Max Planck Institute for Biological Cybernetics, Max Planck Society;
Max Planck Institute for Biological Cybernetics, Max Planck Society;

Externe Ressourcen
Es sind keine externen Ressourcen hinterlegt
Volltexte (beschränkter Zugriff)
Für Ihren IP-Bereich sind aktuell keine Volltexte freigegeben.
Volltexte (frei zugänglich)
Es sind keine frei zugänglichen Volltexte in PuRe verfügbar
Ergänzendes Material (frei zugänglich)
Es sind keine frei zugänglichen Ergänzenden Materialien verfügbar
Zitation

Volkova, E. (2010). Virtual Storytelling of Fairy Tales: Towards Simulation of Emotional Perception of Text. Poster presented at 11th Conference of Junior Neuroscientists of Tübingen (NeNa 2010), Heiligkreuztal, Germany.


Zitierlink: https://hdl.handle.net/11858/00-001M-0000-0013-BE10-0
Zusammenfassung
Emotion analysis (EA) is a rapidly developing area in computational linguistics. For most EA systems, the number of emotion classes is very limited and the text units the classes are assigned to are discrete and predefined. The question we address is whether the set of emotion
categories can be enriched and whether the units to which the categories are assigned can be more flexibly defined. Six untrained participants annotated a corpus of eight texts having no predetermined annotation units and using fifteen emotional categories. The inter-annotator
agreement rates were considerably high for this difficult task: 0.55 (moderate) on average, reaching 0.82 (almost perfect) with some annotator pairs. The final application of the intended EA system is predominantly in the emotion enhancement of human-computer interaction in virtual reality. The system is meant to be a bridge between unprocessed input text and visual and auditory information, like generated speech, facial expressions and body language. The first steps towards integrating text-based information annotated for emotion categories and
simulation of human emotional perception of texts in story telling scenarios for virtual reality are already made. We have created a virtual character, whose animation of face and body is driven by annotations in text.