AFZAL, S., SEZGIN, T. M., GAO, Y., AND ROBINSON, P. 2009. Perception of emotional expressions in different representations using facial feature points. In Affective Computing and Intelligent Interaction and Workshops, 2009. ACII 2009. 3rd International Conference on, IEEE, 1-6.
AUBERGÉ, V., AND BAILLY, G. 1995. Generation of intonation: a global approach. In EUROSPEECH.
BAILLY, G., AND HOLM, B. 2005. Sfc: a trainable prosodic model. Speech Communication 46, 3, 348-364.
BAILLY, G., BARBE, T., AND WANG, H.-D. 1991. Automatic labeling of large prosodic databases: Tools, methodology and links with a text-to-speech system. In The ESCA Workshop on Speech Synthesis.
BARBOSA, P., AND BAILLY, G. 1994. Characterisation of rhythmic patterns for text-to-speech synthesis. Speech Communication 15, 1, 127-137.
BARBULESCU, A., HUEBER, T., BAILLY, G., RONFARD, R., ET AL. 2013. Audio-visual speaker conversion using prosody features. In International Conference on Auditory-Visual Speech Processing.
BARON-COHEN, S. 2003. Mind reading: the interactive guide to emotions. Jessica Kingsley Publishers.
BEN YOUSSEF, A., SHIMODAIRA, H., AND BRAUDE, D. A. 2013. Articulatory features for speech-driven head motion synthesis. Proceedings of Interspeech, Lyon, France.
BERNDT, D. J., AND CLIFFORD, J. 1994. Using dynamic time warping to find patterns in time series. In KDD workshop, vol. 10, Seattle, WA, 359-370.
BOERSMA, P. 2002. Praat, a system for doing phonetics by computer. Glot international 5, 9/10, 341-345.
BOLINGER, D. 1989. Intonation and its uses: Melody in grammar and discourse. Stanford University Press.
BUSSO, C., DENG, Z., GRIMM, M., NEUMANN, U., AND NARAYANAN, S. 2007. Rigid head motion in expressive speech animation: Analysis and synthesis. Audio, Speech, and Language Processing, IEEE Transactions on 15, 3, 1075-1086.
CASSELL, J., PELACHAUD, C., BADLER, N., STEEDMAN, M., ACHORN, B., BECKET, T., DOUVILLE, B., PREVOST, S., AND STONE, M. 1994. Animated conversation: Rule-based generation of facial expression, gesture & spoken intonation for multiple conversational agents. In Proceedings of the 21st Annual Conference on Computer Graphics and Interactive Techniques, ACM, New York, NY, USA, SIGGRAPH '94, 413-420.
CHUANG, E., AND BREGLER, C. 2005. Mood swings: expressive speech animation. ACM Transactions on Graphics (TOG) 24, 2, 331-347.
DE MORAES, J. A., RILLIARD, A., DE OLIVEIRA MOTA, B. A., AND SHOCHI, T. 2010. Multimodal perception and production of attitudinal meaning in brazilian portuguese. Proc. Speech Prosody, paper 340.
EBERLY, D. H. 2001. 3D game engine design. San Francisco: Morgan Kaufmann Publishers, Inc.
EKMAN, P. 1992. An argument for basic emotions. Cognition & Emotion 6, 3-4, 169-200.
FÓNAGY, I., BÉRARD, E., AND FÓNAGY, J. 1983. Clichés mélodiques. Folia linguistica 17, 1-4, 153-186.
INANOGLU, Z., AND YOUNG, S. 2007. A system for transforming the emotion in speech: combining data-driven conversion techniques for prosody and voice quality. In INTERSPEECH, 490-493.
KRAHMER, E., AND SWERTS, M. 2009. Audiovisual prosody-introduction to the special issue. Language and speech 52, 2-3, 129-133.
MADDEN, M. 2005. 99 ways to tell a story: exercises in style. Chamberlain Bros.
MORLEC, Y., BAILLY, G., AND AUBERGÉ, V. 2001. Generating prosodic attitudes in french: data, model and evaluation. Speech Communication 33, 4, 357-371.
MOULINES, E., AND CHARPENTIER, F. 1990. Pitch-synchronous waveform processing techniques for text-to-speech synthesis using diphones. Speech communication 9, 5, 453-467.
QUENEAU, R. 2013. Exercises in style. New Directions Publishing.
RILLIARD, A., MARTIN, J.-C., AUBERGÉ, V., SHOCHI, T., ET AL. 2008. Perception of french audio-visual prosodic attitudes. Speech Prosody, Campinas, Brasil.
SCHERER, K. R., AND ELLGRING, H. 2007. Multimodal expression of emotion: Affect programs or componential appraisal patterns? Emotion 7, 1, 158.
SCHNITZLER, A. 1993. Reigen. Invito alla lettura. Impresor.
SHOEMAKE, K. 1985. Animating rotation with quaternion curves. In ACM SIGGRAPH computer graphics, vol. 19, ACM, 245-254.
SUMMERFIELD, Q. 1992. Lipreading and audio-visual speech perception. Philosophical Transactions of the Royal Society of London. Series B: Biological Sciences 335, 1273, 71-78.
TAO, J., KANG, Y., AND LI, A. 2006. Prosody conversion from neutral speech to emotional speech. Audio, Speech, and Language Processing, IEEE Transactions on 14, 4, 1145-1154.
VROOMEN, J., COLLIER, R., AND MOZZICONACCI, S. J. 1993. Duration and intonation in emotional speech. In Eurospeech.
YEHIA, H., KURATATE, T., AND VATIKIOTIS-BATESON, E. 2000. Facial animation and head motion driven by speech acoustics. In 5th Seminar on Speech Production: Models and Data, Kloster Seeon, Germany, 265-268.
ZENG, Z., PANTIC, M., ROISMAN, G. I., AND HUANG, T. S. 2009. A survey of affect recognition methods: Audio, visual, and spontaneous expressions. Pattern Analysis and Machine Intelligence, IEEE Transactions on 31, 1, 39-58.