@InProceedings{Hönemann2013_123,
author = {Angelika Hönemann and Hansjörg Mixdorff and Sascha Fagel},
booktitle = {Studientexte zur Sprachkommunikation: Elektronische Sprachsignalverarbeitung 2013.},
title = {Alignment between rigid head movements and prosodic landmarks},
year = {2013},
editor = {Petra Wagner},
month = mar,
pages = {181--188},
publisher = {TUDpress, Dresden},
abstract = {In our study we recorded and analyzed an audiovisual speech corpus to develop a model which predicts head and facial non-verbal movements accompanying speech. The model is intended to improve the naturalness of avatars. Our previous paper already gives a preliminary analysis of our speech corpus which includes acoustic and visual recordings of seven individual speakers who talk about three minutes about their last vacation. We showed that for each speaker 20-30% of events in each motion class are aligned with prominent syllables in phrase-initial or -medial position and that the speakers moved most often at the end of an intonation phrase. We also observe that the speakers differ in strength and frequency of visible events. However, there is also a great ratio of about 60% of motion events which are not assigned to the target syllables. In order to account for this result, further analyses had to be carried out. The present paper shows further analyses of the relationship between speech and movements. Therefore, we extracted the fundamental frequency (FO) and the intensity of the acoustic signals using Praat. By marking the prominent syllables we obtained a description of the course of FO. We use the Principle Component Analysis (PCA) to determine the linear combinations of the visual parameters that constitute the main head movements.},
isbn = {978-3-944331-03-4},
issn = {0940-6832},
keywords = {Prosodischer und multimodaler Ausdruck in der Mensch-Maschine Interaktion},
url = {https://www.essv.de/pdf/2013_181_188.pdf},
}