A Bilingual Multimodal Speech Communication Corpus incorporating acoustic data as well as visual data related to face, hands and arms gestures during speech, is presented in this paper. This corpus comprises different speaking modalities, including scripted text speech, natural conversation, and free speech. The corpus has been compiled in two different languages, viz., French and Spanish. The different experimental setups for the recording of the corpus, the acquisition protocols, and the employed equipment are described. Statistics regarding the number and gender of the speakers, number of words, number of sentences, and duration of the recording sessions, are also provided. Preliminary results from the analysis of the correlation among speech, head and hand movements during spontaneous speech are also presented in this paper, showing that acoustic prosodic features are related with head and hand gestures.
To download please contact mauriciocerda@med.uchile.cl
1 Lucas D. Terissi, 1 Gonzalo Sad, 2 Mauricio Cerda, 3 Slim Ouni, 2 Rodrigo Galvez, 1Juan C. Gómez, 3 Bernard Girau, 4 Nancy Hitschfeld-Kahler
1 CIFASIS-CONICET, Universidad Nacional de Rosario, Argentina.
2 SCIAN-Lab, Faculty of Medicine, Universidad de Chile, Santiago, Chile.
3 Université de Lorraine, LORIA, UMR 7503, Villers-lès-Nancy, France,
4 Computer Science Department, FCFyM, Universidad de Chile, Santiago, Chile.
Download Camera Ready Interspeech 2018 paper