The 14th International Conference on Auditory-Visual Speech Processing (AVSP2017)
- ISSN 2308-975X (Online)
Editors: Slim Ouni, Chris Davis, Alexandra Jesse, Jonas Beskow - Publisher: KTH
- August 25-26, 2017, Stockholm, Sweden.
Special session : Remembering Eric Vatikiotis-Bateson
- Eric Vatikiotis-Bateson and the Birth of AVSP – Denis Burnham, with input from Laurie Fais, Ruth Campbell, Kevin Munhall, and Phillip Rubin and Editor-in-Chief, Chris Davis – D1.SP1
S1 – Gaze and Handedness
- Acoustic cue variability affects eye movement behaviour during non-native speech perception: a GAMM model – Jessie S. Nixon and Catherine T. Best -D1.S1.1
- The effect of age and hearing loss on partner-directed gaze in a communicative task – Chris Davis, Jeesun Kim, Outi Tuomainen and Valerie Hazan – D1.S1.2
- Referential Gaze Makes a Difference in Spoken Language Comprehension: Human Speaker vs. Virtual Agent Listener Gaze – Eva Maria Nunnemann, Kirsten Bergmann, Helene Kreysa and Pia Knoeferle – D1.S1.3
- The influence of handedness and pointing direction on deictic gestures and speech interaction: Evidence from motion capture data on Polish counting-out rhymes – Katarzyna Stoltmann and Susanne Fuchs – D1.S1.4
- The Influence of Familial Sinistrality on Audiovisual Speech Perception – Sandhya Vinay and Dawn Behne – D1.S1.5
S2 – AV by machines
- Using deep neural networks to estimate tongue movements from speech face motion – Christian Kroos, Rikke Bundgaard-Nielsen, Catherine Best and Mark D. Plumbley – D1.S2.1
- End-to-End Audiovisual Fusion with LSTMs – Stavros Petridis, Yujiang Wang, Zuwei Li and Maja Pantic – D1.S2.2
- Using visual speech information and perceptually motivated loss functions for binary mask estimation – Danny Websdale and Ben Milner – D1.S2.3
- Combining Multiple Views for Visual Speech Recognition – Marina Zimmermann, Mostafa Mehdipour Ghazi, Hazim Kemal Ekenel and Jean-Philippe Thiran – D1.S2.4
- On the quality of an expressive audiovisual corpus: a case study of acted speech – Slim Ouni, Sara Dahmani and Vincent Colotte – D1.S2.5
- Thin slicing to predict viewer impressions of TED Talks – Ailbhe Cullen and Naomi Harte – D1.S2.6
S3 – Lip reading by machines
- Exploring ROI size in deep learning based lipreading – Alexandros Koumparoulis, Gerasimos Potamianos, Youssef Mroueh and Steven J. Rennie – D1.S3.1
- Towards Lipreading Sentences with Active Appearance Models – George Sterpu and Naomi Harte – D1.S3.2
- Lipreading using deep bottleneck features for optical and depth images Satoshi Tamura, Koichi Miyazaki and Satoru Hayamizu – D1.S3.3
- Inner Lips Parameter Estimation based on Adaptive Ellipse Model – Li Liu, Gang Feng and Denis Beautemps – D1.S3.4
S4 – Prosody and Timing
- Processing of visuo-auditory prosodic information in cochlear-implanted patients deaf patients – Pascal Barone, Mathieu Marx and Anne Lasfargues-Delannoy – D2.S4.1
- Acoustic features of multimodal prominences: Do visual beat gestures affect verbal pitch accent realization? – Gilbert Ambrazaitis and David House – D2.S4.2
- Contribution of visual rhythmic information to speech perception in noise – Vincent Aubanel, Cassandra Masters, Jeesun Kim and Chris Davis – D2.S4.3
- Perceived Audiovisual Simultaneity in Speech by Musicians and Nonmusicians: Preliminary Behavioral and Event-Related Potential (ERP) Findings – Dawn Behne, Marzieh Sorati and Magnus Alm – D2.S4.4
S5 – Emotion & Attitudes
- The developmental path of multisensory perception of emotion and phoneme in Japanese speakers – Hisako W. Yamamoto, Misako Kawahara and Akihiro Tanaka – D2.S5.1
- Impact of Culture on the Development of Multisensory Emotion Perception – Misako Kawahara, Disa Sauter and Akihiro Tanaka – D2.S5.2
- Multisensory Perception of Emotion for Human and Chimpanzee Expressions by Humans – Marina Kawase, Ikuma Adachi and Akihiro Tanaka – D2.S5.3
- Cross-Language Perception of Audio-visual Attitudinal Expressions – Hansjörg Mixdorff, Angelika Hönemann, Albert Rilliard, Tan Lee and Matthew Ma – D2.S5.4
- Facial activity of attitudinal speech in German – Angelika Hoenemann and Petra Wagner – D2.S5.5
S6 – AV Perception
- The McGurk Effect: Auditory Visual Speech Perception’s Piltdown Man – Dominic Massaro – D2.S6.1
- Impact of early bilingualism on infants’ ability to process talking and non-talking faces: new data from 9-month-old infants – Mathilde Fort and Núria Sebastián-Gallés – D2.S6.2
- Atypical phonemic discrimination but not audiovisual speech integration in children with autism and the broader autism phenotype – Julia Irwin, Trey Avery, Jacqueline Turcios, Lawrence Brancazio, Barbara Cook and Nicole Landi – D2.S6.3
- Learning to recognize unfamiliar talkers from the word-level dynamics of visual speech – Alexandra Jesse and Paul Saba – D2.S6.4
- Applying the summation model in audiovisual speech perception – Kaisa Tiippana, Ilmari Kurki and Tarja Peromaa – D2.S6.5