An artistic and tool-driven approach for believable digital characters
Volker Helzle
How to create a look-a-like avatar pipeline using low-cost equipment
Verónica Costa Orvalho
Audiovisual binding in speech perception
Jean-Luc Schwartz
From text-to-speech (TTS) to talking head - a machine learning approach to A/V speech modeling and rendering.
Frank Soong, Lijuan Wang
Children’s spontaneous emotional expressions while receiving (un)wanted prizes in the presence of peers
Mandy Visser, Emiel Krahmer, Marc Swerts
You can raise your eyebrows, I don’t mind: are monolingual and bilingual infants equally good at learning from the eyes region of a talking face?
Mathilde Fort, Anira Escrichs, Alba Ayneto-Gimeno, Núria Sebastián-Gallés
Comparison of visual speech perception of sampled-based talking heads: adults and children with and without developmental dyslexia
Paula D. Paro Costa, Daniella Batista, Mayara Toffoli, Keila A. Baraldi Knobel, Cíntia Alves Salgado, José Mario De Martino
Cross-modality matching of linguistic prosody in older and younger adults
Simone Simonetti, Jeesun Kim, Chris Davis
“I don't see what you are saying”: reduced visual influence on audiovisual speech integration in children with Specific Language Impairment
Aurélie Huyse, Frédéric Berthommier, Jacqueline Leybaert
Message vs. messenger effects on cross-modal matching for spoken phrases
Catherine T. Best, Christian H. Kroos, Karen E. Mulak, Shaun Halovic, Mathilde Fort, Christine Kitamura
Audiovisual generation of social attitudes from neutral stimuli
Adela Barbulescu, Gérard Bailly, Rémi Ronfard, Maël Pouget
Classification of auditory-visual attitudes in German
Angelika Hönemann, Hansjörg Mixdorff, Albert Rilliard
Delayed auditory feedback with static and dynamic visual feedback
Elizabeth Stelle, Caroline L. Smith, Eric Vatikiotis-Bateson
Visual vs. auditory emotion information: how language and culture affect our bias towards the different modalities
Chee Seng Chong, Jeesun Kim, Chris Davis
Comparison of multisensory display rules in expressing complex emotions between cultures
Sachiko Takagi, Shiho Miyazawa, Elisabeth Huis In 't Veld, Beatrice de Gelder, Akihiro Tanaka
Towards the development of facial and vocal expression database in east Asian and Western cultures
Akihiro Tanaka, Sachiko Takagi, Saori Hiramatsu, Elisabeth Huis In 't Veld, Beatrice de Gelder
The effect of modality and speaking style on the discrimination of non-native phonological and phonetic contrasts in noise
Sarah Fenwick, Chris Davis, Catherine T. Best, Michael D. Tyler
Audio-visual perception of Mandarin lexical tones in AX same-different judgment task
Rui Wang, Biao Zeng, Simon Thompson
Lip animation synthesis: a unified framework for speaking and laughing virtual agent
Yu Ding, Catherine Pelachaud
Comparison of dialect models and phone mappings in HSMM-based visual dialect speech synthesis
Dietmar Schabus, Michael Pucher
HMM-based visual speech synthesis using dynamic visemes
Ausdang Thangthai, Barry-John Theobald
Investigating the impact of artificial enhancement of lip visibility on the intelligibility of spectrally-distorted speech
Najwa Alghamdi, Steve Maddock, Guy J. Brown, Jon Barker
The stability of mouth movements for multiple talkers over multiple sessions
Chris Davis, Jeesun Kim, Vincent Aubanel, Greg Zelic, Yatin Mahajan
Voicing classification of visual speech using convolutional neural networks
Thomas Le Cornu, Ben Milner
Comparison of single-model and multiple-model prediction-based audiovisual fusion
Stavros Petridis, Varun Rajgarhia, Maja Pantic
Finding phonemes: improving machine lip-reading
Helen L. Bear, Richard Harvey, Yuxuan Lan
Discovering patterns in visual speech
Stephen Cox
Improving lip-reading performance for robust audiovisual speech recognition using DNNs
Kwanchiva Thangthai, Richard Harvey, Stephen Cox, Barry-John Theobald
Integration of auditory, labial and manual signals in cued speech perception by deaf adults: an adaptation of the McGurk paradigm
Clémence Bayard, Cécile Colin, Jacqueline Leybart
Explaining the visual and masked-visual advantage in speech perception in noise: the role of visual phonetic cues
Vincent Aubanel, Chris Davis, Jeesun Kim
Analysing the importance of different visual feature coefficients
Danny Websdale, Ben Milner
Auditory and audiovisual close-shadowing in normal and cochlear-implanted hearing impaired subjects
Lucie Scarbel, Denis Beautemps, Jean-Luc Schwartz, Marc Sato
4D Cardiff Conversation Database (4D CCDb): a 4D database of natural, dyadic conversations
Jason Vandeventer, Andrew J. Aubrey, Paul L. Rosin, David Marshall
Improved visual speech synthesis using dynamic viseme k-means clustering and decision trees
Christiaan Rademan, Thomas Niesler
Scattering vs. discrete cosine transform features in visual speech processing
Etienne Marcheret, Gerasimos Potamianos, Josef Vopicka, Vaibhava Goel
Speaker-independent machine lip-reading with speaker-dependent viseme classifiers
Helen L. Bear, Stephen J. Cox, Richard W. Harvey
Stream weight estimation using higher order statistics in multi-modal speech recognition
Kazuto Ukai, Satoshi Tamura, Satoru Hayamizu
Combining acoustic and visual features to detect laughter in adults’ speech
Hrishikesh Rao, Zhefan Ye, Yin Li, Mark A. Clements, Agata Rozga, James M. Rehg
Optimal timing of audio-visual text presentation: the role of attention
Maiko Takahashi, Akihiro Tanaka
Anticipation of turn-Switching in auditory-visual dialogs
Hansjörg Mixdorff, Angelika Hönemann, Jeesun Kim, Chris Davis
The multi-modal nature of trustworthiness perception
Elena Tsankova, Eva Krumhuber, Andrew J. Aubrey, Arvid Kappas, Guido Möllering, David Marshall, Paul L. Rosin
Face-speech sensor fusion for non-invasive stress detection
Vasudev Bethamcherla, Will Paul, Cecilia Ovesdotter Alm, Reynold Bailey, Joe Geigel, Linwei Wang
The development of patterns of gaze to a speaking face
Julia Irwin, Lawrence Brancazio
| Article |
|---|