Cascading appearance-based features for visual voice activity detection
Rajitha Navarathna, David Dean, Patrick Lucey, Sridha Sridharan, Clinton Fookes
Audio-visual speech recognition system for a robot
Takami Yoshida, Kazuhiro Nakadai
Audio-visual television broadcast programs processing, transcription, indexing and searching
Josef Chaloupka, Jan Nouza
Decision fusion by boosting method for multi-modal voice activity detection
Shin'ichi Takeuchi, Takashi Hashiba, Satoshi Tamura, Satoru Hayamizu
A study of influence of word lip reading by change of frame rate
Takeshi Saitoh, Ryosuke Konishi
Detection of specific mispronunciations using audiovisual features
Sébastien Picard, G. Ananthakrishnan, Preben Wik, Olov Engwall, Sherif Abdou
Improving visual features for lip-reading
Yuxuan Lan, Barry-John Theobald, Richard Harvey, Eng-Jon Ong, Richard Bowden
Disentangling unisensory from fusion effects in the attentional modulation of Mcgurk effects: a Bayesian modeling study suggests that fusion is attention-dependent
Jean-Luc Schwartz, Kaisa Tiippana, Tobias S. Andersen
Is there a mcgurk effect for tongue reading?
Olov Engwall
The Mcgurk illusion in the oddity task
Tobias S. Andersen
Abstracting visual prosody across speakers and face areas
Erin Cvejic, Jeesun Kim, Chris Davis
Emotion perception by eye and ear and halves and wholes
Jeesun Kim, Chris Davis
Cross-cultural differences in the multisensory perception of emotion
Akihiro Tanaka, Ai Koizumi, Hisato Imai, Saori Hiramatsu, Eriko Hiramoto, Beatrice de Gelder
Long-term cochlear implant users have resistance to noise, but short-term users don’t
Yori Kanekama, Satoko Hisanaga, Kaoru Sekiyama, Narihiro Kodama, Yasuhiro Samejima, Takao Yamada, Eiji Yumoto
Production of Mandarin lexical tones: auditory and visual components
Virginie Attina, Guillaume Gibert, Eric Vatikiotis-Bateson, Denis Burnham
Limitations of visual speech recognition
Jacob L. Newman, Barry-John Theobald, Stephen J. Cox
Investigating the role of the Lombard reflex in visual and audiovisual speech recognition
Panikos Heracleous, Miki Sato, Carlos T. Ishi, Norihiro Hagita
Acoustic head gesture recognition and its applications
Akira Sasou, Yasuharu Hashimoto, Katsuhiko Sakaue
Evaluation of real-time audio-visual speech recognition
Peng Shen, Satoshi Tamura, Satoru Hayamizu
Real-time audio-visual voice activity detection for speech recognition in noisy environments
Carlos T. Ishi, Miki Sato, Norihiro Hagita, Shihong Lao
CENSREC-1-AV: an audio-visual corpus for noisy bimodal speech recognition
Satoshi Tamura, Chiyomi Miyajima, Norihide Kitaoka, Takeshi Yamada, Satoru Tsuge, Tetsuya Takiguchi, Kazumasa Yamamoto, Takanobu Nishiura, Masato Nakayama, Yuki Denda, Masakiyo Fujimoto, Shigeki Matsuda, Tetsuji Ogawa, Shingo Kuroiwa, Kazuya Takeda, Satoshi Nakamura
Exploring visual features through Gabor representations for facial expression detection
Sien W. Chew, Patrick Lucey, Sridha Sridharan, Clinton Fookes
Towards a true acoustic-visual speech synthesis
Asterios Toutios, Utpala Musti, Slim Ouni, Vincent Colotte, Brigitte Wrobel-Dautcourt, Marie-Odile Berger
Building speaker-specific lip models for talking heads from 3d face data
Takaaki Kuratate, Marcia Riley
Brain regions differentially involved with multisensory and visual only speech gesture information
Daniel E. Callan
Impact of language on audiovisual speech perception examined by fMRI
Jun Shinozaki, Kaoru Sekiyama, obuo Hiroe, Taku Yoshioka, Masa-aki Sato
An ERP examination of audiovisual speech perception in Japanese younger and older adults
Satoko Hisanaga, Kaoru Sekiyama, Tomohiko Igasaki, Nobuki Murayama
Infants match auditory and visual speech in schematic point-light displays
Christine Kitamura, Jeesun Kim
I can see what you said: infant sensitivity to articulator congruency between audio-only and silent-video presentations of native and nonnative consonants
Catherine T. Best, Christian Kroos, Julia Irwin
Optimized photorealistic audiovisual speech synthesis using active appearance modeling
Wesley Mattheyses, Lukas Latacz, Werner Verhelst
In pursuit of visemes
Sarah Hilder, Barry-John Theobald, Richard Harvey
Acoustic-to-articulatory inversion in speech based on statistical models
Atef Ben Youssef, Pierre Badin, Gérard Bailly
The Mcgurk effect at various auditory signal-to-noise ratios in american and Finnish listeners
Kaisa Tiippana, Erin Hayes, Riikka Möttönen, Nina Kraus, Mikko Sams
Binding and unbinding in audiovisual speech fusion: removing the Mcgurk effect by an incoherent preceding audiovisual context
Olha Nahorna, Frédéric Berthommier, Jean-Luc Schwartz
Role of form and motion information in auditory-visual speech perception of Mcgurk combinations and fusions
Guillaume Gibert, Andrew Fordyce, Catherine J. Stevens
Speech-specificity of two audiovisual integration effects
Kasper Eskelund, Jyrki Tuomainen, Tobias S. Andersen
The multimodal analysis for understanding child behavior focused on attention-catching
Shogo Ishikawa, Shinya Kiriyama, Yoichi Takebayashi, Shigeyoshi Kitazawa
A study of speech interface for living space adapting to user environment by considering scenery situation
Kenichi Shibata, Shinya Kiriyama, Tomohiro Haraikawa, Yoichi Takebayashi, Shigeyoshi Kitazawa
Effects of speech-rate conversion on asynchrony perception of audio-visual speech
Shiho Miyazawa, Akihiro Tanaka, Shuichi Sakamoto, Takehiko Nishimoto
The effects of anxiety on the perception of emotion in the face and voice
Ai Koizumi, Akihiro Tanaka, Hisato Imai, Saori Hiramatsu, Eriko Hiramoto, Takao Sato, Beatrice de Gelder
d-o-e-s-not-c-o-m-p-u-t-e: vowel hyperarticulation in speech to an auditory-visual avatar
Denis Burnham, Sebastian Joeffry, Lauren Rice
| Article |
|---|