首页> 外文期刊>PLoS Computational Biology >No, There Is No 150 ms Lead of Visual Speech on Auditory Speech, but a Range of Audiovisual Asynchronies Varying from Small Audio Lead to Large Audio Lag
【24h】

No, There Is No 150 ms Lead of Visual Speech on Auditory Speech, but a Range of Audiovisual Asynchronies Varying from Small Audio Lead to Large Audio Lag

机译:不,听觉语音没有150 ms的视觉语音引导,但是视听异步范围从小音频导致大音频滞后

获取原文
           

摘要

An increasing number of neuroscience papers capitalize on the assumption published in this journal that visual speech would be typically 150 ms ahead of auditory speech. It happens that the estimation of audiovisual asynchrony in the reference paper is valid only in very specific cases, for isolated consonant-vowel syllables or at the beginning of a speech utterance, in what we call “preparatory gestures”. However, when syllables are chained in sequences, as they are typically in most parts of a natural speech utterance, asynchrony should be defined in a different way. This is what we call “comodulatory gestures” providing auditory and visual events more or less in synchrony. We provide audiovisual data on sequences of plosive-vowel syllables (pa, ta, ka, ba, da, ga, ma, na) showing that audiovisual synchrony is actually rather precise, varying between 20 ms audio lead and 70 ms audio lag. We show how more complex speech material should result in a range typically varying between 40 ms audio lead and 200 ms audio lag, and we discuss how this natural coordination is reflected in the so-called temporal integration window for audiovisual speech perception. Finally we present a toy model of auditory and audiovisual predictive coding, showing that visual lead is actually not necessary for visual prediction.
机译:越来越多的神经科学论文利用了该杂志上发布的假设,即视觉语音通常比听觉语音提前150毫秒。碰巧,参考文件中的视听异步估计仅在非常特殊的情况下有效,对于孤立的辅音元音节或语音发声的开始,即所谓的“准备手势”。但是,当音节按顺序链接时,正如它们通常在自然语音发声的大多数部分一样,应以不同的方式定义异步。这就是我们所说的“共调手势”,它或多或少地同步提供听觉和视觉事件。我们提供了有关音节音节音节序列(pa,ta,ka,ba,da,ga,ma,na)的视听数据,表明视听同步实际上相当精确,在20 ms音频超前和70 ms音频滞后之间变化。我们展示了更复杂的语音材料如何导致通常在40 ms音频超前和200 ms音频滞后之间的范围内变化,并且我们讨论了这种自然协调如何在所谓的用于视听语音感知的时间积分窗口中反映出来。最后,我们提出了听觉和视听预测编码的玩具模型,表明视觉线索实际上并不是视觉预测所必需的。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号