Audiovisual synchrony detection for fluent speech in early childhood: An eye-tracking study

  • Han yu Zhou
  • , Han xue Yang
  • , Zhen Wei
  • , Guo bin Wan
  • , Simon S.Y. Lui
  • , Raymond C.K. Chan*
  • *Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

4 Scopus citations

Abstract

During childhood, the ability to detect audiovisual synchrony gradually sharpens for simple stimuli such as flashbeeps and single syllables. However, little is known about how children perceive synchrony for natural and continuous speech. This study investigated young children's gaze patterns while they were watching movies of two identical speakers telling stories side by side. Only one speaker's lip movements matched the voices and the other one either led or lagged behind the soundtrack by 600 ms. Children aged 3–6 years (n = 94, 52.13% males) showed an overall preference for the synchronous speaker, with no age-related changes in synchrony-detection sensitivity as indicated by similar gaze patterns across ages. However, viewing time to the synchronous speech was significantly longer in the auditory-leading (AL) condition compared with that in the visual-leading (VL) condition, suggesting asymmetric sensitivities for AL versus VL asynchrony have already been established in early childhood. When further examining gaze patterns on dynamic faces, we found that more attention focused on the mouth region was an adaptive strategy to read visual speech signals and thus associated with increased viewing time of the synchronous videos. Attention to detail, one dimension of autistic traits featured by local processing, has been found to be correlated with worse performances in speech synchrony processing. These findings extended previous research by showing the development of speech synchrony perception in young children, and may have implications for clinical populations (e.g., autism) with impaired multisensory integration.

Original languageEnglish
Pages (from-to)409-418
Number of pages10
JournalPsyCh Journal
Volume11
Issue number3
DOIs
StatePublished - Jun 2022
Externally publishedYes

Keywords

  • audiovisual
  • autistic traits
  • eye-tracking
  • speech
  • synchrony detection

Fingerprint

Dive into the research topics of 'Audiovisual synchrony detection for fluent speech in early childhood: An eye-tracking study'. Together they form a unique fingerprint.

Cite this