Category Ranking

98%

Total Visits

921

Avg Visit Duration

2 minutes

Citations

20

Article Abstract

The comprehension of spoken language benefits from visual speech information. One reason for this is the temporal lead of mouth and lip movements over the onset of acoustic speech utterance. Here, we investigated EEG event-related potentials preceding acoustic speech, focusing on a fronto-central contingent negative variation (CNV) prior to the onset of acoustic speech. We explored influences of expectation and visual speech content as well as age-related differences. In a multi-talker two-alternative speech discrimination task, younger and older subjects responded to short words presented simultaneously to competing speech under free-field conditions. Subjects were always presented with audiovisual speech stimuli, while the modality containing the task-relevant information was modulated in a block-wise fashion. Thus, task-relevant speech information was either available as audio-visually congruent stimuli or only in the visual (visual-valid) or the auditory (auditory-valid) modality. Subjects were instructed to fixate a pre-specified position in the left or right hemispace. In each task block, task-relevant stimuli appeared either at the pre-specified position (standard trials, 80%) or at a rare deviant position (20%). Target words were recognized faster and more accurately when visual speech information was available. The CNV prior to the acoustic speech onset was more pronounced with visual-informative than with visually non-informative speech. Especially in the younger group, a less pronounced CNV occurred with purely visual speech in deviant trials, that is, when a task-irrelevant speech stimulus appeared instead of the expected target stimulus. The results indicate that processes preceding the onset of acoustic speech are modulated by expectations and visual speech content, while age differences are rather small.

Download full-text PDF

Source
http://dx.doi.org/10.1016/j.heares.2022.108636DOI Listing

Publication Analysis

Top Keywords

visual speech
24
acoustic speech
20
speech
17
onset acoustic
12
younger older
8
cnv prior
8
speech content
8
pre-specified position
8
visual
7
acoustic
5

Similar Publications

Prior researches on global-local processing have focused on hierarchical objects in the visual modality, while the real-world involves multisensory interactions. The present study investigated whether the simultaneous presentation of auditory stimuli influences the recognition of visually hierarchical objects. We added four types of auditory stimuli to the traditional visual hierarchical letters paradigm:no sound (visual-only), a pure tone, a spoken letter that was congruent with the required response (response-congruent), or a spoken letter that was incongruent with it (response-incongruent).

View Article and Find Full Text PDF

Purpose: Residual speech sound disorder (RSSD) is a high-prevalence condition that can limit children's academic and social participation, with negative consequences for overall well-being. Previous studies have described visual biofeedback as a promising option for RSSD, but results have been inconclusive due to study design limitations and small sample sizes.

Method: In a preregistered randomized controlled trial, 108 children aged 9-15 years with RSSD affecting American English /ɹ/ were randomly assigned to receive treatment incorporating visual biofeedback (subdivided into ultrasound and visual-acoustic types) or a comparison condition of motor-based treatment consistent with current best practices in speech therapy.

View Article and Find Full Text PDF

Mobile Training for Peers to Support Students With Multiple Disabilities and Cortical Visual Impairment in Communicating Choices.

Am J Speech Lang Pathol

September 2025

Department of Communication Sciences and Disorders, The Pennsylvania State University, University Park.

Purpose: The current study investigated the impact of a short mobile training implemented in peer pairs to teach the Communicating Choices-CVI (Peers) strategy to support interactions with students with multiple disabilities.

Method: A pretest-posttest control group design was used to evaluate the effects of the training created on the INSTRUCT app, which used a checklist of steps with video models to teach elementary-age peers a strategy to structure opportunities for students with multiple disabilities to communicate choices. Peers were randomly assigned to the experimental group ( = 10) or control group ( = 10) and then video-recorded while interacting with students with multiple disabilities during one pretest and one posttest interaction in their typical educational settings.

View Article and Find Full Text PDF

Evaluation of Visual Feedback for f and SPL in Subglottal Pressure Measurements-A Methodological Study.

J Voice

September 2025

Department of Clinical Science, Intervention and Technology (CLINTEC), Division of Speech and Language Pathology, Karolinska Institutet, SE-171 76, Stockholm, Sweden.

Objective: Subglottal pressure is a clinically relevant parameter for assessment of voice disorders and correlates to f and sound pressure level (SPL). The aim of the current study was to evaluate the use of a visual target for feedback of f and SPL in subglottal pressure measurements in habitual voice and at phonation threshold level with a syllable string and a phrase for the purpose of improving the reliability of subglottal pressure measurements.

Methods: Data from 12 vocally healthy women (29-61 years) was analyzed.

View Article and Find Full Text PDF

The speech and language rehabilitation are essential to people who have disorders of communication that may occur due to the condition of neurological disorder, developmental delays, or bodily disabilities. With the advent of deep learning, we introduce an improved multimodal rehabilitation pipeline that incorporates audio, video, and text information in order to provide patient-tailored therapy that adapts to the patient. The technique uses a cross-attention fusion multimodal hierarchical transformer architectural model that allows it to jointly design speech acoustics as well as the facial dynamics, lip articulation, and linguistic context.

View Article and Find Full Text PDF