Talker-specificity beyond the lexicon: Recognition memory for spoken sentences.

Psychon Bull Rev

Department of Linguistics, Stanford University, Building 460, Margaret Jacks Hall 450 Jane Stanford Way, Stanford, CA, 94305, USA.

Published: August 2025


Category Ranking

98%

Total Visits

921

Avg Visit Duration

2 minutes

Citations

20

Article Abstract

Over the past 35 years, it has been established that mental representations of language include fine-grained acoustic details stored in episodic memory. The empirical foundations of this fact were established through a series of word recognition experiments showing that participants were better at remembering words repeated by the same talker than words repeated by a different talker (talker-specificity effect). This effect has been widely replicated, but exclusively with isolated, generally monosyllabic, words as the object of study. Whether fine-grained acoustic detail plays a role in the encoding and retrieval of larger structures, such as spoken sentences, has important implications for theories of language understanding in natural communicative contexts. In this study, we extended traditional recognition memory methods to use full spoken sentences rather than individual words as stimuli. Additionally, we manipulated attention at the time of encoding in order to probe the automaticity of fine-grained acoustic encoding. Participants were more accurate for sentences repeated by the same talker than by a different talker. They were also faster and more accurate in the Full Attention than in the Divided Attention condition. The specificity effect was more pronounced for the Divided Attention than the Full Attention group. These findings provide evidence for specificity at the sentence level. They also highlight the implicit, automatic encoding of fine-grained acoustic detail and point to a central role for cognitive resource allocation in shaping memory-based language representations.

Download full-text PDF

Source
http://dx.doi.org/10.3758/s13423-025-02751-0DOI Listing

Publication Analysis

Top Keywords

fine-grained acoustic
16
spoken sentences
12
repeated talker
12
recognition memory
8
acoustic detail
8
full attention
8
divided attention
8
attention
5
talker-specificity lexicon
4
lexicon recognition
4

Similar Publications

The narrowband components of ship-radiated noise are critical for the passive detection and identification of ship targets. However, the intricate underwater environment poses challenges for conventional acoustic signal processing methods, particularly at low signal-to-noise ratios. Previous studies have suggested the use of deep learning for denoising, but there is a significant lack of research on underwater narrowband signals.

View Article and Find Full Text PDF

This study presents a novel privacy-preserving deep learning framework for accurately classifying fine-grained hygiene and water-usage events in restroom environments. Leveraging a comprehensive, curated dataset comprising approximately 460 min of stereo audio recordings from five acoustically diverse bathrooms, our method robustly identifies 11 distinct events, including nuanced variations in faucet counts and flow rates, toilet flushing, and handwashing activities. Stereo audio inputs were transformed into triple-channel Mel spectrograms using an adaptive one-dimensional convolutional neural network (1D-CNN), dynamically synthesizing spatial cues to enhance discriminative power.

View Article and Find Full Text PDF

Talker-specificity beyond the lexicon: Recognition memory for spoken sentences.

Psychon Bull Rev

August 2025

Department of Linguistics, Stanford University, Building 460, Margaret Jacks Hall 450 Jane Stanford Way, Stanford, CA, 94305, USA.

Over the past 35 years, it has been established that mental representations of language include fine-grained acoustic details stored in episodic memory. The empirical foundations of this fact were established through a series of word recognition experiments showing that participants were better at remembering words repeated by the same talker than words repeated by a different talker (talker-specificity effect). This effect has been widely replicated, but exclusively with isolated, generally monosyllabic, words as the object of study.

View Article and Find Full Text PDF

This research presents DuSAFNet, a lightweight deep neural network for fine-grained bird audio classification. DuSAFNet combines dual-path feature fusion, spectral-temporal attention, and a multi-band ArcMarginProduct classifier to enhance inter-class separability and capture both local and global spectro-temporal cues. Unlike single-feature approaches, DuSAFNet captures both local spectral textures and long-range temporal dependencies in Mel-spectrogram inputs and explicitly enhances inter-class separability across low, mid, and high frequency bands.

View Article and Find Full Text PDF

Acoustic features are crucial behavioral indicators for depression detection. However, prior speech-based depression detection methods often overlook the variability of emotional patterns across samples, leading to interference from speaker identity and hindering the effective extraction of emotional changes. To address this limitation, we developed the Emotional Word Reading Experiment (EWRE) and introduced a method combining self-supervised and supervised learning for depression detection from speech called MFE-Former.

View Article and Find Full Text PDF