98%
921
2 minutes
20
Over the past 35 years, it has been established that mental representations of language include fine-grained acoustic details stored in episodic memory. The empirical foundations of this fact were established through a series of word recognition experiments showing that participants were better at remembering words repeated by the same talker than words repeated by a different talker (talker-specificity effect). This effect has been widely replicated, but exclusively with isolated, generally monosyllabic, words as the object of study. Whether fine-grained acoustic detail plays a role in the encoding and retrieval of larger structures, such as spoken sentences, has important implications for theories of language understanding in natural communicative contexts. In this study, we extended traditional recognition memory methods to use full spoken sentences rather than individual words as stimuli. Additionally, we manipulated attention at the time of encoding in order to probe the automaticity of fine-grained acoustic encoding. Participants were more accurate for sentences repeated by the same talker than by a different talker. They were also faster and more accurate in the Full Attention than in the Divided Attention condition. The specificity effect was more pronounced for the Divided Attention than the Full Attention group. These findings provide evidence for specificity at the sentence level. They also highlight the implicit, automatic encoding of fine-grained acoustic detail and point to a central role for cognitive resource allocation in shaping memory-based language representations.
Download full-text PDF |
Source |
---|---|
http://dx.doi.org/10.3758/s13423-025-02751-0 | DOI Listing |
J Acoust Soc Am
September 2025
IMVIA UR 7535, Université Bourgogne Europe, 21000 Dijon, France.
The narrowband components of ship-radiated noise are critical for the passive detection and identification of ship targets. However, the intricate underwater environment poses challenges for conventional acoustic signal processing methods, particularly at low signal-to-noise ratios. Previous studies have suggested the use of deep learning for denoising, but there is a significant lack of research on underwater narrowband signals.
View Article and Find Full Text PDFSci Rep
September 2025
Sanko School, Gaziantep, Turkey.
This study presents a novel privacy-preserving deep learning framework for accurately classifying fine-grained hygiene and water-usage events in restroom environments. Leveraging a comprehensive, curated dataset comprising approximately 460 min of stereo audio recordings from five acoustically diverse bathrooms, our method robustly identifies 11 distinct events, including nuanced variations in faucet counts and flow rates, toilet flushing, and handwashing activities. Stereo audio inputs were transformed into triple-channel Mel spectrograms using an adaptive one-dimensional convolutional neural network (1D-CNN), dynamically synthesizing spatial cues to enhance discriminative power.
View Article and Find Full Text PDFPsychon Bull Rev
August 2025
Department of Linguistics, Stanford University, Building 460, Margaret Jacks Hall 450 Jane Stanford Way, Stanford, CA, 94305, USA.
Over the past 35 years, it has been established that mental representations of language include fine-grained acoustic details stored in episodic memory. The empirical foundations of this fact were established through a series of word recognition experiments showing that participants were better at remembering words repeated by the same talker than words repeated by a different talker (talker-specificity effect). This effect has been widely replicated, but exclusively with isolated, generally monosyllabic, words as the object of study.
View Article and Find Full Text PDFAnimals (Basel)
July 2025
College of Information Engineering, Sichuan Agriculture University, Ya'an 625014, China.
This research presents DuSAFNet, a lightweight deep neural network for fine-grained bird audio classification. DuSAFNet combines dual-path feature fusion, spectral-temporal attention, and a multi-band ArcMarginProduct classifier to enhance inter-class separability and capture both local and global spectro-temporal cues. Unlike single-feature approaches, DuSAFNet captures both local spectral textures and long-range temporal dependencies in Mel-spectrogram inputs and explicitly enhances inter-class separability across low, mid, and high frequency bands.
View Article and Find Full Text PDFIEEE J Biomed Health Inform
August 2025
Acoustic features are crucial behavioral indicators for depression detection. However, prior speech-based depression detection methods often overlook the variability of emotional patterns across samples, leading to interference from speaker identity and hindering the effective extraction of emotional changes. To address this limitation, we developed the Emotional Word Reading Experiment (EWRE) and introduced a method combining self-supervised and supervised learning for depression detection from speech called MFE-Former.
View Article and Find Full Text PDF