98%
921
2 minutes
20
Background: Depression is major global public health problems among university students. Currently, the evaluation and monitoring of depression predominantly depend on subjective and self-reported methods. There is an urgent necessity to develop objective means of identifying depression. Acoustic features, which convey emotional information, have the potential to enhance the objectivity of depression assessments. This study aimed to investigate the feasibility of utilizing acoustic features for the objective and automated identification and characterization of depression among Chinese university students.
Methods: A cross-sectional study was undertaken involving 103 students with depression and 103 controls matched for age, gender, and education. Participants' voices were recorded using a smartphone as they read neutral texts. Acoustic analysis and feature extraction were performed using the OpenSMILE toolkit, yielding 523 features encompassing spectral, glottal, and prosodic characteristics. These extracted acoustic features were utilized for discriminant analysis between depression and control groups. Pearson correlation analyses were conducted to evaluate the relationship between acoustic features and Patient Health Questionnaire-9 (PHQ-9) scores. Five machine learning algorithms including Linear Discriminant Analysis (LDA), Logistic Regression, Support Vector Classification, Naive Bayes, and Random Forest were used to perform the classification. For training and testing, ten-fold cross-validation was employed. Model performance was assessed using receiver operating characteristic (ROC) curve, area under the curve (AUC), precision, accuracy, recall, and F1 score. Shapley Additive exPlanations (SHAP) method was used for model interpretation.
Results: In depression group, 32 acoustic features (25 spectral features, 5 prosodic features and 2 glottal features) showed significant alterations compared with controls. Further, 27 acoustic features (10 spectral features, 3 prosodic features, and 1 glottal features) were significantly correlated with depression severity. Among five machine learning algorithms, LDA model demonstrated the highest classification performance, with an AUC of 0.771. SHAP analysis suggested that Mel-frequency cepstral coefficients (MFCC) features contributed most to the model's classification efficacy.
Conclusions: The integration of acoustic features and LDA model demonstrates a high accuracy in distinguishing depression among Chinese university students, suggesting its potential utility in rapid and large-scale depression screening. MFCC may serve as objective and valid features for the automated identification of depression on Chinese university campuses.
Download full-text PDF |
Source |
---|---|
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC12119278 | PMC |
http://dx.doi.org/10.3389/fpubh.2025.1561332 | DOI Listing |
Sci Rep
September 2025
Center for Northeast Asian Studies, Tohoku University, 41 Kawauchi, Sendai Aoba-ku, Miyagi, 980-8576, Japan.
Petit-spot volcanism plays a critical role in the metasomatism of oceanic plates prior to subduction and in their recycling into the deep mantle. The extent of metasomatism depends on the number and volume of petit-spot volcanic edifices and intrusions, making precise identification of petit-spot volcanic fields essential. However, conventional methods based on seafloor topography and acoustic backscatter intensity alone have limitations in accurately delineating these features.
View Article and Find Full Text PDFEur J Neurosci
September 2025
The Tampa Human Neurophysiology Lab, Department of Neurosurgery, Brain and Spine, Morsani College of Medicine, University of South Florida, Tampa, Florida, USA.
Sensory areas exhibit modular selectivity to stimuli, but they can also respond to features outside of their basic modality. Several studies have shown cross-modal plastic modifications between visual and auditory cortices; however, the exact mechanisms of these modifications are yet not completely known. To this aim, we investigated the effect of 12 min of visual versus sound adaptation (referring to forceful application of an optimal/nonoptimal stimulus to a neuron[s] under observation) on the infragranular and supragranular primary visual neurons (V1) of the cat (Felis catus).
View Article and Find Full Text PDFJ R Soc Interface
September 2025
ENES Bioacoustics Research Lab, CRNL, CNRS, Inserm, University of Saint-Etienne, Saint-Etienne, France.
Getting caregivers to respond to their pain cries is vital for the human baby. Previous studies have shown that certain features of baby cries-the nonlinear phenomena (NLP)-enable caregivers to assess the pain felt by the baby. However, the extent to which these NLP mobilize the autonomic nervous system of an adult listener remains unexplored.
View Article and Find Full Text PDFJ Acoust Soc Am
September 2025
Instituto Aqualie, Juiz de Fora, MG 36036-330, Brazil.
Beaked whales, deep-diving cetaceans from the family Ziphiidae, exhibit cryptic behaviors, and data on these species in Brazilian waters are limited to strandings and isolated sightings. This study characterizes the occurrence and acoustic behavior of beaked whales in the Foz do Amazonas Basin using combined visual and passive acoustic monitoring along the Brazilian Equatorial Margin. Audio files were analyzed to identify clicks with frequency-modulated pulses, a diagnostic characteristic of beaked whales.
View Article and Find Full Text PDFPLoS One
September 2025
College of Information Engineering, Sichuan Agricultural University, Ya'an, Sichuan Province, China.
Animals communicate information primarily via their calls, and directly using their vocalizations proves essential for executing species conservation and tracking biodiversity. Conventional visual approaches are frequently limited by distance and surroundings, while call-based monitoring concentrates solely on the animals themselves, proving more effective and straightforward than visual techniques. This paper introduces an animal sound classification model named SeqFusionNet, integrating the sequential encoding of Transformer with the global perception of MLP to achieve robust global feature extraction.
View Article and Find Full Text PDF