98%
921
2 minutes
20
This study presents a novel privacy-preserving deep learning framework for accurately classifying fine-grained hygiene and water-usage events in restroom environments. Leveraging a comprehensive, curated dataset comprising approximately 460 min of stereo audio recordings from five acoustically diverse bathrooms, our method robustly identifies 11 distinct events, including nuanced variations in faucet counts and flow rates, toilet flushing, and handwashing activities. Stereo audio inputs were transformed into triple-channel Mel spectrograms using an adaptive one-dimensional convolutional neural network (1D-CNN), dynamically synthesizing spatial cues to enhance discriminative power. Extensive experimentation identified the RegNetY-008 architecture as the most effective backbone, further improved by employing a semi-supervised learning strategy via pseudo-labeling and targeted data augmentation techniques such as XY masking and horizontal CutMix. The proposed ensemble model, combining RegNetY-008 networks with complementary third-channel generation strategies, achieved outstanding generalization performance, yielding an accuracy of 97.8% and macro-averaged F1-score of 0.966 across acoustically distinct test environments. Our publicly available dataset addresses critical gaps in existing resources, promoting future research in intelligent, privacy-conscious restroom monitoring.
Download full-text PDF |
Source |
---|---|
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC12402066 | PMC |
http://dx.doi.org/10.1038/s41598-025-18154-z | DOI Listing |
Sci Rep
September 2025
Sanko School, Gaziantep, Turkey.
This study presents a novel privacy-preserving deep learning framework for accurately classifying fine-grained hygiene and water-usage events in restroom environments. Leveraging a comprehensive, curated dataset comprising approximately 460 min of stereo audio recordings from five acoustically diverse bathrooms, our method robustly identifies 11 distinct events, including nuanced variations in faucet counts and flow rates, toilet flushing, and handwashing activities. Stereo audio inputs were transformed into triple-channel Mel spectrograms using an adaptive one-dimensional convolutional neural network (1D-CNN), dynamically synthesizing spatial cues to enhance discriminative power.
View Article and Find Full Text PDFSci Rep
August 2025
School of Electrical and Electronic Engineering, Universidad del Valle, Cali, Colombia.
IEEE Trans Pattern Anal Mach Intell
August 2025
This paper introduces Stereo-Talker, a novel one-shot audio-driven human video synthesis system that generates 3D talking videos with precise lip synchronization, expressive body gestures, temporally consistent photo-realistic quality, and continuous viewpoint control. The process follows a two-stage approach. In the first stage, the system maps audio input to high-fidelity motion sequences, encompassing upper-body gestures and facial expressions.
View Article and Find Full Text PDFJ Vis Exp
May 2025
Department of Life Sciences, School of Agriculture, Meiji University;
In Arabidopsis seeds, the endosperm, a single layer of living cells located between the embryo and the testa, plays a critical role in regulating seed maturation, dormancy, and germination. Microscopic analysis of intact endosperm cells is essential for understanding the physiological functions of the endosperm at cellular and molecular levels. However, sample preparation has been challenging due to the small size of Arabidopsis seeds and the location of the endosperm cell layer beneath the testa.
View Article and Find Full Text PDFSci Rep
May 2025
School of Electrical and Electronic Engineering, Universidad del Valle, Cali, Colombia.
This work presents an embedded solution for detecting and classifying head-level objects using stereo vision to assist blind individuals. A custom dataset was created, featuring five classes of head-level objects, selected based on a survey of visually impaired users. Object detection and classification were achieved using deep-neural networks such as YoloV5.
View Article and Find Full Text PDF