98%
921
2 minutes
20
Recent advances in unsupervised domain adaptation have shown that mitigating the domain divergence by extracting the domain-invariant features could significantly improve the generalization of a model with respect to a new data domain. However, current methodologies often neglect to retain domain private information, which is the unique information inherent to the unlabeled new domain, compromising generalization. This paper presents a novel method that utilizes mutual information to protect this domain-specific information, ensuring that the latent features of the unlabeled data not only remain domain-invariant but also reflect the unique statistics of the unlabeled domain. We show that simultaneous maximization of mutual information and reduction of domain divergence can effectively preserve domain-private information. We further illustrate that a neural estimator can aptly estimate the mutual information between the unlabeled input space and its latent feature space. Both theoretical analysis and empirical results validate the significance of preserving such unique information of the unlabeled domain for cross-domain generalization. Comparative evaluations reveal our method's superiority over existing state-of-the-art techniques across multiple benchmark datasets.
Download full-text PDF |
Source |
---|---|
http://dx.doi.org/10.1016/j.neunet.2024.106112 | DOI Listing |
Disabil Rehabil Assist Technol
September 2025
School of Foreign Languages, Ningbo University of Technology, Ningbo, China.
The speech and language rehabilitation are essential to people who have disorders of communication that may occur due to the condition of neurological disorder, developmental delays, or bodily disabilities. With the advent of deep learning, we introduce an improved multimodal rehabilitation pipeline that incorporates audio, video, and text information in order to provide patient-tailored therapy that adapts to the patient. The technique uses a cross-attention fusion multimodal hierarchical transformer architectural model that allows it to jointly design speech acoustics as well as the facial dynamics, lip articulation, and linguistic context.
View Article and Find Full Text PDFComput Med Imaging Graph
September 2025
Shanghai Key Laboratory of Trustworthy Computing, East China Normal University, Shanghai 200062, China.
An increasing number of publicly available datasets have facilitated the exploration of building universal medical segmentation models. Existing approaches address partially labeled problem of each dataset by harmonizing labels across datasets and independently focusing on the labeled foreground regions. However, significant challenges persist, particularly in the form of cross-site domain shifts and the limited utilization of partially labeled datasets.
View Article and Find Full Text PDFbioRxiv
August 2025
Department of Electrical Engineering and Computer Science, Florida Atlantic University, Boca Raton, FL, USA.
Identifying genomic regions shaped by natural selection is a central goal in evolutionary genomics. Existing machine learning methods for this task are typically trained using simulated genomic data labeled according to specific evolutionary scenarios. While effective in controlled settings, these models are limited by their reliance on explicit class labels.
View Article and Find Full Text PDFInt J Neural Syst
October 2025
College of Computer Science, Sichuan University, Section 4, Southern 1st Ring Rd, Chengdu, Sichuan 610065, P. R. China.
Semi-supervised semantic segmentation for medical images has evolved through time. While it can leverage the unlabeled data to significantly improve the segmentation performance, it still suffers the problems of intra-class variance and the consequent class-domain distribution misalignment along with costly training. In this paper, a stability-aware dual-head architecture is proposed to synergize prototype-based and Fully Convolutional Network (FCN) methodologies.
View Article and Find Full Text PDFIEEE Trans Pattern Anal Mach Intell
September 2025
Prompt tuning, a recently emerging paradigm, adapts vision-language pre-trained models to new tasks efficiently by learning "soft prompts" for frozen models. However, in few-shot scenarios, its effectiveness is limited by sensitivity to the initialization and the time-consuming search for optimal initialization, hindering rapid adaptation. Additionally, prompt tuning risks reducing the models' generalizability due to overfitting on scarce training samples.
View Article and Find Full Text PDF