Preserving domain private information via mutual information maximization.

Neural Netw

Department of Mechanical Engineering, University of British Columbia, Vancouver, BC, Canada. Electronic address:

Published: April 2024


Category Ranking

98%

Total Visits

921

Avg Visit Duration

2 minutes

Citations

20

Article Abstract

Recent advances in unsupervised domain adaptation have shown that mitigating the domain divergence by extracting the domain-invariant features could significantly improve the generalization of a model with respect to a new data domain. However, current methodologies often neglect to retain domain private information, which is the unique information inherent to the unlabeled new domain, compromising generalization. This paper presents a novel method that utilizes mutual information to protect this domain-specific information, ensuring that the latent features of the unlabeled data not only remain domain-invariant but also reflect the unique statistics of the unlabeled domain. We show that simultaneous maximization of mutual information and reduction of domain divergence can effectively preserve domain-private information. We further illustrate that a neural estimator can aptly estimate the mutual information between the unlabeled input space and its latent feature space. Both theoretical analysis and empirical results validate the significance of preserving such unique information of the unlabeled domain for cross-domain generalization. Comparative evaluations reveal our method's superiority over existing state-of-the-art techniques across multiple benchmark datasets.

Download full-text PDF

Source
http://dx.doi.org/10.1016/j.neunet.2024.106112DOI Listing

Publication Analysis

Top Keywords

unlabeled domain
12
domain private
8
domain
8
domain divergence
8
unlabeled
5
preserving domain
4
mutual
4
private mutual
4
mutual maximization
4
maximization advances
4

Similar Publications

The speech and language rehabilitation are essential to people who have disorders of communication that may occur due to the condition of neurological disorder, developmental delays, or bodily disabilities. With the advent of deep learning, we introduce an improved multimodal rehabilitation pipeline that incorporates audio, video, and text information in order to provide patient-tailored therapy that adapts to the patient. The technique uses a cross-attention fusion multimodal hierarchical transformer architectural model that allows it to jointly design speech acoustics as well as the facial dynamics, lip articulation, and linguistic context.

View Article and Find Full Text PDF

Towards Generic Abdominal Multi-Organ Segmentation with multiple partially labeled datasets.

Comput Med Imaging Graph

September 2025

Shanghai Key Laboratory of Trustworthy Computing, East China Normal University, Shanghai 200062, China.

An increasing number of publicly available datasets have facilitated the exploration of building universal medical segmentation models. Existing approaches address partially labeled problem of each dataset by harmonizing labels across datasets and independently focusing on the labeled foreground regions. However, significant challenges persist, particularly in the form of cross-site domain shifts and the limited utilization of partially labeled datasets.

View Article and Find Full Text PDF

Identifying genomic regions shaped by natural selection is a central goal in evolutionary genomics. Existing machine learning methods for this task are typically trained using simulated genomic data labeled according to specific evolutionary scenarios. While effective in controlled settings, these models are limited by their reliance on explicit class labels.

View Article and Find Full Text PDF

A Stability-Aware Dual-Head Network with Prototype-Based Consistency for Semi-Supervised Medical Image Segmentation.

Int J Neural Syst

October 2025

College of Computer Science, Sichuan University, Section 4, Southern 1st Ring Rd, Chengdu, Sichuan 610065, P. R. China.

Semi-supervised semantic segmentation for medical images has evolved through time. While it can leverage the unlabeled data to significantly improve the segmentation performance, it still suffers the problems of intra-class variance and the consequent class-domain distribution misalignment along with costly training. In this paper, a stability-aware dual-head architecture is proposed to synergize prototype-based and Fully Convolutional Network (FCN) methodologies.

View Article and Find Full Text PDF

Prompt tuning, a recently emerging paradigm, adapts vision-language pre-trained models to new tasks efficiently by learning "soft prompts" for frozen models. However, in few-shot scenarios, its effectiveness is limited by sensitivity to the initialization and the time-consuming search for optimal initialization, hindering rapid adaptation. Additionally, prompt tuning risks reducing the models' generalizability due to overfitting on scarce training samples.

View Article and Find Full Text PDF