Category Ranking

98%

Total Visits

921

Avg Visit Duration

2 minutes

Citations

20

Article Abstract

Information Bottleneck (IB) provides an information-theoretic principle for multi-view learning by revealing the various components contained in each viewpoint. This highlights the necessity to capture their distinct roles to achieve view-invariance and predictive representations but remains under-explored due to the technical intractability of modeling and organizing innumerable mutual information (MI) terms. Recent studies show that sufficiency and consistency play such key roles in multi-view representation learning, and could be preserved via a variational distillation framework. But when it generalizes to arbitrary viewpoints, such strategy fails as the mutual information terms of consistency become complicated. This paper presents Multi-View Variational Distillation (MV D), tackling the above limitations for generalized multi-view learning. Uniquely, MV D can recognize useful consistent information and prioritize diverse components by their generalization ability. This guides an analytical and scalable solution to achieving both sufficiency and consistency. Additionally, by rigorously reformulating the IB objective, MV D tackles the difficulties in MI optimization and fully realizes the theoretical advantages of the information bottleneck principle. We extensively evaluate our model on diverse tasks to verify its effectiveness, where the considerable gains provide key insights into achieving generalized multi-view representations under a rigorous information-theoretic principle.

Download full-text PDF

Source
http://dx.doi.org/10.1109/TPAMI.2023.3343717DOI Listing

Publication Analysis

Top Keywords

variational distillation
12
multi-view learning
12
information-theoretic principle
8
mutual terms
8
sufficiency consistency
8
generalized multi-view
8
multi-view
6
distillation multi-view
4
learning
4
learning bottleneck
4

Similar Publications

Rapid progress in aberration corrected electron microscopy necessitates development of robust methods for the identification of phases, ferroic variants, and other pertinent aspects of materials structure from imaging data. While unsupervised methods for clustering and classification are widely used for these tasks, their performance can be sensitive to hyperparameter selection in the analysis workflow. In this study, the effects of descriptors and hyperparameters are explored on the capability of unsupervised ML methods to distill local structural information, exemplified by the discovery of polarization and lattice distortion in Sm - dopped BiFeO (BFO) thin films.

View Article and Find Full Text PDF

Recent advancements in food image recognition have underscored its importance in dietary monitoring, which promotes a healthy lifestyle and aids in the prevention of diseases such as diabetes and obesity. While mainstream food recognition methods excel in scenarios with large-scale annotated datasets, they falter in few-shot regimes where data is limited. This paper addresses this challenge by introducing a variational generative method, the Multivariate Knowledge-guided Variational AutoEncoder (MK-VAE), for few-shot food recognition.

View Article and Find Full Text PDF

Deep learning analyses have offered sensitivity leaps in detection of cognition-related variables from functional MRI (fMRI) measurements of brain responses. Yet, as deep models perform hierarchical nonlinear transformations on fMRI data, interpreting the association between individual brain regions and the detected variables is challenging. Among explanation approaches for deep fMRI classifiers, attribution methods show poor specificity and perturbation methods show limited sensitivity.

View Article and Find Full Text PDF

Direct Distillation: A Novel Approach for Efficient Diffusion Model Inference.

J Imaging

February 2025

Division of Control and Instrumentation (CI), Nanyang Technological University, Singapore 639798, Singapore.

Diffusion models are among the most common techniques used for image generation, having achieved state-of-the-art performance by implementing auto-regressive algorithms. However, multi-step inference processes are typically slow and require extensive computational resources. To address this issue, we propose the use of an information bottleneck to reschedule inference using a new sampling strategy, which employs a lightweight distilled neural network to map intermediate stages to the final output.

View Article and Find Full Text PDF

Improving functional protein generation via foundation model-derived latent space likelihood optimization.

bioRxiv

January 2025

Machine Biology Group, Departments of Psychiatry and Microbiology, Institute for Biomedical Informatics, Institute for Translational Medicine and Therapeutics, Perelman School of Medicine, University of Pennsylvania, Philadelphia, Pennsylvania, United States of America.

A variety of deep generative models have been adopted to perform functional protein generation. Compared to 3D protein design, sequence-based generation methods, which aim to generate amino acid sequences with desired functions, remain a major approach for functional protein generation due to the abundance and quality of protein sequence data, as well as the relatively low modeling complexity for training. Although these models are typically trained to match protein sequences from the training data, exact matching of every amino acid is not always essential.

View Article and Find Full Text PDF