TY - JOUR
T1 - Variational Distillation for Multi-View Learning
AU - Tian, Xudong
AU - Zhang, Zhizhong
AU - Wang, Cong
AU - Zhang, Wensheng
AU - Qu, Yanyun
AU - Ma, Lizhuang
AU - Wu, Zongze
AU - Xie, Yuan
AU - Tao, Dacheng
N1 - Publisher Copyright:
© 1979-2012 IEEE.
PY - 2024/7/1
Y1 - 2024/7/1
N2 - Information Bottleneck (IB) provides an information-theoretic principle for multi-view learning by revealing the various components contained in each viewpoint. This highlights the necessity to capture their distinct roles to achieve view-invariance and predictive representations but remains under-explored due to the technical intractability of modeling and organizing innumerable mutual information (MI) terms. Recent studies show that sufficiency and consistency play such key roles in multi-view representation learning, and could be preserved via a variational distillation framework. But when it generalizes to arbitrary viewpoints, such strategy fails as the mutual information terms of consistency become complicated. This paper presents Multi-View Variational Distillation (MV$^{2}$2D), tackling the above limitations for generalized multi-view learning. Uniquely, MV$^{2}$2D can recognize useful consistent information and prioritize diverse components by their generalization ability. This guides an analytical and scalable solution to achieving both sufficiency and consistency. Additionally, by rigorously reformulating the IB objective, MV$^{2}$2D tackles the difficulties in MI optimization and fully realizes the theoretical advantages of the information bottleneck principle. We extensively evaluate our model on diverse tasks to verify its effectiveness, where the considerable gains provide key insights into achieving generalized multi-view representations under a rigorous information-theoretic principle.
AB - Information Bottleneck (IB) provides an information-theoretic principle for multi-view learning by revealing the various components contained in each viewpoint. This highlights the necessity to capture their distinct roles to achieve view-invariance and predictive representations but remains under-explored due to the technical intractability of modeling and organizing innumerable mutual information (MI) terms. Recent studies show that sufficiency and consistency play such key roles in multi-view representation learning, and could be preserved via a variational distillation framework. But when it generalizes to arbitrary viewpoints, such strategy fails as the mutual information terms of consistency become complicated. This paper presents Multi-View Variational Distillation (MV$^{2}$2D), tackling the above limitations for generalized multi-view learning. Uniquely, MV$^{2}$2D can recognize useful consistent information and prioritize diverse components by their generalization ability. This guides an analytical and scalable solution to achieving both sufficiency and consistency. Additionally, by rigorously reformulating the IB objective, MV$^{2}$2D tackles the difficulties in MI optimization and fully realizes the theoretical advantages of the information bottleneck principle. We extensively evaluate our model on diverse tasks to verify its effectiveness, where the considerable gains provide key insights into achieving generalized multi-view representations under a rigorous information-theoretic principle.
KW - Multi-view learning
KW - information bottleneck
KW - knowledge distillation
KW - mutual information
KW - variational inference
UR - https://www.scopus.com/pages/publications/85182372226
U2 - 10.1109/TPAMI.2023.3343717
DO - 10.1109/TPAMI.2023.3343717
M3 - 文章
C2 - 38133979
AN - SCOPUS:85182372226
SN - 0162-8828
VL - 46
SP - 4551
EP - 4566
JO - IEEE Transactions on Pattern Analysis and Machine Intelligence
JF - IEEE Transactions on Pattern Analysis and Machine Intelligence
IS - 7
ER -