Wednesday, November 6, 2024 - 4:00pm
Event Calendar Category
Other LIDS Events
Speaker Name
Xiangxiang Xu
Affiliation
RLE
Building and Room number
32-D650
Building and Room Number
LIDS Lounge
“Dependence Induced Representation Learning”
Despite the vast progress in deep learning practice, theoretical understandings of learned feature representations remain limited. In this talk, we discuss three fundamental questions from a unified statistical perspective:
(1) What representations carry useful information?
(2) How are representations learned from distinct algorithms related?
(3) Can we separate representation learning from solving specific tasks?
We formalize representations that extract statistical dependence from data, termed dependence-induced representations. We prove that representations are dependence-induced if and only if they can be learned from specific features defined by Hirschfeld–Gebelein–Rényi (HGR) maximal correlation. This separation theorem signifies the key role of HGR features in representation learning and enables a modular design of learning algorithms. Specifically, we demonstrate the optimality of HGR features in simultaneously achieving different design objectives, including minimal sufficiency (Tishby's information bottleneck), information maximization, enforcing uncorrelated features (VICReg), and encoding information at various granularities (Matryoshka representation learning). We further illustrate that by adapting HGR features, we can obtain representations learned by distinct practices, from cross-entropy or hinge loss minimization, non-negative feature learning, and neural density ratio estimators to their regularized variants. Our development provides a statistical interpretation of the neural collapse phenomenon observed in deep classifiers. We also discuss the applications of our analyses in understanding existing learning designs and obtaining more flexible solutions, e.g., inference-time hyperparameter tuning.
Xiangxiang Xu received the B.Eng. and Ph.D. degrees in electronic engineering from Tsinghua University, Beijing, China, in 2014 and 2020, respectively. He is a postdoctoral associate in the Department of EECS at MIT. His research focuses on information theory and statistical learning, with applications in understanding and developing learning algorithms. He is a recipient of the 2016 IEEE PES Student Prize Paper Award in Honor of T. Burke Hayes and the 2024 ITA (Information Theory and Applications) Workshop Sand Award.
ABOUT LIDS and STATS TEA TALKS:
Tea talks are 20-minute informal talks for the purpose of sharing ideas and creating awareness about topics of interest to the LIDS and Stats communities. Talks are followed by light refreshments and stimulating conversation.
Email lids_stats_teas[at]mit[dot]edu for information about LIDS & Stats Tea Talks
Sign-up to present at LIDS & Stats Tea Talks
LIDS & Stats Tea Talks Committee: Maison Clouatre, Subham Saha, Ashkan Soleymani, Jia Wan