Monday, June 12, 2023 - 2:00pm to 3:00pm
Event Calendar Category
LIDS Thesis Defense
Speaker Name
Joshua David Robinson
Affiliation
CSAIL
Building and Room number
32-D463
Join Zoom meeting
https://mit.zoom.us/j/95814105619
Abstract: Neural networks automate the process of representing objects and their relations on a computer, spanning everything from household items to molecules. They achieve this by transforming different instances into a shared representation space, where variations in data can be measured using simple geometric quantities such as Euclidean distances. This talk studies the geometric structure of this space and its influence on key properties of the learning process, including how much data is needed to acquire new skills, when predictions will fail, and the computational cost of learning. We examine two foundational aspects of the geometry of neural network representations. Part I designs and studies learning algorithms that take into account the location of data in representation space. Focusing on contrastive self-supervised learning, we design a) hard instance sampling strategies and b) methods for controlling what features models learn. Each produces improvements in key characteristics such as training speed, generalization, and model reliability. Part II studies how to use non-Euclidean geometries to build network architectures that respect symmetries and structures arising in physical data. Specifically, we use geometric spaces such as the real projective plane and the spectraplex to build a) provably powerful neural networks that respect the symmetries of (Laplacian) eigenvectors, which is important for building Transformers on graph structured data, and b) neural networks that solve combinatorial optimization problems on graphs such as finding big cliques or small cuts, which arise in molecular engineering and network science.
Thesis committee: Stefanie Jegelka, Suvrit Sra, and Phillip Isola
Thesis advisors: Stefanie Jegelka and Suvrit Sra