Learning Deep Low-Dimensional Models from High-Dimensional Data: From Theory
to Practice
ICCV 2025 Tutorial
Date: TBD (full day tutorial)
Location: TBD
Overview
Over the past decade, the advent of machine learning and large-scale computing
has immeasurably changed the ways we process, interpret, and predict with data
in imaging and computer vision. The “traditional” approach to algorithm
design, based around parametric models for specific structures of signals and
measurements—say sparse and low-rank models—and the associated optimization
toolkit, is now significantly enriched with data-driven learning-based
techniques, where large-scale networks are pre-trained and then adapted to a
variety of specific tasks. Nevertheless, the successes of both modern
data-driven and classic model-based paradigms rely crucially on correctly
identifying the low-dimensional structures present in real-world data, to the
extent that we see the roles of learning and compression of data processing
algorithms—whether explicit or implicit, as with deep networks—as
inextricably linked.
As such, this tutorial provides a timely tutorial that
uniquely bridges low-dimensional models with deep learning in imaging and
vision. This tutorial will show how:
Low-dimensional models and principles provide a valuable lens for
formulating problems and understanding the behavior of modern deep models in
imaging and computer vision; and how
Ideas from low-dimensional models can provide valuable guidance for
designing new parameter efficient, robust, and interpretable deep learning
models for computer vision problems in practice.
We will begin by introducing
fundamental low-dimensional models (e.g., basic sparse and low-rank models)
with motivating computer vision applications.
Based on these developments, we
will discuss strong conceptual, algorithmic, and theoretical connections
between low-dimensional structures and deep models, providing new perspectives
to understand state-of-the-art deep models in terms of learned representations
and generative models.
Finally, we will demonstrate that these
connections can lead to new principles for designing deep networks and learning
low-dimensional structures in computer vision, with both clear interpretability
and practical benefits.
We will conclude with a panel discussion with expert researchers from academia
and industry on what role low-dimensional models can and should play in our
current age of generative AI and foundation models for computer
vision.
The first part will introduce fundamental properties and results for sensing, processing, analyzing, and learning low-dimensional structures from high-dimensional data. We will first discuss classical low-dimensional models, such as sparse coding and low-rank matrix sensing, and motivate these models by applications in computer vision. Based on convex relaxation, we will characterize the conditions, in terms of sample/data complexity, under which the learning problems of recovering such low-dimensional structures become tractable and can be solved efficiently, with guaranteed correctness or accuracy.
Introduction of Low-dimensional Models in Deep Learning
(Lecture Abstract)
Next, we delve into the core principles of low-dimensional models within various neural network architectures. Specifically, we will introduce unrolled optimization as a design principle for interpretable deep networks. As a simple special case, we will examine several unrolled optimization algorithms for sparse coding, and show that they exhibit striking similarities to current deep network architectures. These unrolled networks are white-box and interpretable ab initio.
Session II: Understanding Low-Dimensional Structures in Representation Learning
Understanding Low-Dimensional Representation in Last-layer
(Lecture Abstract)
The session focuses on the strong conceptual connections between low-dimensional structures and deep models in terms of learned representation. We start with the introduction of an intriguing Neural Collapse phenomenon in the last-layer representation and its universality in deep network, and lay out the mathematical foundations of understanding its cause by studying its optimization landscapes. We then generalize and explain this phenomenon and its implications under data imbalanceness. Furthermore, we demonstrate the practical algorithmic implications of Neural Collapse on training deep neural networks.
Understanding Low-Dimensional Representation in Intermediate Layer
(Lecture Abstract)
Second, we show how low-dimensional structures also emerge in intermediate layers of learned deep neural networks. Specifically, we will characterize the progressive feature compression and discrimination from shallow to deep layers. Such a phenomenon is strongly correlated with the low-dimensional structure in learning dynamics, where the evolution of gradient descent only affects a minimal portion of singular vector spaces across all weight matrices.
Session III: Understanding Low-Dimensional Structures in Diffusion Generative Models
Low-Dimensional Models for Understanding Generalization in Diffusion Models
(Lecture Abstract)
We will establish a thorough understanding of the generalization of diffusion models by investigating when and why they are capable of learning the low-dimensional structure of target distributions. We will study the sample complexity to learn these low-dimensional distributions and how they scale with the intrinsic dimensionality of data. We will also show how diffusion models transit from memorization to generalization in terms of model capacity and data size.
Exploring Low-Dimensional Structures for Controlling Diffusion Models
(Lecture Abstract)
We will enhance diffusion models for scientific imaging by improving their efficiency, robustness, and controllability in solving general inverse problems. Our approach focuses on latent-space and patch-based models for efficiency, novel techniques for data consistency in reverse sampling—especially for 3D imaging—and controllable sampling to enforce desired constraints while maintaining sample quality.
Session IV: Designing Deep Networks for Pursuing Low-Dimensional Structures
ReduNet: A White-box Deep Network from the Principle of Maximizing Rate Reduction
(Lecture Abstract)
We focus on learning data distribution and transforming it into a linear discriminative representation (LDR). Using information-theoretic and statistical principles, we design a loss function called coding rate reduction, which is optimized at such a representation. By unrolling gradient ascent on this loss, we construct ReduNet, a deep network where each operator has a mathematically precise, interpretable role in transforming data toward an LDR. Notably, ReduNet can be built layer-wise through forward propagation, eliminating the need for back-propagation.
Finally, we introduce sparse linear discriminative representations by combining sparse coding with rate reduction, optimizing an objective called sparse rate reduction. This leads to CRATE, a deep network derived by unrolling the optimization and parameterizing feature distributions layer-wise. CRATE's operators are mathematically interpretable, with each layer corresponding to an optimization step, making it a white-box model. Despite its distinct design from ReduNet, both share a common goal, highlighting the flexibility of unrolled optimization. Interestingly, CRATE closely resembles transformer architectures, suggesting that such interpretability advances could enhance our understanding of modern deep networks.