In this post, we connect the Siegel upper half-space to denoising diffusion probabilistic models (DDPMs), and explore a way of navigating the Siegel upper half-space.
Denoising diffusion probabilistic models
Denoising diffusion probabilistic models (DDPMs) generate data by iteratively removing small amounts of Gaussian noise. This is typically done over hundreds of time steps. This generation process is the time-reversal of the learning process. The idea is to start with a given datapoint from the ground truth data distribution, and destroy the structure in
until all that is left is random noise. During this process, we use a parameterized model (a deep network) to learn how the noise has been destroyed, in order to learn how to reverse the noise and generate the original image.
In other words, with DDPMs, we attempt to learn a generative model that can generate samples as close as possible to the true distribution
of the ground truth dataset. To do this, we start with a point
in the dataset, and then iteratively destroy the structure in
, over
timesteps from
to
, until the final output
is isotropic Gaussian noise,
. Generation is then the reverse process, where we start with
, and then iteratively compute
until we arrive at the output generated datapoint
.
The Siegel upper half-space and multivariate normal distributions
The diffusion process as described above can be viewed as a path through (a subspace of) the parameter space of multivariate normal distributions. In other words, it is a trajectory through the moduli space of multivariate normal distributions. We can view this moduli space as the Siegel upper half-space for genus
, which is the space of
symmetric matrices with positive-definite imaginary part. The boundary
contains the positive semi-definite matrices.
The reasoning is as follows. Covariance matrices are symmetric. When there is no linear relationship between features (ie, uncorrelated features), a covariance matrix is positive-definite; when there is a linear relationship between features (ie, has correlated features), a covariance matrix is positive semi-definite. Then we can view the covariance matrix as the imaginary part of a matrix
and the mean vector
as a diagonal matrix which is the real part of
. In other words,
and
. The space of these matrices
, which we will denote by
, is then a subspace
(specifically,
is the subspace of matrices with diagonal real part). Then, changes in the multivariate normal distribution’s parameters are exactly changes in the coordinates in this subspace
of
. Thus the diffusion process is a trajectory through this subspace
of
.
Diffusion models typically use isotropic covariance matrices. The reasons include computational efficiency and tractability, ease of optimization, and theoretical conventions. However, if we can capture more of the overall noise by using general (not isotropic) covariance matrices, and if we can navigate between these distributions intentionally, then we should be able to use fewer time steps in the diffusion process. This would allow us to considerably speed up diffusion modeling.
How can we learn trajectories through the Siegel upper half-space ? Here is one idea.
Consider the Siegel-Jacobi space of complex dimension , which we will denote by
. It is the product of the Siegel upper half-space
and the Heisenberg group
, denoted
. The Heisenberg group
consists of upper-triangular matrices with 3 free parameters
and 1’s on the diagonal. We can think of the Heisenberg group as defining a position, and the Siegel upper half-space as defining the content that is located at the location defined by the Heisenberg group.
Now consider the natural group action on the Siegel-Jacobi space by the Jacobi group
, which is the semi-direct product of the symplectic group (the group of isometries of
) and the Heisenberg group, denoted
. This action is given by the mapping:
where for
and
We now have a way to navigate the moduli space of normal distributions and, in principle, to run diffusion models, with deep connections to number theory, automorphic forms, conformal field theory, and many other areas of math and physics. The form of this action is also strikingly similar to that of structured state space models (S4) for very long sequence learning. Given the roots of diffusion models in thermodynamics, these are intriguing results.
Leave a comment