Learning normalized image densities via dual score matching

Florentin Guth, Zahra Kadkhodaie and Eero P. Simoncelli.

arXiv.org e-prints, Technical Report 2506.05310, Jun 2025.

DOI: 10.48550/arXiv.2506.05310

Download:

  • Reprint (pdf)

  • Learning probability models from data is at the heart of many machine learning endeavors, but is notoriously difficult due to the curse of dimensionality. We introduce a new framework for learning normalized energy (log probability) models that is inspired from diffusion generative models, which rely on networks optimized to estimate the score. We modify a score network architecture to compute an energy while preserving its inductive biases. The gradient of this energy network with respect to its input image is the score of the learned density, which can be optimized using a denoising objective. Importantly, the gradient with respect to the noise level provides an additional score that can be optimized with a novel secondary objective, ensuring consistent and normalized energies across noise levels. We train an energy network with this dual score matching objective on the ImageNet64 dataset, and obtain a cross-entropy (negative log likelihood) value comparable to the state of the art. We further validate our approach by showing that our energy model strongly generalizes: estimated log probabilities are nearly independent of the specific images in the training set. Finally, we demonstrate that both image probability and dimensionality of local neighborhoods vary significantly with image content, in contrast with traditional assumptions such as concentration of measure or support on a low-dimensional manifold.
  • Related Publications: Kadkhodaie24a, Kadkhodaie22a, Raphan10, Raphan06b
  • Listing of all publications