list_alt
  • Diffusion models and minimax rates: scores, functionals, and tests

    On February 6, 2026 at 11:00 am till 12:00 pm
    Subhodh Kotekal, MIT

    Abstract: While score-based diffusion models have achieved remarkable success in high-dimensional generative modeling, some basic theoretical questions have not been precisely resolved. In this talk, we address minimax optimality of density estimation, functional estimation, and hypothesis testing. First, we show diffusion models achieve the optimal density estimation rate over Holder balls. This result is a consequence of our sharp characterization of minimax score estimation across all noising levels. A key contribution is our lower bound argument which involves a slight twist to a classical construction. The squared norm of the score function is the Fisher information, which is related to important functionals like mutual information and entropy. Leveraging well-known information theoretic relations (such as the I-MMSE and de Bruijn identities), we furnish estimators of these functionals by studying noised Fisher information estimation; our results establish that simple plug-in estimators can achieve parametric rates. Finally, we study the applicability of diffusion models to hypothesis testing. By aggregating information across noise scales, we demonstrate diffusion models can achieve the sharp minimax separation rate despite going against the natural intuition that noising makes discrimination harder.

    Bio: Subhodh Kotekal is a Norbert Wiener Fellow in the Statistics and Data Science Center at MIT. He obtained a Ph.D. from the Department of Statistics at the University of Chicago in June 2025 and was advised by Chao Gao.

    Find out more »: Diffusion models and minimax rates: scores, functionals, and tests
  • Statistics and Data Science Seminar

    On February 13, 2026 at 11:00 am till 12:00 pm
    Max Simchowitz, Carnegie Mellon University
  • When do spectral gradient updates help in deep learning?

    On February 20, 2026 at 11:00 am till 12:00 pm
    Dmitriy Drusvyatskiy, University of California – San Diego

    Abstract: Spectral gradient methods, such as the recently proposed Muon optimizer, are a promising alternative to standard gradient descent for training deep neural networks and transformers. Yet, it remains unclear in which regimes these spectral methods are expected to perform better. In this talk, I will present a simple condition that predicts when a spectral update yields a larger decrease in the loss than a standard gradient step. Informally, this criterion holds when, on the one hand, the gradient of the loss with respect to each parameter block has a nearly uniform spectrum—measured by its nuclear-to-Frobenius ratio—while, on the other hand, the incoming activation matrix has low stable rank. It is this mismatch in the spectral behavior of the gradient and the propagated data that underlies the advantage of spectral updates. Reassuringly, this condition naturally arises in a variety of settings, including random feature models, neural networks, and transformer architectures. I will conclude by showing that these predictions align with empirical results in synthetic regression problems and in small-scale language model training.

    Biosketch: Dmitriy Drusvyatskiy received his PhD from Cornell University in 2013, followed by a post doctoral appointment at University of Waterloo, 2013-2014. He joined the Mathematics department at University of Washington as an Assistant Professor in 2014 and was promoted to Full Professor in 2022. Since 2025, Dmitriy is a Professor at the Halıcıoğlu Data Science Institute (HDSI) at UC San Diego. Dmitriy’s research broadly focuses on designing and analyzing algorithms for large-scale optimization problems, primarily motivated by applications in data science. Dmitriy has received a number of awards, including the Air Force Office of Scientific Research (AFOSR) Young Investigator Program (YIP) Award, NSF CAREER, SIAG/OPT Best Paper Prize 2023, Paul Tseng Faculty fellowship 2022-2026, INFORMS Optimization Society Young Researcher Prize 2019, and finalist citations for the Tucker Prize 2015 and the Young Researcher Best Paper Prize at ICCOPT 2019.

    Find out more »: When do spectral gradient updates help in deep learning?
  • Statistics and Data Science Seminar

    On February 27, 2026 at 11:00 am till 12:00 pm
    Michael Albergo, Harvard University
  • Statistics and Data Science Seminar

    On March 6, 2026 at 11:00 am till 12:00 pm
    Piotr Indyk, MIT
  • Statistics and Data Science Seminar

    On March 13, 2026 at 11:00 am till 12:00 pm
    Sifan Liu, Duke University
  • Statistics and Data Science Seminar

    On March 20, 2026 at 11:00 am till 12:00 pm
    Dan Alistarh, Institute of Science and Technology, Austria – ISTA
  • Statistics and Data Science Seminar

    On April 3, 2026 at 11:00 am till 12:00 pm
    Surbhi Goel, University of Pennsylvania
  • SDSCon26

    On April 10, 2026
  • Statistics and Data Science Seminar

    On April 17, 2026 at 11:00 am till 12:00 pm
    Hamed Hassani, University of Pennsylvania