Daniil Dmitriev

I am a postdoc at the University of Pennsylvania, hosted by Yuting Wei in the Department of Statistics and Data Science at the Wharton School.

My research interests lie at the intersection of mathematics, machine learning, and theoretical computer science, with a current focus on understanding the interplay between structure and randomness in algorithm analysis.

I completed my PhD at ETH Zurich, advised by Afonso Bandeira and Fanny Yang, and supported by the ETH AI Center and ETH FDS initiative.

In Fall 2024, I visited Simons Institute for the Theory of Computing at UC Berkeley, for the program Modern paradigms in generalization.

I have a MS from EPFL, where I worked with Lenka Zdeborová and Martin Jaggi, and a BS from MIPT, where I worked with Maksim Zhukovskii.

Google Scholar  /  CV  /  GitHub  /  LinkedIn  /  X  /  Blue sky

Contact me at: dmitrievdaniil97@gmail.com.
profile photo

Publications & Preprints

project image

Efficient Sampling with Discrete Diffusion Models: Sharp and Adaptive Guarantees


DD*, Zhihan Huang*, Yuting Wei
preprint, 2026
arxiv /

We analyze two noising processes for score-based discrete diffusion models: uniform and absorbing. For the uniform noising process, we prove an upper bound for the number of sampling steps scaling linearly with the ambient dimension and logarithmically with the vocabulary size. This is achieved with a tau-leaping sampler, for which we show a matching lower bound.
For the absorbing noising process, we show an upper bound that scales (up to logarithmic factors) with the effective total correlation of the distribution, which potentially can be sublinear in the ambient dimension, or even constant for certain structured distributions.

project image

Learning in an echo chamber: online learning with replay adversary


DD, Harald Eskelund Franck, Carolin Heinzler, Amartya Sanyal (αβ order)
SODA, 2026
arxiv /

We propose a new online learning variant in which the learner might receive samples labelled via its own past models. Our results include upper and lower bounds against both adaptive and stochastic adversaries.

project image

The Lovász number of random circulant graphs


Afonso S. Bandeira, Jarosław Błasiok, DD, Ulysse Faure, Anastasia Kireeva, Dmitriy Kunisky (αβ order)
SampTA, 2025
arxiv /

We provide lower and upper bounds for the expected value of the Lovász theta number of random circulant graphs. Circulant graphs are a class of structured graphs where connectivity depends only on the difference between the vertex labels.

project image

Robust mixture learning when outliers overwhelm small groups


DD*, Rares-Darius Buhai*, Stefan Tiegel, Alexander Wolters, Gleb Novikov, Amartya Sanyal,
David Steurer, Fanny Yang
NeurIPS, 2024
arxiv / poster /

We propose an efficient meta-algorithm for recoving means of a mixture model in the presence of large additive contaminations.

project image

On the growth of mistakes in differentially private online learning: a lower bound perspective


DD, Kristof Szabo, Amartya Sanyal
COLT, 2024
arxiv / poster /

We prove that for a certain class of algorithms, the number of mistakes for online learning under differential privacy constraint must grow logarithmically with the time horizon.

project image

Asymptotics of learning with deep structured (random) features


Dominik Schröder*, DD*, Hugo Cui*, Bruno Loureiro
ICML, 2024
arxiv / poster /

We derive a deterministic equivalent for the generalization error of general Lipschitz functions. Furthermore, we prove a linearization formula for a sample covariance matrix of a structured random feature model with two hidden layers.

project image

Greedy heuristics and linear relaxations for the random hitting set problem


Gabriel Arpino, DD, Nicolo Grometto (αβ order)
APPROX, 2024
arxiv /

We prove that the standard greedy algorithm is order-optimal for the hitting set problem in the random bernoulli case.

project image

Deterministic equivalent and error universality of deep random features learning


Dominik Schröder, Hugo Cui, DD, Bruno Loureiro
ICML, 2023
arxiv / video / poster /

We rigorously establish Gaussian universality for the test error in ridge regression in deep networks with frozen intermediate layers.

project image

On monotonicity of Ramanujan function for binomial random variables


DD, Maksim Zhukovskii (αβ order)
Statistics & Probability Letters, 2021
arxiv /

We analyze properties of the CDF of binomial random variable near its median.

project image

Dynamic model pruning with feedback


Tao Lin, Sebastian U Stich, Luis Barba, DD, Martin Jaggi
ICLR, 2020
arxiv /

We propose a novel model compression method that generates a sparse trained model without an additional overhead.



Invited Talks

DACO seminar, ETH Zurich, 2024 website
Youth in High Dimensions, ICTP, Trieste, 2024 video
Delta seminar, University of Copenhagen, 2024 website
Graduate seminar in probability, ETH Zurich, 2023
Workshop on Spin Glasses, Les Diablerets, 2022 video


Teaching and Service

ETH Zurich, TA: Mathematics of Data Science (Fall 2021), Mathematics of Machine Learning (Spring 2022)

EPFL, TA: Artificial Neural Networks (Spring 2020, Spring 2021)

Supervising MSc theses at ETH Zurich: Carolin Heinzler (Fall 2023), Krish Agrawal, Ulysse Faure (Spring 2024)

Berkeley Math Circle (Fall 2024)

Reviewer: NeurIPS 2024 (Top reviewer), ICLR 2025



Design and source code from Leonid Keselman's website