Theory-Focused Research

Foundations of ML, GNN theory, generative modeling theory, and complexity under dependence

My theory-focused work studies the foundations of machine learning under dependence, with emphasis on statistical complexity, minimax analysis, lower bounds, and structure-aware learning theory. A recurring goal is to identify regimes where classical i.i.d.-based intuition breaks down, characterize the resulting statistical limits, and derive consequences for learning, inference, and sampling.

Current themes

  • Minimax learning theory for graph-structured models: I study how graph topology, dependence, and mixing properties affect the effective sample complexity of message-passing architectures. My recent ICLR 2026 paper develops minimax lower bounds and structural regime characterizations for ReLU message-passing GNNs, showing that slowly mixing graphs can induce substantially harder rates than classical sample-size scaling would suggest.

  • Structure-aware complexity in generative modeling: I am extending this line of work to the theory of generative modeling, with current interest in the complexity of score estimation in diffusion-based models, error propagation from score estimation to sampling quality, and statistical-computational tradeoffs in sequence-structured settings. A central question is how dependence structure governs both statistical and algorithmic difficulty.

  • Inference-time complexity and compute-quality tradeoffs: I also study how budget constraints at inference time interact with estimation and decision quality in large models, including confidence-guided stopping and adaptive computation. This provides a bridge between formal complexity questions and deployment-relevant behavior.

Selected papers

  • ICLR 2026 (accepted): Minimax Sample Complexity of Graph Neural Networks: Lower Bounds and Structural Effects
  • CGES (under review / arXiv): Confidence-Guided Early Stopping for Efficient and Accurate Self-Consistency

Methodological perspective

I combine rigorous theorem development with theory-grounded empirical diagnostics designed to distinguish competing scaling laws, identify structural regimes, and test whether observed behavior matches theoretical predictions.