Hi! I'm a fourth-year PhD student in EECS at MIT. My research lies at the intersection of theoretical computer science, statistics, and machine learning: I seek to understand the structural assumptions that enable computationally efficient learning, particularly in the context of interactive learning and decision-making.
- Is a Good Foundation Necessary for Efficient Reinforcement Learning? The Computational Role of the Base Model in Exploration
with Dylan J. Foster and Zakaria Mhammedi
COLT 2025 ♦ 38th Annual Conference on Learning Theory (to appear).
- Computational-Statistical Tradeoffs at the Next-Token Prediction Barrier: Autoregressive and Imitation Learning under Misspecification
with Adam Block, Audrey Huang, Akshay Krishnamurthy, and Dylan J. Foster
COLT 2025 ♦ 38th Annual Conference on Learning Theory (to appear).
- Necessary and Sufficient Oracles: Towards a Computational Taxonomy for Reinforcement Learning
with Dylan J. Foster
COLT 2025 ♦ 38th Annual Conference on Learning Theory (to appear).
- Towards Characterizing the Value of Edge Embeddings in Graph Neural Networks
with Tanya Marwah, Zachary Chase Lipton, Jianfeng Lu, Ankur Moitra, and Andrej Risteski
Preliminary version selected for Oral Presentation at M3L Workshop (NeurIPS 2024)
ICML 2025 ♦ 42nd International Conference on Machine Learning (to appear).
- Self-Improvement in Language Models: The Sharpening Mechanism
with Audrey Huang, Adam Block, Dylan J. Foster, Cyril Zhang, Max Simchowitz, Jordan T. Ash, and Akshay Krishnamurthy
ICLR 2025 (Oral Presentation) ♦ 13th International Conference on Learning Representations (to appear).
- Online Control in Population Dynamics
with Noah Golowich, Elad Hazan, Zhou Lu, and Jennifer Sun
NeurIPS 2024 ♦ 38th Conference on Neural Information Processing Systems.
- Exploration is Harder than Prediction: Cryptographically Separating Reinforcement Learning from Supervised Learning
with Ankur Moitra and Noah Golowich
FOCS 2024 ♦ 65th Annual Symposium on Foundations of Computer Science (to appear).
- Lasso with Latents: Efficient Estimation, Covariate Rescaling, and Computational-Statistical Gaps
with Jonathan Kelner, Frederic Koehler, and Raghu Meka
COLT 2024 ♦ 37th Annual Conference on Learning Theory.
- Exploring and Learning in Sparse Linear MDPs without Computationally Intractable Oracles
with Ankur Moitra and Noah Golowich
STOC 2024 ♦ 56th Annual ACM Symposium on Theory of Computing.
- Provable Benefits of Score Matching
with Chirag Pabbaraju, Anish Sevekari, Holden Lee, Ankur Moitra, and Andrej Risteski
NeurIPS 2023 (Spotlight Presentation) ♦
37th Conference on Neural Information Processing Systems.
- Feature Adaptation for Sparse Linear Regression
with Jonathan Kelner, Frederic Koehler, and Raghu Meka
NeurIPS 2023 (Spotlight Presentation) ♦
37th Conference on Neural Information Processing Systems.
- Learning in Observable POMDPs, without Computationally Intractable Oracles
with Ankur Moitra and Noah Golowich
NeurIPS 2022 ♦ 36th Conference on Neural Information Processing Systems.
- Provably Auditing Ordinary Least Squares in Low Dimensions
with Ankur Moitra
ICLR 2023 ♦ 11th International Conference on Learning Representations.
- Distributional Hardness Against Preconditioned Lasso via Erasure-Robust Designs
with Jonathan Kelner, Frederic Koehler, and Raghu Meka
NeurIPS 2022 ♦ 36th Conference on Neural Information Processing Systems.
- Planning in Observable POMDPs in Quasipolynomial Time
with Ankur Moitra and Noah Golowich
STOC 2023 ♦ 55th Annual ACM Symposium on Theory of Computing.
- Robust Generalized Method of Moments: A Finite Sample Viewpoint
with Vasilis Syrgkanis
Preliminary version selected for Oral Presentation at MLECON Workshop (NeurIPS 2021)
NeurIPS 2022 ♦ 36th Conference on Neural Information Processing Systems.
- On the Power of Preconditioning in Sparse Linear Regression
with Jonathan Kelner, Frederic Koehler, and Raghu Meka
FOCS 2021 ♦ 62nd Annual IEEE Symposium on Foundations of Computer Science.
- Truncated Linear Regression in High Dimensions
with Costis Daskalakis and Manolis Zampetakis
NeurIPS 2020 ♦ 34th Conference on Neural Information Processing Systems.
- Constant-Expansion Suffices for Compressed Sensing with Generative Priors
with Costis Daskalakis and Manolis Zampetakis
NeurIPS 2020 (Spotlight Presentation) ♦
34th Conference on Neural Information Processing Systems.
- Regarding Two Conjectures on Clique and Biclique Partitions
with John C. Urschel and Jake Wellens
E-JC ♦ The Electronic Journal of Combinatorics, 2021.
- Near-Optimal Bounds for Online Caching with Machine-Learned Advice
SODA 2020 ♦ 31st ACM-SIAM Symposium on Discrete Algorithms.
- Conditional Hardness for Approximate Earth Mover Distance
APPROX 2019 ♦ 22nd Intl. Conference on Approximation Algorithms for Combinatorial Optimization Problems.
- Off-diagonal Ordered Ramsey Numbers of Matchings
E-JC ♦ The Electronic Journal of Combinatorics, 2019.