Ashmit Khandelwal

Research Fellow at Microsoft Research

prof_pic.jpg
background: Hallstätter See, Austria

I’m a researcher at Microsoft Research working with Dr. Nagarajan Natarajan and Dr. Amit Sharma on verification and steering for Large Language Model (LLM) agents. I study how these systems search, analyze, and interact with real-world environments, and methods to monitor and steer them.

My recent work interwhen, focuses on test-time verifiers that asynchronously monitor LLM trajectories at runtime, intervening when the agent deviates from a defined specification. We see improvements in both task performance and soundness across code generation, logical reasoning, and agentic settings. I’ve also worked on giving a definition to and evaluating Deep Research, published at ICLR 2026. These are LLM systems that perform structured search over large corpora.

Previously, I finished my Undergraduate at BITS Pilani, India where I studied Computer Science and Data Science. I spent a summer at the Adobe MDSR lab, where I worked with on predicting human behavior at scale using multimodal LLMs. This research was a spotlight publication at ICLR 2024.

publications

* = equal contribution
  1. ICLR
    Spotlight
    Large Content And Behavior Models To Understand, Simulate, And Optimize Content And Behavior
    Ashmit Khandelwal*, Aditya Agrawal*, Aanisha Bhattacharyya*, Yaman K Singla*, and 7 more authors
    In International Conference on Learning Representations, 2024
  2. Characterizing Deep Research: A Benchmark and Formal Definition
    Abhinav Java*, Ashmit Khandelwal*, Sukruta Prakash Midigeshi*, Aaron Halfaker, and 5 more authors
    In International Conference on Learning Representations, 2026
  3. ArXiv preprint
    interwhen: A Generalizable Framework for Verifiable Reasoning with Test-time Monitors
    Vishak K Bhat, Prateek Chanda, Ashmit Khandelwal, Maitreyi Swaroop, and 4 more authors
    In ArXiv, 2026