About me

Hi! I am a second-year CS PhD student jointly at NYU Courant and NYU Shanghai, mentored by Professor Chen Zhao. Previously, I received my B.S. degree in Data Science (AI track) from NYU Shanghai, where I was advised by Professor Wilson Tam and Professor Mathieu Lauriere. My primary research interests are in Natural Language Processing and Deep Learning. Here are some topics that I have been thinking about lately:

  1. How to best empower LLMs with the information retrieval ability, so that they can reliably solve complex real-world agentic tasks (e.g. Deep Researh report generation) that involves knowledge-intensive reasoning?
  2. Can we automate the error characterization process of LLM’s highly complex reasoning chains, or long-form generation in general? Given a set of predictions, can LLM accurately induce and summarize what the error patterns are?
  3. I am always fasinated by novel and smart decoding strategies and embedding learning methods!

If you have any thoughts on these, shoot me an email and I am more than happy to chat with you!

News

  • [February 2026] Check out LLaDA 2.1, large-scale diffusion language models with token editing abilities! Glad to have contributed to the on-policy SFT and editing-specific training strategies.
  • [January 2026] We release an evaluation suite for multi-turn Deep Research report writing, Mr Dre. Check out our paper and code!
  • [December 2025] I am joining Ant Research Center as a Research Intern to work on LLaDA post-training.
  • [May 2025] I joined AWS Bedrock as an Applied Science intern, working on explanability of agents.
  • [May 2025] Our paper on multi-answer QA is accepted to ACL 2025 Findings. Check it out here!
  • [September 2024] Starting my PhD at NYU Courant and NYU Shanghai!

(Last Update: Feb 2026)