About Me
I am a second-year CS PhD student at Cornell University, advised by Prof. Claire Cardie. Previously, I was a research assistant at Carnegie Mellon University supervised by Prof. Zhihao Jia. I obtained my MS at Tsinghua University supervised by Prof. Yong Jiang, and my BS at the CS department of Xidian University, where I was supervised by Prof. Yuesheng Xu.
My research focuses on large language models โ particularly LLM reasoning (self-play, RL post-training), robustness under noisy or adversarial inputs, and efficient training and serving.
I'm always open to collaboration โ feel free to drop me an email!
๐ Bill and Melinda Gates Hall, G23 ยท Campus Rd, Ithaca, NY 14853
News
[Apr 2026] Released Research Arena โ an evaluation of CLI-based AI agents (Claude Code, Codex, Kimi Code) on end-to-end scientific research across 13 CS domains. Read the analysis โ
[Apr 2026] Two papers accepted to ACL 2026 Findings: Better LLM Reasoning via Dual-play and GSM-Noise: Exploring and Enhancing Large Language Models' Reasoning under Noisy Inputs. See you in San Diego!
[Nov 2025] We released Better LLM Reasoning via Dual-play, a framework that trains two nearly identical models adversarially, enabling them to co-evolve through iterative competition without external supervision. Read the blog post โ
Publications
Preprints
-
Towards More Robust Retrieval-Augmented Generation: Evaluating RAG Under Adversarial Poisoning Attacks
-
Graph-based Multi-Task Learning Framework for Model Generalization across Network Topologies
2026
-
GSM-Noise: Exploring and Enhancing Large Language Models' Reasoning under Noisy InputsFindings of ACL, 2026
-
HumanGuideNet: Adapter-Based Alignment of Deep Neural Networks with Human Similarity JudgmentsWACV, Mar 2026
-
HAPO: Training Language Models to Reason Concisely via History-Aware Policy OptimizationAAAI, Jan 2026
2025
-
Learning-Enhanced High-Throughput Pattern Matching Based on Programmable Data PlaneUSENIX ATC, July 2025
2024
-
Quantized Side Tuning: Fast and Efficient Tuning of Quantized Large Language Models โ Outstanding PaperACL, Aug 2024 ยท Oral presentation
-
SpecInfer: Accelerating Large Language Model Serving with Tree-Based Speculative Inference and VerificationASPLOS, July 2024
-
-
2023
* denotes equal contribution
Education
Cornell University
Ph.D. in Computer Science ยท Advisor: Prof. Claire Cardie
2024 โ Present
Carnegie Mellon University
Visiting Scholar ยท Advisor: Prof. Zhihao Jia
2023 โ 2024
Tsinghua University
M.S. in Computer Science and Engineering ยท GPA 3.88/4.0
2021 โ 2024
Xidian University
B.S. in Software Engineering ยท GPA 3.86/4.0 ยท Rank 2/110
2017 โ 2021
Experience
Carnegie Mellon University
Research Assistant ยท Advisor: Prof. Zhihao Jia
Machine learning systems and large language models
Jun 2023 โ Aug 2024
Tencent, Shenzhen
Research Scientist Intern
XDP and eBPF algorithm design
Sep 2022 โ Aug 2023
Awards
Talks
โข Cornell NLP Seminar: Quantized Side Tuning: Fast and Efficient Tuning of Quantized Large Language Models. Sep 2024.
โข Tencent: Regular Expressions in the Data Plane. Dec 2023.
Academic Service
โข ICLR Reviewer 2025
โข AAAI Programme Committee 2025
โข NeurIPS Reviewer 2024, 2025
โข COLM Reviewer 2025