π― About Me
Welcome to my homepage! Currently, I am a M.S. Student at Fudan University, advised by Prof. Tao Gui and Prof. Xuanjing Huang. Previously, I received my B.S. degree from Northeastern University in 2023, where I was an Outstanding Graduate and ranked 3/221 (top 1%) with a GPA of 4.365/5.0.
π My research interests lie in Complex reasoning and long-horizon planning. I am also interested in downstream applications of LLMs, such as deep research and vibe coding.
π If you are interested in my research or would like to collaborate, please feel free to email me. I am in the final year of my masterβs program and actively seeking job opportunities. I welcome any contact!
π₯ News
- Mar 2026: Active on the job market! Looking for research/engineer opportunities.
- Feb 2026: Announcing TDAR-8B-Thinking, the best long CoT block diffusion language models at 8B scale! π
- Jan 2026: R-HORIZON has been accepted by ICLR 2026!
- Jun 2025: Starting my internship at Meituan Longcat Post-train Team!
- May 2025: DPE has been accepted by COLM 2025!
- May 2025: Controlled Study on Long Context Models has been accepted by COLM 2025!
π Selected Publications

Advancing Block Diffusion Language Models for Test-Time Scaling
Yi Lu, Deyang Kong, Jianing Wang, Linsen Guo, Xue Wang, Qi Guo, Tao Gui, Xuanjing Huang, Wei Ye, Shikun Zhang, Wei Wang
- Best long CoT block diffusion language models at 8B scale π
- Proposes a βThink Coarse, Critic Fineβ paradigm for fast exploration and fine-grained reflection in diffusion reasoning models π§

R-Horizon: How Far Can Your Large Reasoning Model Really Go in Breadth and Depth?
Yi Lu, Jianing Wang, Linsen Guo, Wei He, Hongyin Tang, Tao Gui, Xuanjing Huang, Xuezhi Cao, Wei Wang, Xunliang Cai
[Paper] | [Project] | [Code] | [Datasets] | [Models] |
- A benchmark for evaluating long-horizon reasoning π
- Proposes a long-horizon data synthesis pipeline and improves long-horizon reasoning with reinforcement learning π§

A Controlled Study on Long Context Extension and Generalization in LLMs
Yi Lu, Jing Nathan Yan, Songlin Yang, Justin T. Chiu, Siyu Ren, Fei Yuan, Wenting Zhao, Zhiyong Wu, Alexander M. Rush
[Paper] | [Video] | [Code] | [Models and Datasets] |
- Tutorial: Create Long-Context LLM Extension tutorial with Nathan and Sasha Rush [Video] π₯
- The first controlled study that systematically evaluates long-context extension methods π§ͺ

Effective Length Extrapolation via Dimension-Wise Positional Embeddings Manipulation
Yi Lu, Wanxu Zhao, Xin Zhou, Chenxin An, Chenglong Wang, Shuo Li, Yuming Yang, Jun Zhao, Tao Ji, Tao Gui, Qi Zhang, Xuanjing Huang
- Achieves stronger extrapolation performance than YaRN without additional training β‘
- The first training-free extrapolation method that modifies RoPE at the dimension level π§©

LongHeads: Multi-Head Attention is Secretly a Long Context Processor
Yi Lu, Xin Zhou, Wei He, Jun Zhao, Tao Ji, Tao Gui, Qi Zhang, Xuanjing Huang
- The first block-selection method for multi-head attention π
- Closely related to Kimi MoBA and DeepSeek NSA π
- Cited by Kimiβs MoBA π
- ACL 2024 Making Harmful Behaviors Unlearnable for Large Language Models, Xin Zhou*, Yi Lu*, Ruotian Ma, Tao Gui, Qi Zhang, Xuanjing Huang.
- EMNLP 2023 TextMixer: Mixing Multiple Inputs for Privacy-Preserving Inference, Xin Zhou*, Yi Lu*, Ruotian Ma, Tao Gui, Qi Zhang, XuanJing Huang.
- ACL 2023 Textobfuscator: Making pre-trained language model a privacy protector via obfuscating word representations, Xin Zhou*, Yi Lu*, Ruotian Ma, Tao Gui, Yuran Wang, Yong Ding, Yibo Zhang, Qi Zhang, Xuanjing Huang.
- EMNLP 2022 Improved Knowledge Distillation for Pre-trained Language Models via Knowledge Selection, Chenglong Wang, Yi Lu, Yongyu Mu, Yimin Hu, Tong Xiao, Jingbo Zhu.
- EMNLP 2024 LongAgent: Scaling Language Models to 128k Context through Multi-Agent Collaboration, Jun Zhao, Can Zu, Hao Xu, Yi Lu, Wei He, Yiwen Ding, Tao Gui, Qi Zhang, Xuanjing Huang.
- NeurIPS 2025 Understanding Parametric and Contextual Knowledge Reconciliation within Large Language Models, Jun Zhao, Yongzhuo Yang, Xiang Hu, Jingqi Tong, Yi Lu, Wei Wu, Tao Gui, Qi Zhang, Xuanjing Huang.
- Preprint OPE: Overcoming Information Saturation in Parallel Thinking via Outline-Guided Path Exploration, Qi Guo, Jianing Wang, Deyang Kong, Xiangyu Xi, Jianfei Zhang, Yi Lu, Jingang Wang, Wei Wang, Shikun Zhang, Wei Ye.
- EMNLP 2025 Mitigating Object Hallucinations in MLLMs via Multi-Frequency Perturbations, Shuo Li, Jiajun Sun, Guodong Zheng, Xiaoran Fan, Yujiong Shen, Yi Lu, Zhiheng Xi, Yuming Yang, Wenming Tan, Tao Ji, Tao Gui, Qi Zhang, Xuanjing Huang.
- NAACL 2024 Self-Demos: Eliciting Out-of-Demonstration Generalizability in Large Language Models, Wei He, Shichun Liu, Jun Zhao, Yiwen Ding, Yi Lu, Zhiheng Xi, Tao Gui, Qi Zhang, Xuanjing Huang.
*Denotes equal contribution.
π» Internships
- 2025.06 - 2026.02, Beidou Research Intern, Post Train Team at Meituan LongCat, advised by Jianing Wang and Wei Wang.
- 2023.04 - 2023.11, Research Intern, Shanghai AI Lab, advised by Zhiyong Wu.
- 2022.10 - 2023.06, Research Intern, NLP Lab of Fudan University, advised by Xin Zhou and Prof. Tao Gui.
- 2021.06 - 2022.07, Research Intern, NLP Lab of Northeastern University, advised by Chenglong Wang and Prof. Tong Xiao.
π Honors and Awards
- 2024.10 Huawei Scholarship of Fudan University
- 2023.10 Datagrand Scholarship of Fudan University
- 2023.06 President Scholarship of Northeastern University (awarded to 10 students across the entire university)
- 2023.06 Outstanding Graduate of Northeastern University
- 2019 - 2021 National Scholarships of Northeastern University (two times)
- 2019 - 2023 Outstanding Students of Northeastern University (four times)
- 2019 - 2022 First Prize Scholarship of Northeastern University (three times)
- 2022.02 Meritorious in MCM/ICM February 2022
- 2021.03 Second Prize in China Undergraduate Mathematical Contest
π Educations
- 2023.09 - Present, M.S. Student, School of Computer Science, Fudan University
, Shanghai, China. - 2019.09 - 2023.06, B.S. Student, School of Computer Science, Northeastern University
, Shenyang, China.
βWhat doesnβt kill me makes me stronger.β