I am a junior student at School of Electronics Engineering and Computer Science, Peking University. My research interests lie in the development of human-like intelligent agents, with a particular focus on understanding and interpreting their behavior to enhance model control and optimize performance. While my prior work has predominantly been rooted in theoretical research, I am equally passionate about exploring the engineering aspects, aiming to bridge the gap between theoretical advancements and practical implementations to drive more robust, efficient, and human-aligned AI systems.

I am currently looking for 2025 summer internship opportunities!!!

🔥 News

  • 2025.03:  🎉 Our paper “When More is Less: Understanding Chain-of-Thought Length in LLMs” is accepted to ICLR 2025 Workshop on Reasoning and Planning for Large Language Models!
  • 2024.12:  🍁 I attended NuerIPS 2024 at Vancouver and illustrated our poster.
  • 2024.10:  🎉 Our paper “A Theoretical Understanding of Self-Correction through In-context Alignment” has been accepted to NeurIPS 2024!
  • 2024.06:  🏆 “A Theoretical Understanding of Self-Correction through In-context Alignment” received the Best Paper Award at ICML Workshop on In-context Learning!

📝 Publications

(*: Equal Contribution)

ICLR-W'25
cot_length

When More is Less: Understanding Chain-of-Thought Length in LLMs

Yuyang Wu*, Yifei Wang*, Tianqi Du, Stefanie Jegelka, Yisen Wang

  • I mathematically model the CoT process as a task decomposition and subtask-solving procedure and demonstrate that a longer CoT is not necessarily better.
  • Additionally, I conduct both synthetic and real-world experiments, revealing a U-shaped curve and the scaling behavior of the optimal CoT length.
NeurIPS 2024
self_correction

A Theoretical Understanding of Self-Correction through In-context Alignment

Yifei Wang*, Yuyang Wu*, Zeming Wei, Stefanie Jegelka, Yisen Wang

  • Best Paper Award at ICML 2024 Workshop on In-context Learning
  • I established the first rigorous understanding of LLMs’ self-correction ability and developed a simple and efficient self-correction algorithm (CaC) that shows significant improvements across different tasks.

🎖 Honors and Awards

  • 2024.06 Best Paper Award at ICML 2024 Workshop on In-context Learning
  • 2021.12 Silver Medal, Chinese Mathematical Olympiad

📖 Education

  • 2022.09 - Present, Peking University, BS in Computer Science

💻 Research Experience

  • 2023.10 - Present, Research Intern at ZERO Lab, Peking University
    • Researching the in-context abilities in LLMs, including self-correction and chain-of-thought.
    • Collaborating with Postdoc Yifei Wang (MIT) and advised by Prof. Yisen Wang (PKU)

💪 Skills

  • Programming Languages: Python(proficient), C++(proficient), C#, Core Skills (Git/Linux/TeX/etc.)
  • Deep Learning Technologies: Pytorch(proficient), CUDA parallel programming