勿過度思考:高效R1風格大型推理模型綜述
Don't Overthink It: A Survey of Efficient R1-style Large Reasoning Models
August 4, 2025
作者: Linan Yue, Yichao Du, Yizhi Wang, Weibo Gao, Fangzhou Yao, Li Wang, Ye Liu, Ziyu Xu, Qi Liu, Shimin Di, Min-Ling Zhang
cs.AI
摘要
近来,大型推理模型(Large Reasoning Models, LRMs)因其在处理复杂任务中的卓越表现,逐渐成为研究热点。其中,DeepSeek R1凭借其出色的性能及开源特性,引起了广泛关注,推动了R1型LRM研究的进步。与传统的语言大模型(Large Language Models, LLMs)不同,这些模型通过引入长链思维、自我反思等机制,结合强化学习,在推理过程中增强了逻辑推理与决策能力。然而,随着这些模型的广泛应用,过度思考的问题逐渐显现。具体而言,在生成答案时,这些模型往往构建过长的推理链条,包含冗余或重复的步骤,这不仅降低了推理效率,还可能影响最终答案的准确性。为此,多种高效推理方法被提出,旨在不损害模型性能与推理能力的前提下,缩短推理路径。通过系统梳理当前高效推理方法领域的研究进展,我们以单模型优化与模型协作的视角,将现有工作归纳为两大方向:(1)单模型高效推理,专注于提升单个模型的推理效率;(2)模型协作高效推理,探索通过多模型协作优化推理路径。此外,我们维护了一个公开的GitHub仓库,用于追踪高效推理方法的最新进展。
English
Recently, Large Reasoning Models (LRMs) have gradually become a research
hotspot due to their outstanding performance in handling complex tasks. Among
them, DeepSeek R1 has garnered significant attention for its exceptional
performance and open-source nature, driving advancements in the research of
R1-style LRMs. Unlike traditional Large Language Models (LLMs), these models
enhance logical deduction and decision-making capabilities during reasoning by
incorporating mechanisms such as long chain-of-thought and self-reflection
through reinforcement learning. However, with the widespread application of
these models, the problem of overthinking has gradually emerged. Specifically,
when generating answers, these models often construct excessively long
reasoning chains with redundant or repetitive steps, which leads to reduced
reasoning efficiency and may affect the accuracy of the final answer. To this
end, various efficient reasoning methods have been proposed, aiming to reduce
the length of reasoning paths without compromising model performance and
reasoning capability. By reviewing the current research advancements in the
field of efficient reasoning methods systematically, we categorize existing
works into two main directions based on the lens of single-model optimization
versus model collaboration: (1) Efficient Reasoning with Single Model, which
focuses on improving the reasoning efficiency of individual models; and (2)
Efficient Reasoning with Model Collaboration, which explores optimizing
reasoning paths through collaboration among multiple models. Besides, we
maintain a public GitHub repository that tracks the latest progress in
efficient reasoning methods.