ChatPaper.aiChatPaper

BARREL:面向事实性与可靠性的边界感知推理大语言模型

BARREL: Boundary-Aware Reasoning for Factual and Reliable LRMs

May 18, 2025
作者: Junxiao Yang, Jinzhe Tu, Haoran Liu, Xiaoce Wang, Chujie Zheng, Zhexin Zhang, Shiyao Cui, Caishun Chen, Tiantian He, Hongning Wang, Yew-Soon Ong, Minlie Huang
cs.AI

摘要

近期,大型推理模型(LRMs)在数学与逻辑推理方面展现出了令人瞩目的能力。然而,当前的LRMs极少承认无知或回应“我不知道”,反而常常在表现出过度自信的同时给出错误答案,这引发了对其事实可靠性的担忧。在本研究中,我们识别出两种由过度思考导致的病态推理模式,它们助长了这种过度自信与错误答案:最后一刻的猜测和反复纠结的螺旋思维。为解决这些问题,我们提出了BARREL——一个旨在促进简洁且边界意识强的事实推理的新框架。实验表明,经过BARREL训练后,DeepSeek-R1-Distill-Llama-8B的可靠性从39.33%提升至61.48%,同时保持了与基于R1生成推理数据进行微调的模型相当的准确度。这些结果证明,我们的初步研究对于构建更加可靠且基于事实的系统二型LRMs具有启发意义。
English
Recent advances in Large Reasoning Models (LRMs) have shown impressive capabilities in mathematical and logical reasoning. However, current LRMs rarely admit ignorance or respond with "I don't know". Instead, they often produce incorrect answers while showing undue confidence, raising concerns about their factual reliability. In this work, we identify two pathological reasoning patterns characterized by overthinking that contribute to the overconfident and incorrect answers: last-minute guessing and second-thought spiraling. To address these issues, we propose BARREL-a novel framework that promotes concise and boundary-aware factual reasoning. Our experiments show that BARREL-training increases the reliability of DeepSeek-R1-Distill-Llama-8B from 39.33% to 61.48%, while still achieving accuracy comparable to models finetuned on reasoning data generated by R1. These results demonstrate that our pilot study is inspiring to build more reliable and factual System 2 LRMs.

Summary

AI-Generated Summary

PDF102May 22, 2025