ChatPaper.aiChatPaper

DianJin-R1:评估与增强大语言模型中的金融推理能力

DianJin-R1: Evaluating and Enhancing Financial Reasoning in Large Language Models

April 22, 2025
作者: Jie Zhu, Qian Chen, Huaixia Dou, Junhui Li, Lifan Guo, Feng Chen, Chi Zhang
cs.AI

摘要

在金融领域,有效推理仍然是大型语言模型(LLMs)面临的核心挑战,该领域的任务通常需要特定领域的知识、精确的数值计算以及严格遵守合规规则。我们提出了DianJin-R1,一个推理增强框架,旨在通过推理增强监督和强化学习来应对这些挑战。我们方法的核心是DianJin-R1-Data,这是一个从CFLUE、FinQA和专有合规语料库(中文合规检查,CCC)构建的高质量数据集,结合了多样化的金融推理场景和经过验证的注释。我们的模型DianJin-R1-7B和DianJin-R1-32B,是从Qwen2.5-7B-Instruct和Qwen2.5-32B-Instruct微调而来,采用了一种结构化格式,既生成推理步骤也生成最终答案。为了进一步提升推理质量,我们应用了群体相对策略优化(GRPO),这是一种强化学习方法,结合了双重奖励信号:一个鼓励结构化输出,另一个奖励答案的正确性。我们在五个基准上评估了我们的模型:三个金融数据集(CFLUE、FinQA和CCC)和两个通用推理基准(MATH-500和GPQA-Diamond)。实验结果表明,DianJin-R1模型在复杂金融任务上持续优于非推理模型。此外,在现实世界的CCC数据集上,我们的单次调用推理模型匹配甚至超越了需要显著更多计算成本的多代理系统的性能。这些发现证明了DianJin-R1通过结构化监督和奖励对齐学习在增强金融推理方面的有效性,为现实世界应用提供了一个可扩展且实用的解决方案。
English
Effective reasoning remains a core challenge for large language models (LLMs) in the financial domain, where tasks often require domain-specific knowledge, precise numerical calculations, and strict adherence to compliance rules. We propose DianJin-R1, a reasoning-enhanced framework designed to address these challenges through reasoning-augmented supervision and reinforcement learning. Central to our approach is DianJin-R1-Data, a high-quality dataset constructed from CFLUE, FinQA, and a proprietary compliance corpus (Chinese Compliance Check, CCC), combining diverse financial reasoning scenarios with verified annotations. Our models, DianJin-R1-7B and DianJin-R1-32B, are fine-tuned from Qwen2.5-7B-Instruct and Qwen2.5-32B-Instruct using a structured format that generates both reasoning steps and final answers. To further refine reasoning quality, we apply Group Relative Policy Optimization (GRPO), a reinforcement learning method that incorporates dual reward signals: one encouraging structured outputs and another rewarding answer correctness. We evaluate our models on five benchmarks: three financial datasets (CFLUE, FinQA, and CCC) and two general reasoning benchmarks (MATH-500 and GPQA-Diamond). Experimental results show that DianJin-R1 models consistently outperform their non-reasoning counterparts, especially on complex financial tasks. Moreover, on the real-world CCC dataset, our single-call reasoning models match or even surpass the performance of multi-agent systems that require significantly more computational cost. These findings demonstrate the effectiveness of DianJin-R1 in enhancing financial reasoning through structured supervision and reward-aligned learning, offering a scalable and practical solution for real-world applications.

Summary

AI-Generated Summary

PDF92April 28, 2025