ChatPaper.aiChatPaper

VRAG-RL:透過強化學習的迭代推理,強化基於視覺感知的檢索增強生成技術,以深入理解視覺豐富信息

VRAG-RL: Empower Vision-Perception-Based RAG for Visually Rich Information Understanding via Iterative Reasoning with Reinforcement Learning

May 28, 2025
作者: Qiuchen Wang, Ruixue Ding, Yu Zeng, Zehui Chen, Lin Chen, Shihang Wang, Pengjun Xie, Fei Huang, Feng Zhao
cs.AI

摘要

有效檢索、推理並理解視覺豐富的信息,對於RAG方法而言仍是一大挑戰。傳統基於文本的方法無法處理與視覺相關的信息。另一方面,現有的基於視覺的RAG方法常受限於固定流程,且由於模型基礎能力激活不足,往往難以有效進行推理。鑑於RL已被證明對模型推理有益,我們引入了VRAG-RL,這是一個專為跨視覺豐富信息的複雜推理而設計的新穎RL框架。在此框架下,VLMs與搜索引擎互動,借助視覺感知標記自主採樣單輪或多輪推理軌跡,並基於這些樣本進行持續優化。我們的方法凸顯了RL在RAG領域的關鍵限制:(i) 先前的多模態RAG方法往往僅將圖像融入上下文,導致推理標記分配不足,忽視了視覺特有的感知;(ii) 當模型與搜索引擎互動時,其查詢常因無法清晰表達需求而未能檢索到相關信息,從而導致性能欠佳。為應對這些挑戰,我們定義了一個專為視覺豐富輸入量身定制的動作空間,包括裁剪和縮放等動作,使模型能從粗到細的視角收集信息。此外,為彌合用戶原始查詢與檢索器之間的差距,我們採用了一種簡單而有效的獎勵機制,該機制整合了查詢重寫與檢索性能,並結合了基於模型的獎勵。我們的VRAG-RL利用專門設計的RL策略優化VLMs以適應RAG任務,使模型與實際應用場景對齊。代碼可在https://github.com/Alibaba-NLP/VRAG{https://github.com/Alibaba-NLP/VRAG}獲取。
English
Effectively retrieving, reasoning and understanding visually rich information remains a challenge for RAG methods. Traditional text-based methods cannot handle visual-related information. On the other hand, current vision-based RAG approaches are often limited by fixed pipelines and frequently struggle to reason effectively due to the insufficient activation of the fundamental capabilities of models. As RL has been proven to be beneficial for model reasoning, we introduce VRAG-RL, a novel RL framework tailored for complex reasoning across visually rich information. With this framework, VLMs interact with search engines, autonomously sampling single-turn or multi-turn reasoning trajectories with the help of visual perception tokens and undergoing continual optimization based on these samples. Our approach highlights key limitations of RL in RAG domains: (i) Prior Multi-modal RAG approaches tend to merely incorporate images into the context, leading to insufficient reasoning token allocation and neglecting visual-specific perception; and (ii) When models interact with search engines, their queries often fail to retrieve relevant information due to the inability to articulate requirements, thereby leading to suboptimal performance. To address these challenges, we define an action space tailored for visually rich inputs, with actions including cropping and scaling, allowing the model to gather information from a coarse-to-fine perspective. Furthermore, to bridge the gap between users' original inquiries and the retriever, we employ a simple yet effective reward that integrates query rewriting and retrieval performance with a model-based reward. Our VRAG-RL optimizes VLMs for RAG tasks using specially designed RL strategies, aligning the model with real-world applications. The code is available at https://github.com/Alibaba-NLP/VRAG{https://github.com/Alibaba-NLP/VRAG}.

Summary

AI-Generated Summary

PDF103May 29, 2025