ChatPaper.aiChatPaper

Prism-Δ:面向大语言模型提示强调的差分子空间引导技术

Prism-Δ: Differential Subspace Steering for Prompt Highlighting in Large Language Models

March 11, 2026
作者: Yuyao Ge, Shenghua Liu, Yiwei Wang, Tianyu Liu, Baolong Bi, Lingrui Mei, Jiayu Yao, Jiafeng Guo, Xueqi Cheng
cs.AI

摘要

提示词高亮技术通过引导大型语言模型在生成过程中优先处理用户指定的文本片段,其核心挑战在于提取能够区分相关与无关语境的引导方向,而非两者共有的结构模式。我们提出PRISM-Δ(基于投影的相关性感知引导方法),该方法通过解构正负交叉协方差矩阵的差异,在消除共享方向的同时最大化判别能量。每个注意力头会获得连续的softplus重要性权重,使弱效但有用的注意力头能以降低的强度参与计算。该框架可自然扩展至Value表征,捕获仅使用Key的方法所忽略的内容通道信号。在四个基准测试和五种模型上的实验表明,PRISM-Δ在20种配置中的19项达到或超越现有最佳方法,相对增益最高达+10.6%,同时将引导的流畅性损耗降低一半。PRISM-Δ还能适应长上下文检索任务,以最高+4.8%的相对增益优于现有最佳方法。该方法兼容FlashAttention技术,且仅增加可忽略的内存开销。
English
Prompt highlighting steers a large language model to prioritize user-specified text spans during generation. A key challenge is extracting steering directions that capture the difference between relevant and irrelevant contexts, rather than shared structural patterns common to both. We propose PRISM-Δ (Projection-based Relevance-Informed Steering Method), which decomposes the difference between positive and negative cross-covariance matrices to maximize discriminative energy while eliminating shared directions. Each attention head receives a continuous softplus importance weight, letting weak-but-useful heads contribute at reduced strength. The framework extends naturally to Value representations, capturing content-channel signal that Key-only methods leave unused. Across four benchmarks and five models, PRISM-Δ matches or exceeds the best existing method on 19 of 20 configurations, with relative gains up to +10.6%, while halving the fluency cost of steering. PRISM-Δ also scales to long-context retrieval, outperforming the best existing method by up to +4.8% relative gain. PRISM-Δ is compatible with FlashAttention and adds negligible memory overhead.
PDF101March 13, 2026