### 抛物线原理:一种基于视觉中心位置编码的原则性方法
Where to Attend: A Principled Vision-Centric Position Encoding with Parabolas
February 1, 2026
作者: Christoffer Koo Øhrstrøm, Rafael I. Cabral Muchacho, Yifei Dong, Filippos Moumtzidellis, Ronja Güldenring, Florian T. Pokorny, Lazaros Nalpantidis
cs.AI
摘要
我们提出抛物线位置编码(PaPE),一种基于抛物线、适用于注意力架构中视觉模态的位置编码方法。针对图像、点云、视频或事件相机流等视觉标记集合,我们的目标是在编码其位置时兼顾视觉模态的特性。现有研究大多将语言模型中一维序列的位置编码扩展至视觉领域的n维结构,但仅部分考虑了视觉特性。我们通过从先前研究中提炼的设计原则来填补这一空白:平移不变性、旋转不变性(PaPE-RI)、距离衰减、方向性和上下文感知。我们在涵盖4种模态的8个数据集上评估PaPE,发现PaPE或PaPE-RI在7个数据集上均取得最优性能。ImageNet-1K的外推实验表明,PaPE具有卓越的外推能力,其绝对性能较次优位置编码提升最高达10.5%。代码已发布于https://github.com/DTU-PAS/parabolic-position-encoding。
English
We propose Parabolic Position Encoding (PaPE), a parabola-based position encoding for vision modalities in attention-based architectures. Given a set of vision tokens-such as images, point clouds, videos, or event camera streams-our objective is to encode their positions while accounting for the characteristics of vision modalities. Prior works have largely extended position encodings from 1D-sequences in language to nD-structures in vision, but only with partial account of vision characteristics. We address this gap by designing PaPE from principles distilled from prior work: translation invariance, rotation invariance (PaPE-RI), distance decay, directionality, and context awareness. We evaluate PaPE on 8 datasets that span 4 modalities. We find that either PaPE or PaPE-RI achieves the top performance on 7 out of 8 datasets. Extrapolation experiments on ImageNet-1K show that PaPE extrapolates remarkably well, improving in absolute terms by up to 10.5% over the next-best position encoding. Code is available at https://github.com/DTU-PAS/parabolic-position-encoding.