认知代价论:在边缘原生SLM中消解系统1与系统2推理以达成去中心化共识 (注:根据计算语言学原则,标题翻译采用"认知代价论"对应"Cognitive Penalty",既保留学术概念又符合中文表达习惯;"Ablating"译为"消解"更准确体现认知科学中抑制特定思维过程的含义;"Edge-Native SLMs"采用"边缘原生SLM"保持技术术语一致性;"Decentralized Consensus"译为"去中心化共识"符合区块链领域术语规范)
The Cognitive Penalty: Ablating System 1 and System 2 Reasoning in Edge-Native SLMs for Decentralized Consensus
April 18, 2026
作者: Syed Muhammad Aqdas Rizvi
cs.AI
摘要
去中心化自治组织(DAO)正倾向于采用小型语言模型(SLM)作为边缘原生宪制防火墙,用以审查提案并防范语义社会工程攻击。虽然扩展推理时计算(系统2)能增强形式逻辑能力,但其在高对抗性的加密经济治理环境中的有效性仍待深入探究。为此,我们推出Sentinel-Bench——一个包含840次推理的实证框架,对Qwen-3.5-9B模型执行严格的模型内消融实验。通过控制冻结权重下的潜在推理过程,我们分离出推理时计算在对抗性Optimism DAO数据集上的独立影响。研究结果揭示了严重的计算精度倒挂现象:自回归基线(系统1)在13秒内实现了100%的对抗鲁棒性、100%司法一致性和状态终局性;相反地,系统2推理引发了灾难性不稳定,其根本原因在于26.7%的推理不收敛率(认知崩溃)。这种崩溃使试验间共识稳定性降至72.6%,并产生17倍的延迟开销,为治理可提取价值(GEV)和硬件中心化埋下重大隐患。尽管罕见(仅占对抗试验的1.5%),我们实证捕捉到"推理诱导的谄媚现象"——模型生成超长内部独白(平均25,750字符)以合理化其落入对抗陷阱的失败行为。我们得出结论:对于在拜占庭容错(BFT)约束下运行的边缘原生SLM,系统1的参数化直觉在架构效率与经济性上均优于系统2的迭代审议机制,更适用于去中心化共识场景。
代码与数据集:https://github.com/smarizvi110/sentinel-bench
English
Decentralized Autonomous Organizations (DAOs) are inclined explore Small Language Models (SLMs) as edge-native constitutional firewalls to vet proposals and mitigate semantic social engineering. While scaling inference-time compute (System 2) enhances formal logic, its efficacy in highly adversarial, cryptoeconomic governance environments remains underexplored. To address this, we introduce Sentinel-Bench, an 840-inference empirical framework executing a strict intra-model ablation on Qwen-3.5-9B. By toggling latent reasoning across frozen weights, we isolate the impact of inference-time compute against an adversarial Optimism DAO dataset. Our findings reveal a severe compute-accuracy inversion. The autoregressive baseline (System 1) achieved 100% adversarial robustness, 100% juridical consistency, and state finality in under 13 seconds. Conversely, System 2 reasoning introduced catastrophic instability, fundamentally driven by a 26.7% Reasoning Non-Convergence (cognitive collapse) rate. This collapse degraded trial-to-trial consensus stability to 72.6% and imposed a 17x latency overhead, introducing critical vulnerabilities to Governance Extractable Value (GEV) and hardware centralization. While rare (1.5% of adversarial trials), we empirically captured "Reasoning-Induced Sycophancy," where the model generated significantly longer internal monologues (averaging 25,750 characters) to rationalize failing the adversarial trap. We conclude that for edge-native SLMs operating under Byzantine Fault Tolerance (BFT) constraints, System 1 parameterized intuition is structurally and economically superior to System 2 iterative deliberation for decentralized consensus.
Code and Dataset: https://github.com/smarizvi110/sentinel-bench