ChatPaper.aiChatPaper

潜在后验因子的理论基础:多证据推理的形式化保证

Theoretical Foundations of Latent Posterior Factors: Formal Guarantees for Multi-Evidence Reasoning

March 13, 2026
作者: Aliyu Agboola Alege
cs.AI

摘要

我们提出对潜在后验因子(LPF)的完整理论刻画,该框架为概率预测任务中聚合多源异构证据提供了原则性方法。多证据推理普遍存在于高风险领域,如医疗诊断、金融风险评估、法律案例分析和监管合规等,然而现有方法要么缺乏理论保证,要么在架构上无法处理多证据场景。LPF通过变分自编码器将每个证据项编码为高斯潜在后验分布,利用蒙特卡洛边际化将后验转化为软因子,并借助精确的和积网络推理(LPF-SPN)或习得神经网络聚合器(LPF-Learned)实现因子聚合。 我们证明了涵盖可信人工智能关键需求的七项形式化保证:校准保持性(预期校准误差≤ε+C/√K_eff);蒙特卡洛误差以O(1/√M)速率衰减;在N=4200时获得训练-测试差距为0.0085的非平凡PAC-贝叶斯界;运行效率达信息理论下界的1.12倍;在半数证据被对抗替换时仍保持88%性能的优雅退化特性(退化速率O(εδ√K));校准衰减速率O(1/√K)且决定系数R²=0.849;以及误差低于0.002%的精确认知-偶然不确定性分解。所有定理均在包含4200个训练样本的受控数据集上得到实证验证。我们的理论框架确立了LPF作为安全关键应用中可信多证据人工智能的基础方法。
English
We present a complete theoretical characterization of Latent Posterior Factors (LPF), a principled framework for aggregating multiple heterogeneous evidence items in probabilistic prediction tasks. Multi-evidence reasoning arises pervasively in high-stakes domains including healthcare diagnosis, financial risk assessment, legal case analysis, and regulatory compliance, yet existing approaches either lack formal guarantees or fail to handle multi-evidence scenarios architecturally. LPF encodes each evidence item into a Gaussian latent posterior via a variational autoencoder, converting posteriors to soft factors through Monte Carlo marginalization, and aggregating factors via exact Sum-Product Network inference (LPF-SPN) or a learned neural aggregator (LPF-Learned). We prove seven formal guarantees spanning the key desiderata for trustworthy AI: Calibration Preservation (ECE <= epsilon + C/sqrt(K_eff)); Monte Carlo Error decaying as O(1/sqrt(M)); a non-vacuous PAC-Bayes bound with train-test gap of 0.0085 at N=4200; operation within 1.12x of the information-theoretic lower bound; graceful degradation as O(epsilon*delta*sqrt(K)) under corruption, maintaining 88% performance with half of evidence adversarially replaced; O(1/sqrt(K)) calibration decay with R^2=0.849; and exact epistemic-aleatoric uncertainty decomposition with error below 0.002%. All theorems are empirically validated on controlled datasets spanning up to 4,200 training examples. Our theoretical framework establishes LPF as a foundation for trustworthy multi-evidence AI in safety-critical applications.
PDF12March 19, 2026