ChatPaper.aiChatPaper

功率变换再探:数值稳定性与联邦化

Power Transform Revisited: Numerically Stable, and Federated

October 6, 2025
作者: Xuefeng Xu, Graham Cormode
cs.AI

摘要

幂变换是一种流行的参数化技术,用于使数据更接近高斯分布,并广泛用作统计分析和机器学习中的预处理步骤。然而,我们发现直接实现的幂变换存在严重的数值不稳定性,可能导致错误结果甚至系统崩溃。本文深入分析了这些不稳定性的来源,并提出了有效的解决方案。此外,我们将幂变换扩展到联邦学习环境中,解决了在此背景下出现的数值和分布挑战。在真实世界数据集上的实验表明,我们的方法既有效又稳健,与现有方法相比显著提高了稳定性。
English
Power transforms are popular parametric techniques for making data more Gaussian-like, and are widely used as preprocessing steps in statistical analysis and machine learning. However, we find that direct implementations of power transforms suffer from severe numerical instabilities, which can lead to incorrect results or even crashes. In this paper, we provide a comprehensive analysis of the sources of these instabilities and propose effective remedies. We further extend power transforms to the federated learning setting, addressing both numerical and distributional challenges that arise in this context. Experiments on real-world datasets demonstrate that our methods are both effective and robust, substantially improving stability compared to existing approaches.
PDF02October 7, 2025