ChatPaper.aiChatPaper

基於正交等價變換的LLM訓練重參數化方法

Reparameterized LLM Training via Orthogonal Equivalence Transformation

June 9, 2025
作者: Zeju Qiu, Simon Buchholz, Tim Z. Xiao, Maximilian Dax, Bernhard Schölkopf, Weiyang Liu
cs.AI

摘要

尽管大规模语言模型(LLMs)正推动着人工智能的飞速发展,如何高效且可靠地训练这些大型模型仍是该领域最为重大的挑战之一。为应对这一挑战,我们提出了POET,一种新颖的再参数化训练算法,它利用正交等价变换来优化神经元。具体而言,POET通过两个可学习的正交矩阵和一个固定的随机权重矩阵对每个神经元进行再参数化。由于其在理论上能够保持权重矩阵的谱特性,POET能够稳定地优化目标函数,并提升泛化能力。我们进一步开发了高效的近似方法,使得POET在训练大规模神经网络时既灵活又可扩展。大量实验验证了POET在训练LLMs中的有效性和可扩展性。
English
While large language models (LLMs) are driving the rapid advancement of artificial intelligence, effectively and reliably training these large models remains one of the field's most significant challenges. To address this challenge, we propose POET, a novel reParameterized training algorithm that uses Orthogonal Equivalence Transformation to optimize neurons. Specifically, POET reparameterizes each neuron with two learnable orthogonal matrices and a fixed random weight matrix. Because of its provable preservation of spectral properties of weight matrices, POET can stably optimize the objective function with improved generalization. We further develop efficient approximations that make POET flexible and scalable for training large-scale neural networks. Extensive experiments validate the effectiveness and scalability of POET in training LLMs.
PDF22June 12, 2025