ChatPaper.aiChatPaper

現代機器翻譯的新趨勢:基於大型推理模型

New Trends for Modern Machine Translation with Large Reasoning Models

March 13, 2025
作者: Sinuo Liu, Chenyang Lyu, Minghao Wu, Longyue Wang, Weihua Luo, Kaifu Zhang
cs.AI

摘要

大型推理模型(LRMs)的最新進展,尤其是那些利用思維鏈推理(CoT)的模型,為機器翻譯(MT)開闢了全新的可能性。本立場文件主張,LRMs通過將翻譯重新定義為一項需要上下文、文化和語言理解與推理的動態推理任務,從根本上轉變了傳統神經機器翻譯以及基於大型語言模型(LLMs)的翻譯範式。我們識別出三大基礎性轉變:1)上下文連貫性,LRMs通過對跨句子和複雜上下文甚至缺乏上下文的顯式推理來解決歧義並保持話語結構;2)文化意圖性,使模型能夠通過推斷說話者意圖、受眾期望和社會語言規範來調整輸出;3)自我反思,LRMs能夠在推理過程中進行自我反思,以糾正翻譯中的潛在錯誤,特別是在極端噪聲情況下,相比於簡單的X->Y映射翻譯,展現出更好的魯棒性。我們通過展示實證案例,探討了翻譯中的各種場景,包括風格化翻譯、文檔級翻譯和多模態翻譯,這些案例展示了LRMs在翻譯中的優越性。我們還識別了LRMs在機器翻譯中的幾個有趣現象,如自動樞紐翻譯,以及關鍵挑戰,如翻譯中的過度本地化和推理效率。總之,我們認為LRMs重新定義了翻譯系統,使其不僅僅是文本轉換器,而是能夠推理文本之外意義的多語言認知代理。這一範式轉變提醒我們,在更廣泛的背景下,利用LRMs思考翻譯問題——我們在此基礎上能實現什麼。
English
Recent advances in Large Reasoning Models (LRMs), particularly those leveraging Chain-of-Thought reasoning (CoT), have opened brand new possibility for Machine Translation (MT). This position paper argues that LRMs substantially transformed traditional neural MT as well as LLMs-based MT paradigms by reframing translation as a dynamic reasoning task that requires contextual, cultural, and linguistic understanding and reasoning. We identify three foundational shifts: 1) contextual coherence, where LRMs resolve ambiguities and preserve discourse structure through explicit reasoning over cross-sentence and complex context or even lack of context; 2) cultural intentionality, enabling models to adapt outputs by inferring speaker intent, audience expectations, and socio-linguistic norms; 3) self-reflection, LRMs can perform self-reflection during the inference time to correct the potential errors in translation especially extremely noisy cases, showing better robustness compared to simply mapping X->Y translation. We explore various scenarios in translation including stylized translation, document-level translation and multimodal translation by showcasing empirical examples that demonstrate the superiority of LRMs in translation. We also identify several interesting phenomenons for LRMs for MT including auto-pivot translation as well as the critical challenges such as over-localisation in translation and inference efficiency. In conclusion, we think that LRMs redefine translation systems not merely as text converters but as multilingual cognitive agents capable of reasoning about meaning beyond the text. This paradigm shift reminds us to think of problems in translation beyond traditional translation scenarios in a much broader context with LRMs - what we can achieve on top of it.

Summary

AI-Generated Summary

PDF232March 14, 2025