OpenGame:面向遊戲的開放式智慧體程式設計
OpenGame: Open Agentic Coding for Games
April 20, 2026
作者: Yilei Jiang, Jinyuan Hu, Qianyin Xiao, Yaozhi Zheng, Ruize Ma, Kaituo Feng, Jiaming Han, Tianshuo Peng, Kaixuan Fan, Manyuan Zhang, Xiangyu Yue
cs.AI
摘要
遊戲開發坐落於創意設計與複雜軟體工程的交匯點,需要統籌協調遊戲引擎、即時循環系統以及跨越多個文件的緊密耦合狀態。儘管大型語言模型和代碼智能體現已能輕鬆解決孤立的編程任務,但在根據高層次設計要求生成完整可玩遊戲時,它們往往因跨文件不一致性、場景連線斷裂和邏輯混亂而潰敗。我們通過OpenGame彌合這一鴻溝——首個專為端到端網頁遊戲創作設計的開源智能體框架。其核心是Game Skill技術,這種可演進的複用能力由兩部分構成:通過經驗積累形成項目骨架庫的模板技能,以及維護實時驗證修復方案的調試技能,二者協同使智能體能夠搭建穩定架構並系統性修復集成錯誤,而非僅修補孤立語法問題。驅動該框架的是GameCoder-27B,這個專精遊戲引擎的代碼大模型通過持續預訓練、監督微調與執行錨定強化學習的三階段管道打造而成。由於驗證交互可玩性本質上比檢查靜態代碼更困難,我們進一步推出OpenGame-Bench評估管道,通過無頭瀏覽器執行和視覺語言模型判分,從構建健康度、視覺可用性與意圖對齊度三個維度評測智能體遊戲生成能力。在150個多樣化遊戲提示詞測試中,OpenGame樹立了全新標杆。我們期待OpenGame能推動代碼智能體突破離散式軟體工程問題,邁向構建複雜交互式現實應用的新境界。本框架將完全開源。
English
Game development sits at the intersection of creative design and intricate software engineering, demanding the joint orchestration of game engines, real-time loops, and tightly coupled state across many files. While Large Language Models (LLMs) and code agents now solve isolated programming tasks with ease, they consistently stumble when asked to produce a fully playable game from a high-level design, collapsing under cross-file inconsistencies, broken scene wiring, and logical incoherence. We bridge this gap with OpenGame, the first open-source agentic framework explicitly designed for end-to-end web game creation. At its core lies Game Skill, a reusable, evolving capability composed of a Template Skill that grows a library of project skeletons from experience and a Debug Skill that maintains a living protocol of verified fixes - together enabling the agent to scaffold stable architectures and systematically repair integration errors rather than patch isolated syntax bugs. Powering this framework is GameCoder-27B, a code LLM specialized for game engine mastery through a three-stage pipeline of continual pre-training, supervised fine-tuning, and execution-grounded reinforcement learning. Since verifying interactive playability is fundamentally harder than checking static code, we further introduce OpenGame-Bench, an evaluation pipeline that scores agentic game generation along Build Health, Visual Usability, and Intent Alignment via headless browser execution and VLM judging. Across 150 diverse game prompts, OpenGame establishes a new state-of-the-art. We hope OpenGame pushes code agents beyond discrete software engineering problems and toward building complex, interactive real-world applications. Our framework will be fully open-sourced.