ChatPaper.aiChatPaper

EXAONE 3.0 7.8B 指令調校語言模型

EXAONE 3.0 7.8B Instruction Tuned Language Model

August 7, 2024
作者: LG AI Research, Soyoung An, Kyunghoon Bae, Eunbi Choi, Stanley Jungkyu Choi, Yemuk Choi, Seokhee Hong, Yeonjung Hong, Junwon Hwang, Hyojin Jeon, Gerrard Jeongwon Jo, Hyunjik Jo, Jiyeon Jung, Yountae Jung, Euisoon Kim, Hyosang Kim, Joonkee Kim, Seonghwan Kim, Soyeon Kim, Sunkyoung Kim, Yireun Kim, Youchul Kim, Edward Hwayoung Lee, Haeju Lee, Honglak Lee, Jinsik Lee, Kyungmin Lee, Moontae Lee, Seungjun Lee, Woohyung Lim, Sangha Park, Sooyoun Park, Yongmin Park, Boseong Seo, Sihoon Yang, Heuiyeen Yeen, Kyungjae Yoo, Hyeongu Yun
cs.AI

摘要

我們介紹了 EXAONE 3.0 指令調校語言模型,這是由 LG AI 研究開發的大型語言模型(LLMs)家族中首個開放模型。在不同的模型尺寸中,我們公開釋出了 78 億個指令調校模型,以促進開放研究和創新。通過在廣泛的公共和內部基準測試中進行詳盡評估,EXAONE 3.0 展示了與其他同等尺寸的最新開放模型相比具有高度競爭力的真實世界表現,並具有遵循指令的能力。我們的比較分析顯示,EXAONE 3.0 在韓文方面表現出色,同時在一般任務和複雜推理方面取得了引人注目的表現。憑藉其強大的真實世界效能和雙語能力,我們希望 EXAONE 繼續為專家級 AI 的進步做出貢獻。我們的 EXAONE 3.0 指令調校模型可在以下網址獲得:https://huggingface.co/LGAI-EXAONE/EXAONE-3.0-7.8B-Instruct
English
We introduce EXAONE 3.0 instruction-tuned language model, the first open model in the family of Large Language Models (LLMs) developed by LG AI Research. Among different model sizes, we publicly release the 7.8B instruction-tuned model to promote open research and innovations. Through extensive evaluations across a wide range of public and in-house benchmarks, EXAONE 3.0 demonstrates highly competitive real-world performance with instruction-following capability against other state-of-the-art open models of similar size. Our comparative analysis shows that EXAONE 3.0 excels particularly in Korean, while achieving compelling performance across general tasks and complex reasoning. With its strong real-world effectiveness and bilingual proficiency, we hope that EXAONE keeps contributing to advancements in Expert AI. Our EXAONE 3.0 instruction-tuned model is available at https://huggingface.co/LGAI-EXAONE/EXAONE-3.0-7.8B-Instruct

Summary

AI-Generated Summary

PDF363November 28, 2024