ChatPaper.aiChatPaper

WebNovelBench:将LLM小说家置于网络小说分发平台

WebNovelBench: Placing LLM Novelists on the Web Novel Distribution

May 20, 2025
作者: Leon Lin, Jun Zheng, Haidong Wang
cs.AI

摘要

稳健评估大型语言模型(LLMs)的长篇叙事能力仍面临重大挑战,现有基准往往在规模、多样性或客观衡量标准上有所欠缺。为此,我们推出了WebNovelBench,这是一个专为评估长篇小说生成而设计的新颖基准。WebNovelBench利用了一个包含超过4000部中文网络小说的大规模数据集,将评估任务设定为从梗概到故事的生成。我们提出了一套多维度框架,涵盖八个叙事质量指标,通过LLM-as-Judge方法自动评估。采用主成分分析法汇总得分,并将其映射至与人类创作作品相比的百分位排名。实验表明,WebNovelBench能有效区分人类杰作、热门网络小说及LLM生成内容。我们对24个前沿LLM进行了全面分析,排序其叙事能力,并为未来发展提供洞见。该基准为评估和推进LLM驱动的叙事生成提供了一种可扩展、可复制且数据驱动的方法论。
English
Robustly evaluating the long-form storytelling capabilities of Large Language Models (LLMs) remains a significant challenge, as existing benchmarks often lack the necessary scale, diversity, or objective measures. To address this, we introduce WebNovelBench, a novel benchmark specifically designed for evaluating long-form novel generation. WebNovelBench leverages a large-scale dataset of over 4,000 Chinese web novels, framing evaluation as a synopsis-to-story generation task. We propose a multi-faceted framework encompassing eight narrative quality dimensions, assessed automatically via an LLM-as-Judge approach. Scores are aggregated using Principal Component Analysis and mapped to a percentile rank against human-authored works. Our experiments demonstrate that WebNovelBench effectively differentiates between human-written masterpieces, popular web novels, and LLM-generated content. We provide a comprehensive analysis of 24 state-of-the-art LLMs, ranking their storytelling abilities and offering insights for future development. This benchmark provides a scalable, replicable, and data-driven methodology for assessing and advancing LLM-driven narrative generation.

Summary

AI-Generated Summary

PDF32May 22, 2025