AIDB Daily Papers
物語の迷宮:LLMによる長編ストーリー生成における一貫性の欠如
※ 日本語タイトル・ポイントはAIによる自動生成です。正確な内容は原論文をご確認ください。
ポイント
- 長編ストーリー生成におけるLLMの一貫性欠如を評価するベンチマークConStory-Benchを提案。
- 既存研究が plot の質に偏る中、物語の一貫性という重要な課題に焦点を当てている点が新しい。
- 事実や時間に関する矛盾が頻発し、物語の中盤や複雑な箇所でエラーが起こりやすいことが判明。
Abstract
What happens when a storyteller forgets its own story? Large Language Models (LLMs) can now generate narratives spanning tens of thousands of words, but they often fail to maintain consistency throughout. When generating long-form narratives, these models can contradict their own established facts, character traits, and world rules. Existing story generation benchmarks focus mainly on plot quality and fluency, leaving consistency errors largely unexplored. To address this gap, we present ConStory-Bench, a benchmark designed to evaluate narrative consistency in long-form story generation. It contains 2,000 prompts across four task scenarios and defines a taxonomy of five error categories with 19 fine-grained subtypes. We also develop ConStory-Checker, an automated pipeline that detects contradictions and grounds each judgment in explicit textual evidence. Evaluating a range of LLMs through five research questions, we find that consistency errors show clear tendencies: they are most common in factual and temporal dimensions, tend to appear around the middle of narratives, occur in text segments with higher token-level entropy, and certain error types tend to co-occur. These findings can inform future efforts to improve consistency in long-form narrative generation. Our project page is available at https://picrew.github.io/constory-bench.github.io/.
Paper AI Chat
この論文のPDF全文を対象にAIに質問できます。
質問の例: