AIDB Daily Papers
LLMは物語をどう捉える?小説の要約から読み解く注意の流れ
※ 日本語タイトル・ポイントはAIによる自動生成です。正確な内容は原論文をご確認ください。
ポイント
- 本研究では、LLMが長文テキストを理解する能力を、小説の要約生成を通して評価しました。
- 人間とLLMが生成した要約を比較することで、物語の重要な部分に対する認識の違いを明らかにしました。
- LLMは物語の終盤を重視する傾向があり、人間の物語理解とのずれが示唆されました。
Abstract
Although LLM context lengths have grown, there is evidence that their ability to integrate information across long-form texts has not kept pace. We evaluate one such understanding task: generating summaries of novels. When human authors of summaries compress a story, they reveal what they consider narratively important. Therefore, by comparing human and LLM-authored summaries, we can assess whether models mirror human patterns of conceptual engagement with texts. To measure conceptual engagement, we align sentences from 150 human-written novel summaries with the specific chapters they reference. We demonstrate the difficulty of this alignment task, which indicates the complexity of summarization as a task. We then generate and align additional summaries by nine state-of-the-art LLMs for each of the 150 reference texts. Comparing the human and model-authored summaries, we find both stylistic differences between the texts and differences in how humans and LLMs distribute their focus throughout a narrative, with models emphasizing the ends of texts. Comparing human narrative engagement with model attention mechanisms suggests explanations for degraded narrative comprehension and targets for future development. We release our dataset to support future research.
Paper AI Chat
この論文のPDF全文を対象にAIに質問できます。
質問の例: