AIDB Daily Papers
政治・法律分野特化型LLM「PoliLegalLM」:技術報告
※ 日本語タイトル・ポイントはAIによる自動生成です。正確な内容は原論文をご確認ください。
ポイント
- 政治・法律分野に特化した大規模言語モデルPoliLegalLMを開発し、法的引用の誤りや推論能力の課題を解決しました。
- 大規模な法律コーパスと独自の学習フレームワークにより、ドメイン知識の習得とタスク適応能力を向上させました。
- 3つのベンチマーク評価で高い性能を示し、実世界の法律応用におけるPoliLegalLMの実用性を証明しました。
Abstract
Large language models (LLMs) have achieved remarkable success in general-domain tasks, yet their direct application to the legal domain remains challenging due to hallucinated legal citations, incomplete knowledge coverage, and weak structured reasoning. To address these issues, we propose PoliLegalLM, a domain-specific large language model tailored for political and legal applications. Our approach adopts a unified training framework that integrates continued pretraining, progressive supervised fine-tuning, and preference-based reinforcement learning to jointly enhance legal knowledge grounding, task alignment, and reasoning capability. We construct a large-scale, high-quality legal corpus and design a structured post-training pipeline, enabling the model to effectively learn domain-specific knowledge and adapt to diverse legal tasks. We evaluate PoliLegalLM on three representative benchmarks, including LawBench, LexEval, and a real-world dataset, PoliLegal. Experimental results demonstrate that PoliLegalLM achieves strong and consistent performance, outperforming competitive models of similar scale and remaining highly competitive with significantly larger models, while achieving the best results on real-world legal scenarios. These results highlight the effectiveness of our training paradigm and the practical value of domain-specific LLMs for real-world legal applications.
Paper AI Chat
この論文のPDF全文を対象にAIに質問できます。
質問の例: