AIDB Daily Papers
宇宙状況認識LLM:認知階層化データ合成によるドメイン適応
※ 日本語タイトル・ポイントはAIによる自動生成です。正確な内容は原論文をご確認ください。
ポイント
- 宇宙状況認識(SSA)へのLLM適用のため、高品質な教師あり微調整(SFT)データセットを構築するフレームワークを提案。
- 知識構造化、認知階層化質問モデリング、自動品質管理により、知識網羅性、認知深度、品質制御の課題を克服し、ドメイン知識を効率的にLLMへ注入。
- SSA-SFTデータセットでQwen3-8Bを微調整した結果、BLEUスコアが大幅に向上し、汎用性能を維持しつつ、ドメイン特化LLMの有効性を示した。
Abstract
Large language models (LLMs) demonstrate exceptional performance on general-purpose tasks. however, transferring them to complex engineering domains such as space situational awareness (SSA) remains challenging owing to insufficient structural alignment with mission chains, the absence of higher-order cognitive supervision, and poor correspondence between data quality criteria and engineering specifications. The core bottleneck is the construction of high-quality supervised fine-tuning (SFT) datasets. To this end, we propose BD-FDG (Bloom's Taxonomy-based Domain-specific Fine-tuning Data Generation), a framework that addresses incomplete knowledge coverage, shallow cognitive depth, and limited quality controllability through three mechanisms: structured knowledge organization, cognitively layered question modeling, and automated quality control. The framework uses a knowledge tree to ensure structured corpus coverage, designs a question generation scheme spanning nine categories and six cognitive levels from Remember to Create to produce samples with a continuous difficulty gradient, and applies a multidimensional scoring pipeline to enforce domain rigor and consistency. Using BD-FDG, we construct SSA-SFT, a domain dataset of approximately 230K samples, and fine-tune Qwen3-8B to obtain SSA-LLM-8B. Experiments show that SSA-LLM-8B achieves relative BLEU-1 improvements of 144% (no-think) and 176% (think) on the domain test set and a win rate of 82.21% over the baseline in arena comparisons, while largely preserving general benchmark performance (MMLU-Pro, MATH-500). These results validate SFT data construction driven by cognitive layering as an effective paradigm for complex engineering domains and provide a transferable framework for domain-specific LLM adaptation.
Paper AI Chat
この論文のPDF全文を対象にAIに質問できます。
質問の例: