AIDB Daily Papers
AnomalyGen:コード誘導型データ拡張によるログベース異常検知の強化
※ 日本語タイトル・ポイントはAIによる自動生成です。正確な内容は原論文をご確認ください。
ポイント
- AnomalyGenは、ソースコードからラベル付きログシーケンスを合成し、学習データのスパース性を解消する。
- ログ指向静的解析とLLM推論を組み合わせ、実行パスの検証と現実的なランタイムパラメータ生成を実現した点が新しい。
- HDFSとZookeeperの評価で、深層学習モデルのF1スコアが平均2.18%と1.69%向上し、Transformerでは0.818から0.970へ飛躍した。
Abstract
Log-based anomaly detection is fundamentally constrained by training data sparsity. Our empirical study reveals that public benchmark datasets cover less than 10% of source code log templates. Consequently, models frequently misclassify unseen but valid execution paths as anomalies, leading to false alarms. To address this, we propose AnomalyGen, a novel framework that augments training data by synthesizing labeled log sequences from source code. AnomalyGen combines log-oriented static analysis with Large Language Model (LLM) reasoning in three stages: (1) building Log-Oriented Control Flow Graphs (LCFGs) to enumerate structurally valid execution paths; (2) applying LLM Chain-of-Thought (CoT) reasoning to verify logical consistency and generate realistic runtime parameters (e.g., block IDs, IP addresses); and (3) labeling generated sequences with domain heuristics. Evaluations on HDFS and Zookeeper across 12 diverse anomaly detection models show AnomalyGen consistently improves performance. Deep learning models achieved average F1-score gains of 2.18% (HDFS) and 1.69% (Zookeeper), with an unsupervised Transformer on HDFS jumping from 0.818 to 0.970. Ablation results show that both static analysis and LLM-based verification are necessary: removing them reduces F1 by up to 8.7 and 10.7 percentage points, respectively. Our framework and datasets are publicly available to facilitate future research.
Paper AI Chat
この論文のPDF全文を対象にAIに質問できます。
質問の例: