AIDB Daily Papers
Bitboard Tetris AI:高速化と強化学習アルゴリズムの最適化
※ 日本語タイトル・ポイントはAIによる自動生成です。正確な内容は原論文をご確認ください。
ポイント
- Bitboardを用いてTetrisのゲームエンジンを再設計し、衝突判定やライン消去などの処理を高速化するAIフレームワークを提案した。
- Afterstate評価型アクターネットワークを導入し、状態価値の推定を簡略化、パラメータ数を削減しつつ、既存手法を上回る性能を実現した点が新しい。
- 最適化されたPPOアルゴリズムにより、3分以内に平均3,829点のスコアを達成し、大規模な逐次的意思決定研究への貢献が期待される。
Abstract
The efficiency of game engines and policy optimization algorithms is crucial for training reinforcement learning (RL) agents in complex sequential decision-making tasks, such as Tetris. Existing Tetris implementations suffer from low simulation speeds, suboptimal state evaluation, and inefficient training paradigms, limiting their utility for large-scale RL research. To address these limitations, this paper proposes a high-performance Tetris AI framework based on bitboard optimization and improved RL algorithms. First, we redesign the Tetris game board and tetrominoes using bitboard representations, leveraging bitwise operations to accelerate core processes (e.g., collision detection, line clearing, and Dellacherie-Thiery Features extraction) and achieve a 53-fold speedup compared to OpenAI Gym-Tetris. Second, we introduce an afterstate-evaluating actor network that simplifies state value estimation by leveraging Tetris afterstate property, outperforming traditional action-value networks with fewer parameters. Third, we propose a buffer-optimized Proximal Policy Optimization (PPO) algorithm that balances sampling and update efficiency, achieving an average score of 3,829 on 10x10 grids within 3 minutes. Additionally, we develop a Python-Java interface compliant with the OpenAI Gym standard, enabling seamless integration with modern RL frameworks. Experimental results demonstrate that our framework enhances Tetris's utility as an RL benchmark by bridging low-level bitboard optimizations with high-level AI strategies, providing a sample-efficient and computationally lightweight solution for scalable sequential decision-making research.
Paper AI Chat
この論文のPDF全文を対象にAIに質問できます。
質問の例: