AIDB Daily Papers
道徳的にプログラムされたLLMが人間の道徳観を再構築する
※ 日本語タイトル・ポイントはAIによる自動生成です。正確な内容は原論文をご確認ください。
ポイント
- 本研究では、道徳的原則を組み込んだLLMとのインタラクションが、人間の道徳的傾向に与える影響を調査しました。
- 倫理的指針をLLMに組み込むことで、その行動を制限するだけでなく、人間の道徳観を形成する可能性があるという重要な問題提起です。
- 実験の結果、義務論的または功利主義的な原則を持つLLMとの対話により、人間の道徳的判断がLLMの原則に沿って変化することが示されました。
Abstract
As large language models (LLMs) increasingly participate in high-stakes decision-making, a central societal debate has revolved around which moral frameworks-deontological or utilitarian-should guide machine behavior. However, a largely overlooked question is whether the moral principles that humans encode in LLMs could, through repeated interactions, reshape human moral inclinations. We developed two LLMs programmed with either deontological principles (D-LLM) or utilitarian principles (U-LLM) and conducted two pre-registered experiments involving extensive human-LLM interactions, comprising 15,985 total exchanges across the two experiments. Results show that interacting with these morally programmed LLMs systematically shifted human moral inclinations to align with the principles embedded in these systems. These effects remained strong two weeks after the interaction, with only slight decay, suggesting deep internalization rather than superficial agreement. Further, LLM-induced shifts in human moral inclinations translated into meaningful changes in socio-political policy evaluations, shaping how individuals approach contentious social issues. Overall, these results demonstrate that morally programmed LLMs can shape-not merely reflect-human morality, revealing a critical design paradox: embedding moral principles in LLMs not only restricts their behavior but also poses the risk of shaping human morality, raising important ethical and policy questions about who determines which principles intelligent machines should adhere to.
Paper AI Chat
この論文のPDF全文を対象にAIに質問できます。
質問の例: