论文标题
ClarifyDelphi:在社会和道德情况下,加强澄清问题具有降解性奖励
ClarifyDelphi: Reinforced Clarification Questions with Defeasibility Rewards for Social and Moral Situations
论文作者
论文摘要
上下文是一切,即使是常识性道德推理。改变背景可以改变行动的道德判断;一般而言,“向朋友撒谎”是错误的,但如果旨在保护他们的生活,可能在道德上是可以接受的。 我们提出了ClarifyDelphi,这是一个交互式系统,该系统学会了提出澄清问题(例如,为什么对您的朋友撒谎?),以引起社会或道德情况的其他显着背景。我们认为,其潜在答案导致道德判断的问题是最有用的。因此,我们提出了一个增强性学习框架,并具有贬低性奖励,旨在最大程度地提高对问题的假设答案的道德判断。人类评估表明,与竞争基线相比,我们的系统产生了更相关,信息性和不稳定的问题。我们的工作最终受到认知科学的研究的启发,这些研究研究了道德认知的灵活性(即,道德规则可以弯曲的多种背景),我们希望朝这个方向进行研究可以帮助对道德判断的认知和计算调查。
Context is everything, even in commonsense moral reasoning. Changing contexts can flip the moral judgment of an action; "Lying to a friend" is wrong in general, but may be morally acceptable if it is intended to protect their life. We present ClarifyDelphi, an interactive system that learns to ask clarification questions (e.g., why did you lie to your friend?) in order to elicit additional salient contexts of a social or moral situation. We posit that questions whose potential answers lead to diverging moral judgments are the most informative. Thus, we propose a reinforcement learning framework with a defeasibility reward that aims to maximize the divergence between moral judgments of hypothetical answers to a question. Human evaluation demonstrates that our system generates more relevant, informative and defeasible questions compared to competitive baselines. Our work is ultimately inspired by studies in cognitive science that have investigated the flexibility in moral cognition (i.e., the diverse contexts in which moral rules can be bent), and we hope that research in this direction can assist both cognitive and computational investigations of moral judgments.