论文标题

值得信赖的AI

Trustworthy AI

论文作者

Singh, Richa, Vatsa, Mayank, Ratha, Nalini

论文摘要

现代AI系统正在收获新颖的学习方法的优势。随着使用的日益增加,我们正在意识到这些系统的局限性和缺口。对输入数据的较小对抗性变化,解释决策的能力,解决训练数据的偏见,高度不透明度,在揭示系统的血统,如何经过培训和测试方面以及在哪些参数和条件下,他们可以可靠地保证一定程度的绩效,这是最突出的限制。还需要确保数据的隐私和安全性,向数据源分配适当的信用以及交付体面的输出也是AI系统的功能。我们建议有关值得信赖的AI的教程,以解决增强用户和公众对AI系统的信任的六个关键问题,即:(i)偏见和公平性,(ii)可解释性,(iii)强大的缓解对抗性攻击,(iii)(IV)改进模型构建中的隐私和安全性,包括适当的模型和(VI),以及(VI)的模型,以及(VI)的模型,以及(VI)的模型级别;血统中的透明度。

Modern AI systems are reaping the advantage of novel learning methods. With their increasing usage, we are realizing the limitations and shortfalls of these systems. Brittleness to minor adversarial changes in the input data, ability to explain the decisions, address the bias in their training data, high opacity in terms of revealing the lineage of the system, how they were trained and tested, and under which parameters and conditions they can reliably guarantee a certain level of performance, are some of the most prominent limitations. Ensuring the privacy and security of the data, assigning appropriate credits to data sources, and delivering decent outputs are also required features of an AI system. We propose the tutorial on Trustworthy AI to address six critical issues in enhancing user and public trust in AI systems, namely: (i) bias and fairness, (ii) explainability, (iii) robust mitigation of adversarial attacks, (iv) improved privacy and security in model building, (v) being decent, and (vi) model attribution, including the right level of credit assignment to the data sources, model architectures, and transparency in lineage.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源