论文标题
贝叶斯深度学习的近似后期推断的影响很大
Ramifications of Approximate Posterior Inference for Bayesian Deep Learning in Adversarial and Out-of-Distribution Settings
论文作者
论文摘要
深层神经网络在各种歧视性分类任务中已经成功,尽管它们的校准较差通常为错误分类的预测分配了很高的可能性。当部署在实际应用程序中时,潜在的后果可能会导致模型的可信赖性和责任,在这些应用程序中,根据其置信度得分对预测进行评估。现有的解决方案表明,通过结合深层神经网络和贝叶斯推断,可以对模型模棱两可的数据点进行量化不确定性,从而获得了好处。在这项工作中,我们建议验证和测试基于似然模型在未分布检测任务(OOD)任务中的功效。在不同的数据集和指标中,我们表明贝叶斯深度学习模型在某些情况下略优先超过传统的神经网络,并且在IN/OUT分布类别之间的最小重叠时,即使是最佳模型也显示出检测OOD数据的AUC分数降低。初步研究表明,由于选择初始化,体系结构或激活功能,偏倚的潜在固有作用。我们假设,神经网络对看不见的输入的敏感性可能是由不同的建筑设计选择引起的多因素现象,通常会因维度的诅咒而增强。此外,我们进行了一项研究,以找到对抗性噪声方法对分布性能的影响的影响,还研究了贝叶斯深度学习者的对抗性噪声稳健性。
Deep neural networks have been successful in diverse discriminative classification tasks, although, they are poorly calibrated often assigning high probability to misclassified predictions. Potential consequences could lead to trustworthiness and accountability of the models when deployed in real applications, where predictions are evaluated based on their confidence scores. Existing solutions suggest the benefits attained by combining deep neural networks and Bayesian inference to quantify uncertainty over the models' predictions for ambiguous datapoints. In this work we propose to validate and test the efficacy of likelihood based models in the task of out of distribution detection (OoD). Across different datasets and metrics we show that Bayesian deep learning models on certain occasions marginally outperform conventional neural networks and in the event of minimal overlap between in/out distribution classes, even the best models exhibit a reduction in AUC scores in detecting OoD data. Preliminary investigations indicate the potential inherent role of bias due to choices of initialisation, architecture or activation functions. We hypothesise that the sensitivity of neural networks to unseen inputs could be a multi-factor phenomenon arising from the different architectural design choices often amplified by the curse of dimensionality. Furthermore, we perform a study to find the effect of the adversarial noise resistance methods on in and out-of-distribution performance, as well as, also investigate adversarial noise robustness of Bayesian deep learners.