论文标题
证据有条件神经过程
Evidential Conditional Neural Processes
论文作者
论文摘要
有条件的神经过程(CNP)模型家族为解决了更好的可扩展性和竞争性预测性能,为解决几个问题提供了一个有希望的方向。但是,当前的CNP模型仅捕获目标数据点上的预测的总体不确定性。他们在不同的不确定性来源上缺乏系统的细粒度量化,这对于在几次射击设置下对于模型训练和决策至关重要。我们提出了证据的条件神经过程(ECNP),该过程通过证据学习取代了CNP使用的标准高斯分布,从而取代了更丰富的层次贝叶斯结构,以实现认识论的不确定性分解。证据层次结构还导致了对嘈杂训练任务的理论上合理的鲁棒性。对拟议的ECNP的理论分析建立了与CNP的关系,同时对证据参数的作用有了更深入的见解。对合成和现实世界数据进行的广泛实验证明了我们在各种少量设置中提出的模型的有效性。
The Conditional Neural Process (CNP) family of models offer a promising direction to tackle few-shot problems by achieving better scalability and competitive predictive performance. However, the current CNP models only capture the overall uncertainty for the prediction made on a target data point. They lack a systematic fine-grained quantification on the distinct sources of uncertainty that are essential for model training and decision-making under the few-shot setting. We propose Evidential Conditional Neural Processes (ECNP), which replace the standard Gaussian distribution used by CNP with a much richer hierarchical Bayesian structure through evidential learning to achieve epistemic-aleatoric uncertainty decomposition. The evidential hierarchical structure also leads to a theoretically justified robustness over noisy training tasks. Theoretical analysis on the proposed ECNP establishes the relationship with CNP while offering deeper insights on the roles of the evidential parameters. Extensive experiments conducted on both synthetic and real-world data demonstrate the effectiveness of our proposed model in various few-shot settings.