论文标题
与概率解释的对比分类和表示学习
Contrastive Classification and Representation Learning with Probabilistic Interpretation
论文作者
论文摘要
交叉熵损失已成为基于分类任务的主要目标函数。它广泛部署在学习神经网络分类器中,既显示出有效性又显示出概率的解释。最近,在自我监督的对比表示学习方法的成功之后,与仅通过跨熵损失进行训练相比,已经提出了有监督的对比方法来学习表示形式并表现出更高,更健壮的性能。但是,仍需要跨熵损失来训练最终分类层。在这项工作中,我们使用一个目标函数研究了表示表示和分类器的可能性,该目标函数结合了对比度学习的鲁棒性和跨熵损失的概率解释。首先,我们重新访问先前提出的基于对比的目标函数,该目标函数近似交叉熵损失,并提供一个简单的扩展,以共同学习分类器。其次,我们提出了一个新版本的监督对比培训,该培训共同学习了分类器的参数和网络的骨干。我们从经验上表明,我们提出的目标功能在各种挑战的环境中具有更大的训练稳定性和鲁棒性,对标准横熵损失有了显着改善。
Cross entropy loss has served as the main objective function for classification-based tasks. Widely deployed for learning neural network classifiers, it shows both effectiveness and a probabilistic interpretation. Recently, after the success of self supervised contrastive representation learning methods, supervised contrastive methods have been proposed to learn representations and have shown superior and more robust performance, compared to solely training with cross entropy loss. However, cross entropy loss is still needed to train the final classification layer. In this work, we investigate the possibility of learning both the representation and the classifier using one objective function that combines the robustness of contrastive learning and the probabilistic interpretation of cross entropy loss. First, we revisit a previously proposed contrastive-based objective function that approximates cross entropy loss and present a simple extension to learn the classifier jointly. Second, we propose a new version of the supervised contrastive training that learns jointly the parameters of the classifier and the backbone of the network. We empirically show that our proposed objective functions show a significant improvement over the standard cross entropy loss with more training stability and robustness in various challenging settings.