论文标题
自PU:自我增强和校准的积极未标记的培训
Self-PU: Self Boosted and Calibrated Positive-Unlabeled Training
论文作者
论文摘要
许多现实世界中的应用程序必须解决积极的(PU)学习问题,即从大量未标记的数据和一些标记为正面示例中学习二进制分类器。尽管当前的最新方法采用重要的重新加权来设计各种风险估计器,但他们忽略了模型本身的学习能力,这可以提供可靠的监督。这促使我们提出了一个新颖的自PU学习框架,该框架无缝地整合了PU学习和自我训练。自PU强调了三个面向“自我”的构件:一种自定进度的培训算法,随着培训的进行,可以自适应地发现并增强自信的正面/负面示例;自我校准的实例感知损失;以及一种自我依据的计划,介绍了教师学习作为PU学习的有效正规化。我们演示了自PU在普通PU学习基准(MNIST和CIFAR-10)上的最新表现,这与最新的竞争对手相比有利。此外,我们研究了PU学习的现实应用,即对阿尔茨海默氏病的大脑形象进行分类。 Self-PU在现有方法上,Self-PU在著名的阿尔茨海默氏病神经影像倡议(ADNI)数据库中获得了显着改善的结果。该代码可公开可用:https://github.com/tamu-vita/self-pu。
Many real-world applications have to tackle the Positive-Unlabeled (PU) learning problem, i.e., learning binary classifiers from a large amount of unlabeled data and a few labeled positive examples. While current state-of-the-art methods employ importance reweighting to design various risk estimators, they ignored the learning capability of the model itself, which could have provided reliable supervision. This motivates us to propose a novel Self-PU learning framework, which seamlessly integrates PU learning and self-training. Self-PU highlights three "self"-oriented building blocks: a self-paced training algorithm that adaptively discovers and augments confident positive/negative examples as the training proceeds; a self-calibrated instance-aware loss; and a self-distillation scheme that introduces teacher-students learning as an effective regularization for PU learning. We demonstrate the state-of-the-art performance of Self-PU on common PU learning benchmarks (MNIST and CIFAR-10), which compare favorably against the latest competitors. Moreover, we study a real-world application of PU learning, i.e., classifying brain images of Alzheimer's Disease. Self-PU obtains significantly improved results on the renowned Alzheimer's Disease Neuroimaging Initiative (ADNI) database over existing methods. The code is publicly available at: https://github.com/TAMU-VITA/Self-PU.