论文标题
零射击学习的迭代共同训练转导框架
An Iterative Co-Training Transductive Framework for Zero Shot Learning
论文作者
论文摘要
在零射击学习(ZSL)社区中,通常认识到,转导学习的性能要比归纳阶段更好,因为在其训练阶段也使用了看不见的级别样本。如何为看不见的样本生成伪标签以及如何使用这种通常嘈杂的伪标签是跨导性学习的两个关键问题。在这项工作中,我们引入了一个迭代共同训练框架,该框架包含两个不同的基本ZSL模型和一个交换模块。在每次迭代中,将两个不同的ZSL模型共同训练,以分别预测未见级别样品的伪标签,而交换模块交换了预测的伪标签,然后将交换的伪标记样品添加到下一个迭代的训练集中。这样,我们的框架可以通过充分利用这两个模型的分类功能的潜在互补性来逐渐提高ZSL性能。此外,我们的共同训练框架还应用于广义ZSL(GZSL),在该ZSL(GZSL)中,提议在同类级别的分类之前选择一个语义引导的OOD检测器,以减轻GZSL中最有可能看不见的级别样本。三个基准测试的广泛实验表明,我们提出的方法的表现可能会大约超过31美元的最先进的方法。
In zero-shot learning (ZSL) community, it is generally recognized that transductive learning performs better than inductive one as the unseen-class samples are also used in its training stage. How to generate pseudo labels for unseen-class samples and how to use such usually noisy pseudo labels are two critical issues in transductive learning. In this work, we introduce an iterative co-training framework which contains two different base ZSL models and an exchanging module. At each iteration, the two different ZSL models are co-trained to separately predict pseudo labels for the unseen-class samples, and the exchanging module exchanges the predicted pseudo labels, then the exchanged pseudo-labeled samples are added into the training sets for the next iteration. By such, our framework can gradually boost the ZSL performance by fully exploiting the potential complementarity of the two models' classification capabilities. In addition, our co-training framework is also applied to the generalized ZSL (GZSL), in which a semantic-guided OOD detector is proposed to pick out the most likely unseen-class samples before class-level classification to alleviate the bias problem in GZSL. Extensive experiments on three benchmarks show that our proposed methods could significantly outperform about $31$ state-of-the-art ones.