论文标题
多阶级的几阶段学习多晶级层次结构
Many-Class Few-Shot Learning on Multi-Granularity Class Hierarchy
论文作者
论文摘要
我们在监督学习和元学习环境中研究了许多级别的几杆(MCF)问题。与研究良好的多级次数和几乎没有类别的问题相比,MCFS问题通常发生在实际应用中,但在以前的文献中很少研究。它带来了新的挑战,可以区分每个课程的许多班级。在本文中,我们利用类层次结构作为先验知识来训练可以在两种情况下对MCFS问题产生准确预测的粗到最佳分类器。提出的模型“记忆增强的分层分类网络(Mahinet)”进行了粗到1的分类,每个粗糙类都可以覆盖多个精美的类。由于每个类别的数据很少,直接区分各种精美的类是一个挑战,因此Mahinet从学习分类器上的粗级别培训数据开始,并提供了更多的培训数据,其标签的便宜得多。粗糙的分类器减少了精细类别的搜索范围,从而减轻了“许多类别”的挑战。在体系结构上,Mahinet首先部署卷积神经网络(CNN)来提取功能。然后,它集成了一个记忆启动的注意模块和多层感知器(MLP),以在粗糙和精细的类别上产生概率。当MLP扩展线性分类器时,注意模块扩展了KNN分类器,两者都将针对“少量射击”问题的问题共同。我们设计了Mahinet的几种培训策略,用于监督学习和元学习。此外,我们提出了两个新颖的基准数据集“ MCFSIMAGENET”和“ MCFSOMNIGLOT”,专为MCFS问题而设计。在实验中,我们表明,Mahinet在监督学习和元学习中的MCFS问题都优于几个最新模型。
We study many-class few-shot (MCFS) problem in both supervised learning and meta-learning settings. Compared to the well-studied many-class many-shot and few-class few-shot problems, the MCFS problem commonly occurs in practical applications but has been rarely studied in previous literature. It brings new challenges of distinguishing between many classes given only a few training samples per class. In this paper, we leverage the class hierarchy as a prior knowledge to train a coarse-to-fine classifier that can produce accurate predictions for MCFS problem in both settings. The propose model, "memory-augmented hierarchical-classification network (MahiNet)", performs coarse-to-fine classification where each coarse class can cover multiple fine classes. Since it is challenging to directly distinguish a variety of fine classes given few-shot data per class, MahiNet starts from learning a classifier over coarse-classes with more training data whose labels are much cheaper to obtain. The coarse classifier reduces the searching range over the fine classes and thus alleviates the challenges from "many classes". On architecture, MahiNet firstly deploys a convolutional neural network (CNN) to extract features. It then integrates a memory-augmented attention module and a multi-layer perceptron (MLP) together to produce the probabilities over coarse and fine classes. While the MLP extends the linear classifier, the attention module extends the KNN classifier, both together targeting the "few-shot" problem. We design several training strategies of MahiNet for supervised learning and meta-learning. In addition, we propose two novel benchmark datasets "mcfsImageNet" and "mcfsOmniglot" specially designed for MCFS problem. In experiments, we show that MahiNet outperforms several state-of-the-art models on MCFS problems in both supervised learning and meta-learning.