论文标题

跨级实例组歧视的无监督功能学习

Unsupervised Feature Learning by Cross-Level Instance-Group Discrimination

论文作者

Wang, Xudong, Liu, Ziwei, Yu, Stella X.

论文摘要

无监督的功能学习在基于实例歧视和不变型映射的对比学习方面取得了长足的进步,这是在策划的类平衡数据集中标准的。但是,自然数据可以高度相关且长尾分布。自然之间的相似性与假定实例的区别冲突,导致训练不稳定和性能不佳。 我们的想法是发现并集成在实例学习之间,而不是直接通过实例分组,而是通过实例和本地实例组之间的跨层次歧视(CLD)。虽然每个实例的不变映射是通过吸引力在其增强视图中施加的,但之间的相似性之间的相似性可能会来自对实例组的共同排斥。 我们的批处理和跨视图比较也大大提高了对比度学习的正/负样本比率,并实现了更好的不变映射。为了实现分组和歧视目标,我们将它们强加于从共享表示形式分别得出的特征。此外,我们首次提出了归一化的投影头和无监督的高参数调整。 我们广泛的实验表明,CLD是现有方法的精益而有力的附加组件,例如NPID,MOCO,INFOMIN和BYOL在高度相关,长尾或平衡的数据集上。它不仅在自我统治,半义务和转移学习基准方面实现了新的最新,而且还击败了Moco V2和Simc​​lr,这在每个报告的表现上都具有更大的计算。 CLD有效地使无监督的学习更接近自然数据和现实世界应用。我们的代码可在以下网址公开获取:https://github.com/frank-xwang/cld-unsupervisedlearning。

Unsupervised feature learning has made great strides with contrastive learning based on instance discrimination and invariant mapping, as benchmarked on curated class-balanced datasets. However, natural data could be highly correlated and long-tail distributed. Natural between-instance similarity conflicts with the presumed instance distinction, causing unstable training and poor performance. Our idea is to discover and integrate between-instance similarity into contrastive learning, not directly by instance grouping, but by cross-level discrimination (CLD) between instances and local instance groups. While invariant mapping of each instance is imposed by attraction within its augmented views, between-instance similarity could emerge from common repulsion against instance groups. Our batch-wise and cross-view comparisons also greatly improve the positive/negative sample ratio of contrastive learning and achieve better invariant mapping. To effect both grouping and discrimination objectives, we impose them on features separately derived from a shared representation. In addition, we propose normalized projection heads and unsupervised hyper-parameter tuning for the first time. Our extensive experimentation demonstrates that CLD is a lean and powerful add-on to existing methods such as NPID, MoCo, InfoMin, and BYOL on highly correlated, long-tail, or balanced datasets. It not only achieves new state-of-the-art on self-supervision, semi-supervision, and transfer learning benchmarks, but also beats MoCo v2 and SimCLR on every reported performance attained with a much larger compute. CLD effectively brings unsupervised learning closer to natural data and real-world applications. Our code is publicly available at: https://github.com/frank-xwang/CLD-UnsupervisedLearning.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源