论文标题
体重共享神经体系结构搜索:缩小优化差距的战斗
Weight-Sharing Neural Architecture Search: A Battle to Shrink the Optimization Gap
论文作者
论文摘要
神经建筑搜索(NAS)吸引了学术界和行业的注意力越来越高。在很小的时候,研究人员主要采用单个搜索方法,这些方法分别采样和评估候选体系结构,从而产生大量的计算开销。为了减轻负担,提出了体重分担的方法,在同一超级网络中,许多体系结构呈指数级共享权重,并且仅执行一次昂贵的培训程序。这些方法虽然快得多,但通常会遇到不稳定的问题。本文提供了有关NAS的文献综述,尤其是重量分享方法,并指出主要挑战来自超级网络和子构造之间的优化差距。从这个角度来看,我们根据它们在弥合差距方面的努力,将现有方法汇总为多种类别,并分析这些方法的优势和缺点。最后,我们在NAS和AUTOML的未来方向上分享我们的意见。由于作者的专业知识,本文主要关注NAS在计算机视觉问题上的应用,并可能偏向我们小组中的工作。
Neural architecture search (NAS) has attracted increasing attentions in both academia and industry. In the early age, researchers mostly applied individual search methods which sample and evaluate the candidate architectures separately and thus incur heavy computational overheads. To alleviate the burden, weight-sharing methods were proposed in which exponentially many architectures share weights in the same super-network, and the costly training procedure is performed only once. These methods, though being much faster, often suffer the issue of instability. This paper provides a literature review on NAS, in particular the weight-sharing methods, and points out that the major challenge comes from the optimization gap between the super-network and the sub-architectures. From this perspective, we summarize existing approaches into several categories according to their efforts in bridging the gap, and analyze both advantages and disadvantages of these methodologies. Finally, we share our opinions on the future directions of NAS and AutoML. Due to the expertise of the authors, this paper mainly focuses on the application of NAS to computer vision problems and may bias towards the work in our group.