论文标题

统一和增强基于梯度的无培训神经体系结构搜索

Unifying and Boosting Gradient-Based Training-Free Neural Architecture Search

论文作者

Shu, Yao, Dai, Zhongxiang, Wu, Zhaoxuan, Low, Bryan Kian Hsiang

论文摘要

神经架构搜索(NAS)由于其自动化神经建筑设计的能力而获得了极大的知名度。最近,提议许多无培训指标实现NAS,而无需培训,从而使NAS更具可扩展性。尽管具有竞争性的经验表现,但仍缺乏对这些无培训指标的统一理论理解。结果,(a)这些指标之间的关系尚不清楚,(b)对其经验表现没有理论解释,并且(c)在现有的无培训NAS中可能存在未开发的潜力,这可能可以通过统一的理论理解来揭幕。 To this end, this paper presents a unified theoretical analysis of gradient-based training-free NAS, which allows us to (a) theoretically study their relationships, (b) theoretically guarantee their generalization performances, and (c) exploit our unified theoretical understanding to develop a novel framework named hybrid NAS (HNAS) which consistently boosts training-free NAS in a principled way.值得注意的是,HNA可以享受无训练(即优越的搜索效率)和基于培训的(即,出色的搜索效率)NAS的优势,我们通过广泛的实验证明了这一点。

Neural architecture search (NAS) has gained immense popularity owing to its ability to automate neural architecture design. A number of training-free metrics are recently proposed to realize NAS without training, hence making NAS more scalable. Despite their competitive empirical performances, a unified theoretical understanding of these training-free metrics is lacking. As a consequence, (a) the relationships among these metrics are unclear, (b) there is no theoretical interpretation for their empirical performances, and (c) there may exist untapped potential in existing training-free NAS, which probably can be unveiled through a unified theoretical understanding. To this end, this paper presents a unified theoretical analysis of gradient-based training-free NAS, which allows us to (a) theoretically study their relationships, (b) theoretically guarantee their generalization performances, and (c) exploit our unified theoretical understanding to develop a novel framework named hybrid NAS (HNAS) which consistently boosts training-free NAS in a principled way. Remarkably, HNAS can enjoy the advantages of both training-free (i.e., the superior search efficiency) and training-based (i.e., the remarkable search effectiveness) NAS, which we have demonstrated through extensive experiments.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源