论文标题
用于加速神经体系结构跨模态的硬件感知框架
A Hardware-Aware Framework for Accelerating Neural Architecture Search Across Modalities
论文作者
论文摘要
神经体系结构搜索(NAS)的最新进展,例如单发NAS,可以从特定于任务的超级网络中提取专门的硬件子网络配置。尽管已经为改进第一阶段而采取了巨大努力,即对超级网络的培训,但寻找衍生性高性能子网络的搜索仍然尚未探索。流行方法将子网络搜索的超级网络培训与绩效预测变量分发,以减少在不同硬件平台上搜索的计算负担。我们提出了一个灵活的搜索框架,该框架自动有效地找到了针对不同性能指标和硬件配置进行优化的最佳子网。具体而言,我们展示了如何将进化算法与迭代循环中训练有素的客观预测变量配对,以在多目标设置中加速架构搜索,以供各种模态,包括机器翻译和图像分类。
Recent advances in Neural Architecture Search (NAS) such as one-shot NAS offer the ability to extract specialized hardware-aware sub-network configurations from a task-specific super-network. While considerable effort has been employed towards improving the first stage, namely, the training of the super-network, the search for derivative high-performing sub-networks is still under-explored. Popular methods decouple the super-network training from the sub-network search and use performance predictors to reduce the computational burden of searching on different hardware platforms. We propose a flexible search framework that automatically and efficiently finds optimal sub-networks that are optimized for different performance metrics and hardware configurations. Specifically, we show how evolutionary algorithms can be paired with lightly trained objective predictors in an iterative cycle to accelerate architecture search in a multi-objective setting for various modalities including machine translation and image classification.