论文标题
FBNETV3:使用预测器预处理的联合体系结构搜索
FBNetV3: Joint Architecture-Recipe Search using Predictor Pretraining
论文作者
论文摘要
神经架构搜索(NAS)产生最先进的神经网络,其表现优于其最佳手动设计的对应物。但是,以前的NAS方法在一组训练超参数(即培训食谱)下搜索体系结构,俯瞰着卓越的体系结构组合。为了解决这个问题,我们提出神经体系结构搜索(NARS),以同时搜索(a)架构和(b)其相应的培训食谱。 NARS利用了共同分数和培训配方的精度预测指标,从而指导样本选择和排名。此外,为了弥补扩大的搜索空间,我们利用“自由”体系结构统计(例如,失败计数)预处理预测因子,从而显着提高了其样品效率和预测可靠性。通过受约束的迭代优化训练预测变量后,我们仅在CPU分钟内运行快速进化搜索,以生成架构配对对,以实现称为FBNETV3的各种资源约束。 FBNETV3构成了一个最先进的紧凑型神经网络家族,它们的表现都超过了自动和手动设计的竞争对手。例如,FBNETV3分别匹配ImageNet上的有效网和重新精确度,分别少于2.0倍和7.1倍的拖鞋。此外,FBNETV3可为下游对象检测任务带来显着的性能增长,尽管拖失板少了18%,而参数少了34%,但比基于高效网络的等效物少了34%。
Neural Architecture Search (NAS) yields state-of-the-art neural networks that outperform their best manually-designed counterparts. However, previous NAS methods search for architectures under one set of training hyper-parameters (i.e., a training recipe), overlooking superior architecture-recipe combinations. To address this, we present Neural Architecture-Recipe Search (NARS) to search both (a) architectures and (b) their corresponding training recipes, simultaneously. NARS utilizes an accuracy predictor that scores architecture and training recipes jointly, guiding both sample selection and ranking. Furthermore, to compensate for the enlarged search space, we leverage "free" architecture statistics (e.g., FLOP count) to pretrain the predictor, significantly improving its sample efficiency and prediction reliability. After training the predictor via constrained iterative optimization, we run fast evolutionary searches in just CPU minutes to generate architecture-recipe pairs for a variety of resource constraints, called FBNetV3. FBNetV3 makes up a family of state-of-the-art compact neural networks that outperform both automatically and manually-designed competitors. For example, FBNetV3 matches both EfficientNet and ResNeSt accuracy on ImageNet with up to 2.0x and 7.1x fewer FLOPs, respectively. Furthermore, FBNetV3 yields significant performance gains for downstream object detection tasks, improving mAP despite 18% fewer FLOPs and 34% fewer parameters than EfficientNet-based equivalents.