Authors
Jiahui Yu, Pengchong Jin, Hanxiao Liu, Gabriel Bender, Pieter-Jan Kindermans, Mingxing Tan, Thomas Huang, Xiaodan Song, Ruoming Pang, Quoc Le
Publication date
2020/3/24
Journal
European Conference on Computer Vision (ECCV), 2020
Description
Neural architecture search (NAS) has shown promising results discovering models that are both accurate and fast. For NAS, training a one-shot model has become a popular strategy to rank the relative quality of different architectures (child models) using a single set of shared weights. However, while one-shot model weights can effectively rank different network architectures, the absolute accuracies from these shared weights are typically far below those obtained from stand-alone training. To compensate, existing methods assume that the weights must be retrained, finetuned, or otherwise post-processed after the search is completed. These steps significantly increase the compute requirements and complexity of the architecture search and model deployment. In this work, we propose BigNAS, an approach that challenges the conventional wisdom that post-processing of the weights is necessary to get …
Total citations
20192020202120222023202441576749050
Scholar articles
J Yu, P Jin, H Liu, G Bender, PJ Kindermans, M Tan… - Computer Vision–ECCV 2020: 16th European …, 2020
J Yu, P Jin, H Liu, G Bender, PJ Kindermans, M Tan… - URL https://openreview. net/forum