ASAP: Architecture Search, Anneal and Prune

Asaf Noy, Sivan Doveh, Niv Nayman, Itamar Friedman, Tal Ridnik, Raja Giryes, Nadav Zamir, Lihi Zelnik-Manor

Research output: Contribution to journalConference articlepeer-review

Abstract

Automatic methods for Neural Architecture Search (NAS) have been shown to produce state-of-the-art network models. Yet, their main drawback is the computational complexity of the search process. As some primal methods optimized over a discrete search space, thousands of days of GPU were required for convergence. A recent approach is based on constructing a differentiable search space that enables gradient-based optimization, which reduces the search time to a few days. While successful, it still includes some noncontinuous steps, e.g., the pruning of many weak connections at once. In this paper, we propose a differentiable search space that allows the annealing of architecture weights, while gradually pruning inferior operations. In this way, the search converges to a single output network in a continuous manner. Experiments on several vision datasets demonstrate the effectiveness of our method with respect to the search cost and accuracy of the achieved model. Specifically, with 0.2 GPU search days we achieve an error rate of 1.68% on CIFAR-10.

Original languageEnglish
Pages (from-to)493-503
Number of pages11
JournalProceedings of Machine Learning Research
Volume108
StatePublished - 2020
Externally publishedYes
Event23rd International Conference on Artificial Intelligence and Statistics, AISTATS 2020 - Virtual, Online
Duration: 26 Aug 202028 Aug 2020

ASJC Scopus subject areas

  • Software
  • Artificial Intelligence
  • Control and Systems Engineering
  • Statistics and Probability

Fingerprint

Dive into the research topics of 'ASAP: Architecture Search, Anneal and Prune'. Together they form a unique fingerprint.

Cite this