Skip to main navigation Skip to search Skip to main content

Single-Path Mobile AutoML: Efficient ConvNet Design and NAS Hyperparameter Optimization

  • Dimitrios Stamoulis*
  • , Ruizhou Ding
  • , Di Wang
  • , Dimitrios Lymberopoulos
  • , Bodhi Priyantha
  • , Jie Liu
  • , Diana Marculescu
  • *Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

Abstract

Can we reduce the search cost of Neural Architecture Search (NAS) from days down to only a few hours? NAS methods automate the design of Convolutional Networks (ConvNets) under hardware constraints and they have emerged as key components of AutoML frameworks. However, the NAS problem remains challenging due to the combinatorially large design space and the significant search time (at least 200 GPU-hours). In this article, we alleviate the NAS search cost down to less than 3 hours, while achieving state-of-the-art image classification results under mobile latency constraints. We propose a novel differentiable NAS formulation, namely Single-Path NAS, that uses one single-path over-parameterized ConvNet to encode all architectural decisions based on shared convolutional kernel parameters, hence drastically decreasing the search overhead. Single-Path NAS achieves state-of-the-art top-1 ImageNet accuracy (75.62%), hence outperforming existing mobile NAS methods in similar latency settings ($\sim$80 ms). In particular, we enhance the accuracy-runtime trade-off in differentiable NAS by treating the Squeeze-and-Excitation path as a fully searchable operation with our novel single-path encoding. Our method has an overall cost of only 8 epochs (24 TPU-hours), which is up to 5,000× faster compared to prior work. Moreover, we study how different NAS formulation choices affect the performance of the designed ConvNets. Furthermore, we exploit the efficiency of our method to answer an interesting question: instead of empirically tuning the hyperparameters of the NAS solver (as in prior work), can we automatically find the hyperparameter values that yield the desired accuracy-runtime trade-off (e.g., target runtime for different platforms)? We view our extensive experimental results as a valuable exploration for NAS-based cloud AutoML services, and we open-source our entire codebase at: https://github.com/dstamoulis/single-path-nas.

Original languageEnglish
Article number8979335
Pages (from-to)609-622
Number of pages14
JournalIEEE Journal on Selected Topics in Signal Processing
Volume14
Issue number4
DOIs
StatePublished - May 2020

Keywords

  • AutoML
  • ConvNets
  • Neural architecture search (NAS)
  • hardware-aware convnets

Fingerprint

Dive into the research topics of 'Single-Path Mobile AutoML: Efficient ConvNet Design and NAS Hyperparameter Optimization'. Together they form a unique fingerprint.

Cite this