Semantic Scholar Open Access 2018 2978 sitasi

Efficient Neural Architecture Search via Parameter Sharing

Hieu Pham M. Guan Barret Zoph Quoc V. Le J. Dean

Abstrak

We propose Efficient Neural Architecture Search (ENAS), a fast and inexpensive approach for automatic model design. In ENAS, a controller learns to discover neural network architectures by searching for an optimal subgraph within a large computational graph. The controller is trained with policy gradient to select a subgraph that maximizes the expected reward on the validation set. Meanwhile the model corresponding to the selected subgraph is trained to minimize a canonical cross entropy loss. Thanks to parameter sharing between child models, ENAS is fast: it delivers strong empirical performances using much fewer GPU-hours than all existing automatic model design approaches, and notably, 1000x less expensive than standard Neural Architecture Search. On the Penn Treebank dataset, ENAS discovers a novel architecture that achieves a test perplexity of 55.8, establishing a new state-of-the-art among all methods without post-training processing. On the CIFAR-10 dataset, ENAS designs novel architectures that achieve a test error of 2.89%, which is on par with NASNet (Zoph et al., 2018), whose test error is 2.65%.

Penulis (5)

H

Hieu Pham

M

M. Guan

B

Barret Zoph

Q

Quoc V. Le

J

J. Dean

Format Sitasi

Pham, H., Guan, M., Zoph, B., Le, Q.V., Dean, J. (2018). Efficient Neural Architecture Search via Parameter Sharing. https://www.semanticscholar.org/paper/fe9b8aac9fa3bfd9724db5a881a578e471e612d7

Akses Cepat

PDF tidak tersedia langsung

Cek di sumber asli →
Lihat di Sumber
Informasi Jurnal
Tahun Terbit
2018
Bahasa
en
Total Sitasi
2978×
Sumber Database
Semantic Scholar
Akses
Open Access ✓