Differentiable Architecture Search Based on Coordinate Descent

Cited 3 time in webofscience Cited 0 time in scopus
  • Hit : 300
  • Download : 0
DC FieldValueLanguage
dc.contributor.authorAhn, Pyunghwanko
dc.contributor.authorHong, Hyeong Gwonko
dc.contributor.authorKim, Junmoko
dc.date.accessioned2021-04-20T06:50:07Z-
dc.date.available2021-04-20T06:50:07Z-
dc.date.created2021-04-19-
dc.date.created2021-04-19-
dc.date.issued2021-04-
dc.identifier.citationIEEE ACCESS, v.9, pp.48544 - 48554-
dc.identifier.issn2169-3536-
dc.identifier.urihttp://hdl.handle.net/10203/282492-
dc.description.abstractNeural architecture search (NAS) is an automated method searching for the optimal network architecture by optimizing the combinations of edges and operations. For efficiency, recent differentiable architecture search methods adopt a one-shot network, containing all the candidate operations in each edge, instead of sampling and training individual architectures. However, a recent study doubts the effectiveness of differentiable methods by showing that random search can achieve comparable performance with differentiable methods using the same search cost. Therefore, there is a need to reduce the search cost even for previous differentiable methods. For more efficient differentiable architecture search, we propose a differentiable architecture search based on coordinate descent (DARTS-CD) that searches for optimal operation over only one sampled edge per training step. DARTS-CD is proposed based on the coordinate descent algorithm, which is an efficient learning method for resolving large-scale problems by updating only a subset of parameters. In DARTS-CD, one edge is randomly sampled, in which all the operations are performed, whereas only one operation is applied to the other edges. Weight update is also performed only at the sampled edge. By optimizing each edge separately, as in the coordinate descent that optimizes each coordinate individually, DARTS-CD converges much faster than DARTS while using the network architecture similar to that used for evaluation. We experimentally show that DARTS-CD performs comparably to the state-of-the-art efficient architecture search algorithms, with an extremely low search cost of 0.125 GPU days (1/12 of the search cost of DARTS) on CIFAR-10 and CIFAR-100. Furthermore, a warm-up regularization method is introduced to improve the exploration capability, which further enhances the performance.-
dc.languageEnglish-
dc.publisherIEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC-
dc.titleDifferentiable Architecture Search Based on Coordinate Descent-
dc.typeArticle-
dc.identifier.wosid000637182300001-
dc.identifier.scopusid2-s2.0-85103292014-
dc.type.rimsART-
dc.citation.volume9-
dc.citation.beginningpage48544-
dc.citation.endingpage48554-
dc.citation.publicationnameIEEE ACCESS-
dc.identifier.doi10.1109/ACCESS.2021.3068766-
dc.contributor.localauthorKim, Junmo-
dc.description.isOpenAccessY-
dc.type.journalArticleArticle-
dc.subject.keywordAuthorComputer architecture-
dc.subject.keywordAuthorMicroprocessors-
dc.subject.keywordAuthorTraining-
dc.subject.keywordAuthorSearch problems-
dc.subject.keywordAuthorArchitecture-
dc.subject.keywordAuthorTask analysis-
dc.subject.keywordAuthorNetwork architecture-
dc.subject.keywordAuthorAutomatic machine learning (AutoML)-
dc.subject.keywordAuthordifferentiable architecture search (DARTS)-
dc.subject.keywordAuthorneural architecture search (NAS)-
Appears in Collection
EE-Journal Papers(저널논문)
Files in This Item
There are no files associated with this item.
This item is cited by other documents in WoS
⊙ Detail Information in WoSⓡ Click to see webofscience_button
⊙ Cited 3 items in WoS Click to see citing articles in records_button

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0