Practical Block-Wise Neural Network Architecture Generation

Convolutional neural networks have gained a remarkable success in computer vision. However, most usable network architectures are hand-crafted and usually require expertise and elaborate design. In this paper, we provide a block-wise network generation pipeline called BlockQNN which automatically bu...

Full description

Saved in:
Bibliographic Details
Published in2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition pp. 2423 - 2432
Main Authors Zhong, Zhao, Yan, Junjie, Wu, Wei, Shao, Jing, Liu, Cheng-Lin
Format Conference Proceeding
LanguageEnglish
Published IEEE 01.06.2018
Subjects
Online AccessGet full text

Cover

Loading…
Abstract Convolutional neural networks have gained a remarkable success in computer vision. However, most usable network architectures are hand-crafted and usually require expertise and elaborate design. In this paper, we provide a block-wise network generation pipeline called BlockQNN which automatically builds high-performance networks using the Q-Learning paradigm with epsilon-greedy exploration strategy. The optimal network block is constructed by the learning agent which is trained sequentially to choose component layers. We stack the block to construct the whole auto-generated network. To accelerate the generation process, we also propose a distributed asynchronous framework and an early stop strategy. The block-wise generation brings unique advantages: (1) it performs competitive results in comparison to the hand-crafted state-of-the-art networks on image classification, additionally, the best network generated by BlockQNN achieves 3.54% top-1 error rate on CIFAR-10 which beats all existing auto-generate networks. (2) in the meanwhile, it offers tremendous reduction of the search space in designing networks which only spends 3 days with 32 GPUs, and (3) moreover, it has strong generalizability that the network built on CIFAR also performs well on a larger-scale ImageNet dataset.
AbstractList Convolutional neural networks have gained a remarkable success in computer vision. However, most usable network architectures are hand-crafted and usually require expertise and elaborate design. In this paper, we provide a block-wise network generation pipeline called BlockQNN which automatically builds high-performance networks using the Q-Learning paradigm with epsilon-greedy exploration strategy. The optimal network block is constructed by the learning agent which is trained sequentially to choose component layers. We stack the block to construct the whole auto-generated network. To accelerate the generation process, we also propose a distributed asynchronous framework and an early stop strategy. The block-wise generation brings unique advantages: (1) it performs competitive results in comparison to the hand-crafted state-of-the-art networks on image classification, additionally, the best network generated by BlockQNN achieves 3.54% top-1 error rate on CIFAR-10 which beats all existing auto-generate networks. (2) in the meanwhile, it offers tremendous reduction of the search space in designing networks which only spends 3 days with 32 GPUs, and (3) moreover, it has strong generalizability that the network built on CIFAR also performs well on a larger-scale ImageNet dataset.
Author Liu, Cheng-Lin
Wu, Wei
Shao, Jing
Zhong, Zhao
Yan, Junjie
Author_xml – sequence: 1
  givenname: Zhao
  surname: Zhong
  fullname: Zhong, Zhao
– sequence: 2
  givenname: Junjie
  surname: Yan
  fullname: Yan, Junjie
– sequence: 3
  givenname: Wei
  surname: Wu
  fullname: Wu, Wei
– sequence: 4
  givenname: Jing
  surname: Shao
  fullname: Shao, Jing
– sequence: 5
  givenname: Cheng-Lin
  surname: Liu
  fullname: Liu, Cheng-Lin
BookMark eNotjEtLw0AURkdRsNasXbjJH0i888wMrmqoVSi1iI9lmdze4tiYyGSK-O8N6OLjwOHwnbOTru-IsUsOJefgruvX9VMpgNsSQOjqiGWuslxLa4wS4I7ZhIORhXHcnbFsGD5g7IyVVukJu1lHjymgb_Pbtsd98RYGyld0iKNZUfru4z6fRXwPiTAdIuUL6ij6FPrugp3ufDtQ9s8pe7mbP9f3xfJx8VDPlkUQiqdCklcKySjbCI1Kj3MVarnVzsMWFReEoLnyViJKhaJRiESysRZ2DSg5ZVd_v4GINl8xfPr4s7G6slJr-QtC7Enq
CODEN IEEPAD
ContentType Conference Proceeding
DBID 6IE
6IH
CBEJK
RIE
RIO
DOI 10.1109/CVPR.2018.00257
DatabaseName IEEE Electronic Library (IEL) Conference Proceedings
IEEE Proceedings Order Plan (POP) 1998-present by volume
IEEE Xplore All Conference Proceedings
IEEE Electronic Library (IEL)
IEEE Proceedings Order Plans (POP) 1998-present
DatabaseTitleList
Database_xml – sequence: 1
  dbid: RIE
  name: IEEE Electronic Library (IEL)
  url: https://proxy.k.utb.cz/login?url=https://ieeexplore.ieee.org/
  sourceTypes: Publisher
DeliveryMethod fulltext_linktorsrc
Discipline Applied Sciences
EISBN 9781538664209
1538664208
EISSN 1063-6919
EndPage 2432
ExternalDocumentID 8578355
Genre orig-research
GroupedDBID 6IE
6IH
6IL
6IN
AAWTH
ABLEC
ADZIZ
ALMA_UNASSIGNED_HOLDINGS
BEFXN
BFFAM
BGNUA
BKEBE
BPEOZ
CBEJK
CHZPO
IEGSK
IJVOP
OCL
RIE
RIL
RIO
ID FETCH-LOGICAL-i241t-3ea44ce648b25c455c497c53d59a0dc412ec0514a83cc34c2b4ccee3b880fb043
IEDL.DBID RIE
IngestDate Wed Aug 27 02:52:16 EDT 2025
IsPeerReviewed false
IsScholarly true
Language English
LinkModel DirectLink
MergedId FETCHMERGED-LOGICAL-i241t-3ea44ce648b25c455c497c53d59a0dc412ec0514a83cc34c2b4ccee3b880fb043
PageCount 10
ParticipantIDs ieee_primary_8578355
PublicationCentury 2000
PublicationDate 2018-06
PublicationDateYYYYMMDD 2018-06-01
PublicationDate_xml – month: 06
  year: 2018
  text: 2018-06
PublicationDecade 2010
PublicationTitle 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition
PublicationTitleAbbrev CVPR
PublicationYear 2018
Publisher IEEE
Publisher_xml – name: IEEE
SSID ssj0002683845
ssj0003211698
Score 2.5784826
Snippet Convolutional neural networks have gained a remarkable success in computer vision. However, most usable network architectures are hand-crafted and usually...
SourceID ieee
SourceType Publisher
StartPage 2423
SubjectTerms Computer architecture
Convolutional codes
Convolutional neural networks
Indexes
Network architecture
Task analysis
Title Practical Block-Wise Neural Network Architecture Generation
URI https://ieeexplore.ieee.org/document/8578355
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
link http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV1LS8NAEB7anjxVbcU3e_Dotsk-kg2etFiK0FLEam9ldzOBUmjFJhd_vbtJrA88eAgkyxyGHZYvM_PNtwBXMQvSJENNhTOggmWS6oghzRLMfOMLE-2nkceTaDQTD3M5b8D1bhYGEUvyGfb8a9nLTze28KWyvpK-TiGb0HSJWzWrtaunsEhxVXfI_Dd3mU2UqFrNJwyS_uB5-ui5XCV5Uv68TqVEk2Ebxp9-VCSSVa_ITc--_5Jo_K-j-9D9mtsj0x0iHUAD14fQrn80SX2Mtx24qVSKXHjInQOzFX1ZbpF4nQ63MqmI4eT2W4uBVOrUPohdmA3vnwYjWt-iQJcOnXPKUQthMRLKMGmFdE8SW8lTmeggtSJkaL0IulbcWi4sM8I6P7lxJzszgeBH0Fpv1ngMJOBxLNMw1D6LZDrUsTOJI2etjBfWO4GO34vFayWUsai34fTv5TPY89GoeFfn0MrfCrxwCJ-byzK0H9sfpNw
linkProvider IEEE
linkToHtml http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV1LT8JAEJ4gHvSECsa3e_DoQruPthtPSiSoQIgB9Ua222lCSMBIufjr3W0rPuLBW7uZw2Qmm2935ptvAS5C5iUqRU2FNaCCpZLqgCFNFaau8YVKu2nk_iDojsX9i3ypwOV6FgYRc_IZNt1n3stPFmblSmWtSLo6hdyATYv70i-mtdYVFRZEPCp7ZO6f27tNoKJSz8f3VKv9NHx0bK6cPil_PqiS40mnBv1PTwoayay5yuKmef8l0vhfV3eg8TW5R4ZrTNqFCs73oFYeNUm5kZd1uCp0imyCyI2Fsxl9ni6ROKUOuzIoqOHk-luTgRT61C6NDRh3bkftLi3fUaBTi88Z5aiFMBiIKGbS2MgZoUIjeSKV9hIjfIbGyaDriBvDhWGxMNZPHtu9ncae4PtQnS_meADE42EoE9_X7h7JtK9DaxIG1jqKnbTeIdRdLCavhVTGpAzD0d_L57DVHfV7k97d4OEYtl1mChbWCVSztxWeWrzP4rM8zR-o86gl
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&rft.genre=proceeding&rft.title=2018+IEEE%2FCVF+Conference+on+Computer+Vision+and+Pattern+Recognition&rft.atitle=Practical+Block-Wise+Neural+Network+Architecture+Generation&rft.au=Zhong%2C+Zhao&rft.au=Yan%2C+Junjie&rft.au=Wu%2C+Wei&rft.au=Shao%2C+Jing&rft.date=2018-06-01&rft.pub=IEEE&rft.eissn=1063-6919&rft.spage=2423&rft.epage=2432&rft_id=info:doi/10.1109%2FCVPR.2018.00257&rft.externalDocID=8578355