Autonomous deep learning: A genetic DCNN designer for image classification

•Search deep convolutional neural networks automatically for image classification.•Encode a deep convolutional neural network’s architecture into an integer string.•Evolve a population of DCNN architectures using genetic evolutionary operations.•Obtained DCNNs achieved satisfying results with less l...

Full description

Saved in:
Bibliographic Details
Published inNeurocomputing (Amsterdam) Vol. 379; pp. 152 - 161
Main Authors Ma, Benteng, Li, Xiang, Xia, Yong, Zhang, Yanning
Format Journal Article
LanguageEnglish
Published Elsevier B.V 28.02.2020
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:•Search deep convolutional neural networks automatically for image classification.•Encode a deep convolutional neural network’s architecture into an integer string.•Evolve a population of DCNN architectures using genetic evolutionary operations.•Obtained DCNNs achieved satisfying results with less layers than pre-trained models.•Image classification using neither handcrafted features nor handcrafted networks. Recent years have witnessed the breakthrough success of deep convolutional neural networks (DCNNs) in image classification and other vision applications. DCNNs have distinct advantages over traditional solutions in providing a uniform feature extraction-classification framework to free users from troublesome handcrafted feature extraction. However, DCNNs are far from autonomous, since their performance relies heavily on the handcrafted architectures, which also requires a lot expertise and experience to design, and cannot be continuously improved once the tuning of hyper-parameters converges. In this paper, we propose an autonomous and continuous learning (ACL) algorithm to generate automatically a DCNN architecture for each given vision task. We first partition a DCNN into multiple stacked meta convolutional blocks and fully connected blocks, each of which may contain the operations of convolution, pooling, fully connection, batch normalization, activation and drop out, and thus convert the architecture into an integer code. Then, we use genetic evolutionary operations, including selection, mutation and crossover to evolve a population of DCNN architectures. We have evaluated this algorithm on six image classification tasks, i.e., MNIST, Fashion-MNIST, EMNIST-Letters, EMNIST-Digits, CIFAR10 and CIFAR100. Our results indicate that the proposed ACL algorithm is able to evolve the DCNN architecture continuously if more time cost is allowed and can find a suboptimal DCNN architecture, whose performance is comparable to the state of the art.
ISSN:0925-2312
1872-8286
DOI:10.1016/j.neucom.2019.10.007