Efficient evolutionary neural architecture search by modular inheritable crossover

Research output: Journal article publicationJournal articleAcademic researchpeer-review

25 Citations (Scopus)

Abstract

Deep neural networks are widely used in the domain of image classification, and a large number of excellent deep neural networks have been proposed in recent years. However, hand-crafted neural networks often require human experts for elaborate designs, which can be time-consuming and error-prone. Hence, neural architecture search (NAS) methods have been proposed to design model architecture automatically. The evolutionary NAS methods have achieved encouraging results due to the global search capability of evolutionary algorithms. Nevertheless, most existing evolutionary NAS methods use only the mutation operator to generate offspring architectures. Consequently, the generated architectures could be pretty different from their parent architectures, failing to inherit the modular information to accelerate the convergence rate. We propose an efficient evolutionary method using a tailored crossover operator to address this deficiency, which enables the offspring architectures to inherit from their parent architectures. Moreover, we combine it with mutation operators under the framework of the evolutionary algorithm. Experimental results on both the CIFAR-10 and CIFAR-100 tasks show that our proposed evolutionary NAS method has achieved state-of-the-art results.

Original languageEnglish
Article number100894
JournalSwarm and Evolutionary Computation
Volume64
DOIs
Publication statusPublished - Jul 2021
Externally publishedYes

Keywords

  • Evolutionary algorithm
  • Image classification
  • Inheritable crossover
  • Neural architecture search

ASJC Scopus subject areas

  • General Computer Science
  • General Mathematics

Fingerprint

Dive into the research topics of 'Efficient evolutionary neural architecture search by modular inheritable crossover'. Together they form a unique fingerprint.

Cite this