TY - GEN
T1 - Efficient Evolutionary Deep Neural Architecture Search (NAS) by Noisy Network Morphism Mutation
AU - Chen, Yiming
AU - Pan, Tianci
AU - He, Cheng
AU - Cheng, Ran
N1 - Publisher Copyright:
© 2020, Springer Nature Singapore Pte Ltd.
PY - 2020
Y1 - 2020
N2 - Deep learning has achieved enormous breakthroughs in the field of image recognition. However, due to the time-consuming and error-prone process in discovering novel neural architecture, it remains a challenge for designing a specific network in handling a particular task. Hence, many automated neural architecture search methods are proposed to find suitable deep neural network architecture for a specific task without human experts. Nevertheless, these methods are still computationally/economically expensive, since they require a vast amount of computing resource and/or computational time. In this paper, we propose several network morphism mutation operators with extra noise, and further redesign the macro-architecture based on the classical network. The proposed methods are embedded in an evolutionary algorithm and tested on CIFAR-10 classification task. Experimental results indicate the capability of our proposed method in discovering powerful neural architecture which has achieved a classification error 2.55% with only 4.7M parameters on CIFAR-10 within 12 GPU-hours.
AB - Deep learning has achieved enormous breakthroughs in the field of image recognition. However, due to the time-consuming and error-prone process in discovering novel neural architecture, it remains a challenge for designing a specific network in handling a particular task. Hence, many automated neural architecture search methods are proposed to find suitable deep neural network architecture for a specific task without human experts. Nevertheless, these methods are still computationally/economically expensive, since they require a vast amount of computing resource and/or computational time. In this paper, we propose several network morphism mutation operators with extra noise, and further redesign the macro-architecture based on the classical network. The proposed methods are embedded in an evolutionary algorithm and tested on CIFAR-10 classification task. Experimental results indicate the capability of our proposed method in discovering powerful neural architecture which has achieved a classification error 2.55% with only 4.7M parameters on CIFAR-10 within 12 GPU-hours.
KW - Evolutionary algorithm
KW - Network morphism
KW - Neural architecture search
UR - http://www.scopus.com/inward/record.url?scp=85083954872&partnerID=8YFLogxK
U2 - 10.1007/978-981-15-3415-7_41
DO - 10.1007/978-981-15-3415-7_41
M3 - Conference article published in proceeding or book
AN - SCOPUS:85083954872
SN - 9789811534140
T3 - Communications in Computer and Information Science
SP - 497
EP - 508
BT - Bio-inspired Computing
A2 - Pan, Linqiang
A2 - Liang, Jing
A2 - Qu, Boyang
PB - Springer
T2 - 14th International Conference on Bio-inspired Computing: Theories and Applications, BIC-TA 2019
Y2 - 22 November 2019 through 25 November 2019
ER -