TY - JOUR
T1 - Convolutional Neural Networks with Dynamic Regularization
AU - Wang, Yi
AU - Bian, Zhen Peng
AU - Hou, Junhui
AU - Chau, Lap Pui
N1 - Funding Information:
Manuscript received September 25, 2019; revised March 25, 2020 and April 13, 2020; accepted May 19, 2020. Date of publication June 8, 2020; date of current version May 3, 2021. This work was supported in part by the Hong Kong Research Grants Council under Grant 9048123 (CityU 21211518) and Grant 9042820 (CityU 11219019). (Yi Wang and Zhen-Peng Bian contributed equally to this work). (Corresponding author: Lap-Pui Chau.) Yi Wang and Lap-Pui Chau are with the School of Electrical and Electronics Engineering, Nanyang Technological University, Singapore 639798 (e-mail: [email protected], [email protected]).
Publisher Copyright:
© 2012 IEEE.
PY - 2021/5
Y1 - 2021/5
N2 - Regularization is commonly used for alleviating overfitting in machine learning. For convolutional neural networks (CNNs), regularization methods, such as DropBlock and Shake-Shake, have illustrated the improvement in the generalization performance. However, these methods lack a self-adaptive ability throughout training. That is, the regularization strength is fixed to a predefined schedule, and manual adjustments are required to adapt to various network architectures. In this article, we propose a dynamic regularization method for CNNs. Specifically, we model the regularization strength as a function of the training loss. According to the change of the training loss, our method can dynamically adjust the regularization strength in the training procedure, thereby balancing the underfitting and overfitting of CNNs. With dynamic regularization, a large-scale model is automatically regularized by the strong perturbation, and vice versa. Experimental results show that the proposed method can improve the generalization capability on off-the-shelf network architectures and outperform state-of-the-art regularization methods.
AB - Regularization is commonly used for alleviating overfitting in machine learning. For convolutional neural networks (CNNs), regularization methods, such as DropBlock and Shake-Shake, have illustrated the improvement in the generalization performance. However, these methods lack a self-adaptive ability throughout training. That is, the regularization strength is fixed to a predefined schedule, and manual adjustments are required to adapt to various network architectures. In this article, we propose a dynamic regularization method for CNNs. Specifically, we model the regularization strength as a function of the training loss. According to the change of the training loss, our method can dynamically adjust the regularization strength in the training procedure, thereby balancing the underfitting and overfitting of CNNs. With dynamic regularization, a large-scale model is automatically regularized by the strong perturbation, and vice versa. Experimental results show that the proposed method can improve the generalization capability on off-the-shelf network architectures and outperform state-of-the-art regularization methods.
KW - Convolutional neural network (CNN)
KW - generalization
KW - image classification
KW - overfitting
KW - regularization
UR - http://www.scopus.com/inward/record.url?scp=85105605068&partnerID=8YFLogxK
U2 - 10.1109/TNNLS.2020.2997044
DO - 10.1109/TNNLS.2020.2997044
M3 - Journal article
C2 - 32511095
AN - SCOPUS:85105605068
SN - 2162-237X
VL - 32
SP - 2299
EP - 2304
JO - IEEE Transactions on Neural Networks and Learning Systems
JF - IEEE Transactions on Neural Networks and Learning Systems
IS - 5
M1 - 9110754
ER -