TY - JOUR
T1 - Hyperparameter optimization of neural networks based on Q-learning
AU - Qi, Xin
AU - Xu, Bing
N1 - Publisher Copyright:
© 2022, The Author(s), under exclusive licence to Springer-Verlag London Ltd., part of Springer Nature.
PY - 2022
Y1 - 2022
N2 - Machine learning algorithms are sensitive to hyperparameters, and hyperparameter optimization techniques are often computationally expensive, especially for complex deep neural networks. In this paper, we use Q-learning algorithm to search for good hyperparameter configurations for neural networks, where the learning agent searches for the optimal hyperparameter configuration by continuously updating the Q-table to optimize hyperparameter tuning strategy. We modify the initial states and termination conditions of Q-learning to improve search efficiency. The experimental results on hyperparameter optimization of a convolutional neural network and a bidirectional long short-term memory network show that our method has higher search efficiency compared with tree of Parzen estimators, random search and genetic algorithm and can find out the optimal or near-optimal hyperparameter configuration of neural network models with minimum number of trials.
AB - Machine learning algorithms are sensitive to hyperparameters, and hyperparameter optimization techniques are often computationally expensive, especially for complex deep neural networks. In this paper, we use Q-learning algorithm to search for good hyperparameter configurations for neural networks, where the learning agent searches for the optimal hyperparameter configuration by continuously updating the Q-table to optimize hyperparameter tuning strategy. We modify the initial states and termination conditions of Q-learning to improve search efficiency. The experimental results on hyperparameter optimization of a convolutional neural network and a bidirectional long short-term memory network show that our method has higher search efficiency compared with tree of Parzen estimators, random search and genetic algorithm and can find out the optimal or near-optimal hyperparameter configuration of neural network models with minimum number of trials.
KW - Hyperparameter optimization
KW - Markov decision process
KW - Neural networks
KW - Q-learning
UR - http://www.scopus.com/inward/record.url?scp=85139617264&partnerID=8YFLogxK
U2 - 10.1007/s11760-022-02377-y
DO - 10.1007/s11760-022-02377-y
M3 - Journal article
AN - SCOPUS:85139617264
SN - 1863-1703
JO - Signal, Image and Video Processing
JF - Signal, Image and Video Processing
ER -