TY - JOUR
T1 - Robust Online Learning Method Based on Dynamical Linear Quadratic Regulator
AU - Ning, Hanwen
AU - Zhang, Jiaming
AU - Jing, Xingjian
AU - Tian, Tianhai
N1 - Funding Information:
This work was supported in part by the National Natural Science Foundation of China under Project 11301544, Project 61773401, and Project 11571368, in part by the National Social Science Foundation of China under Project 19BTJ025, in part by the China Scholarship Council under Project 201707085011, in part by the Research Grants Council, University Grants Committee, Hong Kong, through the General Research Fund under Project 15206717, and in part by the Internal Research Grants through The Hong Kong Polytechnic University.
Publisher Copyright:
© 2013 IEEE.
PY - 2019
Y1 - 2019
N2 - In this paper, a novel algorithm is proposed for inferring online learning tasks efficiently. By a carefully designed scheme, the online learning problem is first formulated as a state feedback control problem for a series of finite-dimensional systems. Then, the online linear quadratic regulator (OLQR) learning algorithm is developed to obtain the optimal parameter updating. Solid mathematical analysis on the convergence and rationality of our method is also provided. Compared with the conventional learning methods, our learning framework represents a completely different approach with optimal control techniques, but does not introduce any assumption on the characteristics of noise or learning rate. The proposed method not only guarantees the fast and robust convergence but also achieves better performance in learning efficiency and accuracy, especially for the data streams with complex noise disturbances. In addition, under the proposed framework, new robust algorithms can be potentially developed for various machine learning tasks by using the powerful optimal control techniques. Numerical results on benchmark datasets and practical applications confirm the advantages of our new method.
AB - In this paper, a novel algorithm is proposed for inferring online learning tasks efficiently. By a carefully designed scheme, the online learning problem is first formulated as a state feedback control problem for a series of finite-dimensional systems. Then, the online linear quadratic regulator (OLQR) learning algorithm is developed to obtain the optimal parameter updating. Solid mathematical analysis on the convergence and rationality of our method is also provided. Compared with the conventional learning methods, our learning framework represents a completely different approach with optimal control techniques, but does not introduce any assumption on the characteristics of noise or learning rate. The proposed method not only guarantees the fast and robust convergence but also achieves better performance in learning efficiency and accuracy, especially for the data streams with complex noise disturbances. In addition, under the proposed framework, new robust algorithms can be potentially developed for various machine learning tasks by using the powerful optimal control techniques. Numerical results on benchmark datasets and practical applications confirm the advantages of our new method.
KW - complex noise disturbances
KW - linear quadratic regulator
KW - Online machine learning
KW - optimal control
UR - http://www.scopus.com/inward/record.url?scp=85088864240&partnerID=8YFLogxK
U2 - 10.1109/ACCESS.2019.2936537
DO - 10.1109/ACCESS.2019.2936537
M3 - Journal article
AN - SCOPUS:85088864240
SN - 2169-3536
VL - 7
SP - 117780
EP - 117795
JO - IEEE Access
JF - IEEE Access
M1 - 8807177
ER -