TY - JOUR
T1 - Deep stacked least square support matrix machine with adaptive multi-layer transfer for EEG classification
AU - Hang, Wenlong
AU - Li, Zengguang
AU - Yin, Mingbo
AU - Liang, Shuang
AU - Shen, Hang
AU - Wang, Qiong
AU - Qin, Jin
AU - Choi, Kup Sze
N1 - Funding Information:
This work was supported in part by the National Natural Science Foundation of China ( 61802177 , 61902197 ), by the Postgraduate Research & Practice Innovation Program of Jiangsu Province (SJCX22_0417), by the Natural Science Foundation of Jiangsu Province ( BK20201357 ), by the Six Talent Peaks Project in Jiangsu Province ( RJFW-020 ), by the Inter-Faculty Collaboration Scheme for FH, FHSS and FENG in The Hong Kong Polytechnic University (P0043025), and by the Hong Kong Research Grants Council (PolyU 152006/19E ).
Publisher Copyright:
© 2023
PY - 2023/4
Y1 - 2023/4
N2 - The recent extraordinary success of deep neural networks for electroencephalogram (EEG) decoding can be mainly attributed to the availability of large-scale labeled datasets. However, it requires prolonged EEG calibration, which is especially inconvenient for people with disabilities. Training deep neural networks with insufficient labeled EEG data will make the model suffer from limited generalization capability. In view of this, a new deep stacked network (DSN), called deep stacked transfer least square support matrix machine (DST-LSSMM), is proposed to alleviate the above issue. Following the philosophy of stacked generalization, the proposed DST-LSSMM utilizes least square support matrix machine (LSSMM) as its basic building unit and the random projection of previous layers as its stacking element. The adopted matrix classifier LSSMM directly models the matrix-form data, which is able to exploit the structural information between the columns or rows in EEG feature matrices to further improve the generalization capability of the model. Besides, we design an adaptive multi-layer model knowledge transfer learning scheme to guarantee consistency across the interrelated layers. The proposed adaptive multi-layer transfer scheme allows the use of model knowledge of previous related layers to facilitate the model construction at higher layers, consequently enhancing the generalization capability of the deep stacked model. DST-LSSMM is carried out in an efficient feed-forward way without parameter pre-training and fine-tuning, which can simplify the model optimization process. We extensively evaluate our method on three EEG datasets. Experimental results demonstrate that our method achieves promising performances on EEG classification.
AB - The recent extraordinary success of deep neural networks for electroencephalogram (EEG) decoding can be mainly attributed to the availability of large-scale labeled datasets. However, it requires prolonged EEG calibration, which is especially inconvenient for people with disabilities. Training deep neural networks with insufficient labeled EEG data will make the model suffer from limited generalization capability. In view of this, a new deep stacked network (DSN), called deep stacked transfer least square support matrix machine (DST-LSSMM), is proposed to alleviate the above issue. Following the philosophy of stacked generalization, the proposed DST-LSSMM utilizes least square support matrix machine (LSSMM) as its basic building unit and the random projection of previous layers as its stacking element. The adopted matrix classifier LSSMM directly models the matrix-form data, which is able to exploit the structural information between the columns or rows in EEG feature matrices to further improve the generalization capability of the model. Besides, we design an adaptive multi-layer model knowledge transfer learning scheme to guarantee consistency across the interrelated layers. The proposed adaptive multi-layer transfer scheme allows the use of model knowledge of previous related layers to facilitate the model construction at higher layers, consequently enhancing the generalization capability of the deep stacked model. DST-LSSMM is carried out in an efficient feed-forward way without parameter pre-training and fine-tuning, which can simplify the model optimization process. We extensively evaluate our method on three EEG datasets. Experimental results demonstrate that our method achieves promising performances on EEG classification.
KW - Deep stacked network
KW - Electroencephalography
KW - Generalization capability
KW - Matrix classifier
KW - Motor imagery
KW - Stacked generalization
UR - http://www.scopus.com/inward/record.url?scp=85146673895&partnerID=8YFLogxK
U2 - 10.1016/j.bspc.2023.104579
DO - 10.1016/j.bspc.2023.104579
M3 - Journal article
AN - SCOPUS:85146673895
SN - 1746-8094
VL - 82
JO - Biomedical Signal Processing and Control
JF - Biomedical Signal Processing and Control
M1 - 104579
ER -