TY - JOUR
T1 - A Residual Based Attention Model for EEG Based Sleep Staging
AU - Qu, Wei
AU - Wang, Zhiyong
AU - Hong, Hong
AU - Chi, Zheru
AU - Feng, David Dagan
AU - Grunstein, Ron
AU - Gordon, Christopher
N1 - Funding Information:
Manuscript received December 1, 2019; revised February 19, 2020; accepted February 26, 2020. Date of publication March 3, 2020; date of current version October 5, 2020. This work was supported in part under Grant ARC DP170104304. Wei Qu was supported with a scholarship by the CRC for Alertness, Safety and Productivity. (Corresponding author: Qu Wei.) Wei Qu, Zhiyong Wang, and David Dagan Feng are with the School of Computer Science, The University of Sydney, Sydney, NSW 2006, Australia (e-mail: [email protected]; [email protected]; [email protected]).
Publisher Copyright:
© 2013 IEEE.
Copyright:
Copyright 2020 Elsevier B.V., All rights reserved.
PY - 2020/10
Y1 - 2020/10
N2 - Sleep staging is to score the sleep state of a subject into different sleep stages such as Wake and Rapid Eye Movement (REM). It plays an indispensable role in the diagnosis and treatment of sleep disorders. As manual sleep staging through well-trained sleep experts is time consuming, tedious, and subjective, many automatic methods have been developed for accurate, efficient, and objective sleep staging. Recently, deep learning based methods have been successfully proposed for electroencephalogram (EEG) based sleep staging with promising results. However, most of these methods directly take EEG raw signals as input of convolutional neural networks (CNNs) without considering the domain knowledge of EEG staging. Apart from that, to capture temporal information, most of the existing methods utilize recurrent neural networks such as LSTM (Long Short Term Memory) which are not effective for modelling global temporal context and difficult to train. Therefore, inspired by the clinical guidelines of sleep staging such as AASM (American Academy of Sleep Medicine) rules where different stages are generally characterized by EEG waveforms of various frequencies, we propose a multi-scale deep architecture by decomposing an EEG signal into different frequency bands as input to CNNs. To model global temporal context, we utilize the multi-head self-attention module of the transformer model to not only improve performance, but also shorten the training time. In addition, we choose residual based architecture which makes training end-to-end. Experimental results on two widely used sleep staging datasets, Montreal Archive of Sleep Studies (MASS) and sleep-EDF datasets, demonstrate the effectiveness and significant efficiency (up to 12 times less training time) of our proposed method over the state-of-the-art.
AB - Sleep staging is to score the sleep state of a subject into different sleep stages such as Wake and Rapid Eye Movement (REM). It plays an indispensable role in the diagnosis and treatment of sleep disorders. As manual sleep staging through well-trained sleep experts is time consuming, tedious, and subjective, many automatic methods have been developed for accurate, efficient, and objective sleep staging. Recently, deep learning based methods have been successfully proposed for electroencephalogram (EEG) based sleep staging with promising results. However, most of these methods directly take EEG raw signals as input of convolutional neural networks (CNNs) without considering the domain knowledge of EEG staging. Apart from that, to capture temporal information, most of the existing methods utilize recurrent neural networks such as LSTM (Long Short Term Memory) which are not effective for modelling global temporal context and difficult to train. Therefore, inspired by the clinical guidelines of sleep staging such as AASM (American Academy of Sleep Medicine) rules where different stages are generally characterized by EEG waveforms of various frequencies, we propose a multi-scale deep architecture by decomposing an EEG signal into different frequency bands as input to CNNs. To model global temporal context, we utilize the multi-head self-attention module of the transformer model to not only improve performance, but also shorten the training time. In addition, we choose residual based architecture which makes training end-to-end. Experimental results on two widely used sleep staging datasets, Montreal Archive of Sleep Studies (MASS) and sleep-EDF datasets, demonstrate the effectiveness and significant efficiency (up to 12 times less training time) of our proposed method over the state-of-the-art.
KW - attention model
KW - deep learning
KW - EEG signal
KW - Hilbert transform
KW - Sleep staging
UR - http://www.scopus.com/inward/record.url?scp=85092750052&partnerID=8YFLogxK
U2 - 10.1109/JBHI.2020.2978004
DO - 10.1109/JBHI.2020.2978004
M3 - Journal article
C2 - 32149700
AN - SCOPUS:85092750052
SN - 2168-2194
VL - 24
SP - 2833
EP - 2843
JO - IEEE Journal of Biomedical and Health Informatics
JF - IEEE Journal of Biomedical and Health Informatics
IS - 10
M1 - 9022981
ER -