TY - GEN
T1 - Neural-MCRL: Neural Multimodal Contrastive Representation Learning for EEG-based Visual Decoding
AU - Li, Yueyang
AU - Kang, Zijian
AU - Gong, Shengyu
AU - Dong, Wenhao
AU - Zeng, Weiming
AU - Yan, Hongjie
AU - Siok, Wai Ting
AU - Wang, Nizhuan
N1 - Publisher Copyright:
© 2025 IEEE.
PY - 2025/10/30
Y1 - 2025/10/30
N2 - Decoding neural visual representations from electroencephalogram (EEG)-based brain activity is crucial for advancing brain-machine interfaces (BMI) and has transformative potential for neural sensory rehabilitation. While multimodal contrastive representation learning (MCRL) has shown promise in neural decoding, existing methods often overlook semantic consistency and completeness within modalities and lack effective semantic alignment across modalities. This limits their ability to capture the complex representations of visual neural responses. We propose Neural-MCRL, a novel framework that achieves multimodal alignment through semantic bridging and cross-attention mechanisms, while ensuring completeness within modalities and consistency across modalities. Our framework also features the Neural Encoder with Spectral-Temporal Adaptation (NESTA), a EEG encoder that adaptively captures spectral patterns and learns subject-specific transformations. Experimental results demonstrate significant improvements in visual decoding accuracy and model generalization compared to state-of-the-art methods, advancing the field of EEG-based neural visual representation decoding in BMI. Code will be available at: https://github.com/NZWANG/Neural-MCRL.
AB - Decoding neural visual representations from electroencephalogram (EEG)-based brain activity is crucial for advancing brain-machine interfaces (BMI) and has transformative potential for neural sensory rehabilitation. While multimodal contrastive representation learning (MCRL) has shown promise in neural decoding, existing methods often overlook semantic consistency and completeness within modalities and lack effective semantic alignment across modalities. This limits their ability to capture the complex representations of visual neural responses. We propose Neural-MCRL, a novel framework that achieves multimodal alignment through semantic bridging and cross-attention mechanisms, while ensuring completeness within modalities and consistency across modalities. Our framework also features the Neural Encoder with Spectral-Temporal Adaptation (NESTA), a EEG encoder that adaptively captures spectral patterns and learns subject-specific transformations. Experimental results demonstrate significant improvements in visual decoding accuracy and model generalization compared to state-of-the-art methods, advancing the field of EEG-based neural visual representation decoding in BMI. Code will be available at: https://github.com/NZWANG/Neural-MCRL.
KW - EEG-based visual decoding
KW - Multimodal contrastive representation learning
KW - Semantic consistency and completion
KW - Multimodal semantic alignment
UR - https://www.scopus.com/pages/publications/105022635770
U2 - 10.1109/ICME59968.2025.11210130
DO - 10.1109/ICME59968.2025.11210130
M3 - Conference article published in proceeding or book
T3 - Proceedings - IEEE International Conference on Multimedia and Expo
SP - 1
EP - 6
BT - 2025 IEEE International Conference on Multimedia and Expo
PB - Institute of Electrical and Electronics Engineers Inc.
ER -