Residential Collegefalse
Status即將出版Forthcoming
Subject-specific CNN model with parameter-based transfer learning for SSVEP detection
Ji, Zhouyu1; Xu, Tao2; Chen, Chuangquan1; Yin, Haojun1; Wan, Feng3,4; Wang, Hongtao1
2025-05-01
Source PublicationBiomedical Signal Processing and Control
ISSN1746-8094
Volume103Pages:107404
Abstract

Steady-state visual evoked potentials (SSVEP)-based brain–computer interfaces (BCIs) leverage machine learning methods to enhance performance. However, these methods require a sufficiently long time window to achieve high accuracy and information transfer rate (ITR), which restricts their applications in real-world scenarios, particularly for user-specific decoding. To address this issue, we propose a parameter-based transfer learning CNN (PTL-CNN) approach for the SSVEP-BCI system, which can automatically fuse and extract both inter- and intra-subject features in EEG signals. Specifically, we first introduce a shallow CNN architecture and adopt a short time-window to train a pretrained model on a dataset comprising numerous subjects, aiming to explore the universal features across subjects. Subsequently, a new user is utilized to fine-tune the model, calibrating it to this specific user. Experimental results demonstrate that PTL-CNN achieves remarkable performance and significantly outperforms the compared algorithms under short time windows. For instance, in a time window of 0.4 s, PTL-CNN achieves an average accuracy of 80.60% with an average ITR of 247.77 bits/min on the Benchmark dataset, and an average accuracy of 66.91% with an average ITR of 185.90 bits/min on the Beta dataset. This performance is significantly better than that of Ensemble-TRCA (Benchmark: 71.21%, 209.12 bits/min; Beta: 53.04%, 135.53 bits/min). In summary, our proposed PTL-CNN achieves the highest average accuracy with the fastest average ITR and is of implications for the development of real-time BCI applications, as well as inspiration for other application paradigms.

KeywordBrain–computer Interface Deep Learning Electroencephalogram (Eeg) Steady-state Visual Evoked Potential (Ssvep) Transfer Learning
DOI10.1016/j.bspc.2024.107404
URLView the original
Language英語English
PublisherElsevier Ltd
Scopus ID2-s2.0-85213080153
Fulltext Access
Citation statistics
Document TypeJournal article
CollectionDEPARTMENT OF ELECTRICAL AND COMPUTER ENGINEERING
Affiliation1.School of Electronics and Information Engineering at Wuyi University, Jiangmen, China
2.Department of Biomedical Engineering at Shantou University, Shantou, China
3.Department of Electrical and Computer Engineering, Faculty of Science and Engineering, University of Macau, Macao
4.Centre for Cognitive and Brain Sciences, Institute of Collaborative Innovation, University of Macau, Macao
Recommended Citation
GB/T 7714
Ji, Zhouyu,Xu, Tao,Chen, Chuangquan,et al. Subject-specific CNN model with parameter-based transfer learning for SSVEP detection[J]. Biomedical Signal Processing and Control, 2025, 103, 107404.
APA Ji, Zhouyu., Xu, Tao., Chen, Chuangquan., Yin, Haojun., Wan, Feng., & Wang, Hongtao (2025). Subject-specific CNN model with parameter-based transfer learning for SSVEP detection. Biomedical Signal Processing and Control, 103, 107404.
MLA Ji, Zhouyu,et al."Subject-specific CNN model with parameter-based transfer learning for SSVEP detection".Biomedical Signal Processing and Control 103(2025):107404.
Files in This Item:
There are no files associated with this item.
Related Services
Recommend this item
Bookmark
Usage statistics
Export to Endnote
Google Scholar
Similar articles in Google Scholar
[Ji, Zhouyu]'s Articles
[Xu, Tao]'s Articles
[Chen, Chuangquan]'s Articles
Baidu academic
Similar articles in Baidu academic
[Ji, Zhouyu]'s Articles
[Xu, Tao]'s Articles
[Chen, Chuangquan]'s Articles
Bing Scholar
Similar articles in Bing Scholar
[Ji, Zhouyu]'s Articles
[Xu, Tao]'s Articles
[Chen, Chuangquan]'s Articles
Terms of Use
No data!
Social Bookmark/Share
All comments (0)
No comment.
 

Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.