Projects per year
Abstract
Objective. Deep learning is increasingly used for brain-computer interfaces (BCIs). However, the quantity of available data is sparse, especially for invasive BCIs. Data augmentation (DA) methods, such as generative models, can help to address this sparseness. However, all the existing studies on brain signals were based on convolutional neural networks and ignored the temporal dependence. This paper attempted to enhance generative models by capturing the temporal relationship from a time-series perspective. Approach. A conditional generative network (conditional transformer-based generative adversarial network (cTGAN)) based on the transformer model was proposed. The proposed method was tested using a stereo-electroencephalography (SEEG) dataset which was recorded from eight epileptic patients performing five different movements. Three other commonly used DA methods were also implemented: noise injection (NI), variational autoencoder (VAE), and conditional Wasserstein generative adversarial network with gradient penalty (cWGANGP). Using the proposed method, the artificial SEEG data was generated, and several metrics were used to compare the data quality, including visual inspection, cosine similarity (CS), Jensen-Shannon distance (JSD), and the effect on the performance of a deep learning-based classifier. Main results. Both the proposed cTGAN and the cWGANGP methods were able to generate realistic data, while NI and VAE outputted inferior samples when visualized as raw sequences and in a lower dimensional space. The cTGAN generated the best samples in terms of CS and JSD and outperformed cWGANGP significantly in enhancing the performance of a deep learning-based classifier (each of them yielding a significant improvement of 6% and 3.4%, respectively). Significance. This is the first time that DA methods have been applied to invasive BCIs based on SEEG. In addition, this study demonstrated the advantages of the model that preserves the temporal dependence from a time-series perspective.
Original language | English |
---|---|
Article number | 016026 |
Journal | Journal of Neural Engineering |
Volume | 21 |
Issue number | 1 |
Early online date | 22 Feb 2024 |
DOIs | |
Publication status | Published - 29 Feb 2024 |
Data Availability Statement
The data cannot be made publicly available upon publication due to legal restrictions preventing unrestricted public distribution. The data that support the findings of this study are available upon reasonable request from the authors.Funding
This work is supported by the EPSRC New Horizons Grant of UK (EP/X018342/1), the National Natural Science Foundation of China (Nos. 91848112, 52105030 and 82272116), the China Postdoctoral Science Foundation (No. 20Z102060158), and the Medical & Engineering Cross Foundation of SJTU (No. AH0200003).
Funders | Funder number |
---|---|
Medical & Engineering Cross Foundation of SJTU | AH0200003 |
Engineering and Physical Sciences Research Council | EP/X018342/1 |
National Natural Science Foundation of China | 52105030, 91848112, 82272116 |
China Postdoctoral Science Foundation | 20Z102060158 |
Keywords
- brain-computer interface (BCI)
- data augmentation
- deep learning
- stereo-electroencephalography (SEEG)
- transformer
ASJC Scopus subject areas
- Cellular and Molecular Neuroscience
- Biomedical Engineering
Fingerprint
Dive into the research topics of 'Data augmentation for invasive brain-computer interfaces based on stereo-electroencephalography (SEEG)'. Together they form a unique fingerprint.Projects
- 1 Active
-
New Horizons 21 - Decoding Speech using Invasive Brain-Computer Interfaces based on Intracranial Brain Signals (dSPEECH)
Zhang, D. (PI)
Engineering and Physical Sciences Research Council
1/01/23 → 31/12/24
Project: Research council