NM000196: eeg dataset, 12 subjects#
c-VEP dataset from Thielen et al. (2015)
Access recordings and metadata through EEGDash.
Citation: Jordy Thielen, Philip van den Broek, Jason Farquhar, Peter Desain (2015). c-VEP dataset from Thielen et al. (2015).
Modality: eeg Subjects: 12 Recordings: 36 License: CC0-1.0 Source: nemar
Metadata: Complete (90%)
Quickstart#
Install
pip install eegdash
Access the data
from eegdash.dataset import NM000196
dataset = NM000196(cache_dir="./data")
# Get the raw object of the first recording
raw = dataset.datasets[0].raw
print(raw.info)
Filter by subject
dataset = NM000196(cache_dir="./data", subject="01")
Advanced query
dataset = NM000196(
cache_dir="./data",
query={"subject": {"$in": ["01", "02"]}},
)
Iterate recordings
for rec in dataset:
print(rec.subject, rec.raw.info['sfreq'])
If you use this dataset in your research, please cite the original authors.
BibTeX
@dataset{nm000196,
title = {c-VEP dataset from Thielen et al. (2015)},
author = {Jordy Thielen and Philip van den Broek and Jason Farquhar and Peter Desain},
}
About This Dataset#
c-VEP dataset from Thielen et al. (2015)
c-VEP dataset from Thielen et al. (2015)
Dataset Overview
Code: Thielen2015
Paradigm: cvep
DOI: 10.34973/1ecz-1232
View full README
c-VEP dataset from Thielen et al. (2015)
c-VEP dataset from Thielen et al. (2015)
Dataset Overview
Code: Thielen2015
Paradigm: cvep
DOI: 10.34973/1ecz-1232
Subjects: 12
Sessions per subject: 1
Events: 1.0=101, 0.0=100
Trial interval: (0, 0.3) s
Runs per session: 3
File format: mat
Data preprocessed: True
Acquisition
Sampling rate: 2048.0 Hz
Number of channels: 64
Channel types: eeg=64
Channel names: AF3, AF4, AF7, AF8, AFz, C1, C2, C3, C4, C5, C6, CP1, CP2, CP3, CP4, CP5, CP6, CPz, Cz, F1, F2, F3, F4, F5, F6, F7, F8, FC1, FC2, FC3, FC4, FC5, FC6, FCz, FT7, FT8, Fp1, Fp2, Fpz, Fz, Iz, O1, O2, Oz, P1, P10, P2, P3, P4, P5, P6, P7, P8, P9, PO3, PO4, PO7, PO8, POz, Pz, T7, T8, TP7, TP8
Montage: standard_1020
Hardware: Biosemi ActiveTwo
Reference: CMS/DRL
Sensor type: EEG
Line frequency: 50.0 Hz
Electrode type: active
Participants
Number of subjects: 12
Health status: patients
Clinical population: Healthy
Age: mean=24.0, std=2.3
Gender distribution: male=4, female=8
BCI experience: naive
Species: human
Experimental Protocol
Paradigm: cvep
Number of classes: 2
Class labels: 1.0, 0.0
Trial duration: 4.2 s
Study design: 6x6 matrix speller BCI using modulated Gold codes for visual stimulation; participants focused on target symbols while cells flashed according to pseudo-random bit-sequences
Feedback type: visual
Stimulus type: pseudo-random noise-code
Stimulus modalities: visual
Primary modality: visual
Synchronicity: synchronous
Mode: online
Training/test split: False
Instructions: participants visually attended cells containing target symbols during stimulation
HED Event Annotations
Schema: HED 8.4.0 | Browse: https://www.hedtags.org/hed-schema-browser
1.0
├─ Sensory-event
├─ Experimental-stimulus
├─ Visual-presentation
└─ Label/intensity_1_0
0.0
├─ Sensory-event
├─ Experimental-stimulus
├─ Visual-presentation
└─ Label/intensity_0_0
Paradigm-Specific Parameters
Detected paradigm: cvep
Code type: modulated Gold codes
Code length: 126
Number of targets: 36
Data Structure
Trials: 108
Trials context: 108 total per subject: 3 fixed-length copy-spelling runs x 36 trials per run, each trial 4.2 seconds (4 code cycles)
Preprocessing
Data state: preprocessed
Preprocessing applied: True
Steps: downsampling from 2048 Hz to 360 Hz, linear de-trending, common average referencing, spectral filtering
Highpass filter: 5 Hz
Lowpass filter: 100 Hz
Bandpass filter: {‘band1’: [5, 48], ‘band2’: [52, 100]}
Re-reference: car
Downsampled to: 360.0 Hz
Signal Processing
Classifiers: template matching, CCA
Feature extraction: correlation
Spatial filters: Canonical Correlation Analysis
Cross-Validation
Method: training-testing split
Evaluation type: within-subject
Performance (Original Study)
Accuracy Fixed Length: 86.0
Itr Fixed Length: 38.12
Spm Fixed Length: 6.93
Accuracy Early Stopping: 86.0
Itr Early Stopping: 48.37
Spm Early Stopping: 8.99
BCI Application
Applications: speller
Environment: laboratory
Online feedback: True
Tags
Pathology: Healthy
Modality: Visual
Type: Research
Documentation
DOI: 10.1371/journal.pone.0133797
License: CC0-1.0
Investigators: Jordy Thielen, Philip van den Broek, Jason Farquhar, Peter Desain
Senior author: Peter Desain
Contact: jordy.thielen@gmail.com; info@donders.ru.nl
Institution: Radboud University Nijmegen
Department: Donders Center for Cognition
Country: NL
Repository: GitHub
Data URL: https://public.data.ru.nl/dcc/DSC_2018.00047_553_v3
Publication year: 2015
Funding: BrainGain Smart Mix Program of the Netherlands Ministry of Economic Affairs; Netherlands Ministry of Education, Culture and Science (SSM06011)
Ethics approval: Ethical Committee of the Faculty of Social Sciences at the Radboud University Nijmegen
Keywords: Brain-Computer Interface, BCI, Broad-Band Visually Evoked Potentials, BBVEP, Gold codes, reconvolution, speller, visual stimulation
Abstract
Brain-Computer Interfaces (BCIs) allow users to control devices and communicate by using brain activity only. BCIs based on broad-band visual stimulation can outperform BCIs using other stimulation paradigms. Visual stimulation with pseudo-random bit-sequences evokes specific Broad-Band Visually Evoked Potentials (BBVEPs) that can be reliably used in BCI for high-speed communication in speller applications. In this study, we report a novel paradigm for a BBVEP-based BCI that utilizes a generative framework to predict responses to broad-band stimulation sequences. In this study we designed a BBVEP-based BCI using modulated Gold codes to mark cells in a visual speller BCI. We defined a linear generative model that decomposes full responses into overlapping single-flash responses. These single-flash responses are used to predict responses to novel stimulation sequences, which in turn serve as templates for classification. The linear generative model explains on average 50% and up to 66% of the variance of responses to both seen and unseen sequences. In an online experiment, 12 participants tested a 6 × 6 matrix speller BCI. On average, an online accuracy of 86% was reached with trial lengths of 3.21 seconds. This corresponds to an Information Transfer Rate of 48 bits per minute (approximately 9 symbols per minute). This study indicates the potential to model and predict responses to broad-band stimulation. These predicted responses are proven to be well-suited as templates for a BBVEP-based BCI, thereby enabling communication and control by brain activity only.
Methodology
The study implements a novel BBVEP-based BCI using modulated Gold codes with a reconvolution approach for template generation. The reconvolution model decomposes responses into single-flash responses (short and long pulses) and predicts responses to unseen sequences. Two sets of Gold codes were used: set V for training (65 sequences) and set U for testing (65 sequences). Each sequence had 126 bits with duration of 1.05s. The classifier uses template matching with correlation, combined with Canonical Correlation Analysis for spatial filtering. Subset optimization (Platinum subset) selects the most distinguishable codes, and layout optimization arranges codes on the 6x6 grid to minimize cross-talk. An early stopping algorithm was implemented to reduce trial duration. Online experiments were conducted with 12 participants using a synchronous BCI paradigm.
References
Thielen, J. (Jordy), Jason Farquhar, Desain, P.W.M. (Peter) (2023): Broad-Band Visually Evoked Potentials: Re(con)volution in Brain-Computer Interfacing. Version 2. Radboud University. (dataset). DOI: https://doi.org/10.34973/1ecz-1232 Thielen, J., Van Den Broek, P., Farquhar, J., & Desain, P. (2015). Broad-Band visually evoked potentials: re(con)volution in brain-computer interfacing. PLOS ONE, 10(7), e0133797. DOI: https://doi.org/10.1371/journal.pone.0133797 Notes .. versionadded:: 1.0.0 Appelhoff, S., Sanderson, M., Brooks, T., Vliet, M., Quentin, R., Holdgraf, C., Chaumon, M., Mikulan, E., Tavabi, K., Hochenberger, R., Welke, D., Brunner, C., Rockhill, A., Larson, E., Gramfort, A. and Jas, M. (2019). MNE-BIDS: Organizing electrophysiological data into the BIDS format and facilitating their analysis. Journal of Open Source Software 4: (1896). https://doi.org/10.21105/joss.01896 Pernet, C. R., Appelhoff, S., Gorgolewski, K. J., Flandin, G., Phillips, C., Delorme, A., Oostenveld, R. (2019). EEG-BIDS, an extension to the brain imaging data structure for electroencephalography. Scientific Data, 6, 103. https://doi.org/10.1038/s41597-019-0104-8 Generated by MOABB 1.5.0 (Mother of All BCI Benchmarks) https://github.com/NeuroTechX/moabb
Dataset Information#
Dataset ID |
|
Title |
c-VEP dataset from Thielen et al. (2015) |
Author (year) |
|
Canonical |
— |
Importable as |
|
Year |
2015 |
Authors |
Jordy Thielen, Philip van den Broek, Jason Farquhar, Peter Desain |
License |
CC0-1.0 |
Citation / DOI |
Unknown |
Source links |
OpenNeuro | NeMAR | Source URL |
Found an issue with this dataset?
If you encounter any problems with this dataset (missing files, incorrect metadata, loading errors, etc.), please let us know!
Technical Details#
Subjects: 12
Recordings: 36
Tasks: 1
Channels: 64
Sampling rate (Hz): 2048.0
Duration (hours): 2.6154667154947915
Pathology: Healthy
Modality: Visual
Type: Attention
Size on disk: 3.5 GB
File count: 36
Format: BIDS
License: CC0-1.0
DOI: —
API Reference#
Use the NM000196 class to access this dataset programmatically.
- class eegdash.dataset.NM000196(cache_dir: str, query: dict | None = None, s3_bucket: str | None = None, **kwargs)[source]#
Bases:
EEGDashDatasetc-VEP dataset from Thielen et al. (2015)
- Study:
nm000196(NeMAR)- Author (year):
Thielen2015- Canonical:
—
Also importable as:
NM000196,Thielen2015.Modality:
eeg; Experiment type:Attention; Subject type:Healthy. Subjects: 12; recordings: 36; tasks: 1.- Parameters:
cache_dir (str | Path) – Directory where data are cached locally.
query (dict | None) – Additional MongoDB-style filters to AND with the dataset selection. Must not contain the key
dataset.s3_bucket (str | None) – Base S3 bucket used to locate the data.
**kwargs (dict) – Additional keyword arguments forwarded to
EEGDashDataset.
- data_dir#
Local dataset cache directory (
cache_dir / dataset_id).- Type:
Path
- query#
Merged query with the dataset filter applied.
- Type:
dict
- records#
Metadata records used to build the dataset, if pre-fetched.
- Type:
list[dict] | None
Notes
Each item is a recording; recording-level metadata are available via
dataset.description.querysupports MongoDB-style filters on fields inALLOWED_QUERY_FIELDSand is combined with the dataset filter. Dataset-specific caveats are not provided in the summary metadata.References
OpenNeuro dataset: https://openneuro.org/datasets/nm000196 NeMAR dataset: https://nemar.org/dataexplorer/detail?dataset_id=nm000196
Examples
>>> from eegdash.dataset import NM000196 >>> dataset = NM000196(cache_dir="./data") >>> recording = dataset[0] >>> raw = recording.load()
See Also#
eegdash.dataset.EEGDashDataseteegdash.dataset