DS007591: eeg dataset, 3 subjects#
Delineating neural contributions to EEG-based speech decoding
Access recordings and metadata through EEGDash.
Citation: Motoshige Sato, Yasuo Kabe, Sensho Nobe, Akito Yoshida, Masakazu Inoue, Mayumi Shimizu, Kenichi Tomeoka, Shuntaro Sasai (2026). Delineating neural contributions to EEG-based speech decoding. 10.18112/openneuro.ds007591.v1.0.1
Modality: eeg Subjects: 3 Recordings: 21 License: CC0 Source: openneuro
Metadata: Complete (100%)
Quickstart#
Install
pip install eegdash
Access the data
from eegdash.dataset import DS007591
dataset = DS007591(cache_dir="./data")
# Get the raw object of the first recording
raw = dataset.datasets[0].raw
print(raw.info)
Filter by subject
dataset = DS007591(cache_dir="./data", subject="01")
Advanced query
dataset = DS007591(
cache_dir="./data",
query={"subject": {"$in": ["01", "02"]}},
)
Iterate recordings
for rec in dataset:
print(rec.subject, rec.raw.info['sfreq'])
If you use this dataset in your research, please cite the original authors.
BibTeX
@dataset{ds007591,
title = {Delineating neural contributions to EEG-based speech decoding},
author = {Motoshige Sato and Yasuo Kabe and Sensho Nobe and Akito Yoshida and Masakazu Inoue and Mayumi Shimizu and Kenichi Tomeoka and Shuntaro Sasai},
doi = {10.18112/openneuro.ds007591.v1.0.1},
url = {https://doi.org/10.18112/openneuro.ds007591.v1.0.1},
}
About This Dataset#
Delineating neural contributions to EEG-based speech decoding
Overview
128-channel EEG recordings during speech production tasks. Participants produced one of 5 color words (green, magenta, orange, violet, yellow) under three speech conditions: overt, minimally overt, and covert. Each trial consists of 5 repetitions of the same word (1.25 sec per repetition).
View full README
Delineating neural contributions to EEG-based speech decoding
Overview
128-channel EEG recordings during speech production tasks. Participants produced one of 5 color words (green, magenta, orange, violet, yellow) under three speech conditions: overt, minimally overt, and covert. Each trial consists of 5 repetitions of the same word (1.25 sec per repetition).
Channel layout (139 channels total)
Channels 1-128: EEG
Channels 129-130: DISPLAY (bipolar pair, misc)
Channels 131-132: MIC (bipolar pair, misc)
Channels 133-134: EOG (bipolar pair)
Channels 135-136: EMG upper orbicularis oris (bipolar pair)
Channels 137-138: EMG lower orbicularis oris (bipolar pair)
Channel 139: TRIGGER (marks trial onsets)
Session types
calibration: Offline data collection for decoder training
online: Real-time decoding with trained decoder
Preprocessing note
The EEG channels were recorded with a 10x preamp gain. Raw values have been converted to Volts (×1e-6).
Code
Code for data loading, preprocessing, and decoding models is available at: https://github.com/arayabrain/uhd-gmail-public
Dataset Information#
Dataset ID |
|
Title |
Delineating neural contributions to EEG-based speech decoding |
Author (year) |
|
Canonical |
|
Importable as |
|
Year |
2026 |
Authors |
Motoshige Sato, Yasuo Kabe, Sensho Nobe, Akito Yoshida, Masakazu Inoue, Mayumi Shimizu, Kenichi Tomeoka, Shuntaro Sasai |
License |
CC0 |
Citation / DOI |
|
Source links |
OpenNeuro | NeMAR | Source URL |
Copy-paste BibTeX
@dataset{ds007591,
title = {Delineating neural contributions to EEG-based speech decoding},
author = {Motoshige Sato and Yasuo Kabe and Sensho Nobe and Akito Yoshida and Masakazu Inoue and Mayumi Shimizu and Kenichi Tomeoka and Shuntaro Sasai},
doi = {10.18112/openneuro.ds007591.v1.0.1},
url = {https://doi.org/10.18112/openneuro.ds007591.v1.0.1},
}
Found an issue with this dataset?
If you encounter any problems with this dataset (missing files, incorrect metadata, loading errors, etc.), please let us know!
Technical Details#
Subjects: 3
Recordings: 21
Tasks: 3
Channels: 139
Sampling rate (Hz): 256.0
Duration (hours): 6.775833333333333
Pathology: Healthy
Modality: —
Type: Motor
Size on disk: 1.6 GB
File count: 21
Format: BIDS
License: CC0
DOI: doi:10.18112/openneuro.ds007591.v1.0.1
API Reference#
Use the DS007591 class to access this dataset programmatically.
- class eegdash.dataset.DS007591(cache_dir: str, query: dict | None = None, s3_bucket: str | None = None, **kwargs)[source]#
Bases:
EEGDashDatasetDelineating neural contributions to EEG-based speech decoding
- Study:
ds007591(OpenNeuro)- Author (year):
Sato2026_Delineating- Canonical:
Sato2025
Also importable as:
DS007591,Sato2026_Delineating,Sato2025.Modality:
eeg; Experiment type:Motor; Subject type:Healthy. Subjects: 3; recordings: 21; tasks: 3.- Parameters:
cache_dir (str | Path) – Directory where data are cached locally.
query (dict | None) – Additional MongoDB-style filters to AND with the dataset selection. Must not contain the key
dataset.s3_bucket (str | None) – Base S3 bucket used to locate the data.
**kwargs (dict) – Additional keyword arguments forwarded to
EEGDashDataset.
- data_dir#
Local dataset cache directory (
cache_dir / dataset_id).- Type:
Path
- query#
Merged query with the dataset filter applied.
- Type:
dict
- records#
Metadata records used to build the dataset, if pre-fetched.
- Type:
list[dict] | None
Notes
Each item is a recording; recording-level metadata are available via
dataset.description.querysupports MongoDB-style filters on fields inALLOWED_QUERY_FIELDSand is combined with the dataset filter. Dataset-specific caveats are not provided in the summary metadata.References
OpenNeuro dataset: https://openneuro.org/datasets/ds007591 NeMAR dataset: https://nemar.org/dataexplorer/detail?dataset_id=ds007591 DOI: https://doi.org/10.18112/openneuro.ds007591.v1.0.1
Examples
>>> from eegdash.dataset import DS007591 >>> dataset = DS007591(cache_dir="./data") >>> recording = dataset[0] >>> raw = recording.load()
See Also#
eegdash.dataset.EEGDashDataseteegdash.dataset