DS006914#
Visual Naming EC
Access recordings and metadata through EEGDash.
Citation: Ryuzaburo Kochi, Aya Kanno, Hiroshi Uda, Keisuke Hatano, Masaki Sonoda, Hidenori Endo, Michael Cools, Robert Rothermel, Aimee F. Luat, Eishi Asano (2025). Visual Naming EC. 10.18112/openneuro.ds006914.v1.0.3
Modality: ieeg Subjects: 110 Recordings: 1770 License: CC0 Source: openneuro
Metadata: Complete (100%)
Quickstart#
Install
pip install eegdash
Access the data
from eegdash.dataset import DS006914
dataset = DS006914(cache_dir="./data")
# Get the raw object of the first recording
raw = dataset.datasets[0].raw
print(raw.info)
Filter by subject
dataset = DS006914(cache_dir="./data", subject="01")
Advanced query
dataset = DS006914(
cache_dir="./data",
query={"subject": {"$in": ["01", "02"]}},
)
Iterate recordings
for rec in dataset:
print(rec.subject, rec.raw.info['sfreq'])
If you use this dataset in your research, please cite the original authors.
BibTeX
@dataset{ds006914,
title = {Visual Naming EC},
author = {Ryuzaburo Kochi and Aya Kanno and Hiroshi Uda and Keisuke Hatano and Masaki Sonoda and Hidenori Endo and Michael Cools and Robert Rothermel and Aimee F. Luat and Eishi Asano},
doi = {10.18112/openneuro.ds006914.v1.0.3},
url = {https://doi.org/10.18112/openneuro.ds006914.v1.0.3},
}
About This Dataset#
This dataset, used in the analysis reported by Kochi et al., (2025), contains intracranial EEG recordings from 110 individuals who performed an visual‑naming task. Electrode coordinates are provided in MNI‑305 space.
Each EDF file is tagged for the visual naming task with the following event codes: 401 – stimulus onset 501 – response onset
Reference: Ryuzaburo Kochi, Aya Kanno, Hiroshi Uda, Keisuke Hatano, Masaki Sonoda, Hidenori Endo, Michael Cools, Robert Rothermel, Aimee F. Luat, Eishi Asano. Whole-Brain Millisecond-Scale Effective Connectivity Atlases of Speech
Dataset Information#
Dataset ID |
|
Title |
Visual Naming EC |
Year |
2025 |
Authors |
Ryuzaburo Kochi, Aya Kanno, Hiroshi Uda, Keisuke Hatano, Masaki Sonoda, Hidenori Endo, Michael Cools, Robert Rothermel, Aimee F. Luat, Eishi Asano |
License |
CC0 |
Citation / DOI |
|
Source links |
OpenNeuro | NeMAR | Source URL |
Copy-paste BibTeX
@dataset{ds006914,
title = {Visual Naming EC},
author = {Ryuzaburo Kochi and Aya Kanno and Hiroshi Uda and Keisuke Hatano and Masaki Sonoda and Hidenori Endo and Michael Cools and Robert Rothermel and Aimee F. Luat and Eishi Asano},
doi = {10.18112/openneuro.ds006914.v1.0.3},
url = {https://doi.org/10.18112/openneuro.ds006914.v1.0.3},
}
Found an issue with this dataset?
If you encounter any problems with this dataset (missing files, incorrect metadata, loading errors, etc.), please let us know!
Technical Details#
Subjects: 110
Recordings: 1770
Tasks: 1
Channels: 128 (490), 138 (38), 136 (38), 140 (16), 112 (12), 110 (12), 150 (10), 156 (10), 164 (8), 148 (8), 130 (8), 134 (8), 84 (6), 96 (6), 154 (6), 144 (6), 118 (6), 152 (6), 160 (6), 64 (4), 58 (2)
Sampling rate (Hz): 1000.0
Duration (hours): 0.0
Pathology: Epilepsy
Modality: Visual
Type: Other
Size on disk: 17.5 GB
File count: 1770
Format: BIDS
License: CC0
DOI: doi:10.18112/openneuro.ds006914.v1.0.3
API Reference#
Use the DS006914 class to access this dataset programmatically.
- class eegdash.dataset.DS006914(cache_dir: str, query: dict | None = None, s3_bucket: str | None = None, **kwargs)[source]#
Bases:
EEGDashDatasetOpenNeuro dataset
ds006914. Modality:ieeg; Experiment type:Other; Subject type:Epilepsy. Subjects: 110; recordings: 353; tasks: 1.- Parameters:
cache_dir (str | Path) – Directory where data are cached locally.
query (dict | None) – Additional MongoDB-style filters to AND with the dataset selection. Must not contain the key
dataset.s3_bucket (str | None) – Base S3 bucket used to locate the data.
**kwargs (dict) – Additional keyword arguments forwarded to
EEGDashDataset.
- data_dir#
Local dataset cache directory (
cache_dir / dataset_id).- Type:
Path
- query#
Merged query with the dataset filter applied.
- Type:
dict
- records#
Metadata records used to build the dataset, if pre-fetched.
- Type:
list[dict] | None
Notes
Each item is a recording; recording-level metadata are available via
dataset.description.querysupports MongoDB-style filters on fields inALLOWED_QUERY_FIELDSand is combined with the dataset filter. Dataset-specific caveats are not provided in the summary metadata.References
OpenNeuro dataset: https://openneuro.org/datasets/ds006914 NeMAR dataset: https://nemar.org/dataexplorer/detail?dataset_id=ds006914
Examples
>>> from eegdash.dataset import DS006914 >>> dataset = DS006914(cache_dir="./data") >>> recording = dataset[0] >>> raw = recording.load()
See Also#
eegdash.dataset.EEGDashDataseteegdash.dataset