NM000132: eeg dataset, 40 subjects#

ERP CORE

Access recordings and metadata through EEGDash.

Citation: Emily S. Kappenman, Jaclyn L. Farrens, Wendy Zhang, Andrew X. Stewart, Steven J. Luck. (2004). ERP CORE. 10.82901/nemar.nm000132

Modality: eeg Subjects: 40 Recordings: 240 License: CC-BY-4.0 Source: nemar

Metadata: Complete (100%)

Quickstart#

Install

pip install eegdash

Access the data

from eegdash.dataset import NM000132

dataset = NM000132(cache_dir="./data")
# Get the raw object of the first recording
raw = dataset.datasets[0].raw
print(raw.info)

Filter by subject

dataset = NM000132(cache_dir="./data", subject="01")

Advanced query

dataset = NM000132(
    cache_dir="./data",
    query={"subject": {"$in": ["01", "02"]}},
)

Iterate recordings

for rec in dataset:
    print(rec.subject, rec.raw.info['sfreq'])

If you use this dataset in your research, please cite the original authors.

BibTeX

@dataset{nm000132,
  title = {ERP CORE},
  author = {Emily S. Kappenman and Jaclyn L. Farrens and Wendy Zhang and Andrew X. Stewart and Steven J. Luck.},
  doi = {10.82901/nemar.nm000132},
  url = {https://doi.org/10.82901/nemar.nm000132},
}

About This Dataset#

DOI

ERP CORE (Compendium of Open Resources and Experiments)

Emily S. Kappenman(1,2), Jaclyn L. Farrens(1), Wendy Zhang(1,2), Andrew X. Stewart(3), & Steven J. Luck(3) 1. San Diego State University, Department of Psychology, San Diego, CA, 92120 2. SDSU/UC San Diego Joint Doctoral Program in Clinical Psychology, San Diego, CA, 92120 3. University of California, Davis, Center for Mind & Brain and Department of Psychology, Davis, CA, 95616

The ERP CORE (https://doi.org/10.18115/D5JW4R) is a freely available online resource consisting of optimized paradigms, experiment control scripts, example data from 40 neurotypical adults, data processing pipelines and analysis scripts, and a broad set of results for 7 widely used ERP components: N170, mismatch negativity (MMN), N2pc, N400, P3, lateralized readiness potential (LRP), and error-related negativity (ERN).

View full README

DOI

ERP CORE (Compendium of Open Resources and Experiments)

Emily S. Kappenman(1,2), Jaclyn L. Farrens(1), Wendy Zhang(1,2), Andrew X. Stewart(3), & Steven J. Luck(3) 1. San Diego State University, Department of Psychology, San Diego, CA, 92120 2. SDSU/UC San Diego Joint Doctoral Program in Clinical Psychology, San Diego, CA, 92120 3. University of California, Davis, Center for Mind & Brain and Department of Psychology, Davis, CA, 95616

The ERP CORE (https://doi.org/10.18115/D5JW4R) is a freely available online resource consisting of optimized paradigms, experiment control scripts, example data from 40 neurotypical adults, data processing pipelines and analysis scripts, and a broad set of results for 7 widely used ERP components: N170, mismatch negativity (MMN), N2pc, N400, P3, lateralized readiness potential (LRP), and error-related negativity (ERN).

What’s Included

  1. Raw EEG data for 6 task paradigms (yielding 7 ERP components) from 40 participants, located in subject folders sub-001 through sub-040

  2. The event code schemes for all experiment paradigms

  3. The task stimuli used for eliciting N170, MMN, and N400, located in the stimuli/ folder

  4. Demographic information for all 40 participants (participants.tsv and participants.json)

Tasks

| Task     | ERP Component(s) | Description                                                   |
| -------- | ---------------- | ------------------------------------------------------------- |
| flankers | ERN, LRP         | Eriksen flanker task (compatible/incompatible arrow flankers) |
| MMN      | MMN              | Passive auditory oddball                                      |
| N170     | N170             | Face perception (faces, cars, scrambled)                      |
| N2pc     | N2pc             | Visual search                                                 |
| N400     | N400             | Word pair judgment (related/unrelated)                        |
| P3       | P3               | Active visual oddball                                         |

The following more detailed description of the tasks was taken from Kappenman et al. 2021 and its supplement. In all tasks the interstimulus image refers to a gray screen with a white central fixation point. The order of the six tasks was randomized across participants, and the task parameters were counterbalanced within participants wherever possible (see individual task descriptions below for details). Interstimulus intervals (ISIs) were jittered using a rectangular distribution to prevent phase-locking of alpha-band EEG oscillations to the stimulus sequence. Each block of trials began with a sequence of “Ready,” “Set,” “Go” screens. Participant-controlled rest breaks were provided between blocks.

Eriksen flanker task

The lateralized readiness potential (LRP) and the error-related negativity (ERN) were elicited using a variant of the Eriksen flanker task ( Eriksen and Eriksen, 1974 ; see Fig. 1 F). A central arrowhead pointing to the left (<) or right (>) was flanked on both sides by 2 arrowheads that pointed in the same direction (congruent trials) or the opposite direction (incongruent trials). Participants indicated the direction of the central arrowhead on each trial with a left or right-hand buttonpress. The arrowheads were displayed for 200 ms followed by the interstimulus image displayed from 1200-1400 ms.

Passive auditory oddball

The MMN was elicited using a passive auditory oddball task modeled on Näätänen et al. (2004). Standard tones (presented at 80 dB, with p = .8) and deviant tones (presented at 70 dB, with p = .2) were presented over speakers while participants watched a silent video and ignored the tones. The stimuli were 1000 Hz pure auditory tones, 100 ms in duration (including 5 ms rise and fall times), separated by a silent ISI of 450-550 ms. To establish the auditory context for the standard, a stream of 15 standard stimuli was presented at the start of the task; all remaining stimuli were presented in a random order with the specified probabilities, except with the constraint that no two deviant tones could be presented sequentially. A total of 1000 tones were presented (including the initial stream of 15 standards), consisting of 800 standards and 200 deviants. Participants watched a silent movie of sand drawings by artist Kseniya Simonova.

Face perception

The N170 was elicited in a face perception task modified from Rossion and Caharel (2011) using their stimuli. In this task, an image of a face, car, scrambled face, or scrambled car was presented on each trial in the center of the screen, and participants responded whether the stimulus was an ‘object’ (face or car) or a ‘texture’ (scrambled face or scrambled car). The cropped images were displayed on a gray background for 300 ms followed by the interstimulus image for 1100-1300 ms. Each participant completed a total of 320 trials. Each category included 40 exemplars, each of which was presented twice, yielding 80 total trials of each stimulus category. The stimuli were presented in a randomly shuffled sequence, with the constraint that a given exemplar was presented only once in the first half and once in the second half of the session. The task was divided into blocks of 40 trials.

Visual search

The N2pc was elicited using a simple visual search task based on Luck et al. (2006). Participants were given a target color of pink or blue at the beginning of a trial block, and responded on each trial whether the gap in the target color square was on the top or bottom. On each trial, 12 items were presented in the left visual field and 12 items were presented in the right visual field. Each item was an outlined square with a gap on one side. Eleven of the items in each visual field were black and had either a left-side or right-side gap (randomly and independently determined). In addition, one blue square and one pink square with a gap on either the top or bottom (randomly and independently determined) were presented on each trial. The blue and pink were equally distant from the gray background in the CIE (1976) color space. The pink item and blue item were always presented in opposite visual fields, with the location randomized across trials. Each stimulus array was presented for 500 ms, separated by the interstimulus image of 900-1100 ms. The fixation point was always visible. The blue item was the target for half of the task, and the pink item was the target for the other half; the order was counterbalanced across participants. Participants pressed a button to indicate the location of the gap in the target square using the index finger (top gap) and middle finger (bottom gap) of the dominant hand. Because the response mapping was natural (e.g., an upper buttonpress for a gap on the top), and the data were eventually collapsed across top and bottom gap positions, the stimulus-response mapping was held consistent across participants. Each participant completed a total of 320 trials, divided into blocks of 40.

Word pair judgement

The N400 was elicited using a word pair judgment task adapted from Holcomb & Kutas (1990). On each trial, a red prime word was followed by a green target word. Participants responded whether the target word was semantically related or unrelated to the prime word. Each target word was presented twice in the experiment, once preceded by an associated prime word and once preceded by an unassociated prime word. Participants received a total of 60 associated word pairs and 60 unassociated word pairs. The order of word pairs was randomized with the constraint that each target word occurred only once within each half of the experiment. The prime word on each trial was presented for 200 ms in red (x = 0.65, y = 0.33, 5.8 cd/m2 ), followed by an ISI of 900-1100 ms. The target word was then presented for 200 ms in green, followed by an intertrial interval (ITI) of 1400-1600 ms. The prime and target words were presented in different colors to make it easier for participants to determine which word in the pair required a response. Participants were instructed to press one button if the target word was semantically related to the prime word and another button if the words were unrelated. Each participant completed 120 trials, divided into blocks of 20.

Active visual search

The P3 was elicited in an active visual oddball task adapted from Luck et al. (2009). The letters A, B, C, D, and E were presented in random order ( p = .2 for each letter). One letter was designated the target for a given block of trials, and the other 4 letters were non-targets. Thus, the probability of the target category was .2, but the same physical stimulus served as a target in some blocks and a nontarget in others. Participants pressed one button for targets and another button for non-targets. Each of the five letters served as a target in one block of the experiment and as a non-target in the other four blocks, with the order of blocks randomized across participants. Each letter was presented with equal probability within a block of trials (p = .2), such that the target category was rare (p = .2) and the non-target category was frequent (p = .8). This design eliminates any possible sensory differences between the target and non-target stimuli, including differential sensory adaptation of the individual target and non-target stimuli (see Luck, 2014).

BIDS Restructuring Notes

The original dataset from NEMAR used ses-* (session) directories for each ERP paradigm. This was corrected because all 6 paradigms were administered in a single recording session (Kappenman et al., 2021). The following changes were made: - Removed session-level directories; data is now organized as sub-XXX/eeg/sub-XXX_task-YYY_* - Merged task-ERN and task-LRP into task-flankers, since both ERP components come from the same flanker paradigm recording (identical raw data) - Fixed EEGCoordinateSystem from invalid "ARS" to "EEGLAB" (Anterior-Left-Superior orientation) - Fixed task-N170_events.json: replaced invalid range-based Levels keys (e.g., "1-40") with a Description field - Fixed task-MMN_events.json and task-N2pc_events.json: added missing event value definitions - Consolidated duplicate electrodes.tsv and coordsystem.json files using BIDS inheritance (one per subject instead of one per task) - Fixed dataset_description.json name from "ERP CORE N400" to "ERP CORE"

Contact

Experiment control files, archival copies of the EEGLAB/ERPLAB scripts and accessory files used for analyzing the data, and the processed data can all be found at https://doi.org/10.18115/D5JW4R. To download the latest versions of the scripts or to report bugs, visit https://github.com/lucklab/ERP_CORE. For other questions about the ERP CORE, contact us at erpbootcamp@gmail.com. Additional ERP resources are available on the ERP Info website (https://erpinfo.org).

Dataset Information#

Dataset ID

NM000132

Title

ERP CORE

Author (year)

Kappenman2021

Canonical

ERPCORE, ERP_CORE

Importable as

NM000132, Kappenman2021, ERPCORE, ERP_CORE

Year

2004

Authors

Emily S. Kappenman, Jaclyn L. Farrens, Wendy Zhang, Andrew X. Stewart, Steven J. Luck.

License

CC-BY-4.0

Citation / DOI

10.82901/nemar.nm000132

Source links

OpenNeuro | NeMAR | Source URL

Copy-paste BibTeX
@dataset{nm000132,
  title = {ERP CORE},
  author = {Emily S. Kappenman and Jaclyn L. Farrens and Wendy Zhang and Andrew X. Stewart and Steven J. Luck.},
  doi = {10.82901/nemar.nm000132},
  url = {https://doi.org/10.82901/nemar.nm000132},
}

Found an issue with this dataset?

If you encounter any problems with this dataset (missing files, incorrect metadata, loading errors, etc.), please let us know!

Report an Issue on GitHub

Technical Details#

Subjects & recordings
  • Subjects: 40

  • Recordings: 240

  • Tasks: 6

Channels & sampling rate
  • Channels: 33

  • Sampling rate (Hz): 1024

  • Duration (hours): 37.74583333333333

Tags
  • Pathology: Not specified

  • Modality: —

  • Type: —

Files & format
  • Size on disk: 17.5 GB

  • File count: 240

  • Format: BIDS

License & citation
  • License: CC-BY-4.0

  • DOI: 10.82901/nemar.nm000132

Provenance

API Reference#

Use the NM000132 class to access this dataset programmatically.

class eegdash.dataset.NM000132(cache_dir: str, query: dict | None = None, s3_bucket: str | None = None, **kwargs)[source]#

Bases: EEGDashDataset

ERP CORE

Study:

nm000132 (NeMAR)

Author (year):

Kappenman2021

Canonical:

ERPCORE, ERP_CORE

Also importable as: NM000132, Kappenman2021, ERPCORE, ERP_CORE.

Modality: eeg. Subjects: 40; recordings: 240; tasks: 6.

Parameters:
  • cache_dir (str | Path) – Directory where data are cached locally.

  • query (dict | None) – Additional MongoDB-style filters to AND with the dataset selection. Must not contain the key dataset.

  • s3_bucket (str | None) – Base S3 bucket used to locate the data.

  • **kwargs (dict) – Additional keyword arguments forwarded to EEGDashDataset.

data_dir#

Local dataset cache directory (cache_dir / dataset_id).

Type:

Path

query#

Merged query with the dataset filter applied.

Type:

dict

records#

Metadata records used to build the dataset, if pre-fetched.

Type:

list[dict] | None

Notes

Each item is a recording; recording-level metadata are available via dataset.description. query supports MongoDB-style filters on fields in ALLOWED_QUERY_FIELDS and is combined with the dataset filter. Dataset-specific caveats are not provided in the summary metadata.

References

OpenNeuro dataset: https://openneuro.org/datasets/nm000132 NeMAR dataset: https://nemar.org/dataexplorer/detail?dataset_id=nm000132 DOI: https://doi.org/10.82901/nemar.nm000132

Examples

>>> from eegdash.dataset import NM000132
>>> dataset = NM000132(cache_dir="./data")
>>> recording = dataset[0]
>>> raw = recording.load()
__init__(cache_dir: str, query: dict | None = None, s3_bucket: str | None = None, **kwargs)[source]#
save(path, overwrite=False)[source]#

Save the dataset to disk.

Parameters:
  • path (str or Path) – Destination file path.

  • overwrite (bool, default False) – If True, overwrite existing file.

Return type:

None

See Also#