NM000112: eeg dataset, 123 subjects#

FACED - Finer-grained Affective Computing EEG Dataset

Access recordings and metadata through EEGDash.

Citation: Yisi Liu, Olga Sourina, Minh Khoa Nguyen (2023). FACED - Finer-grained Affective Computing EEG Dataset. 10.82901/nemar.nm000112

Modality: eeg Subjects: 123 Recordings: 123 License: CC-BY-4.0 Source: nemar

Metadata: Complete (100%)

Quickstart#

Install

pip install eegdash

Access the data

from eegdash.dataset import NM000112

dataset = NM000112(cache_dir="./data")
# Get the raw object of the first recording
raw = dataset.datasets[0].raw
print(raw.info)

Filter by subject

dataset = NM000112(cache_dir="./data", subject="01")

Advanced query

dataset = NM000112(
    cache_dir="./data",
    query={"subject": {"$in": ["01", "02"]}},
)

Iterate recordings

for rec in dataset:
    print(rec.subject, rec.raw.info['sfreq'])

If you use this dataset in your research, please cite the original authors.

BibTeX

@dataset{nm000112,
  title = {FACED - Finer-grained Affective Computing EEG Dataset},
  author = {Yisi Liu and Olga Sourina and Minh Khoa Nguyen},
  doi = {10.82901/nemar.nm000112},
  url = {https://doi.org/10.82901/nemar.nm000112},
}

About This Dataset#

DOI

FACED - Finer-grained Affective Computing EEG Dataset

Introduction

The Finer-grained Affective Computing EEG Dataset (FACED) contains scalp EEG recordings from 123 healthy participants who watched 28 emotion-eliciting video clips designed to evoke nine different emotion categories. The dataset includes four negative emotions (anger, fear, disgust, sadness) from Ekman’s basic emotions and four positive emotions (amusement, inspiration, joy, tenderness) selected based on recent psychological and neuroscience progress and application needs. Participants provided detailed self-reported emotion ratings on 12 dimensions: eight emotions, arousal, valence, liking, and familiarity. The dataset is designed to facilitate cross-subject affective computing research and development of EEG-based emotion recognition algorithms for real-world applications.

Overview of the experiment

View full README

DOI

FACED - Finer-grained Affective Computing EEG Dataset

Introduction

The Finer-grained Affective Computing EEG Dataset (FACED) contains scalp EEG recordings from 123 healthy participants who watched 28 emotion-eliciting video clips designed to evoke nine different emotion categories. The dataset includes four negative emotions (anger, fear, disgust, sadness) from Ekman’s basic emotions and four positive emotions (amusement, inspiration, joy, tenderness) selected based on recent psychological and neuroscience progress and application needs. Participants provided detailed self-reported emotion ratings on 12 dimensions: eight emotions, arousal, valence, liking, and familiarity. The dataset is designed to facilitate cross-subject affective computing research and development of EEG-based emotion recognition algorithms for real-world applications.

Overview of the experiment

Participants (123 subjects, 75 female, ages 17-38, mean=23.2 years) were seated 60 cm from a 22-inch LCD monitor in a regular office environment. Each trial consisted of: (1) a 5-second fixation cross, (2) a video clip of varying length (typically 30-60 seconds), and (3) subjective emotional rating on 12 items (anger, fear, disgust, sadness, amusement, inspiration, joy, tenderness, valence, arousal, liking, familiarity) on a continuous 0-7 scale, followed by at least 30 seconds rest. Video clips were presented in blocks: three positive blocks, three negative blocks, and one neutral block, with 20 arithmetic problems between blocks to minimize carryover effects. The 28 video clips were designed to target nine emotion categories, with randomized presentation order across participants. EEG was recorded using a 32-channel biosignal recording system sampled at either 1000 Hz (92 subjects) or 250 Hz (31 subjects), with channels positioned according to the International 10-20 system. Signal units were recorded in either Volts or microVolts depending on the hardware configuration used. Video stimulus information: The dataset includes 28 video clips designed to elicit nine emotion categories (Trigger values 1–28): - Anger (Videos 1-3): Durations 73-81 seconds, negative valence - Disgust (Videos 4-6): Durations 69-91 seconds, negative valence - Fear (Videos 7-9): Durations 56-106 seconds, negative valence - Sadness (Videos 10-12): Durations 45-82 seconds, negative valence - Neutral (Videos 13-16): Durations 35-43 seconds, neutral valence - Amusement (Videos 17-19): Durations 56-73 seconds, positive valence - Inspiration (Videos 20-22): Durations 76-129 seconds, positive valence - Joy (Videos 23-25): Durations 34-68 seconds, positive valence - Tenderness (Videos 26-28): Durations 54-77 seconds, positive valence

Metadata for each video (duration, source film, source database, valence, targeted emotion) is read from Stimuli_info.xlsx. Event markers (from evt.bdf annotations): - 100: Task/block start - 101: Video onset - 102: Video offset - 1–28: Video index (appears just before 101, used to link to stimulus metadata) - 201/202: Block boundary markers - “Start Impedance” / “Stop Impedance”: Technical markers (ignored)

The conversion script reads evt.bdf annotations for each subject, parses video presentation spans (from video index + 101 to 102), and creates MNE Annotations with the source film title (video_title) as description. These annotations are exported to BIDS events.tsv with extra columns: - emotion_label: targeted emotion category (Anger, Disgust, Fear, Sadness, Neutral, Amusement, Inspiration, Joy, Tenderness) - binary_label: positive/negative/neutral classification - video_index: 1–28 - Self-reported ratings (Joy, Tenderness, Inspiration, Amusement, Anger, Disgust, Fear, Sadness, Arousal, Valence, Familiarity, Liking)

Description of the preprocessing if any

Raw BDF files from the biosignal recording system have been converted to BIDS format. Channel names are standardized to match the International 10-20 nomenclature. Subjects have been assigned numeric IDs (sub-000 through sub-122) corresponding to their original subject designations in the dataset. Recording dates have been set to a default value (2023-01-01) due to privacy considerations, while time relationships between files are preserved. Subject demographic information (age, sex) has been extracted from the Recording_info.csv file and properly formatted for BIDS. Stimulus timing information from the evt.bdf event files has been parsed and enriched with metadata from Stimuli_info.xlsx. Each video presentation is annotated with the targeted emotion category (Anger, Disgust, Fear, Sadness, Neutral, Amusement, Inspiration, Joy, Tenderness) and includes self-reported ratings from After_remarks.mat when available.

Citation

When using this dataset, please cite: 1. Liu, Y., Sourina, O., & Nguyen, M. K. (2023). Finer-grained Affective Computing EEG Dataset. Scientific Data, 10(1), 809. https://doi.org/10.1038/s41597-023-02650-w 2. Synapse Platform: https://www.synapse.org/#!Synapse:syn50614194 3. The dataset is available at the Synapse platform repository.

Data curators: Pierre Guetschel (BIDS conversion) Original data collection team: - Yisi Liu (Nanyang Technological University) - Olga Sourina (Nanyang Technological University)

- Minh Khoa Nguyen (Nanyang Technological University)

Automatic report

Report automatically generated by ``mne_bids.make_report()``.

The FACED - Finer-grained Affective Computing EEG Dataset dataset was created

by Yisi Liu, Olga Sourina, and Minh Khoa Nguyen and conforms to BIDS version 1.7.0. This report was generated with MNE-BIDS (https://doi.org/10.21105/joss.01896). The dataset consists of 123 participants (comprised of 48 male and 75 female participants; handedness were all unknown; ages ranged from 17.0 to 38.0 (mean = 22.94, std = 4.66)) . Data was recorded using an EEG system (Biosemi) sampled at 1000.0, and 250.0 Hz with line noise at n/a Hz. There were 123 scans in total. Recording durations ranged from 3468.0 to 6743.0 seconds (mean = 4544.83, std = 647.24), for a total of 559013.71 seconds of data recorded over all scans. For each dataset, there were on average 32.0 (std = 0.0) recording channels per scan, out of which 32.0 (std = 0.0) were used in analysis (0.0 +/- 0.0 were removed from analysis).

Dataset Information#

Dataset ID

NM000112

Title

FACED - Finer-grained Affective Computing EEG Dataset

Author (year)

Liu2024_112

Canonical

FACED

Importable as

NM000112, Liu2024_112, FACED

Year

2023

Authors

Yisi Liu, Olga Sourina, Minh Khoa Nguyen

License

CC-BY-4.0

Citation / DOI

10.82901/nemar.nm000112

Source links

OpenNeuro | NeMAR | Source URL

Copy-paste BibTeX
@dataset{nm000112,
  title = {FACED - Finer-grained Affective Computing EEG Dataset},
  author = {Yisi Liu and Olga Sourina and Minh Khoa Nguyen},
  doi = {10.82901/nemar.nm000112},
  url = {https://doi.org/10.82901/nemar.nm000112},
}

Found an issue with this dataset?

If you encounter any problems with this dataset (missing files, incorrect metadata, loading errors, etc.), please let us know!

Report an Issue on GitHub

Technical Details#

Subjects & recordings
  • Subjects: 123

  • Recordings: 123

  • Tasks: 1

Channels & sampling rate
  • Channels: 32

  • Sampling rate (Hz): 1000 (68), 250 (55)

  • Duration (hours): 155.28158666666664

Tags
  • Pathology: Not specified

  • Modality: —

  • Type: —

Files & format
  • Size on disk: 31.4 GB

  • File count: 123

  • Format: BIDS

License & citation
  • License: CC-BY-4.0

  • DOI: 10.82901/nemar.nm000112

Provenance

API Reference#

Use the NM000112 class to access this dataset programmatically.

class eegdash.dataset.NM000112(cache_dir: str, query: dict | None = None, s3_bucket: str | None = None, **kwargs)[source]#

Bases: EEGDashDataset

FACED - Finer-grained Affective Computing EEG Dataset

Study:

nm000112 (NeMAR)

Author (year):

Liu2024_112

Canonical:

FACED

Also importable as: NM000112, Liu2024_112, FACED.

Modality: eeg. Subjects: 123; recordings: 123; tasks: 1.

Parameters:
  • cache_dir (str | Path) – Directory where data are cached locally.

  • query (dict | None) – Additional MongoDB-style filters to AND with the dataset selection. Must not contain the key dataset.

  • s3_bucket (str | None) – Base S3 bucket used to locate the data.

  • **kwargs (dict) – Additional keyword arguments forwarded to EEGDashDataset.

data_dir#

Local dataset cache directory (cache_dir / dataset_id).

Type:

Path

query#

Merged query with the dataset filter applied.

Type:

dict

records#

Metadata records used to build the dataset, if pre-fetched.

Type:

list[dict] | None

Notes

Each item is a recording; recording-level metadata are available via dataset.description. query supports MongoDB-style filters on fields in ALLOWED_QUERY_FIELDS and is combined with the dataset filter. Dataset-specific caveats are not provided in the summary metadata.

References

OpenNeuro dataset: https://openneuro.org/datasets/nm000112 NeMAR dataset: https://nemar.org/dataexplorer/detail?dataset_id=nm000112 DOI: https://doi.org/10.82901/nemar.nm000112

Examples

>>> from eegdash.dataset import NM000112
>>> dataset = NM000112(cache_dir="./data")
>>> recording = dataset[0]
>>> raw = recording.load()
__init__(cache_dir: str, query: dict | None = None, s3_bucket: str | None = None, **kwargs)[source]#
save(path, overwrite=False)[source]#

Save the dataset to disk.

Parameters:
  • path (str or Path) – Destination file path.

  • overwrite (bool, default False) – If True, overwrite existing file.

Return type:

None

See Also#