{"success":true,"database":"eegdash","data":{"_id":"69d16e05897a7725c66f4cd7","dataset_id":"nm000323","associated_paper_doi":null,"authors":["Min-Ho Lee","O-Yeon Kwon","Yong-Jeong Kim","Hong-Kyung Kim","Young-Eun Lee","John Williamson","Siamac Fazli","Seong-Whan Lee"],"bids_version":"1.9.0","contact_info":null,"contributing_labs":null,"data_processed":false,"dataset_doi":"doi:10.1093/gigascience/giz002","datatypes":["eeg"],"demographics":{"subjects_count":54,"ages":[29,29,29,29,29,29,29,29,29,29,29,29,29,29,29,29,29,29,29,29,29,29,29,29,29,29,29,29,29,29,29,29,29,29,29,29,29,29,29,29,29,29,29,29,29,29,29,29,29,29,29,29,29,29],"age_min":29,"age_max":29,"age_mean":29.0,"species":null,"sex_distribution":null,"handedness_distribution":{"r":54}},"experimental_modalities":null,"external_links":{"source_url":"https://openneuro.org/datasets/nm000323","osf_url":null,"github_url":null,"paper_url":null},"funding":[],"ingestion_fingerprint":"e60a4bf65e7d0cda9abdd8b7a5a8c8e49ac7b3f5c2aa997dba6e761d75bbc917","license":"GPL-3.0","n_contributing_labs":null,"name":"Lee et al. 2019 (ERP) — EEG dataset and OpenBMI toolbox for three BCI paradigms: an investigation into BCI illiteracy","readme":"Lee2019-ERP\n===========\nBMI/OpenBMI dataset for P300.\nDataset Overview\n----------------\n  Code: Lee2019-ERP\n  Paradigm: p300\n  DOI: 10.5524/100542\n  Subjects: 54\n  Sessions per subject: 2\n  Events: Target=1, NonTarget=2\n  Trial interval: [0.0, 1.0] s\n  Runs per session: 2\n  File format: MAT\nAcquisition\n-----------\n  Sampling rate: 1000.0 Hz\n  Number of channels: 62\n  Channel types: eeg=62, emg=4\n  Channel names: AF3, AF4, AF7, AF8, C1, C2, C3, C4, C5, C6, CP1, CP2, CP3, CP4, CP5, CP6, CPz, Cz, EMG1, EMG2, EMG3, EMG4, F10, F3, F4, F7, F8, F9, FC1, FC2, FC3, FC4, FC5, FC6, FT10, FT9, FTT10h, FTT9h, Fp1, Fp2, Fz, O1, O2, Oz, P1, P2, P3, P4, P7, P8, PO10, PO3, PO4, PO9, POz, Pz, T7, T8, TP10, TP7, TP8, TP9, TPP10h, TPP8h, TPP9h, TTP7h\n  Montage: standard_1005\n  Hardware: BrainAmp\n  Software: OpenBMI\n  Reference: nasion\n  Ground: AFz\n  Sensor type: Ag/AgCl\n  Line frequency: 60.0 Hz\n  Impedance threshold: 10 kOhm\n  Cap manufacturer: Brain Products\n  Auxiliary channels: EMG (4 ch)\nParticipants\n------------\n  Number of subjects: 54\n  Health status: healthy\n  Age: mean=29.5, min=24, max=35\n  Gender distribution: female=25, male=29\n  Handedness: right\n  BCI experience: mixed\n  Species: human\nExperimental Protocol\n---------------------\n  Paradigm: p300\n  Task type: copy_spelling\n  Number of classes: 2\n  Class labels: Target, NonTarget\n  Study design: 36-symbol ERP row-column speller with random-set presentation and face stimuli, offline training and online test phases\n  Feedback type: visual\n  Stimulus type: rc_speller\n  Stimulus modalities: visual\n  Primary modality: visual\n  Mode: offline\n  Training/test split: True\n  Instructions: Subjects were asked to copy-spell given sentences by gazing at target characters on screen. In training: 'NEURAL NETWORKS AND DEEP LEARNING' (33 characters), in test: 'PATTERN RECOGNITION MACHINE LEARNING' (36 characters). Participants counted number of times each target character flashed.\nHED Event Annotations\n---------------------\n  Schema: HED 8.4.0 | Browse: https://www.hedtags.org/hed-schema-browser\n  Target\n    ├─ Sensory-event\n    ├─ Experimental-stimulus\n    ├─ Visual-presentation\n    └─ Target\n  NonTarget\n    ├─ Sensory-event\n    ├─ Experimental-stimulus\n    ├─ Visual-presentation\n    └─ Non-target\nParadigm-Specific Parameters\n----------------------------\n  Detected paradigm: p300\n  Number of targets: 36\n  Number of repetitions: 5\n  Inter-stimulus interval: 135.0 ms\n  Stimulus onset asynchrony: 215.0 ms\nData Structure\n--------------\n  Trials: {'training': 1980, 'test': 2160}\n  Trials context: Training: copy-spell 'NEURAL NETWORKS AND DEEP LEARNING' (33 characters). Test: copy-spell 'PATTERN RECOGNITION MACHINE LEARNING' (36 characters). Each character received 5 sequences of 12 flashes (60 flashes total).\nPreprocessing\n-------------\n  Data state: raw\n  Preprocessing applied: False\nSignal Processing\n-----------------\n  Classifiers: LDA\n  Feature extraction: Mean Amplitudes\nCross-Validation\n----------------\n  Method: training-test split\n  Evaluation type: within_session, cross_session\nPerformance (Original Study)\n----------------------------\n  Accuracy: 96.7%\n  Accuracy Std: 0.05\n  Illiteracy Rate: 11.1\nBCI Application\n---------------\n  Applications: speller, communication\n  Online feedback: True\nTags\n----\n  Pathology: Healthy\n  Modality: Visual\n  Type: Perception\nDocumentation\n-------------\n  Description: EEG dataset and OpenBMI toolbox for three BCI paradigms: an investigation into BCI illiteracy\n  DOI: 10.1093/gigascience/giz002\n  License: GPL-3.0\n  Investigators: Min-Ho Lee, O-Yeon Kwon, Yong-Jeong Kim, Hong-Kyung Kim, Young-Eun Lee, John Williamson, Siamac Fazli, Seong-Whan Lee\n  Senior author: Seong-Whan Lee\n  Contact: sw.lee@korea.ac.kr; Tel: +82-2-3290-3197; Fax: +82-2-3290-3583\n  Institution: Korea University\n  Department: Department of Brain and Cognitive Engineering\n  Address: 145 Anam-ro, Seongbuk-gu, Seoul, 02841, Korea\n  Country: KR\n  Repository: GigaDB\n  Publication year: 2019\n  Keywords: EEG datasets, brain-computer interface, event-related potential, steady-state visually evoked potential, motor-imagery, OpenBMI toolbox, BCI illiteracy\nReferences\n----------\nLee, M. H., Kwon, O. Y., Kim, Y. J., Kim, H. K., Lee, Y. E., Williamson, J., … Lee, S. W. (2019). EEG dataset and OpenBMI toolbox for three BCI paradigms: An investigation into BCI illiteracy. GigaScience, 8(5), 1–16. https://doi.org/10.1093/gigascience/giz002\nAppelhoff, S., Sanderson, M., Brooks, T., Vliet, M., Quentin, R., Holdgraf, C., Chaumon, M., Mikulan, E., Tavabi, K., Hochenberger, R., Welke, D., Brunner, C., Rockhill, A., Larson, E., Gramfort, A. and Jas, M. (2019). MNE-BIDS: Organizing electrophysiological data into the BIDS format and facilitating their analysis. Journal of Open Source Software 4: (1896). https://doi.org/10.21105/joss.01896\nPernet, C. R., Appelhoff, S., Gorgolewski, K. J., Flandin, G., Phillips, C., Delorme, A., Oostenveld, R. (2019). EEG-BIDS, an extension to the brain imaging data structure for electroencephalography. Scientific Data, 6, 103. https://doi.org/10.1038/s41597-019-0104-8\n---\nGenerated by MOABB 1.5.0 (Mother of All BCI Benchmarks)\nhttps://github.com/NeuroTechX/moabb","recording_modality":["eeg"],"senior_author":null,"sessions":["1","2"],"size_bytes":41482053813,"source":"openneuro","storage":{"backend":"s3","base":"s3://openneuro.org/nm000323","raw_key":"dataset_description.json","dep_keys":["README","participants.json","participants.tsv"]},"study_design":null,"study_domain":null,"tasks":["p300"],"timestamps":{"digested_at":"2026-04-22T12:52:25.205962+00:00","dataset_created_at":null,"dataset_modified_at":null},"total_files":216,"computed_title":"Lee et al. 2019 (ERP) — EEG dataset and OpenBMI toolbox for three BCI paradigms: an investigation into BCI illiteracy","nchans_counts":[{"val":66,"count":216}],"sfreq_counts":[{"val":1000.0,"count":216}],"stats_computed_at":"2026-04-22T23:16:00.314543+00:00","total_duration_s":209248.784,"tagger_meta":{"config_hash":"3557b68bca409f28","metadata_hash":"2cdac3e1b5b10e62","model":"openai/gpt-5.2","tagged_at":"2026-04-07T09:32:40.872789+00:00"},"tags":{"pathology":["Healthy"],"modality":["Visual"],"type":["Attention"],"confidence":{"pathology":0.9,"modality":0.9,"type":0.8},"reasoning":{"few_shot_analysis":"Most similar few-shot conventions are the oddball/target-detection style datasets. Example: the “Cross-modal Oddball Task” (Parkinson’s) uses standard vs oddball cues and maps that paradigm to an attentional target-detection style task (even though its Type is set to Clinical/Intervention due to clinical focus). For selecting Type within a non-clinical cohort, the “EEG: DPX Cog Ctl Task in Acute Mild TBI” example shows that tasks emphasizing cue/target processing and cognitive control are labeled as Attention rather than Perception. In contrast, the “Meta-rdk: Preprocessed EEG data” example is a visual discrimination task (motion direction) and is labeled Perception—this helps distinguish discrimination from target-detection/oddball paradigms like P300 spellers.","metadata_analysis":"Key metadata indicates a healthy BCI P300 speller with visual stimuli: (1) Population: \"Health status: healthy\" and also \"Tags\\n----\\n  Pathology: Healthy\". (2) Paradigm/task: \"Paradigm: p300\" and \"Task type: copy_spelling\" with \"36-symbol ERP row-column speller\". (3) Visual modality: \"Stimulus modalities: visual\" and \"Primary modality: visual\"; HED annotations also include \"Visual-presentation\" under both Target and NonTarget events. (4) Target detection structure: \"Events: Target=1, NonTarget=2\" and instructions: \"Participants counted number of times each target character flashed.\"","paper_abstract_analysis":"No useful paper information.","evidence_alignment_check":"Pathology: Metadata says healthy (\"Health status: healthy\"; \"Pathology: Healthy\"). Few-shot pattern suggests using explicit recruited diagnosis when present; here it aligns with Healthy. Modality: Metadata explicitly says visual (\"Stimulus modalities: visual\"; \"Primary modality: visual\"; HED \"Visual-presentation\"); few-shot convention aligns (stimulus channel determines Modality). Type: Metadata describes a P300 row-column speller with Target vs NonTarget flashes and counting targets (\"Paradigm: p300\"; \"Events: Target... NonTarget\"; \"counted number of times each target character flashed\"). Few-shot conventions: discrimination tasks map to Perception (Meta-rdk), while target/cue-driven paradigms map to Attention (DPX) and oddball-style target detection is consistent with Attention. Alignment: favors Attention over Perception; no conflict with explicit metadata facts.","decision_summary":"Top-2 candidates — Pathology: (1) Healthy (quotes: \"Health status: healthy\"; \"Pathology: Healthy\"; \"Subjects: 54\" with no disorder mentioned) vs (2) Unknown (would apply if no population info). Select Healthy. Modality: (1) Visual (quotes: \"Stimulus modalities: visual\"; \"Primary modality: visual\"; HED includes \"Visual-presentation\") vs (2) Multisensory/Other (no evidence of non-visual stimuli). Select Visual. Type: (1) Attention (evidence: P300 target detection \"Events: Target=1, NonTarget=2\"; instruction to attend/count targets \"counted number of times each target character flashed\"; few-shot convention from DPX/oddball-like target paradigms) vs (2) Perception (because it is visual stimulus processing, and dataset tag suggests Perception). Head-to-head: P300 speller primarily indexes attentional target selection/oddball processing rather than sensory discrimination, so Attention is stronger."}},"canonical_name":null,"name_confidence":0.72,"name_meta":{"suggested_at":"2026-04-14T10:18:35.344Z","model":"openai/gpt-5.2 + openai/gpt-5.4-mini + deterministic_fallback"},"name_source":"canonical","author_year":"Lee2019_ERP"}}