{"success":true,"database":"eegdash","data":{"_id":"69d16e05897a7725c66f4cb2","dataset_id":"nm000222","associated_paper_doi":null,"authors":["Jongmin Lee","Minju Kim","Dojin Heo","Jongsu Kim","Min-Ki Kim","Taejun Lee","Jongwoo Park","HyunYoung Kim","Minho Hwang","Laehyun Kim","Sung-Phil Kim"],"bids_version":"1.9.0","contact_info":null,"contributing_labs":null,"data_processed":false,"dataset_doi":null,"datatypes":["eeg"],"demographics":{"subjects_count":10,"ages":[22,22,22,22,22,22,22,22,22,22],"age_min":22,"age_max":22,"age_mean":22.0,"species":null,"sex_distribution":null,"handedness_distribution":null},"experimental_modalities":null,"external_links":{"source_url":"https://nemar.org/dataexplorer/detail/nm000222","osf_url":null,"github_url":null,"paper_url":null},"funding":[],"ingestion_fingerprint":"54aabffa84f92e7be8a26d1ac2dedb5a3d871c937c42c178dbc94f14de86a018","license":"CC-BY-4.0","n_contributing_labs":null,"name":"Air conditioner control experiment (10 subjects, 4 classes, 25 EEG ch)","readme":"# Air conditioner control experiment (10 subjects, 4 classes, 25 EEG ch)\nAir conditioner control experiment (10 subjects, 4 classes, 25 EEG ch).\n## Dataset Overview\n- **Code**: Lee2024-AC\n- **Paradigm**: p300\n- **DOI**: 10.3389/fnhum.2024.1320457\n- **Subjects**: 10\n- **Sessions per subject**: 1\n- **Events**: Target=2, NonTarget=1\n- **Trial interval**: [0, 1] s\n- **File format**: MATLAB\n## Acquisition\n- **Sampling rate**: 500.0 Hz\n- **Number of channels**: 25\n- **Channel types**: eeg=25\n- **Channel names**: Fp1, Fpz, Fp2, F7, F3, Fz, F4, F8, FC5, FC1, FC2, FC6, C3, Cz, C4, CP5, CP1, CP2, CP6, P3, Pz, P4, O1, Oz, O2\n- **Montage**: standard_1020\n- **Hardware**: actiCHamp (Brain Products)\n- **Reference**: linked mastoids\n- **Sensor type**: active\n- **Line frequency**: 60.0 Hz\n## Participants\n- **Number of subjects**: 10\n- **Health status**: healthy\n- **Age**: mean=22.4, std=2.59\n- **Gender distribution**: male=6, female=4\n- **Species**: human\n## Experimental Protocol\n- **Paradigm**: p300\n- **Number of classes**: 2\n- **Class labels**: Target, NonTarget\n- **Trial duration**: 1.0 s\n- **Study design**: P300 BCI for AC home appliance control; 4-class oddball; LCD display\n- **Feedback type**: visual\n- **Stimulus type**: flash\n- **Stimulus modalities**: visual\n- **Primary modality**: visual\n- **Mode**: online\n## HED Event Annotations\nSchema: HED 8.4.0 | Browse: https://www.hedtags.org/hed-schema-browser\n```\n  Target\n    ├─ Sensory-event\n    ├─ Experimental-stimulus\n    ├─ Visual-presentation\n    └─ Target\n  NonTarget\n    ├─ Sensory-event\n    ├─ Experimental-stimulus\n    ├─ Visual-presentation\n    └─ Non-target\n```\n## Paradigm-Specific Parameters\n- **Detected paradigm**: p300\n- **Stimulus onset asynchrony**: 750.0 ms\n## Data Structure\n- **Trials**: 50 training + 30 testing blocks per subject\n- **Trials context**: per_subject\n## BCI Application\n- **Applications**: home_appliance_control\n- **Environment**: laboratory\n- **Online feedback**: True\n## Tags\n- **Pathology**: Healthy\n- **Modality**: ERP\n- **Type**: P300\n## Documentation\n- **DOI**: 10.3389/fnhum.2024.1320457\n- **License**: CC-BY-4.0\n- **Investigators**: Jongmin Lee, Minju Kim, Dojin Heo, Jongsu Kim, Min-Ki Kim, Taejun Lee, Jongwoo Park, HyunYoung Kim, Minho Hwang, Laehyun Kim, Sung-Phil Kim\n- **Institution**: Ulsan National Institute of Science and Technology\n- **Country**: KR\n- **Data URL**: https://github.com/jml226/Home-Appliance-Control-Dataset\n- **Publication year**: 2024\n## References\nAppelhoff, S., Sanderson, M., Brooks, T., Vliet, M., Quentin, R., Holdgraf, C., Chaumon, M., Mikulan, E., Tavabi, K., Hochenberger, R., Welke, D., Brunner, C., Rockhill, A., Larson, E., Gramfort, A. and Jas, M. (2019). MNE-BIDS: Organizing electrophysiological data into the BIDS format and facilitating their analysis. Journal of Open Source Software 4: (1896). https://doi.org/10.21105/joss.01896\nPernet, C. R., Appelhoff, S., Gorgolewski, K. J., Flandin, G., Phillips, C., Delorme, A., Oostenveld, R. (2019). EEG-BIDS, an extension to the brain imaging data structure for electroencephalography. Scientific Data, 6, 103. https://doi.org/10.1038/s41597-019-0104-8\n---\nGenerated by MOABB 1.5.0 (Mother of All BCI Benchmarks)\nhttps://github.com/NeuroTechX/moabb","recording_modality":["eeg"],"senior_author":null,"sessions":["0"],"size_bytes":435425386,"source":"nemar","storage":{"backend":"nemar","base":"s3://nemar/nm000222","raw_key":"dataset_description.json","dep_keys":["README.md","participants.json","participants.tsv"]},"study_design":null,"study_domain":null,"tasks":["p300"],"timestamps":{"digested_at":"2026-04-30T14:09:09.249001+00:00","dataset_created_at":null,"dataset_modified_at":"2026-03-25T15:25:19Z"},"total_files":305,"computed_title":"Air conditioner control experiment (10 subjects, 4 classes, 25 EEG ch)","nchans_counts":[{"val":25,"count":305}],"sfreq_counts":[{"val":500.0,"count":305}],"stats_computed_at":"2026-05-01T13:49:34.645858+00:00","total_duration_s":11471.388,"tagger_meta":{"config_hash":"3557b68bca409f28","metadata_hash":"68ecf80760c95138","model":"openai/gpt-5.2","tagged_at":"2026-04-07T09:32:40.872789+00:00"},"tags":{"pathology":["Healthy"],"modality":["Visual"],"type":["Attention"],"confidence":{"pathology":0.8,"modality":0.9,"type":0.85},"reasoning":{"few_shot_analysis":"Most similar few-shot paradigms are the oddball/P300-style tasks: (1) “Cross-modal Oddball Task” (Parkinson’s; labeled Type=Clinical/Intervention, Modality=Multisensory) demonstrates that oddball target vs non-target paradigms map naturally to Attention-type constructs, with modality determined by stimulus channels. (2) “EEG: Three-Stim Auditory Oddball…” (TBI; Modality=Auditory) reinforces that oddball datasets are labeled primarily by stimulus modality (auditory vs visual vs multisensory) and that the recruited clinical group determines Pathology. For the current dataset, the same core structure exists (“Paradigm: p300”, “Target=2, NonTarget=1”) but with healthy participants and visual flashes, so we follow the same convention: Pathology from recruitment (healthy), Modality from stimulus (visual), and Type as attentional target detection (oddball/P300).","metadata_analysis":"Key facts from provided metadata/readme:\n- Population: “Health status: healthy” and “Participants… Number of subjects: 10… Health status: healthy”.\n- Paradigm/task: “Paradigm: p300” and “Events: Target=2, NonTarget=1”.\n- Stimulus modality: “Stimulus type: flash”, “Stimulus modalities: visual”, and “Primary modality: visual”.\n- Context: “Study design: P300 BCI for AC home appliance control; 4-class oddball; LCD display” and “Mode: online”.","paper_abstract_analysis":"No useful paper information.","evidence_alignment_check":"Pathology:\n1) Metadata says: “Health status: healthy”.\n2) Few-shot pattern suggests: Pathology should reflect recruited group (e.g., Parkinson’s, TBI, etc.), otherwise Healthy.\n3) ALIGN.\n\nModality:\n1) Metadata says: “Stimulus modalities: visual” and “Stimulus type: flash”.\n2) Few-shot pattern suggests: modality is defined by the stimulus channel (auditory/visual/multisensory) in oddball/P300 tasks.\n3) ALIGN.\n\nType:\n1) Metadata says: “Paradigm: p300”, “Events: Target=2, NonTarget=1”, and “4-class oddball”.\n2) Few-shot pattern suggests: oddball/P300 target detection paradigms map to attention/target-detection constructs (even when used for BCI), with clinical label used only when pathology is the main cohort focus.\n3) ALIGN (Healthy cohort + P300 oddball implies non-clinical attentional paradigm rather than Clinical/Intervention).","decision_summary":"Top-2 candidates with head-to-head selection:\n\nPathology:\n- Healthy (winner): explicit recruitment/participant status “Health status: healthy”; “Number of subjects: 10… Health status: healthy”.\n- Unknown (runner-up): would apply only if no population info were provided.\nSelection: Healthy.\nConfidence basis: 2 explicit quotes.\n\nModality:\n- Visual (winner): “Stimulus modalities: visual”; “Primary modality: visual”; “Stimulus type: flash”.\n- Multisensory (runner-up): only plausible by analogy to cross-modal oddball, but no auditory/tactile stimulus is described here.\nSelection: Visual.\nConfidence basis: 3 explicit quotes.\n\nType:\n- Attention (winner): P300/oddball target detection implied by “Paradigm: p300” + “Target=2, NonTarget=1” + “4-class oddball”; few-shot oddball examples support labeling around attentional target detection with modality separately captured.\n- Perception (runner-up): could fit generic visual detection, but P300 oddball is more canonically attention/target detection than sensory discrimination.\nSelection: Attention.\nConfidence basis: 3 explicit task/paradigm quotes plus strong few-shot analog (oddball tasks)."}},"canonical_name":null,"name_confidence":0.66,"name_meta":{"suggested_at":"2026-04-14T10:18:35.343Z","model":"openai/gpt-5.2 + openai/gpt-5.4-mini + deterministic_fallback"},"name_source":"author_year","author_year":"Lee2024_Air_conditioner_control"}}