{"success":true,"database":"eegdash","data":{"_id":"69d16e05897a7725c66f4caf","dataset_id":"nm000218","associated_paper_doi":null,"authors":["Boyla Mainsah","Chance Fleeting","Thomas Balmat","Eric Sellers","Leslie Collins"],"bids_version":"1.9.0","contact_info":null,"contributing_labs":null,"data_processed":false,"dataset_doi":null,"datatypes":["eeg"],"demographics":{"subjects_count":16,"ages":[],"age_min":null,"age_max":null,"age_mean":null,"species":null,"sex_distribution":null,"handedness_distribution":null},"experimental_modalities":null,"external_links":{"source_url":"https://nemar.org/dataexplorer/detail/nm000218","osf_url":null,"github_url":null,"paper_url":null},"funding":[],"ingestion_fingerprint":"43e6e44d0a022294768a64bc86aaee4224dcad0723ac88767d6bce7ba4a99b3a","license":"CC-BY-4.0","n_contributing_labs":null,"name":"BigP3BCI Study H — 9x8 checkerboard with gaze conditions (16 healthy subjects)","readme":"# BigP3BCI Study H — 9x8 checkerboard with gaze conditions (16 healthy subjects)\nBigP3BCI Study H — 9x8 checkerboard with gaze conditions (16 healthy subjects).\n## Dataset Overview\n- **Code**: Mainsah2025-H\n- **Paradigm**: p300\n- **DOI**: 10.13026/0byy-ry86\n- **Subjects**: 16\n- **Sessions per subject**: 1\n- **Events**: Target=2, NonTarget=1\n- **Trial interval**: [0, 1.0] s\n## Acquisition\n- **Sampling rate**: 256.0 Hz\n- **Number of channels**: 16\n- **Channel types**: eeg=16\n- **Montage**: standard_1020\n- **Hardware**: g.USBamp (g.tec)\n- **Line frequency**: 60.0 Hz\n## Participants\n- **Number of subjects**: 16\n- **Health status**: healthy\n## Experimental Protocol\n- **Paradigm**: p300\n- **Number of classes**: 2\n- **Class labels**: Target, NonTarget\n## HED Event Annotations\nSchema: HED 8.4.0 | Browse: https://www.hedtags.org/hed-schema-browser\n```\n  Target\n    ├─ Sensory-event\n    ├─ Experimental-stimulus\n    ├─ Visual-presentation\n    └─ Target\n  NonTarget\n    ├─ Sensory-event\n    ├─ Experimental-stimulus\n    ├─ Visual-presentation\n    └─ Non-target\n```\n## Paradigm-Specific Parameters\n- **Detected paradigm**: p300\n## Signal Processing\n- **Feature extraction**: P300_ERP_detection\n## Cross-Validation\n- **Method**: calibration-then-test\n- **Evaluation type**: within_subject\n## BCI Application\n- **Applications**: speller\n- **Environment**: laboratory\n- **Online feedback**: True\n## Tags\n- **Modality**: visual\n- **Type**: perception\n## Documentation\n- **Description**: BigP3BCI: the largest public P300 BCI dataset, containing EEG recordings from ~267 subjects across 20 studies using 6x6 or 9x8 character grids with various stimulus paradigms.\n- **DOI**: 10.13026/0byy-ry86\n- **License**: CC-BY-4.0\n- **Investigators**: Boyla Mainsah, Chance Fleeting, Thomas Balmat, Eric Sellers, Leslie Collins\n- **Institution**: Duke University; East Tennessee State University\n- **Country**: US\n- **Repository**: PhysioNet\n- **Data URL**: https://physionet.org/content/bigp3bci/1.0.0/\n- **Publication year**: 2025\n## References\nAppelhoff, S., Sanderson, M., Brooks, T., Vliet, M., Quentin, R., Holdgraf, C., Chaumon, M., Mikulan, E., Tavabi, K., Hochenberger, R., Welke, D., Brunner, C., Rockhill, A., Larson, E., Gramfort, A. and Jas, M. (2019). MNE-BIDS: Organizing electrophysiological data into the BIDS format and facilitating their analysis. Journal of Open Source Software 4: (1896). https://doi.org/10.21105/joss.01896\nPernet, C. R., Appelhoff, S., Gorgolewski, K. J., Flandin, G., Phillips, C., Delorme, A., Oostenveld, R. (2019). EEG-BIDS, an extension to the brain imaging data structure for electroencephalography. Scientific Data, 6, 103. https://doi.org/10.1038/s41597-019-0104-8\n---\nGenerated by MOABB 1.5.0 (Mother of All BCI Benchmarks)\nhttps://github.com/NeuroTechX/moabb","recording_modality":["eeg"],"senior_author":null,"sessions":["0"],"size_bytes":342380913,"source":"nemar","storage":{"backend":"nemar","base":"s3://nemar/nm000218","raw_key":"dataset_description.json","dep_keys":["README.md","participants.json","participants.tsv"]},"study_design":null,"study_domain":null,"tasks":["p300"],"timestamps":{"digested_at":"2026-04-30T14:09:08.260248+00:00","dataset_created_at":null,"dataset_modified_at":"2026-03-24T06:28:29Z"},"total_files":372,"computed_title":"BigP3BCI Study H — 9x8 checkerboard with gaze conditions (16 healthy subjects)","nchans_counts":[{"val":16,"count":372}],"sfreq_counts":[{"val":256.0,"count":372}],"stats_computed_at":"2026-05-01T13:49:34.645819+00:00","total_duration_s":26741.546875,"tagger_meta":{"config_hash":"3557b68bca409f28","metadata_hash":"1f994168cb7448a0","model":"openai/gpt-5.2","tagged_at":"2026-04-07T09:32:40.872789+00:00"},"tags":{"pathology":["Healthy"],"modality":["Visual"],"type":["Attention"],"confidence":{"pathology":0.9,"modality":0.9,"type":0.7},"reasoning":{"few_shot_analysis":"Most similar few-shot examples by paradigm structure are the oddball-style datasets: (1) “Cross-modal Oddball Task.” (Parkinson’s) uses standard vs oddball cues and is labeled with modality based on stimulus channels (multisensory) and type reflecting the study aim; (2) “EEG: Three-Stim Auditory Oddball and Rest in Acute and Chronic TBI” shows classic target/standard/novel event structure. These examples guide the convention that target vs non-target (oddball/P300-like) paradigms are categorized by the stimulus modality (auditory/visual/multisensory) and the cognitive construct often relates to attention to targets. For modality/type comparison, the “Meta-rdk” example shows a visual stimulus discrimination task labeled Visual + Perception, indicating that when the main requirement is detecting/discriminating sensory stimuli, Perception is a strong candidate label.","metadata_analysis":"Key metadata facts: (1) Population: readme explicitly states “**Health status**: healthy” and the title includes “**(16 healthy subjects)**”. (2) Paradigm/task: “**Paradigm**: p300” and “**Events**: Target=2, NonTarget=1”. (3) Stimulus channel: HED annotations include “**Visual-presentation**” for both Target and NonTarget, and the dataset lists “**Modality**: visual”. (4) Application context: “**BCI Application** — **Applications**: speller” and “**Online feedback**: True”, consistent with a P300 BCI speller (visual oddball target detection).","paper_abstract_analysis":"No useful paper information.","evidence_alignment_check":"Pathology: Metadata says participants are “healthy” (\"Health status: healthy\"; title: \"16 healthy subjects\"). Few-shot convention: when no disorder recruitment is present, label Healthy. ALIGN.\nModality: Metadata says visual stimuli (HED: “Visual-presentation”; tag: “Modality: visual”; checkerboard grid in title). Few-shot convention: modality follows stimulus channel (e.g., oddball example labeled multisensory when visual+auditory cues). ALIGN.\nType: Metadata indicates a P300 target vs non-target paradigm (\"Paradigm: p300\"; \"Target, NonTarget\"; BCI speller with online feedback). Few-shot pattern suggests oddball/P300 tasks commonly index attention to targets; however, a competing convention from the visual discrimination example maps visual stimulus discrimination to Perception. PARTIAL ALIGN/AMBIGUITY (both plausible); no explicit metadata statement that the primary construct is “attention” vs “perception,” so decision is based on paradigm interpretation.","decision_summary":"Top-2 candidates per category:\n\nPathology:\n1) Healthy — supported by “Health status: healthy” and “16 healthy subjects”.\n2) Unknown — only if health status were not explicit (but it is).\nChosen: Healthy. (Alignment: aligns with few-shot Healthy conventions.)\n\nModality:\n1) Visual — supported by HED “Visual-presentation”, tag “Modality: visual”, and checkerboard/speller grid in title.\n2) Multisensory — only if additional auditory/tactile cues existed (not indicated).\nChosen: Visual. (Alignment: aligns with few-shot modality-by-stimulus convention.)\n\nType:\n1) Attention — P300 speller relies on attending to rare targets among non-targets (“Target=2, NonTarget=1”, “Paradigm: p300”, speller with online feedback).\n2) Perception — also plausible because it is stimulus detection/discrimination in the visual domain and metadata even lists “Type: perception”.\nChosen: Attention because the defining cognitive operation in a P300 oddball/speller is selective attention to target flashes rather than general sensory perception alone. Confidence is moderated due to the explicit competing metadata tag (“Type: perception”)."}},"canonical_name":null,"name_confidence":0.66,"name_meta":{"suggested_at":"2026-04-14T10:18:35.343Z","model":"openai/gpt-5.2 + openai/gpt-5.4-mini + deterministic_fallback"},"name_source":"canonical","author_year":"Mainsah2025_BigP3BCI_H"}}