{"success":true,"database":"eegdash","data":{"_id":"6953f4249276ef1ee07a3458","dataset_id":"ds006547","associated_paper_doi":null,"authors":["Sanaz Ghaffari","Arian Yavari","Sara Bonyadian","Arsalan Ghofrani","Russell Butler"],"bids_version":"1.9.0","contact_info":["Russell Butler"],"contributing_labs":null,"data_processed":false,"dataset_doi":"doi:10.18112/openneuro.ds006547.v1.0.0","datatypes":["eeg"],"demographics":{"subjects_count":31,"ages":[],"age_min":null,"age_max":null,"age_mean":null,"species":null,"sex_distribution":null,"handedness_distribution":null},"experimental_modalities":null,"external_links":{"source_url":"https://openneuro.org/datasets/ds006547","osf_url":null,"github_url":null,"paper_url":null},"funding":[],"ingestion_fingerprint":"23d7e17bfadc2586e3d3b42e26c990872d838c013a51ea5106f0b7a0c922729f","license":"CC0","n_contributing_labs":null,"name":"Visual EEG Study (BrainVision → BIDS)","readme":"This dataset contains high-density EEG recordings collected during a visual stimulation task.\nFiles are organized according to the EEG-BIDS specification. Raw data are BrainVision (.vhdr/.vmrk/.eeg).\nTask: visual\nSession: ses-01\nProvenance: Converted from c:/shared/raw_eeg with this helper script.\nNo acquisition-time filters applied (offline preprocessing not included here).","recording_modality":["eeg"],"senior_author":"Russell Butler","sessions":["01"],"size_bytes":18921502382,"source":"openneuro","study_design":null,"study_domain":null,"tasks":["visual"],"timestamps":{"digested_at":"2026-04-22T12:29:32.571581+00:00","dataset_created_at":"2025-08-12T10:31:23.086Z","dataset_modified_at":"2025-08-12T11:19:42.000Z"},"total_files":31,"storage":{"backend":"s3","base":"s3://openneuro.org/ds006547","raw_key":"dataset_description.json","dep_keys":["CHANGES","README","participants.json","participants.tsv","task-visual_events.json"]},"tagger_meta":{"config_hash":"3557b68bca409f28","metadata_hash":"f8f1b0037104400e","model":"openai/gpt-5.2","tagged_at":"2026-04-07T09:32:40.872789+00:00"},"tags":{"pathology":["Healthy"],"modality":["Visual"],"type":["Perception"],"confidence":{"pathology":0.6,"modality":0.8,"type":0.6},"reasoning":{"few_shot_analysis":"Most similar few-shot by task/modality is the schizophrenia visual discrimination example (\"We used a visual discrimination task\"), which is labeled Modality=Visual and Type=Perception—showing the convention that visually presented stimulation/discrimination paradigms map to Visual + Perception. Although that example has a clinical group, it guides the Visual/Perception mapping for a generic \"visual\" task when no higher-level construct (e.g., memory, learning, motor imagery) is specified. Other few-shots (e.g., ABR/music-speech) reinforce that the stimulus channel (auditory vs visual) drives Modality labeling.","metadata_analysis":"Key metadata indicates a generic visual stimulation experiment with no clinical recruitment described: (1) Readme: \"high-density EEG recordings collected during a visual stimulation task.\" (2) Readme/Task: \"Task: visual\" and tasks list includes \"visual\". (3) Participants overview only states \"Subjects: 31\" with no diagnosis/group fields.","paper_abstract_analysis":"No useful paper information.","evidence_alignment_check":"Pathology: Metadata says nothing about any diagnosis or patient group (\"Subjects: 31\" only). Few-shot pattern suggests that when no clinical population is stated, label as Healthy (normative cohort convention). ALIGN (no conflict; inference required).\nModality: Metadata explicitly says \"visual stimulation task\" and \"Task: visual\". Few-shot pattern maps visual tasks to Modality=Visual. ALIGN.\nType: Metadata describes \"visual stimulation\" but does not specify learning/memory/decision-making/motor/rest/sleep. Few-shot convention labels basic sensory stimulation/discrimination under Type=Perception. ALIGN (inference based on generic sensory-stimulation description).","decision_summary":"Top-2 candidates — Pathology: (1) Healthy: supported by absence of any clinical recruitment info (\"Subjects: 31\" only) and typical OpenNeuro convention when no disorder is mentioned; (2) Unknown: possible because metadata never explicitly says \"healthy\". Winner: Healthy (more useful/standard given no clinical facts stated). Confidence=0.6 because it is contextual inference without explicit \"healthy\" wording.\nTop-2 candidates — Modality: (1) Visual: supported by \"visual stimulation task\" and \"Task: visual\"; (2) Unknown: only if task label were ambiguous (it is not). Winner: Visual. Confidence=0.8 (2 explicit quotes + strong few-shot alignment).\nTop-2 candidates — Type: (1) Perception: supported by generic \"visual stimulation task\" implying sensory processing; (2) Attention/Other: possible if the stimulation was meant to probe attentional effects, but no such aim is stated. Winner: Perception. Confidence=0.6 (reasonable inference; no explicit cognitive construct beyond visual stimulation)."}},"computed_title":"Visual EEG Study (BrainVision → BIDS)","nchans_counts":[{"val":64,"count":31}],"sfreq_counts":[{"val":500.0,"count":31}],"stats_computed_at":"2026-04-22T23:16:00.311773+00:00","total_duration_s":null,"canonical_name":null,"name_confidence":0.72,"name_meta":{"suggested_at":"2026-04-14T10:18:35.343Z","model":"openai/gpt-5.2 + openai/gpt-5.4-mini + deterministic_fallback"},"name_source":"author_year","author_year":"Ghaffari2025"}}