{"success":true,"database":"eegdash","data":{"_id":"6953f4249276ef1ee07a3425","dataset_id":"ds005929","associated_paper_doi":null,"authors":["Yücel, Meryem","Selb, Juliette","Boas, David","Cash, Sydney","Cooper, Robert"],"bids_version":"1.7.1","contact_info":["MERYEM YUCEL"],"contributing_labs":null,"data_processed":false,"dataset_doi":"doi:10.18112/openneuro.ds005929.v1.0.1","datatypes":["fnirs"],"demographics":{"subjects_count":7,"ages":[],"age_min":null,"age_max":null,"age_mean":null,"species":null,"sex_distribution":null,"handedness_distribution":null},"experimental_modalities":null,"external_links":{"source_url":"https://openneuro.org/datasets/ds005929","osf_url":null,"github_url":null,"paper_url":null},"funding":["NIH grants P41-RR14075, R01-EB006385 and R90-DA023427"],"ingestion_fingerprint":"7eb66005505a885bcb02792124c91403204e55a463271d6927f4cbee88531ae6","license":"CC0","n_contributing_labs":null,"name":"Motion-Yucel2014","readme":null,"recording_modality":["fnirs"],"senior_author":"Cooper, Robert","sessions":["1"],"size_bytes":71847616,"source":"openneuro","study_design":null,"study_domain":null,"tasks":["MA"],"timestamps":{"digested_at":"2026-04-22T12:28:59.764136+00:00","dataset_created_at":"2025-02-17T15:30:09.154Z","dataset_modified_at":"2025-07-09T15:20:10.000Z"},"total_files":7,"storage":{"backend":"s3","base":"s3://openneuro.org/ds005929","raw_key":"dataset_description.json","dep_keys":["CHANGES","participants.json","participants.tsv"]},"tagger_meta":{"config_hash":"3557b68bca409f28","metadata_hash":"d2bf17379fcfccde","model":"openai/gpt-5.2","tagged_at":"2026-04-07T09:32:40.872789+00:00"},"tags":{"pathology":["Healthy"],"modality":["Motor"],"type":["Motor"],"confidence":{"pathology":0.6,"modality":0.55,"type":0.55},"reasoning":{"few_shot_analysis":"Closest few-shot by paradigm is the motor-task example “EEG Motor Movement/Imagery Dataset”, where the core manipulation is participant movement/imagery (labeled Modality=Visual due to on-screen cues, Type=Motor). The current dataset title emphasizes “Motion”, which more naturally aligns with a movement-focused recording than with a sensory oddball/memory paradigm. A secondary (weaker) analogy is the schizophrenia visual motion-discrimination example (moving dots; Visual/Perception), but the current dataset metadata does not mention discrimination, dots, or perceptual judgments—only “Motion” and a task code (“MA”).","metadata_analysis":"Key available metadata facts are sparse:\n- Title: “Motion-Yucel2014” (explicitly mentions “Motion”).\n- Participants: “Subjects: 7” (no diagnosis/clinical descriptors provided).\n- Task label: tasks include “MA” (task abbreviation not expanded in provided metadata).\nNo readme/task description/abstract is provided to clarify whether “Motion” refers to visual motion stimuli, participant movement, or motion artifacts.","paper_abstract_analysis":"No useful paper information.","evidence_alignment_check":"Pathology:\n- Metadata says: only “Subjects: 7” with no patient group/diagnosis stated.\n- Few-shot pattern suggests: when no clinical recruitment is stated, label as Healthy.\n- Alignment: ALIGN (no clinical facts to override).\n\nModality:\n- Metadata says: title “Motion-Yucel2014” and task code “MA”, but no stimulus channel specified.\n- Few-shot pattern suggests: if movement execution/imagery is central, modality often maps to Motor unless clear external sensory stimuli are described; if motion is visual stimulus, modality would be Visual.\n- Alignment: PARTIAL/UNCLEAR (metadata under-specified; inference required).\n\nType:\n- Metadata says: only “Motion” and task “MA”; no cognitive construct described.\n- Few-shot pattern suggests: movement-focused paradigms map to Type=Motor; purely perceptual motion discrimination would map to Type=Perception.\n- Alignment: PARTIAL/UNCLEAR (insufficient detail; choose most plausible based on title).","decision_summary":"Top-2 candidates and decision per category:\n\nPathology:\n1) Healthy — Evidence: no diagnosis/clinical recruitment mentioned; only “Subjects: 7”.\n2) Unknown — Would be used if even healthy inference is too speculative.\nDecision: Healthy (standard convention when no pathology is described).\n\nModality:\n1) Motor — Evidence: dataset title “Motion-Yucel2014” + task code “MA” plausibly indicating a movement/motion condition; no explicit visual/auditory stimulus described.\n2) Visual — Alternative if “Motion” refers to visual motion stimuli (as in motion-dot perception tasks), but there is no supporting task/stimulus description.\nDecision: Motor.\n\nType:\n1) Motor — Evidence: title emphasis on “Motion” suggests movement is the primary focus/condition.\n2) Perception — Possible if it is a visual motion-perception task, but unsupported by provided metadata.\nDecision: Motor.\n\nConfidence justification:\n- Pathology confidence is moderate (no explicit ‘healthy’ quote, but no clinical group either).\n- Modality/Type confidence is lower because no stimulus/task description is provided beyond title/task code."}},"computed_title":"Motion-Yucel2014","nchans_counts":[{"val":28,"count":7}],"sfreq_counts":[{"val":50.0,"count":7}],"stats_computed_at":"2026-04-22T23:16:00.311085+00:00","total_duration_s":null,"canonical_name":null,"name_confidence":0.74,"name_meta":{"suggested_at":"2026-04-14T10:18:35.343Z","model":"openai/gpt-5.2 + openai/gpt-5.4-mini + deterministic_fallback"},"name_source":"author_year","author_year":"MotionYucel2014"}}