{"success":true,"database":"eegdash","data":{"_id":"6953f4249276ef1ee07a3348","dataset_id":"ds004398","associated_paper_doi":null,"authors":["G. Elliott Wimmer","Yunzhe Liu","Daniel C. McNamee","Raymond J. Dolan"],"bids_version":"","contact_info":["Elliott Wimmer"],"contributing_labs":null,"data_processed":false,"dataset_doi":"doi:10.18112/openneuro.ds004398.v1.0.0","datatypes":["meg"],"demographics":{"subjects_count":1,"ages":[],"age_min":null,"age_max":null,"age_mean":null,"species":null,"sex_distribution":null,"handedness_distribution":null},"experimental_modalities":null,"external_links":{"source_url":"https://openneuro.org/datasets/ds004398","osf_url":null,"github_url":null,"paper_url":null},"funding":["G.E.W. was supported by a fellowship from the Deutsche Forschungsgemeinschaft (DFG) and an MRC Career Development Award (MR/V032429/1)","R.D. is supported by Wellcome Trust Investigator Award 098362/ Z/12/Z.","Y.L. was supported by the Open Research Fund of the State Key Laboratory of Cognitive Neuroscience and Learning.","D.C.M. was supported by a Sir Henry Wellcome Trust Postdoctoral Research Fellowship Q: (110257/Z/15/Z).","The Wellcome Centre for Human Neuroimaging is supported by core funding from the Wellcome Trust (203147/Z/16/Z).","The Max Planck University College London Centre is a joint initiative supported by University College London and the Max Planck Society."],"ingestion_fingerprint":"dbf1c584c13739dc4a5e38d2d65c4b707abef1b193578f60435b2642afe85b24","license":"CC0","n_contributing_labs":null,"name":"planmemreplay","readme":"The MEG files contain a channel with triggers necessary for event marking and timing. Separate event files with onsets are provided in the participant directories for completeness only; the MEG triggers should be used for actual onsets in analysis. The delay between the trigger and the visual onset of an on-screen event sent by the projector is approximately 20 ms, as estimated using a photodiode.\nLocalizer phase triggers: [Info to be added]\nStruct and Rew phase triggers: [Info to be added]\nPost triggers:  [Info to be added]","recording_modality":["meg"],"senior_author":"Raymond J. Dolan","sessions":[],"size_bytes":1373056065,"source":"openneuro","study_design":null,"study_domain":null,"tasks":["loc"],"timestamps":{"digested_at":"2026-04-22T12:26:35.348281+00:00","dataset_created_at":"2023-01-11T14:18:04.348Z","dataset_modified_at":"2023-01-25T04:25:53.000Z"},"total_files":1,"storage":{"backend":"s3","base":"s3://openneuro.org/ds004398","raw_key":"dataset_description.json","dep_keys":["CHANGES","README","task-loc_events.json","task-post_events.json","task-rew_events.json","task-struct_events.json"]},"tagger_meta":{"config_hash":"3557b68bca409f28","metadata_hash":"8922e786cb6715f7","model":"openai/gpt-5.2","tagged_at":"2026-04-07T09:32:40.872789+00:00"},"tags":{"pathology":["Unknown"],"modality":["Visual"],"type":["Unknown"],"confidence":{"pathology":0.4,"modality":0.7,"type":0.4},"reasoning":{"few_shot_analysis":"No close few-shot match on task paradigm: examples cover oddball, motor imagery, digit span, resting-state/sleep, reinforcement learning, etc., while this dataset only lists a generic task name (\"loc\") and provides trigger timing notes. For modality labeling, the few-shot convention is to map explicit stimulus channel statements to Modality (e.g., music/speech -> Auditory; moving dots -> Visual). Here, the metadata explicitly mentions \"visual onset\" and \"projector\" for on-screen events, which follows the same convention as the Visual examples.","metadata_analysis":"Key available facts are sparse but include explicit visual-stimulus timing language: (1) \"The delay between the trigger and the visual onset of an on-screen event sent by the projector is approximately 20 ms\". (2) \"The MEG files contain a channel with triggers necessary for event marking and timing.\" Also: task list contains only \"loc\" (\"tasks\": [\"loc\"]) and participant count is minimal/unspecified clinically (\"Subjects: 1\"). No diagnosis/clinical recruitment terms appear in the provided metadata.","paper_abstract_analysis":"No useful paper information.","evidence_alignment_check":"Pathology: Metadata says only \"Subjects: 1\" with no clinical descriptors; few-shot patterns would label as Healthy when explicitly described as healthy controls/participants, but that explicit fact is absent here. ALIGNMENT: not applicable (insufficient metadata); choose Unknown.\nModality: Metadata says \"visual onset of an on-screen event\" and mentions a \"projector\"; few-shot pattern maps explicit stimulus channel mentions to the corresponding modality label (visual tasks -> Visual). ALIGNMENT: aligns; choose Visual.\nType: Metadata provides only trigger/timing notes and a task label \"loc\" without describing the cognitive construct; few-shot patterns require task description (e.g., working memory, oddball, motor imagery) to map to Type. ALIGNMENT: not applicable (insufficient metadata); choose Unknown.","decision_summary":"Top-2 candidates and decision:\n- Pathology: (1) Unknown—supported by lack of any diagnosis/recruitment description (quote: \"Subjects: 1\"); (2) Healthy—plausible default for non-clinical MEG/EEG studies but not stated. Winner: Unknown (metadata does not explicitly state healthy recruitment). Confidence based on only negative evidence.\n- Modality: (1) Visual—explicitly supported by \"visual onset of an on-screen event\" and \"projector\"; (2) Other/Unknown—possible if stimuli were not sensory, but metadata clearly references on-screen visual events. Winner: Visual.\n- Type: (1) Unknown—no description beyond triggers and task name \"loc\"; (2) Perception—weakly plausible if \"loc\" is a visual localizer, but not explicitly stated. Winner: Unknown due to insufficient task/construct detail.\nConfidence justification: Modality has one direct quote clearly indicating visual stimulation; pathology and type lack explicit statements and rely on absence of information."}},"nemar_citation_count":1,"computed_title":"planmemreplay","nchans_counts":[{"val":305,"count":1}],"sfreq_counts":[{"val":600.0,"count":1}],"stats_computed_at":"2026-04-22T23:16:00.307586+00:00","total_duration_s":null,"canonical_name":null,"name_confidence":0.86,"name_meta":{"suggested_at":"2026-04-14T10:18:35.343Z","model":"openai/gpt-5.2 + openai/gpt-5.4-mini + deterministic_fallback"},"name_source":"author_year","author_year":"Wimmer2023"}}