{"success":true,"database":"eegdash","data":{"_id":"6953f4249276ef1ee07a33b5","dataset_id":"ds005087","associated_paper_doi":null,"authors":["Amanda K Robinson","Tijl Grootswagers","Sophia M Shatek","Marlene Behrmann","Thomas A Carlson"],"bids_version":"1.9.1","contact_info":["Tijl Grootswagers","Amanda K Robinson"],"contributing_labs":null,"data_processed":true,"dataset_doi":"doi:10.18112/openneuro.ds005087.v1.0.1","datatypes":["eeg"],"demographics":{"subjects_count":20,"ages":[22,22,19,59,24,20,23,20,18,25,22,25,22,20,17,23,21,26,20,29],"age_min":17,"age_max":59,"age_mean":23.85,"species":null,"sex_distribution":{"f":15,"m":5},"handedness_distribution":{"r":17,"a":2,"l":1}},"experimental_modalities":null,"external_links":{"source_url":"https://openneuro.org/datasets/ds005087","osf_url":null,"github_url":null,"paper_url":null},"funding":[],"ingestion_fingerprint":"a5d9ed6460cf987a44568c30e6988be1011bdcbe361ed4a12d4ec23dd6dbf703","license":"CC0","n_contributing_labs":null,"name":"rapid-hemifield-object-eeg","readme":"Object and word stimuli presented at 5Hz to the left or right visual fields, or centrally, while participants performed an orthogonal red target detection task\n[PUBLICATION] Robinson A.K., Grootswagers T., Shatek S., Behrmann M., Carlson T.A. (2025). Dynamics of visual object coding within and across the hemispheres: Objects in the periphery. Science Advances, 11, eadq0889, https://doi.org/10.1126/sciadv.adq0889","recording_modality":["eeg"],"senior_author":"Thomas A Carlson","sessions":[],"size_bytes":13146879384,"source":"openneuro","study_design":null,"study_domain":null,"tasks":["restclosed","restopen","rsvp"],"timestamps":{"digested_at":"2026-04-22T12:27:18.091574+00:00","dataset_created_at":"2024-04-15T04:45:59.237Z","dataset_modified_at":"2025-01-23T22:53:56.000Z"},"total_files":60,"storage":{"backend":"s3","base":"s3://openneuro.org/ds005087","raw_key":"dataset_description.json","dep_keys":["CHANGES","README","participants.json","participants.tsv","task-rsvp_eeg.json","task-rsvp_events.json"]},"tagger_meta":{"config_hash":"4a051be509a0e3d0","metadata_hash":"c17258b96d32e8ae","model":"openai/gpt-5.2","tagged_at":"2026-01-20T17:41:03.129378+00:00"},"tags":{"pathology":["Healthy"],"modality":["Visual"],"type":["Perception"],"confidence":{"pathology":0.6,"modality":0.8,"type":0.7},"reasoning":{"few_shot_analysis":"Most similar few-shot conventions are the visually driven perception-style datasets. Example: the schizophrenia-spectrum dataset uses a \"visual discrimination task\" with visually presented dot motion and is labeled Modality=Visual, Type=Perception; this supports mapping visually presented object/word stimuli to Visual+Perception even if an orthogonal detection response is present. Also consistent with the music/speech ABR example being labeled Perception when the focus is stimulus processing rather than decision policy.","metadata_analysis":"Key facts from provided README:\n1) Visual stimulus channel: \"Object and word stimuli presented at 5Hz to the left or right visual fields, or centrally\".\n2) Task context: participants \"performed an orthogonal red target detection task\".\n3) Study focus implied by citation title: \"Dynamics of visual object coding within and across the hemispheres\" (suggests perceptual/representational visual processing emphasis, not clinical intervention).","paper_abstract_analysis":"No useful paper information.","evidence_alignment_check":"Pathology:\n- Metadata says: no diagnosis/clinical recruitment is mentioned; only generic \"participants\".\n- Few-shot pattern suggests: when no disorder is specified and typical cognitive task is described, label as Healthy.\n- Alignment: ALIGN (no conflict).\n\nModality:\n- Metadata says: \"Object and word stimuli... to the left or right visual fields\".\n- Few-shot pattern suggests: visually presented stimuli map to Modality=Visual.\n- Alignment: ALIGN.\n\nType:\n- Metadata says: \"Dynamics of visual object coding\" and visual object/word stimulation; task is \"orthogonal red target detection\".\n- Few-shot pattern suggests: discrimination/detection tasks primarily about processing stimuli are labeled Perception (even with responses), unless decision policy is central.\n- Alignment: Mostly ALIGN; the detection component could suggest Attention, but the stated research focus (visual object coding) aligns more strongly with Perception.","decision_summary":"Pathology top-2:\n1) Healthy — Evidence: no clinical terms/diagnoses given; generic \"participants\" only. Aligns with few-shot convention that non-clinical cognitive/perceptual EEG studies default to Healthy.\n2) Unknown — Competing because recruitment criteria are not explicitly stated.\nFinal: Healthy. Confidence=0.6 because it is inference-by-absence (no explicit healthy/control statement).\n\nModality top-2:\n1) Visual — Evidence: \"Object and word stimuli\" presented to \"visual fields\" and \"centrally\".\n2) Multisensory/Other — Weak alternative because there is also a \"red target detection\" feature, but it is still visual.\nFinal: Visual. Confidence=0.8 (explicit visual stimulus quote + strong few-shot analog for visual tasks).\n\nType top-2:\n1) Perception — Evidence: focus on \"visual object coding\" with \"Object and word stimuli\" in the visual periphery/central field; detection is described as \"orthogonal\" (supporting that it is not the main construct).\n2) Attention — Alternative because there is a \"target detection task\", which can index attentional selection.\nFinal: Perception. Confidence=0.7 (one explicit task description + contextual support from the publication title; moderate separation from Attention)."}},"nemar_citation_count":1,"computed_title":"rapid-hemifield-object-eeg","nchans_counts":[{"val":63,"count":60}],"sfreq_counts":[{"val":1000.0,"count":60}],"stats_computed_at":"2026-04-22T23:16:00.308947+00:00","total_duration_s":null,"author_year":"Robinson2024_rapid","canonical_name":null}}