{"success":true,"database":"eegdash","data":{"_id":"6953f4249276ef1ee07a344a","dataset_id":"ds006394","associated_paper_doi":null,"authors":["En-Lin Leong","Yun Da Chua","Takashi Obana","Christopher L. Asplund"],"bids_version":"1.10.0","contact_info":["En-Lin Leong"],"contributing_labs":null,"data_processed":false,"dataset_doi":"doi:10.18112/openneuro.ds006394.v1.0.3","datatypes":["eeg"],"demographics":{"subjects_count":33,"ages":[],"age_min":null,"age_max":null,"age_mean":null,"species":null,"sex_distribution":null,"handedness_distribution":null},"experimental_modalities":null,"external_links":{"source_url":"https://openneuro.org/datasets/ds006394","osf_url":null,"github_url":null,"paper_url":null},"funding":[],"ingestion_fingerprint":"bbc1aaa9007ce8524d1d396e837d7419658d8cf73f4de5d6077487c825d8938b","license":"CC0","n_contributing_labs":null,"name":"Electrophysiological markers of surprise-induced failures of visual and auditory awareness","readme":"This is the dataset for Leong et al. (in prep). 33 participants completed both a visual and auditory surprise task in counterbalanced order. Methodological details are contained in the manuscript.\nCertain participants were excluded at various stages of the analyses. Their data and event lists are included up to the stage of processing that their data reached.\nDue to incorrect settings specific to OpenBCI GUI v5.0.1, indicated EEG values are 24 times larger than what they should be. The units (also specified in the channels.tsv files) are thus in microvolts / 24.","recording_modality":["eeg"],"senior_author":"Christopher L. Asplund","sessions":[],"size_bytes":560818153,"source":"openneuro","study_design":null,"study_domain":null,"tasks":["SiB","SiD"],"timestamps":{"digested_at":"2026-04-22T12:29:26.649032+00:00","dataset_created_at":"2025-06-26T10:42:24.675Z","dataset_modified_at":"2025-06-27T04:44:21.000Z"},"total_files":60,"storage":{"backend":"s3","base":"s3://openneuro.org/ds006394","raw_key":"dataset_description.json","dep_keys":["CHANGES","README.txt"]},"tagger_meta":{"config_hash":"4a051be509a0e3d0","metadata_hash":"13db283ece7cf0e7","model":"openai/gpt-5.2","tagged_at":"2026-01-20T18:57:30.354805+00:00"},"tags":{"pathology":["Healthy"],"modality":["Multisensory"],"type":["Attention"],"confidence":{"pathology":0.6,"modality":0.7,"type":0.55},"reasoning":{"few_shot_analysis":"Most similar few-shot reference is the \"Cross-modal Oddball Task\" example: it combines auditory+visual cues and uses an oddball-like (surprise/deviant) structure, and it is labeled with Modality=Multisensory. This guides mapping a design that explicitly includes both visual and auditory stimulus tasks into the Multisensory modality label. For Type, oddball/surprise paradigms in the examples are used to study cognitive control/attentional orienting (e.g., cue-evoked rhythms), which supports an Attention-type interpretation when no other primary construct (e.g., reinforcement learning, working memory) is stated.","metadata_analysis":"Key available metadata is only the README. It states: (1) \"33 participants completed both a visual and auditory surprise task in counterbalanced order.\" This directly supports Multisensory stimulus modality (visual + auditory). (2) \"This is the dataset for Leong et al. (in prep).\" and \"Methodological details are contained in the manuscript.\" No clinical recruitment criteria or diagnosis is mentioned, so pathology cannot be directly determined from metadata. Also: \"Certain participants were excluded at various stages of the analyses\" and the OpenBCI scaling note do not add labeling-relevant facts about population/task type beyond confirming this is an experimental task dataset.","paper_abstract_analysis":"No useful paper information.","evidence_alignment_check":"Pathology: Metadata SAYS nothing about diagnosis or clinical recruitment (e.g., only \"33 participants\"; no patient groups). Few-shot pattern SUGGESTS that when no disorder is stated and participants are just 'participants', label as Healthy (as in several healthy-cohort examples). ALIGN (no conflict; inference needed due to missing explicit statement).\nModality: Metadata SAYS \"both a visual and auditory surprise task\". Few-shot pattern SUGGESTS that combined visual+auditory paradigms map to Multisensory (e.g., Cross-modal Oddball Task). ALIGN.\nType: Metadata SAYS \"surprise task\" but gives no explicit construct label (e.g., attention, learning, perception). Few-shot pattern SUGGESTS that surprise/oddball-style tasks are commonly categorized under Attention (orienting/cognitive control) unless explicitly framed as learning or decision-making. PARTIAL ALIGN but weak: the mapping relies on convention rather than explicit metadata statements.","decision_summary":"Pathology top-2: (1) Healthy — evidence: no clinical population described (\"33 participants\" only; no diagnosis terms anywhere). (2) Unknown — evidence: participant health status not explicitly stated. Winner: Healthy because most OpenNeuro experimental task datasets without clinical descriptors are healthy cohorts by convention; however explicit confirmation is absent. \nModality top-2: (1) Multisensory — evidence: explicit \"visual and auditory\" tasks; aligns with few-shot cross-modal example. (2) Visual — evidence: one of the tasks is visual, but metadata clearly includes auditory too. Winner: Multisensory.\nType top-2: (1) Attention — evidence: \"surprise task\" commonly operationalizes attentional orienting/oddball responses; supported by oddball/surprise conventions in few-shot examples. (2) Learning — evidence: 'surprise' can relate to prediction error/statistical learning, but no learning/prediction language is present in metadata. Winner: Attention, with low confidence due to lack of task detail beyond the word 'surprise'."}},"computed_title":"Electrophysiological markers of surprise-induced failures of visual and auditory awareness","nchans_counts":[{"val":16,"count":60}],"sfreq_counts":[{"val":125.0,"count":60}],"stats_computed_at":"2026-04-22T23:16:00.311552+00:00","total_duration_s":null,"author_year":"Leong2025","canonical_name":null}}