{"success":true,"database":"eegdash","data":{"_id":"6953f4249276ef1ee07a3473","dataset_id":"ds006910","associated_paper_doi":null,"authors":["Ryuzaburo Kochi","Aya Kanno","Hiroshi Uda","Keisuke Hatano","Masaki Sonoda","Hidenori Endo","Michael Cools","Robert Rothermel","Aimee F. Luat","Eishi Asano"],"bids_version":"1.7.0","contact_info":["Ryuzaburo Kochi"],"contributing_labs":null,"data_processed":false,"dataset_doi":"doi:10.18112/openneuro.ds006910.v1.0.1","datatypes":["ieeg"],"demographics":{"subjects_count":121,"ages":[16,17,15,8,14,14,16,17,10,8,8,11,16,18,17,8,17,14,10,10,15,19,6,14,11,13,23,10,5,16,16,37,14,5,11,21,17,15,44,37,14,28,20,14,13,41,12,8,10,12,9,28,27,17,15,6,12,5,9,30,21,13,11,17,16,8,13,14,12,13,11,15,14,11,12,17,11,17,10,13,11,16,15,15,10,10,16,12,8,14,12,19,6,16,8,16,19,15,14,5,16,13,16,9,13,11,13,20,49,12,15,8,16,17,13,7,17,19,15,18,8],"age_min":5,"age_max":49,"age_mean":14.826446280991735,"species":null,"sex_distribution":{"m":62,"f":59},"handedness_distribution":{"r":112,"l":8,"a":1}},"experimental_modalities":null,"external_links":{"source_url":"https://openneuro.org/datasets/ds006910","osf_url":null,"github_url":null,"paper_url":null},"funding":["N/A"],"ingestion_fingerprint":"6eb9edcc7c3e0d207d5d2dbd3a3f5653f09829540075971fbfcb8ec77ec8e215","license":"CC0","n_contributing_labs":null,"name":"Auditory Naming EC","readme":"This dataset, used in the analysis reported by Kochi et al., (2025), contains intracranial EEG recordings from 121 individuals who performed an auditory‑naming task. Electrode coordinates are provided in MNI‑305 space.\nEach EDF file is tagged for the auditory naming task with the following event codes:\n401 – stimulus onset\n402 – stimulus offset\n501 – response onset\nReference:\nRyuzaburo Kochi, Aya Kanno, Hiroshi Uda, Keisuke Hatano,  Masaki Sonoda, Hidenori Endo, Michael Cools, Robert Rothermel, Aimee F. Luat, Eishi Asano. Whole-Brain Millisecond-Scale Effective Connectivity Atlases of Speech","recording_modality":["ieeg"],"senior_author":"Eishi Asano","sessions":["1","2","3","4","5","6"],"size_bytes":47891802796,"source":"openneuro","study_design":null,"study_domain":null,"tasks":["auditory"],"timestamps":{"digested_at":"2026-04-22T12:29:46.443869+00:00","dataset_created_at":"2025-11-08T18:07:51.580Z","dataset_modified_at":"2025-11-09T22:31:33.000Z"},"total_files":384,"storage":{"backend":"s3","base":"s3://openneuro.org/ds006910","raw_key":"dataset_description.json","dep_keys":["CHANGES","README","participants.json","participants.tsv"]},"tagger_meta":{"config_hash":"3557b68bca409f28","metadata_hash":"154abde2d00ab710","model":"openai/gpt-5.2","tagged_at":"2026-04-07T09:32:40.872789+00:00"},"tags":{"pathology":["Unknown"],"modality":["Auditory"],"type":["Other"],"confidence":{"pathology":0.6,"modality":0.85,"type":0.7},"reasoning":{"few_shot_analysis":"Closest few-shot matches by paradigm/stimulus channel are the auditory-task examples. In particular: (1) the digit-span dataset (Healthy / Auditory / Memory) shows that when stimuli are spoken sounds/digits presented to participants, the Modality is labeled “Auditory” (stimulus-channel based). (2) the music-vs-speech ABR dataset (Healthy / Auditory / Perception) further reinforces that speech/music listening paradigms map to “Auditory” modality even when the scientific goal differs. These examples guide Modality assignment strongly; they do not provide a direct convention for a ‘naming’/speech-production construct under the provided Type label set, so Type requires mapping to the nearest allowed category (likely “Other”). For Pathology, few-shots show explicit diagnoses (e.g., Parkinson’s, Epilepsy) are used when stated; here no diagnosis is explicitly stated, so we should not infer Epilepsy solely from iEEG.","metadata_analysis":"Key quoted metadata facts:\n- Population/recording type: “contains intracranial EEG recordings from 121 individuals” and “Electrode coordinates are provided in MNI‑305 space.”\n- Task/stimulus channel: participants “performed an auditory‑naming task.”\n- Event structure consistent with stimulus presentation and speech response: “401 – stimulus onset”, “402 – stimulus offset”, “501 – response onset”.\n- Task field: tasks = [\"auditory\"].\nNotably absent: any explicit recruitment diagnosis/condition such as “epilepsy”, “seizure”, “patients”, “surgery candidates”, etc.","paper_abstract_analysis":"No paper abstract text was provided in the input (only a reference citation/title: “Whole-Brain Millisecond-Scale Effective Connectivity Atlases of Speech”), so no additional disambiguating population/task details are available.","evidence_alignment_check":"Pathology:\n- Metadata says: “intracranial EEG recordings from 121 individuals” (no diagnosis stated).\n- Few-shot pattern suggests: when a clinical condition is explicitly named (e.g., “pediatric patients with epilepsy”), use that Pathology label; otherwise do not assume.\n- Alignment: ALIGN (both support not assigning a specific pathology without explicit statement). Therefore Pathology should be Unknown rather than inferred Epilepsy.\n\nModality:\n- Metadata says: “auditory‑naming task” and task label “auditory”.\n- Few-shot pattern suggests: auditory stimuli/listening tasks map to Modality = Auditory (e.g., digit span; music/speech ABR).\n- Alignment: ALIGN. Choose Auditory.\n\nType:\n- Metadata says: “auditory‑naming task” and reference title includes “Atlases of Speech.” This is primarily speech/language processing/production, which is not an explicit allowed Type label.\n- Few-shot pattern suggests: Type should reflect the cognitive construct; when construct doesn’t fit listed categories, use “Other.”\n- Alignment: ALIGN. Choose Other (language/speech naming not cleanly Perception/Memory/Motor under the given label set).","decision_summary":"Top-2 candidates and final selections:\n\nPathology:\n1) Unknown — Evidence: no explicit diagnosis/recruitment condition is stated; only “intracranial EEG recordings from 121 individuals.”\n2) Epilepsy — Weak inference only (iEEG often collected in epilepsy), but this is not explicitly stated anywhere in the provided metadata.\nHead-to-head: Unknown wins because explicit recruitment pathology is absent (and we should not infer a diagnosis).\nConfidence basis: contextual inference only (iEEG implies clinical), but no direct diagnostic quote.\n\nModality:\n1) Auditory — Evidence quotes: “auditory‑naming task”; task list includes “auditory”; events include “stimulus onset/offset” consistent with presented auditory stimuli.\n2) Multisensory — Possible if there were visual prompts, but none are mentioned.\nHead-to-head: Auditory wins with multiple explicit task/stimulus mentions.\nConfidence basis: 2–3 explicit metadata indicators of auditory stimulation.\n\nType:\n1) Other — Evidence: task is naming/speech (“auditory‑naming task”; “Atlases of Speech”), which is not an allowed Type option.\n2) Motor — Naming includes speech production (a motor act), but the paradigm emphasis appears to be speech/language connectivity rather than limb motor control, and the stimulus framing is auditory naming.\nHead-to-head: Other wins because ‘language/speech naming’ does not map cleanly onto the provided Type categories.\nConfidence basis: 1 explicit task quote + contextual mapping to nearest allowed label."}},"computed_title":"Auditory Naming EC","nchans_counts":[{"val":128,"count":269},{"val":138,"count":14},{"val":136,"count":11},{"val":112,"count":9},{"val":140,"count":8},{"val":164,"count":8},{"val":134,"count":7},{"val":110,"count":6},{"val":142,"count":5},{"val":156,"count":5},{"val":150,"count":5},{"val":130,"count":4},{"val":144,"count":4},{"val":132,"count":4},{"val":148,"count":4},{"val":160,"count":3},{"val":152,"count":3},{"val":154,"count":3},{"val":96,"count":3},{"val":118,"count":3},{"val":84,"count":3},{"val":64,"count":2},{"val":58,"count":1}],"sfreq_counts":[{"val":1000.0,"count":384}],"stats_computed_at":"2026-04-22T23:16:00.312135+00:00","total_duration_s":null,"canonical_name":null,"name_confidence":0.74,"name_meta":{"suggested_at":"2026-04-14T10:18:35.343Z","model":"openai/gpt-5.2 + openai/gpt-5.4-mini + deterministic_fallback"},"name_source":"author_year","author_year":"Kochi2025_Auditory_Naming_EC"}}