{"success":true,"database":"eegdash","data":{"_id":"696fdefaac44fa1028dc6319","dataset_id":"ds007028","associated_paper_doi":null,"authors":["Yoshinao Kajikawa","Charles Schroeder"],"bids_version":"1.1.1","contact_info":["Yoshinao Kajikawa"],"contributing_labs":null,"data_processed":true,"dataset_doi":"doi:10.18112/openneuro.ds007028.v1.0.0","datatypes":["eeg"],"demographics":{"subjects_count":3,"ages":[],"age_min":null,"age_max":null,"age_mean":null,"species":null,"sex_distribution":null,"handedness_distribution":null},"experimental_modalities":null,"external_links":{"source_url":"https://openneuro.org/datasets/ds007028","osf_url":null,"github_url":null,"paper_url":null},"funding":[],"ingestion_fingerprint":"d53220f3250df993373cd8c1d09edc38ca747af7c59795ed7dada1820151b749","license":"CC0","n_contributing_labs":null,"name":"Auditory Cortex Macaque Monkey DISC Data","readme":null,"recording_modality":["eeg"],"senior_author":"Charles Schroeder","sessions":["0001"],"size_bytes":14882843764,"source":"openneuro","storage":{"backend":"s3","base":"s3://openneuro.org/ds007028","raw_key":"dataset_description.json","dep_keys":["CHANGES"]},"study_design":null,"study_domain":null,"tasks":["tono14noconcat"],"timestamps":{"digested_at":"2026-04-22T12:29:57.207985+00:00","dataset_created_at":"2025-12-04T03:20:37.542Z","dataset_modified_at":"2025-12-04T12:05:48.000Z"},"total_files":3,"computed_title":"Auditory Cortex Macaque Monkey DISC Data","nchans_counts":[{"val":64,"count":3}],"sfreq_counts":[{"val":20000.0,"count":3}],"stats_computed_at":"2026-04-22T23:16:00.312316+00:00","total_duration_s":null,"tagger_meta":{"config_hash":"3557b68bca409f28","metadata_hash":"57b12fd90a7e6b15","model":"openai/gpt-5.2","tagged_at":"2026-04-07T09:32:40.872789+00:00"},"tags":{"pathology":["Other"],"modality":["Auditory"],"type":["Perception"],"confidence":{"pathology":0.6,"modality":0.75,"type":0.65},"reasoning":{"few_shot_analysis":"Closest few-shot match on stimulus modality is the example titled \"Subcortical responses to music and speech are alike while cortical responses diverge\", which is labeled Modality=Auditory and Type=Perception when participants listen to acoustic stimuli. Another auditory-task example is \"EEG: Three-Stim Auditory Oddball and Rest...\" (also Auditory modality). These examples guide the convention that listening to tones/music/speech is labeled Auditory (modality) and typically Perception (type) when no higher-level construct is specified. None of the few-shot examples cover non-human primate datasets; thus few-shot patterns help mainly for Modality/Type, not Pathology.","metadata_analysis":"Key metadata facts available are sparse but informative: (1) Title explicitly specifies auditory system and non-human subjects: \"Auditory Cortex Macaque Monkey DISC Data\". (2) Participants field indicates a small N typical of animal neurophysiology: \"Subjects: 3\". (3) Task label suggests tone stimulation: task \"tono14noconcat\" (\"tono\" strongly implies tones/tonotopy-like auditory stimuli). No clinical recruitment/diagnosis terms are present in the provided metadata.","paper_abstract_analysis":"No useful paper information.","evidence_alignment_check":"Pathology: Metadata says non-human primates (\"Macaque Monkey\") and gives no diagnosis; few-shot patterns mostly distinguish clinical vs healthy human cohorts, but do not address animal cohorts. This partially CONFLICTS with a naive few-shot-driven tendency to choose Healthy for non-clinical cohorts; however, since the population is not a standard human healthy cohort, labeling as Other better reflects the recruitment population.\nModality: Metadata says \"Auditory Cortex\" and task \"tono...\" implying auditory tone stimulation; few-shot auditory listening datasets map such stimuli to Modality=Auditory. ALIGN.\nType: Metadata implies auditory tone stimulation (task \"tono...\"; auditory cortex focus) with no explicit learning/memory/decision framing; few-shot conventions map basic sensory stimulus processing to Type=Perception. ALIGN.","decision_summary":"Top-2 candidates per category:\n- Pathology: (1) Other vs (2) Healthy. Evidence for Other: non-human cohort in title (\"Macaque Monkey\"), small \"Subjects: 3\", and no clinical recruitment described. Evidence for Healthy: absence of disease terms. Head-to-head: Other wins because the dataset population is not a human healthy cohort; it is an animal physiology dataset.\n- Modality: (1) Auditory vs (2) Unknown. Evidence for Auditory: title \"Auditory Cortex...\" and task name \"tono14noconcat\" suggesting tones; few-shot auditory examples label similar listening paradigms as Auditory. Auditory wins.\n- Type: (1) Perception vs (2) Other. Evidence for Perception: tone/auditory stimulation implied by \"tono...\" and auditory cortex focus; few-shot auditory stimulus datasets (music/speech, oddball) follow Perception when primarily sensory processing is studied. Perception wins.\nConfidence justification: sparse metadata (only title/subjects/task) limits certainty, especially for Pathology and Type (no explicit task description beyond task name)."}},"canonical_name":null,"name_confidence":0.42,"name_meta":{"suggested_at":"2026-04-14T10:18:35.343Z","model":"openai/gpt-5.2 + openai/gpt-5.4-mini + deterministic_fallback"},"name_source":"author_year","author_year":"Kajikawa2025"}}