{"success":true,"database":"eegdash","data":{"_id":"6953f4249276ef1ee07a3450","dataset_id":"ds006465","associated_paper_doi":null,"authors":["Xinyu Ma","Jiang Yi","Ning Jiang"],"bids_version":"1.7.0","contact_info":["Xinyu Ma"],"contributing_labs":null,"data_processed":true,"dataset_doi":"doi:10.18112/openneuro.ds006465.v2.0.0","datatypes":["eeg"],"demographics":{"subjects_count":20,"ages":[25,25,29,27,23,25,30,26,21,23,24,21,27,25,24,25,24,19,24,24],"age_min":19,"age_max":30,"age_mean":24.55,"species":null,"sex_distribution":{"m":9,"f":11},"handedness_distribution":{"r":20}},"experimental_modalities":null,"external_links":{"source_url":"https://openneuro.org/datasets/ds006465","osf_url":null,"github_url":null,"paper_url":null},"funding":["a 1.3.5 project for disciplines of excellence from West China Hospital (#ZYYC22001)"],"ingestion_fingerprint":"56ebe9b8fd91e165c9ec5e398a8324c4856ba74e0446e98a85c10385f9781ce2","license":"CC0","n_contributing_labs":null,"name":"3M-CPSEED：An EEG-based Dataset for Chinese Pinyin Production in Overt, Silent-intended, and Imagined Speech","readme":"Overview\nThis dataset, named 3M-CPSEED, consists of electroencephalogram (EEG) recordings obtained from 20 participants engaged in imagined speech tasks. 3M-CPSEED holds significant implications for speech neurophysiology research, not only facilitating exploration of neural activity differences across pinyin articulations but also enabling robust transfer learning studies for other alphabetic languages.\nData Collection\nParticipants: 20 healthy, right-handed individuals (average age: 24.55 years, standard deviation: 2.58 years; 11 females, 9 males) who are native Chinese speakers.\nMaterials: To strike a balance between comprehensively capturing the articulatory features of the Chinese phonological system and maintaining a concise, controllable set of stimuli, we selected this set of Pinyin sounds: Finals: \"a, i, u, ü\"; Initials: \"m, f, j, l, k, ch\".\nProcedure: Participants read Pinyin displayed on a screen at 'speak', 'Silently articulated' and 'imagined' phase. Each participant completed 4 blocks of 1600 trials in total.\nData Structure\nThe dataset is organized according to the BIDS standard:\nMain Folder:\ndataset_description.json: Description of the dataset.\nparticipants.tsv: Participant information.\nparticipants.json: Details of columns in participants.tsv.\nREADME: General information about the dataset.\ndata_all.mat: Labeled EEG data of all subjects in MAT format.\nDerivative Data:\npreproc/: Preprocessed data, including subfolders for each subject (sub-01, etc.), with data in .mat formats .\nAcknowledgments\nThis work was supported by a 1.3.5 project for disciplines of excellence from West China Hospital (#ZYYC22001).","recording_modality":["eeg"],"senior_author":"Ning Jiang","sessions":["1","2","3","4"],"size_bytes":8768486080,"source":"openneuro","study_design":null,"study_domain":null,"tasks":["imaginedspeech"],"timestamps":{"digested_at":"2026-04-22T12:29:28.188443+00:00","dataset_created_at":"2025-07-12T08:34:49.595Z","dataset_modified_at":"2025-10-29T03:31:08.000Z"},"total_files":80,"storage":{"backend":"s3","base":"s3://openneuro.org/ds006465","raw_key":"dataset_description.json","dep_keys":["CHANGES","README.md","participants.json","participants.tsv"]},"tagger_meta":{"config_hash":"4a051be509a0e3d0","metadata_hash":"a932948f834a8e72","model":"openai/gpt-5.2","tagged_at":"2026-01-20T18:59:34.746706+00:00"},"tags":{"pathology":["Healthy"],"modality":["Visual"],"type":["Motor"],"confidence":{"pathology":0.7,"modality":0.8,"type":0.8},"reasoning":{"few_shot_analysis":"Most similar few-shot convention is the \"EEG Motor Movement/Imagery Dataset\" example (Healthy / Visual / Motor). It shows that when the core paradigm is movement/imagery (even if visually cued by a screen target), the catalog labels Type as \"Motor\" and Modality as \"Visual\" (because stimuli are presented on a screen). The current dataset is also an imagery paradigm (imagined speech) with visual cues (Pinyin on a screen), matching this convention more than perception-focused visual discrimination (Meta-rdk) or auditory-only paradigms.","metadata_analysis":"Key facts from metadata:\n1) Population: \"20 healthy, right-handed individuals\".\n2) Task/paradigm: participants were \"engaged in imagined speech tasks\" and \"read Pinyin displayed on a screen\" across phases including \"'Silently articulated' and 'imagined' phase\".\nThese lines support Healthy participants, Visual stimulus presentation, and a speech production/imagery (motor) research focus.","paper_abstract_analysis":"No useful paper information.","evidence_alignment_check":"Pathology:\n- Metadata says: \"20 healthy, right-handed individuals\".\n- Few-shot pattern suggests: imagery tasks in non-clinical volunteers are labeled Healthy (e.g., motor imagery dataset labeled Healthy).\n- ALIGN.\n\nModality:\n- Metadata says: \"Participants read Pinyin displayed on a screen\".\n- Few-shot pattern suggests: visually presented cues in imagery tasks are labeled Visual (e.g., motor movement/imagery dataset uses on-screen targets and is labeled Visual modality).\n- ALIGN.\n\nType:\n- Metadata says: \"imagined speech tasks\" and phases include \"'Silently articulated' and 'imagined' phase\".\n- Few-shot pattern suggests: imagery/execution paradigms map to Type=Motor (motor movement/imagery example).\n- ALIGN.","decision_summary":"Top-2 candidates per category with head-to-head comparison:\n\nPathology candidates: (1) Healthy vs (2) Unknown\n- Healthy evidence: \"20 healthy, right-handed individuals\".\n- Unknown evidence: none (only would apply if health status absent).\nWinner: Healthy (explicitly stated). Alignment: aligns with few-shot conventions for non-clinical imagery datasets.\nConfidence basis: 1 explicit quote naming healthy participants (clear but only one primary quote).\n\nModality candidates: (1) Visual vs (2) Motor\n- Visual evidence: \"read Pinyin displayed on a screen\" (stimulus channel is visual).\n- Motor evidence: speech articulation/imagery involves motor systems, but modality is defined as stimulus/input channel, not response.\nWinner: Visual (explicit screen-based stimulus presentation). Alignment: aligns with motor imagery few-shot where cues are visual.\nConfidence basis: 1 explicit quote + strong few-shot analog.\n\nType candidates: (1) Motor vs (2) Perception\n- Motor evidence: \"imagined speech tasks\" and \"'Silently articulated' and 'imagined' phase\" indicate speech production/imagery focus.\n- Perception evidence: participants read stimuli, but the goal is not sensory discrimination; it is speech imagery/production.\nWinner: Motor (imagery/production is the central construct). Alignment: aligns with motor imagery few-shot labeling.\nConfidence basis: 2 explicit task-paradigm quotes + strong few-shot analog."}},"computed_title":"3M-CPSEED：An EEG-based Dataset for Chinese Pinyin Production in Overt, Silent-intended, and Imagined Speech","nchans_counts":[{"val":32,"count":58},{"val":126,"count":19},{"val":33,"count":3}],"sfreq_counts":[{"val":500.0,"count":80}],"stats_computed_at":"2026-04-22T23:16:00.311633+00:00","total_duration_s":130319.84,"author_year":"Ma2025","canonical_name":null,"name_source":"canonical"}}