{"id":34622,"date":"2025-07-08T08:27:52","date_gmt":"2025-07-08T06:27:52","guid":{"rendered":"https:\/\/www.graviton.at\/letterswaplibrary\/imagined-and-read-speech-eeg-datasets\/"},"modified":"2025-07-08T08:27:52","modified_gmt":"2025-07-08T06:27:52","slug":"imagined-and-read-speech-eeg-datasets","status":"publish","type":"post","link":"https:\/\/www.graviton.at\/letterswaplibrary\/imagined-and-read-speech-eeg-datasets\/","title":{"rendered":"Imagined And Read Speech EEG Datasets"},"content":{"rendered":"<p><!-- SC_OFF --><\/p>\n<div class=\"md\">\n<h1><strong>Imageind\/Read Speech EEG Datasets<\/strong><\/h1>\n<p>General EEG papers: <a href=\"https:\/\/arxiv.org\/search\/?query=EEG&amp;searchtype=all&amp;source=header\">Arxiv<\/a><\/p>\n<ul>\n<li>\n<p><a href=\"https:\/\/osf.io\/q3zws\/\">ZuCo<\/a> | <a href=\"https:\/\/osf.io\/2urht\/wiki\/home\/\">Data 2<\/a> | <a href=\"https:\/\/www.nature.com\/articles\/sdata2018291\">Paper<\/a> (Imagined\/Read)<\/p>\n<\/li>\n<li>\n<p><a href=\"https:\/\/osf.io\/e82p9\/\">Speech Decoding<\/a> | <a href=\"https:\/\/www.biorxiv.org\/content\/10.1101\/2022.11.15.516461v1.full\">Paper<\/a> (Listened\/Read)<\/p>\n<\/li>\n<li>\n<p><a href=\"https:\/\/lifesciences.datastations.nl\/dataset.xhtml?persistentId=doi%3A10.17026\/dans-xc3-66ze\">DAIS: the Delft Database<\/a> | <a href=\"https:\/\/www.isca-archive.org\/interspeech_2024\/bras24_interspeech.pdf\">Paper<\/a> | <a href=\"https:\/\/github.com\/ChrisSBras\/imagined_vs_articulated_speech\">Code<\/a> (Imagined\/Read) <\/p>\n<\/li>\n<li>\n<p><a href=\"https:\/\/data.ru.nl\/collections\/ru\/cls\/dutch_eeg_speech_register_corpus_dsc_807\">The Dutch EEG Speech Register Corpus<\/a> | <a href=\"https:\/\/mirjamernestus.nl\/Ernestus\/DESRC\/Bentum_tenBosch_vandenBosch_Ernestus_DESRC.pdf\">Paper<\/a> (Listened) <\/p>\n<\/li>\n<li>\n<p><a href=\"https:\/\/www.kaggle.com\/datasets\/ignazio\/kumars-eeg-imagined-speech\">Kumar&#8217;s EEG Imagined Speech<\/a> (Imagined)<\/p>\n<\/li>\n<li>\n<p><a href=\"https:\/\/www.cs.toronto.edu\/~complingweb\/data\/karaOne\/karaOne.html\">KARA ONE<\/a> (Imagined\/Read)<\/p>\n<\/li>\n<li>\n<p><a href=\"https:\/\/openneuro.org\/datasets\/ds005170\/versions\/1.1.2\">Chisco<\/a> | <a href=\"https:\/\/www.nature.com\/articles\/s41597-024-04114-1\">Paper<\/a> | <a href=\"https:\/\/github.com\/zhangzihan-is-good\/Chisco\">Code<\/a> (Imagined)<\/p>\n<\/li>\n<li>\n<p><a href=\"https:\/\/openneuro.org\/datasets\/ds003626\/versions\/2.1.2\">Inner<\/a>\/<a href=\"https:\/\/drive.google.com\/file\/d\/0By7apHbIp8ENZVBLRFVlSFhzbHc\/view?resourcekey=0-JVHv2UiRsxim41Wioro0EA\">Imagined<\/a> Speech Datasets | <a href=\"https:\/\/www.nature.com\/articles\/s41597-022-01147-2\">Paper<\/a> (Imagined)<\/p>\n<\/li>\n<li>\n<p><a href=\"https:\/\/drum.um.edu.mt\/articles\/dataset\/Motor_and_Speech_Imagery_EEG_Dataset\/24465871\">Motor and Speech Imagery EEG Dataset<\/a> | <a href=\"https:\/\/www.researchgate.net\/publication\/382483623_A_combined_EEG_motor_and_speech_imagery_paradigm_with_automated_successive_halving_for_customizable_command_selection\">Paper<\/a> (Imagined) <\/p>\n<\/li>\n<li>\n<p><a href=\"https:\/\/data.mendeley.com\/datasets\/57g8z63tmy\/1\">Gamified Imagined Speech Datasets<\/a> (Imagined)<\/p>\n<\/li>\n<li>\n<p><a href=\"https:\/\/zenodo.org\/records\/3554128\">FEIS<\/a> | <a href=\"https:\/\/www.isca-archive.org\/interspeech_2020\/clayton20_interspeech.pdf\">Paper<\/a> | <a href=\"https:\/\/github.com\/scottwellington\/FEIS\">Code<\/a> (Imagined)<\/p>\n<\/li>\n<li>\n<p><a href=\"https:\/\/openneuro.org\/datasets\/ds004196\/versions\/2.0.0\">iSpeech<\/a> | <a href=\"https:\/\/arxiv.org\/pdf\/2306.10854\">Paper<\/a> | <a href=\"https:\/\/www.nature.com\/articles\/s41597-023-02286-w\">Paper 2<\/a> | <a href=\"https:\/\/github.com\/scottwellington\/iSpeech\">Code<\/a> | <a href=\"https:\/\/github.com\/LTU-Machine-Learning\/Inner_Speech_EEG_FMRI\">Code 2<\/a> (Imagined)<\/p>\n<\/li>\n<li>\n<p><a href=\"https:\/\/data.mendeley.com\/datasets\/73g4fw884c\/1\">EEGIS<\/a> (Imagined) <\/p>\n<\/li>\n<li>\n<p><a href=\"https:\/\/datadryad.org\/dataset\/doi:10.5061\/dryad.070jc\">DRYAD<\/a> | <a href=\"https:\/\/www.sciencedirect.com\/science\/article\/pii\/S0960982218301465\">Paper<\/a> (Listened) <\/p>\n<\/li>\n<li>\n<p><a href=\"https:\/\/github.com\/farhanwadia\/EEGClassification\">Open\/Close<\/a> (Imagined)<\/p>\n<\/li>\n<li>\n<p><a href=\"https:\/\/osf.io\/eyzaq\/\">Replication Recipe Analysis<\/a> | <a href=\"https:\/\/elifesciences.org\/articles\/33468#data\">Paper<\/a> (Read)<\/p>\n<\/li>\n<li>\n<p><a href=\"https:\/\/rdr.kuleuven.be\/dataset.xhtml?persistentId=doi%3A10.48804\/K3VSND\">SparrKULee<\/a> | <a href=\"https:\/\/www.biorxiv.org\/content\/10.1101\/2023.07.24.550310v1\">Paper<\/a> | <a href=\"https:\/\/github.com\/exporl\/auditory-eeg-dataset\">Code<\/a> (Listened) <\/p>\n<\/li>\n<li>\n<p><a href=\"https:\/\/huggingface.co\/datasets\/Alidr79\/cueless_EEG_subject_identification\">Cueless EEG<\/a> | <a href=\"https:\/\/arxiv.org\/abs\/2501.09700\">Paper<\/a> | <a href=\"https:\/\/github.com\/Alidr79\/cueless_EEG_subject_identification?tab=readme-ov-file&amp;1-data-structuring--file-integration\">Code<\/a> (Imagined)<\/p>\n<\/li>\n<\/ul><\/div>\n<p><!-- SC_ON -->   submitted by   <a href=\"https:\/\/www.reddit.com\/user\/copywriterpirate\"> \/u\/copywriterpirate <\/a> <br \/> <span><a href=\"https:\/\/www.reddit.com\/r\/datasets\/comments\/1lugoyb\/imagined_and_read_speech_eeg_datasets\/\">[link]<\/a><\/span>   <span><a href=\"https:\/\/www.reddit.com\/r\/datasets\/comments\/1lugoyb\/imagined_and_read_speech_eeg_datasets\/\">[comments]<\/a><\/span><\/p><div class='watch-action'><div class='watch-position align-right'><div class='action-like'><a class='lbg-style1 like-34622 jlk' href='javascript:void(0)' data-task='like' data-post_id='34622' data-nonce='614a020375' rel='nofollow'><img class='wti-pixel' src='https:\/\/www.graviton.at\/letterswaplibrary\/wp-content\/plugins\/wti-like-post\/images\/pixel.gif' title='Like' \/><span class='lc-34622 lc'>0<\/span><\/a><\/div><\/div> <div class='status-34622 status align-right'><\/div><\/div><div class='wti-clear'><\/div>","protected":false},"excerpt":{"rendered":"<p>Imageind\/Read Speech EEG Datasets General EEG papers: Arxiv ZuCo | Data 2 | Paper (Imagined\/Read) Speech Decoding&#8230;<\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[85],"tags":[],"class_list":["post-34622","post","type-post","status-publish","format-standard","hentry","category-datatards","wpcat-85-id"],"_links":{"self":[{"href":"https:\/\/www.graviton.at\/letterswaplibrary\/wp-json\/wp\/v2\/posts\/34622","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.graviton.at\/letterswaplibrary\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.graviton.at\/letterswaplibrary\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.graviton.at\/letterswaplibrary\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.graviton.at\/letterswaplibrary\/wp-json\/wp\/v2\/comments?post=34622"}],"version-history":[{"count":0,"href":"https:\/\/www.graviton.at\/letterswaplibrary\/wp-json\/wp\/v2\/posts\/34622\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.graviton.at\/letterswaplibrary\/wp-json\/wp\/v2\/media?parent=34622"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.graviton.at\/letterswaplibrary\/wp-json\/wp\/v2\/categories?post=34622"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.graviton.at\/letterswaplibrary\/wp-json\/wp\/v2\/tags?post=34622"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}