{"id":34855,"date":"2025-07-31T19:27:55","date_gmt":"2025-07-31T17:27:55","guid":{"rendered":"https:\/\/www.graviton.at\/letterswaplibrary\/seeking-publicly-available-paired-mri-genomic-structured-data-for-multimodal-ml-human-animal-plant\/"},"modified":"2025-07-31T19:27:55","modified_gmt":"2025-07-31T17:27:55","slug":"seeking-publicly-available-paired-mri-genomic-structured-data-for-multimodal-ml-human-animal-plant","status":"publish","type":"post","link":"https:\/\/www.graviton.at\/letterswaplibrary\/seeking-publicly-available-paired-mri-genomic-structured-data-for-multimodal-ml-human-animal-plant\/","title":{"rendered":"Seeking Publicly Available Paired MRI + Genomic\/Structured Data For Multimodal ML (Human\/Animal\/Plant)"},"content":{"rendered":"<p><!-- SC_OFF --><\/p>\n<div class=\"md\">\n<p>I&#8217;m working on a multimodal machine learning pipeline that combines image data with structured\/genomic-like data for prediction task. I&#8217;m looking for publicly available datasets where MRI\/Image data and Genomic\/Structured data are explicitly paired for the same individual\/subject. My ideal scenario would be human cancer (like Glioblastoma Multiforme, where I know TCGA exists), but given recent data access changes (e.g., TCIA policies), I&#8217;m open to other domains that fit this multimodal structure:<\/p>\n<p>What I&#8217;m looking for (prioritized):<\/p>\n<p>Human Medical Data (e.g., Cancer): MRI\/Image: Brain MRI (T1, T1Gd, T2, FLAIR). Genomic: Gene expression, mutations, methylation. Crucial: Data must be for the same patients, linked by ID (like TCGA IDs).<\/p>\n<p>I&#8217;m aware of TCGA-GBM via TCIA\/GDC, but access to the BraTS-TCGA-GBM imaging seems to be undergoing changes as of July 2025. Any direct links or advice on navigating the updated TCIA\/NIH Data Commons policies for this specific type of paired data would be incredibly helpful.<\/p>\n<p>Animal Data:<\/p>\n<p>Image: Animal MRI, X-rays, photos\/video frames of animals (e.g., for health monitoring, behavior).<\/p>\n<p>Genomic\/Structured: Genetic markers, physiological sensor data (temp, heart rate), behavioral data (activity), environmental data (pen conditions), individual animal ID\/metadata.<\/p>\n<p>Crucial: Paired for the same individual animal.<\/p>\n<p>I understand animal MRI+genomics is rare publicly, so I&#8217;m also open to other imaging (e.g., photos) combined with structured data.<\/p>\n<p>Plant Data:<\/p>\n<p>Image: Photos of plant leaves\/stems\/fruits (e.g., disease symptoms, growth).<\/p>\n<p>Structured: Environmental sensor data (temp, humidity, soil pH), plant species\/cultivar genetics, agronomic metadata. Crucial: Paired for the same plant specimen\/plot.<\/p>\n<p>I&#8217;m aware of PlantVillage for images, but seeking datasets that explicitly combine images with structured non-image data per plant.<\/p>\n<p>What I&#8217;m NOT looking for:<\/p>\n<p>Datasets with only images or only genomic\/structured data.<\/p>\n<p>Datasets where pairing would require significant, unreliable manual matching.<\/p>\n<p>Data that requires extremely complex or exclusive access permissions (unless it&#8217;s the only viable option and the process is clearly outlined).<\/p>\n<p>Any pointers to specific datasets, data repositories, research groups known for sharing such data, or advice on current access methods for TCGA-linked imaging would be immensely appreciated!<\/p>\n<p>Thank you!<\/p>\n<\/div>\n<p><!-- SC_ON -->   submitted by   <a href=\"https:\/\/www.reddit.com\/user\/01kaushikjain01\"> \/u\/01kaushikjain01 <\/a> <br \/> <span><a href=\"https:\/\/www.reddit.com\/r\/datasets\/comments\/1me6ik6\/seeking_publicly_available_paired_mri\/\">[link]<\/a><\/span>   <span><a href=\"https:\/\/www.reddit.com\/r\/datasets\/comments\/1me6ik6\/seeking_publicly_available_paired_mri\/\">[comments]<\/a><\/span><\/p><div class='watch-action'><div class='watch-position align-right'><div class='action-like'><a class='lbg-style1 like-34855 jlk' href='javascript:void(0)' data-task='like' data-post_id='34855' data-nonce='65e0e39b87' rel='nofollow'><img class='wti-pixel' src='https:\/\/www.graviton.at\/letterswaplibrary\/wp-content\/plugins\/wti-like-post\/images\/pixel.gif' title='Like' \/><span class='lc-34855 lc'>0<\/span><\/a><\/div><\/div> <div class='status-34855 status align-right'><\/div><\/div><div class='wti-clear'><\/div>","protected":false},"excerpt":{"rendered":"<p>I&#8217;m working on a multimodal machine learning pipeline that combines image data with structured\/genomic-like data for prediction&#8230;<\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[85],"tags":[],"class_list":["post-34855","post","type-post","status-publish","format-standard","hentry","category-datatards","wpcat-85-id"],"_links":{"self":[{"href":"https:\/\/www.graviton.at\/letterswaplibrary\/wp-json\/wp\/v2\/posts\/34855","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.graviton.at\/letterswaplibrary\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.graviton.at\/letterswaplibrary\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.graviton.at\/letterswaplibrary\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.graviton.at\/letterswaplibrary\/wp-json\/wp\/v2\/comments?post=34855"}],"version-history":[{"count":0,"href":"https:\/\/www.graviton.at\/letterswaplibrary\/wp-json\/wp\/v2\/posts\/34855\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.graviton.at\/letterswaplibrary\/wp-json\/wp\/v2\/media?parent=34855"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.graviton.at\/letterswaplibrary\/wp-json\/wp\/v2\/categories?post=34855"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.graviton.at\/letterswaplibrary\/wp-json\/wp\/v2\/tags?post=34855"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}