{"id":33547,"date":"2025-04-20T13:27:25","date_gmt":"2025-04-20T11:27:25","guid":{"rendered":"https:\/\/www.graviton.at\/letterswaplibrary\/a-dataset-of-annotated-cc0-images-what-to-do-with-it\/"},"modified":"2025-04-20T13:27:25","modified_gmt":"2025-04-20T11:27:25","slug":"a-dataset-of-annotated-cc0-images-what-to-do-with-it","status":"publish","type":"post","link":"https:\/\/www.graviton.at\/letterswaplibrary\/a-dataset-of-annotated-cc0-images-what-to-do-with-it\/","title":{"rendered":"A Dataset Of Annotated CC0 Images, What To Do With It?"},"content":{"rendered":"<p><!-- SC_OFF --><\/p>\n<div class=\"md\">\n<p>years ago (before the current generative AI wave) I&#8217;d seen this person start a website for crowdsourced image annotations, I thought that was a great idea so I tried to support by becoming a user, when I had spare moments I&#8217;d go annotate. Killed a lot of time doing that during pandemic lockdowns etc. There around 300,000 polygonal outlines here accumulated over many years. to view them you must search for specific labels ; there&#8217;s a few hundred listed in the system and a backlog of new label requests hidden from public view. there is an export feature<\/p>\n<p><a href=\"https:\/\/imagemonkey.io\/\">https:\/\/imagemonkey.io<\/a><\/p>\n<p>example .. roads\/pavements in street scenes (&#8220;rework&#8221; mode will show you outlines, you can also go to &#8220;dataset-&gt;explore&#8221; to browse or export)<\/p>\n<p><a href=\"https:\/\/imagemonkey.io\/annotate?mode=browse&amp;view=unified&amp;query=road%7Cpavement&amp;search_option=rework\">https:\/\/imagemonkey.io\/annotate?mode=browse&amp;view=unified&amp;query=road%7Cpavement&amp;search_option=rework<\/a><\/p>\n<p>it&#8217;s also possible to get the annotations out in batches via a python API<\/p>\n<p><a href=\"https:\/\/github.com\/ImageMonkey\/imagemonkey-libs\/blob\/master\/python\/snippets\/export.py\">https:\/\/github.com\/ImageMonkey\/imagemonkey-libs\/blob\/master\/python\/snippets\/export.py<\/a><\/p>\n<p>i&#8217;m worried the owner might get disheartened from a sense of futility (so few contributors, and now there are really powerful foundation models available including image to text)<\/p>\n<p>but I figure &#8220;every little helps&#8221;, it would be useful to get this data out into a format or location where it can feed back into training, maybe even if it&#8217;s obscure and not yet in training sets it could be used for benchmarking or testing other models<\/p>\n<p>When the site was started the author imagined a tool for automatically fine-tuning some vision nets for specific labels, I&#8217;d wanted to broaden it to become more general. the label list did grow and there&#8217;s probably a couple of hundred more that would make sense to make &#8216;live&#8217;<\/p>\n<p>There&#8217;s also an aspect that these generative AI models get accused of theft, so the more deliberate voluntary data there is out there the better. I&#8217;d guess that you could mix image annotations somehow into the pretraining data for multimodal models, right? I&#8217;m also aware that you can reduce the number of images needed to train image-generators if you have polygonal annotations aswell as image\/descriptions-text pairs.<\/p>\n<p>Just before the diffusion craze kicked off I&#8217;d had some attempts at trying to train small vision nets myself from scratch (rtx3080) but could only get so far. When stable diffusion came out I figured my own attemtps to train things were futile.<\/p>\n<p>Here&#8217;s a thread where I documented my training attempt for the site owner<\/p>\n<p><a href=\"https:\/\/github.com\/ImageMonkey\/imagemonkey-core\/issues\/300\">https:\/\/github.com\/ImageMonkey\/imagemonkey-core\/issues\/300<\/a> &#8211; in here you&#8217;ll see some visualisations of the annotations (the usual color coded overlays)<\/p>\n<p>I think these labels today could be generalised by using an NLP model to turn the labels into vector embeddings (cluster similar labels or train image to embedding, etc)<\/p>\n<p>The annotations would probably want to be converted to some better known format that could be loaded into other tools. they are available in his json format.<\/p>\n<p>can anyone advise on how to get this effort fed back into some kind of visible community benefit?<\/p>\n<\/div>\n<p><!-- SC_ON -->   submitted by   <a href=\"https:\/\/www.reddit.com\/user\/dobkeratops\"> \/u\/dobkeratops <\/a> <br \/> <span><a href=\"https:\/\/www.reddit.com\/r\/datasets\/comments\/1k3k39g\/a_dataset_of_annotated_cc0_images_what_to_do_with\/\">[link]<\/a><\/span>   <span><a href=\"https:\/\/www.reddit.com\/r\/datasets\/comments\/1k3k39g\/a_dataset_of_annotated_cc0_images_what_to_do_with\/\">[comments]<\/a><\/span><\/p><div class='watch-action'><div class='watch-position align-right'><div class='action-like'><a class='lbg-style1 like-33547 jlk' href='javascript:void(0)' data-task='like' data-post_id='33547' data-nonce='614a020375' rel='nofollow'><img class='wti-pixel' src='https:\/\/www.graviton.at\/letterswaplibrary\/wp-content\/plugins\/wti-like-post\/images\/pixel.gif' title='Like' \/><span class='lc-33547 lc'>0<\/span><\/a><\/div><\/div> <div class='status-33547 status align-right'><\/div><\/div><div class='wti-clear'><\/div>","protected":false},"excerpt":{"rendered":"<p>years ago (before the current generative AI wave) I&#8217;d seen this person start a website for crowdsourced&#8230;<\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[85],"tags":[],"class_list":["post-33547","post","type-post","status-publish","format-standard","hentry","category-datatards","wpcat-85-id"],"_links":{"self":[{"href":"https:\/\/www.graviton.at\/letterswaplibrary\/wp-json\/wp\/v2\/posts\/33547","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.graviton.at\/letterswaplibrary\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.graviton.at\/letterswaplibrary\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.graviton.at\/letterswaplibrary\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.graviton.at\/letterswaplibrary\/wp-json\/wp\/v2\/comments?post=33547"}],"version-history":[{"count":0,"href":"https:\/\/www.graviton.at\/letterswaplibrary\/wp-json\/wp\/v2\/posts\/33547\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.graviton.at\/letterswaplibrary\/wp-json\/wp\/v2\/media?parent=33547"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.graviton.at\/letterswaplibrary\/wp-json\/wp\/v2\/categories?post=33547"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.graviton.at\/letterswaplibrary\/wp-json\/wp\/v2\/tags?post=33547"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}