In response to Uldis Bojārs (@CaptSolo) talk in #pmod
for a lot of #opendata applications the developers are “scraping the data” – taking data as/is and restructuring it (converting to rdf, mapping to a vocabulary, …)

  • publishers ask, how to ensure the data / representation of data is still valid knowing it was modified?
  • could we do scraping more often to show to publishers why publishing raw, preferably linked data is preferred as it won’t need to be modified in potentially wrong ways?