Resource:
ScraperWiki
ScraperWiki
- Links: Website | On Twitter | Facebook
- Tags: Public Information, Web Scraping
ScraperWiki is a site and platform for building data scrapers that can transform unstructured data such as plain text files or tables on web pages (or even in PDFs) into structured data that can be queried with an API in JSON or XML format. Scrapers can be written in Python, Ruby, or PHP, and can be edited by anyone who has registered for the site (the “wiki” in the name). Non-programmers can request a data set and let the community help by putting together a working scraper. Source: Journalism Accelerator
The Journalism Accelerator is not responsible for the content we post here, as excerpts from the source, or links on those sites. The JA does not endorse these sites or their products outright but we sure are intrigued with what they’re up to.
Topics: Resources Technology
Weigh In: Remember to refresh often to see latest comments!
1 comment so far.
The ScraperWiki API API now has an option to make RSS feeds as a format (i.e. instead of JSON, CSV or HTML tables).
For example, Anna made a scraper that gets alocohol licensing applications for Islington in London. She wanted an RSS feed to keep track of new applications using Google Reader. Read more about it on their blog: http://blog.scraperwiki.com/2011/09/21/make-rss-with-an-sql-query/