Comments on: Knight Foundation finance ScraperWiki for journalism https://blog.scraperwiki.com/2011/06/knight-foundation-finance-scraperwiki-for-journalism/ Extract tables from PDFs and scrape the web Thu, 14 Jul 2016 16:12:42 +0000 hourly 1 https://wordpress.org/?v=4.6 By: $1 million to build a data platform | ScraperWiki Data Blog https://blog.scraperwiki.com/2011/06/knight-foundation-finance-scraperwiki-for-journalism/#comment-662 Sun, 12 May 2013 09:31:07 +0000 http://blog.scraperwiki.com/?p=758215012#comment-662 […] total, provided we hit certain milestones next August, and with the Knight Foundation money, this means we have a cool $1,000,000 […]

]]>
By: $1 million to build a data platform | ScraperWiki Data Blog https://blog.scraperwiki.com/2011/06/knight-foundation-finance-scraperwiki-for-journalism/#comment-661 Thu, 09 Feb 2012 18:26:55 +0000 http://blog.scraperwiki.com/?p=758215012#comment-661 […] total, provided we hit certain milestones next August, and with the Knight Foundation money, this means we have a cool $1,000,000 […]

]]>
By: Tow Center/ScraperWiki Datacamp | Tow Center for Digital Journalism https://blog.scraperwiki.com/2011/06/knight-foundation-finance-scraperwiki-for-journalism/#comment-660 Tue, 27 Dec 2011 21:08:20 +0000 http://blog.scraperwiki.com/?p=758215012#comment-660 […] & 4th, the Tow Center will be hosting a two-day “DataCamp” along with ScraperWiki, a recent Knight News Challenge winner and innovative platform for collecting and accessing online […]

]]>
By: Francis Irving https://blog.scraperwiki.com/2011/06/knight-foundation-finance-scraperwiki-for-journalism/#comment-659 Sun, 10 Jul 2011 22:28:11 +0000 http://blog.scraperwiki.com/?p=758215012#comment-659 Hi Edward – sorry for slow reply! I’ve been on holiday and only just saw this.

Our experience is that right now an attempt at a completely automated tool would always end up frustrating. There are quite a few products that try to do it, and they can be very useful – but they can’t magically scrape anything that is scrapable. You’ll end up programming, or doing programming like things, within them anyway (as happens in, say, Refine).

We are, however, doing several things with the grant money that are in the direction of an automatic tool. One of these is a set of tools that write the first pass of code for you – and in simple cases that will be all you need to do. (Others are journalist specific tutorials, and general usability improvements)

Would love to visit Portland!

]]>
By: LSDI : ScraperWiki, una ‘’ruspa’’ che scava nelle miniere di dati del web https://blog.scraperwiki.com/2011/06/knight-foundation-finance-scraperwiki-for-journalism/#comment-658 Wed, 29 Jun 2011 20:35:04 +0000 http://blog.scraperwiki.com/?p=758215012#comment-658 […] obbiettivo ora è realizzare una serie di nuovi servizi, fra cui, ad esempio, un sistema di embargo che consenta ai giornalisti di creare delle […]

]]>
By: M. Edward Borasky https://blog.scraperwiki.com/2011/06/knight-foundation-finance-scraperwiki-for-journalism/#comment-657 Sat, 25 Jun 2011 14:05:37 +0000 http://blog.scraperwiki.com/?p=758215012#comment-657 P.S.: I think Portland, Oregon would *love* a ScraperWiki camp – at least the hackers would.

]]>
By: M. Edward Borasky https://blog.scraperwiki.com/2011/06/knight-foundation-finance-scraperwiki-for-journalism/#comment-656 Sat, 25 Jun 2011 14:02:41 +0000 http://blog.scraperwiki.com/?p=758215012#comment-656 We have a pretty active hacker community in Portland, Oregon, and an active “Civic Apps” community associated with the City of Portland and to a lesser extent a regional government called “Metro”. I know of at least three of those hackers, myself among them, who have used ScraperWiki.

But nearly all of the journalists I know are dead-set against the kind of “coding” that’s required to acquire data via ScraperWiki. They have deadlines, and, let’s face it, HTML parsing, regular expressions and the mechanics of interrogating sites is hard programming. So the uptake of data journalism here in Portland has been slow, even with available hackers like myself. Right now I know of only one journalist here who’s attempted to use ScraperWiki (with the help of two hackers) and it’s been a frustrating experience.

What I think I’m looking for is some kind of “wireframing tool” – an interactive “drag and drop, point and click, WYSIWIG” user interface so that *journalists* can build scrapers. They have this now for the visualization / storytelling part of storytelling with tools like Tableau. They (or at least I) have this now for the data exploration part with tools like R, GGobi and Mondrian (and Excel, of course). But there’s not a tool I know of for rapidly building a scraper. Is there a chance you can use some of the grant money on a more journalist-friendly user interface?

]]>