top | item 3642774

The uncanny valley of web scraping

172 points| Swizec | 14 years ago |zemanta.com | reply

49 comments

order
[+] ecaron|14 years ago|reply
For the last 7 years I have worked at a company that does specialized job listing web scraping. And on nearly a weekly basis I encounter other programmers who say, "pssssh, I could do that in a weekend."

It does seem disgustingly easy, but once you move from "getting" the data to "understanding" the data it becomes a beastly nightmare. So thank you to the OP for helping raise "give scrapers some credit" awareness.

Except scrapers that don't respect robots.txt or meta noindex - a pox on their houses....

[+] Radim|14 years ago|reply
But there will always be hordes of new, excitable hackers who fail to appreciate the complexity of finishing real-world tasks (and I'm not talking only intelligent scraping). The "awareness" is ephemeral.

Of course, by the time they come around, they will have driven down market prices as well as hackers' image on the whole, with their "Easy! 2 days max, here's the expected cost", invariably followed by "Give me an extra week or two, I'm almost there".

Eventually they'll learn, then come and vent on HN. The circle of life.

[+] ericabiz|14 years ago|reply
Damn. I'm just really glad to know I'm not the only one.

One of my company's biggest sellers is a tool that does web scraping. And since we charge a monthly fee for it (I don't think most people realize that our server fees alone run nearly $2,000 a month for this, and that's with amazing deals on web hosting), we get a lot of "Pfft--how hard could that really be to build?"

The answer is: It's easy--if you're only scraping one site, in small quantities. The minute you try to implement in Unicode (we have) or bundle it all in a report, or keep an ongoing cache of what your scraper found, or heck, scale it to millions of scrapes per month, all cached and indexed--now you are looking at something difficult.

We have a talented team, we're 16 months into this, and we've since seen competitors pop up all over the place. They all get a few customers and then run into scaling issues. It's one of those markets that seems seductively easy, but really isn't. And since there are many competitors lowballing, we've had to focus even more development time on awesome features to stay ahead of the curve.

I wish articles like this woke people up. Unfortunately, once you're committed and have a few customers, it's easy to rationalize going deeper. But web scraping can be a tricky industry to turn a profit in, especially if you are relying on a small and fickle customer base.

[+] cookiecaper|14 years ago|reply
People should realize that search engines are essentially just huge, generic scrapers.

I've written my share of scrapers and they are almost always a major PITA unless you're just doing a drive-by for one or two DOM elements. Scraping often will end up to be extremely time consuming, although there is rarely significant challenge to it. Just a lot of monotonous, mind-numbing, tedious work.

My advice to most freelancers is: avoid scrapers until you get some teammates to share in the agony.

[+] pnathan|14 years ago|reply
I spent some time about a year ago fussing with getting a web spider/scraper going; it was simple enough to download data, but actually putting it into a database that was domain-, content-, and time-aware was impressively complex and I put it on the backburner.
[+] sho_hn|14 years ago|reply
This is only tangentally related to the article, but on scraping HTML in general: If you're a Python user, use lxml for it. I know most content on the web will tell you to use BeautifulSoup, and lxml is something you've only heard of in connection with reading and writing XML, but lxml actually has a lovely HTML sub-package, it's faster than BeautifulSoup, and it's compatible with Python 3. I've gotten lots of good mileage out of it (and no, I'm not a developer on it :)).
[+] waitwhat|14 years ago|reply
FYI, BeautifulSoup is actually dropping its own parser entirely for the next version, in favour of being a wrapper around lxml/html5lib/html.parser http://www.crummy.com/2012/02/02/0
[+] kanzure|14 years ago|reply
I admit to using BeautifulSoup, lxml, scrapy, nokogiri and mechanize. But these days I just stick with PhantomJS. Just load the DOM with WebKit. http://phantomjs.org/
[+] sanxiyn|14 years ago|reply
Another tangentally related recommendation: Use Scrapy if you need something more than one-shot script using BeautifulSoup or lxml.

http://scrapy.org/

[+] chernevik|14 years ago|reply
I need to build / find a tool for parsing EDGAR filings for their financial statements tables (not so bad) and parsing those financial tables into usable information (pretty bad, the tables have very bad and often inconsistent HTML layouts).

Any suggestions to where I should be looking? Python? XSLT?

[+] cuppster|14 years ago|reply
Anyone scraping in Perl? I've found pQuery very useful. (I tried it in node.js to stay cool, but async scraping is an anti-pattern) You can use jQuery selectors, etc... Just posted something related to it on by blog: scrape with pQuery, dump into Redis, reformat into CSV then into mysql...

http://cuppster.com/2012/02/28/utf-8-round-trip-for-perl-mys...

[+] dmn001|14 years ago|reply
I also use Perl for web scraping, never heard of pQuery though, I use HTML::TokeParser or HTML::TreeBuilder::XPath.
[+] Kevin_Marks|14 years ago|reply
Relying on RSS feeds is tricky, as many of them are partial extracts, summaries, or just plain wrong (eg archival standalone pages linking to the current front page, stale feeds, links to now-defunct feed services).

If you want to help people writing these things, using hAtom in your HTML is a really good idea.

http://microformats.org/wiki/hatom

[+] lowglow|14 years ago|reply
I'm currently doing this with http://rtcool.com/

Edit: It's basically a service that abstracts out scraping for those that want to create a readability type project.

I've got one last thing to add and then it will be ready for mass consumption.

[+] joering2|14 years ago|reply
very cool idea, thought about it myself! make sure you got all tech sites in. I have adBlock turned on all the time but I have no idea how can soemone "read" something without ad block.
[+] NameNickHN|14 years ago|reply
The title should have been "The uncanny valley of recognizing content on a website".

Scraping is a technicality and, as such, trivial. As the article points out, processing the scraped content and getting useful results is the hard part.

I'm running some scrapers for customers but the information they want exist in structured form on the various websites. Thank goodness.

[+] AndrewDucker|14 years ago|reply
Yeah, I get this with ReadItLater, which works 95% of the time, and produces very odd results the other 5%.
[+] nreece|14 years ago|reply
We have a home-brewed scraper and parser (written in C#) at Feedity - http://feedity.com and let me tell you - it's one thing to scrape data but to derive information out of it is not as easy.
[+] jacoblyles|14 years ago|reply
Does anyone have experience using diffbot for web scraping? I'm looking for some data points.
[+] paulhauggis|14 years ago|reply
I don't know why, but I love web scraping. Somehow it's fun to me to be able to grab data and organize it in a DB in a meaningful way. Even better is using the data to make money.

I wrote my own scraper framework for various page types (I don't want to go into too many details) and my latest uses approx 500GB of bandwidth/month. I run it on one VPS for around $50/month.

[+] sente|14 years ago|reply
That's awesome. I also love scraping, there's just something about which I find really satisfying.
[+] Tichy|14 years ago|reply
Where do you store the data? Most hosts don't seem to offer a lot of storage on their servers.