How does the fair use doctrine apply to websites in terms of screen-scraping?
The particular example I am thinking of is extraction of the useful data from a website, and re-presentation of the raw data aggregated with data from other similar websites. For example, suppose one was to extract data from a variety of websites to produce a database of structured data in a similar manner to DBpedia, could this be considered fair-use. If not, where does the boundary lie?
Obviously, this differs by jurisdiction, but then so the websites being scraped as well as the location of the scraper. Similarly, ethics may well differ from legalities.
Okay, so in many cases it may well be illegal, but as a content-provider how can it actually be prevented? I've posted a follow-up question on Protection from screen scraping