views:

545

answers:

3

Hi, is there a way to batch copy certain wikipedia articles(about 10,000) to my own mediawiki site?

EDIT: How do I do this without overwriting similarly named articles/pages? Also I don't plan on using illegal means (crawlers etc)

+2  A: 

The Wikipedia database is available for download

victor hugo
+1  A: 

If you're looking to obtain a specific set of articles, then you may be able to use the Export page (http://en.wikipedia.org/wiki/Special:Export) to obtain an XML dump of the pages involved; you can export multiple pages at once, although you may wish to space out your requests.

You can import the XML dumps into MediaWiki using Special:Import or one of the import scripts in maintenance/.

Rob
Does this include images? If not is there an automatic method to obtain images as well?
It doesn't include images because of the copyright.
l2mt
It doesn't include images because it's a straight XML dump of the page information, revision metadata and text, and the format hasn't been expanded to include the binary image data. Strictly speaking, it's a technical limitation. In theory, you could import copyright-infringing text and the software wouldn't stop that.
Rob
A: 

Here's a guide to import a Wikipedia database dump into MediaWiki.

BTW, fetching a large number of articles from Wikipedia using your own code (using a web crawler, for instance) is forbidden, if you're asking about that.

Moayad Mardini