I'm trying to fetch Wikipedia pages using LWP::Simple, but they're not coming back. This code:
#!/usr/bin/perl
use strict;
use LWP::Simple;
print get("http://en.wikipedia.org/wiki/Stack_overflow");
doesn't print anything. But if I use some other webpage, say http://www.google.com, it works fine.
Is there some other name that I shou...
I want to embed a wikipedia article into a page but I don't want all the wrapper (navigation, etc.) that sits around the articles. I saw it done here: http://www.dayah.com/periodic/. Click on an element and the iframe is displayed and links to the article only (no wrapper). So how'd they do that? Seems like JavaScript handles showing...
I try to fetch a Wikipedia article with Phython's urllib:
f = urllib.urlopen("http://en.wikipedia.org/w/index.php?title=Albert_Einstein&printable=yes")
s = f.read()
f.close()
However instead of the html page I get the following response: Error - Wikimedia Foundation:
Request: GET http://en.wikipedia.org/w/index.php?tit...
The subject of data caching within software systems must be one of the oldest ever in the software design community. However, being such a vast subject also makes it hard to locate good resources about it.
Please share your pointers to serious resources and material on this subject. Books, wikipedia articles, forums, etc.
Thanks.
/R
...
Wikipedia allows you to reproduce its content.
If you want to use Wikipedia materials in your own books/articles/websites or other publications, you can do so -- but only in compliance with the GFDL.
See here.
I am considering using some wikipedia content to pad out pages on my website. For example - just taking the first couple of par...
Hi again,
I am trying to execute this SQL command:
SELECT page.page_namespace, pagelinks.pl_namespace, COUNT(*)
FROM page, pagelinks
WHERE
(page.page_namespace <=3 OR page.page_namespace = 12
OR page.page_namespace = 13
)
AND
(pagelinks.pl_namespace <=3 OR pagelinks.pl_namespace ...
When it comes to classifying music by genre, I've found wikipedia to have more interesting genre information than most other data sources.
I seem to remember a database that collected this sort of information from Wikipedia and made it more easily accessible, but I couldn't google anything up today.
If I was to attempt to retrieve this...
I need to implement something similar to wikilinks on my site. The user is entering plain text and will enter [[asdf]] wherever there is an internal link. Only the first five examples are really applicable in the implementation I need.
Would you use regex, what expression would do this? Is there a library out there somewhere that alread...
Hi, I'm trying to use java.util.Scanner to take Wikipedia contents and use it for word based searches.
The fact is that it's all fine but when reading some words it give me errors.
Looking at code and making some check it turned out that with some words it seems
not to recognize the encoding, or so, and the content is no more readable....
I know that the wiki concept is let people from around the world contribute and everything. Now is there a way to add code ("contribute") to Wikipedia's codebase? Open-source?
For eg. I'd like to offer a poll to every Wikipedian (registered user) on their user-page asking whether they're deletionists (rather have less) or additionists (...
I want my own hosted Wiki so its fully under my control, but would like something as advanced as Wikia rather than the plain, feature-lacking Wikipedia style MediaWiki.
What software does Wikia use and can I have that for my own hosted Wiki? Is it just upgraded/added-on MediaWiki.
...
On my Wikipedia user page, I run a Wikipedia script that displays my statistics (number of pages edited, number of new pages, monthly activity etc.).
I'd like to put this info on my blog.
Can anyone recommend any API's that would allow me to do something like this?
...
Does a tool exist that can parse text and output that text, hyper-linked to Wikipedia entries for words of interest?
For example, I'd like a tool that could turn something like:
The most popular search algorithm on a
sorted list is the binary search.
Into:
The most popular search algorithm on a
sorted list is the binary se...
I have a project where in I collect all the wikipedia articles belonging to a particular category, pull out the dump from the wikipedia, and put it into our db. So I should be parsing wikipedia dump file to get the stuff done. Do we have an efficient parser to do this job. I am a python developer. So I prefer any parser in python. If not...
If I have downloaded Wikipedia XML dumps, is there any way of removing all of the internal links from within an XML file?
Thanks
...
Quite hard to describe, but if a user on my site said they were the Wikipedia user "example" what would be the best way of confirming they are indeed that Wikipedia user & not an imposter?
Hope that makes sense?!
...
Hey,
This may be a pathetically simple problem but I cannot seem to format the post webrequest/response to get data from the wikipedia api. I have posted my code below if anyone can help me see my problem.
string pgTitle = txtPageTitle.Text;
Uri address = new Uri("http://en.wikipedia.org/w/api.php");
HttpWebRequest request =...
Hi guys I wish to get information for entries I have in my database from wikipedia like for example some stadiums and country information. I'm using Zend Framework and also how would I be able to handle queries that return multiple ambiguous entries or the like.. I would like all the help I can get here...
...
I'm using curl to retrieve information from wikipedia. So far I've been successful in retrieving basic text information but I really would want to retrieve it in HTML.
Here is my code:
$s = curl_init();
$url = 'http://boss.yahooapis.com/ysearch/web/v1/site:en.wikipedia.org+'.$article_name.'?appid=myID';
curl_setopt($s,CURLOPT_URL,...
Ok guys I've downloaded the wikipedia xml dump and its a whopping 12 GB of data :\ for one table and I wanted to import it into mysql databse on my localhost - however its a humongous file 12GB and obviously navicats taking its sweet time in importing it or its more likely its hanged :(.
Is there a way to include this dump or atleast pa...