I am trying to make a program that can convert a series of manga scans into one pdf file, and I don't want to have to attempt to download the picture to determine if I have the right url. Is there a shell scripting command that I can use to just check if a web page exists?
+5
A:
Under a *NIX, you can use curl
to issue a simple HEAD
request (HEAD
only asks for the headers, not the page body):
curl --head http://myurl/
Then you can take only the first line, which contains the HTTP status code (200 OK, 404 Not Found, etc.):
curl -s --head http://myurl/ | head -n 1
And then check if you got a decent response (status code is 200 or 3**):
curl -s --head http://myurl/ | head -n 1 | grep "HTTP/1.[01] [23].."
This will output the first line if the status code is okay, or nothing if it isn't. You can also pipe that to /dev/null to get no output, and use $?
to determine if it worked or no:
curl -s --head http://myurl/ | head -n 1 | grep "HTTP/1.[01] [23].." > /dev/null
# on success (page exists), $? will be 0; on failure (page does not exist or
# is unreachable), $? will be 1
EDIT -s
simply tells curl
to not show a "progress bar".
zneak
2010-05-27 19:59:17
Brilliant. I wouldn't have thought of just asking for the headers. +1
Norman Ramsey
2010-05-27 22:19:20
A:
wget or cURL will do the job. See here wget or cURL for details and download locations. Supply the URL to these command-line tools and check the response.
jschulenklopper
2010-05-27 20:01:19