views:

251

answers:

7

Can anyone recommend a tool for downloading an entire website. It is my own organisation's website, however I can't get ftp access as important people are on holiday.

I would like to download the entire site exactly as it appears, i.e. same folder structure, image names. I have tried at least 4 different tools using google and download.com, none do what I want. Either they fail to download the files altogether, or they download everything into one file in some proprietary format which is no good.

If you know a good solution, you will save me a few hours.

+3  A: 

Try GNU wget, it's more or less the standard choice.

unwind
+7  A: 

GNU wget is a nice "mirror tool";

 $ wget -m www.target.com

why -m?

‘-m’ ‘--mirror’ Turn on options suitable for mirroring. This option turns on recursion and time-stamping, > sets infinite recursion depth and keeps ftp directory listings. It is currently equivalent > to ‘-r -N -l inf --no-remove-listing’.

dfa
+1  A: 

I've usually used HTTrack. wget has the unfortunate habit of ignoring any URLs in CSS. So essentially you're getting your site sans CSS-defined images.

Joey
at least under windows it worked a lot smoother than wget for downloading a complete wiki. wget ran into problems creating files with special characters
HerdplattenToni
Thanks, I'll try this as well
Ankur
HTTrack works very well
Ankur
+1  A: 

I second HTTrack. You will be able to correctly 'mirror' a site.

sybreon
A: 

Check out PageNest (formerly webstripper). I used the grandfather webstripper some years ago and it saved my bacon many a time

Conrad
A: 

take a look at the curl command under linux -also available for cygwin

Luixv
A: 

Try surfoffline.

Its a good tool.

Regards!

MRFerocius