I need to download all pdf files from a certain domain. There are about 6000 pdf on that domain and most of them don't have an html link (either they have removed the link or they never put one in the first place).
I know there are about 6000 files because I'm googling: filetype:pdf site:*.adomain.com
However, Google lists only the first 1000 results. I believe there are two ways to achieve this:
a) Use Google. However, how I can get all 6000 results from Google? Maybe a scraper? (tried scroogle, no luck) b) Skip Google and search directly on domain for pdf files. How do I do that when most them are not linked?
I found HTTrack to be better than wget. It produces less errors and is more easy to tune.