[NTLUG:Discuss] copying web documents
anindya Chakraborty
anindya.links at gmail.com
Thu May 18 22:16:29 CDT 2006
If you are open to use something apart than wget you can try httrack, it's
really nice for downloading websites.
- anindya
On 5/18/06, steve <sjbaker1 at airmail.net> wrote:
>
> Stuart Johnston wrote:
>
> > Apparently, wget respects the robots.txt file which causes problems with
> > this site. But here's what I did.
>
> wget only respects robots.txt when it's doing recursive loading. For a
> single page - it's OK.
>
> But you have the source code...so you know what to do!
>
>
> _______________________________________________
> http://ntlug.pmichaud.com/mailman/listinfo/discuss
>
More information about the Discuss
mailing list