[NTLUG:Discuss] copying web documents

anindya Chakraborty anindya.links at gmail.com
Thu May 18 22:16:29 CDT 2006


If you are open to use something apart than wget you can try httrack, it's
really nice for downloading websites.

- anindya

On 5/18/06, steve <sjbaker1 at airmail.net> wrote:
>
> Stuart Johnston wrote:
>
> > Apparently, wget respects the robots.txt file which causes problems with
> > this site.  But here's what I did.
>
> wget only respects robots.txt when it's doing recursive loading.  For a
> single page - it's OK.
>
> But you have the source code...so you know what to do!
>
>
> _______________________________________________
> http://ntlug.pmichaud.com/mailman/listinfo/discuss
>


More information about the Discuss mailing list