[NTLUG:Discuss] Use wget to generate all urls from a site?
Wayne Walker
wwalker at bybent.com
Sun Aug 5 16:39:22 CDT 2007
wget --mirror --no-verbose http://www.bybent.com/
or
wget --mirror --no-verbose http://www.bybent.com/ 2>&1 | sed -e 's/^.* URL://' -e 's/ \[.*//'
You must "rm -rf www.bybent.com" before subsequent runs as it won't
retry existing pages.
NOTE: this will make a local copy of the site at the same time.
On Sun, Aug 05, 2007 at 03:51:17PM -0500, Neil Aggarwal wrote:
> Hello:
>
> I have been trying to figure this out but cant:
>
> Is there a way to get a list of all urls in a site
> using wget?
>
> Thanks,
> Neil
>
>
> --
> Neil Aggarwal, (832)245-7314, www.JAMMConsulting.com
> FREE! Eliminate junk email and reclaim your inbox.
> Visit http://www.spammilter.com for details.
>
>
> _______________________________________________
> http://www.ntlug.org/mailman/listinfo/discuss
--
Wayne Walker
www.unwiredbuyer.com - when you just can't be by the computer
wwalker at bybent.com Do you use Linux?!
http://www.bybent.com Get Counted! http://counter.li.org/
Perl - http://www.perl.org/ Perl User Groups - http://www.pm.org/
Jabber: wwalker at jabber.gnumber.com AIM: lwwalkerbybent
IRC: wwalker on freenode.net
More information about the Discuss
mailing list