[NTLUG:Discuss] robots.txt / also Nimda & CodeRed fighting...

Darin W. Smith darin_ext at darinsmith.net
Wed Mar 26 16:35:23 CST 2003


I use:

User-agent: *
Disallow: /

in my robots.txt to not allow robots to search my site at all.  I've seen 
several try.  All of them I've seen in my logs have honored robots.txt.  It 
was their requests for the file that first alerted me to such a file.

I do use passwords on anything privy, or anything that is particularly 
bandwidth-consuming that I don't want the whole world to be pulling 
down...like family photo albums, for both reasons.

Family can get in, and I setup another login for guests I want to let in.  
Only those with a "need to know" are allowed in.

Any other attbi (now Comcast) subscribers noticing a huge upswing in the 
number of Nimda and CodeRed infested machines since the comcast deal was 
completed?  Was the upsurge just co-incidental, since (for my stuff, at 
least) everything is still attbi.com (i.e., no real change in 
administration of the network...just change in ownership)?

I've gone from seeing about a dozen different IPs (mostly other attbi.com 
customers) requesting default.ida in a week's time, to about 3 dozen 
different IPs (also, mostly other attbi.com customers).

I've considered using a 'doze box to do a NET SEND xxx.xxx.xxx.xxx "Please 
scan your computer.  You have a virus/worm." to all those IP's.  Think I'd 
get in trouble?  It would reveal my IP to them.  This is assuming that they 
have the messaging service open...which most probably will since they 
haven't bothered to apply any patches or scan for viruses.

Think it would do any good?

Since I have my webserver (Linux box - Mandrake 8.1 but soon migrating to 
something not French and not bankrupt) setup to automatically find and 
block http traffic from those IPs, I'm not all that concerned about my box. 
 I'm more concerned about those people not knowing that they are 
unwittingly spreading junk like this around to other 'doze users.

Couldn't Comcast automatically block this stuff with packet filters at 
their routers?  Couldn't they also pick off the IP address of infected 
machines and notify the owners that they need to run a virus scan?  It 
seems that would be pretty easy to automate.

In the past, I have sent lists of the attbi.com IP's that are this way to 
abuse at attbi.com or abuse at comcast.com, but I don't think they really care.  
I care, partially because that stuff winds up eating lots of bandwidth.

D!

On 26 Mar 2003 16:08:20 -0600, MadHat <madhat at unspecific.com> wrote:

>
> On Tue, 2003-03-25 at 20:23, Terry Hancock wrote:
>> On Tuesday 25 March 2003 05:04 pm, David Ross wrote:
>> > I understand that this has to do with search engines,but what should a 
>> proper > robots.txt file contain? and what file permissions should it 
>> have (444,400??)
>>
>> I don't think file permissions are relevant -- if your web server can 
>> serve it, that's all that counts.
>>
>> Here's an example of what one looks like -- it's basically a list of 
>> things you don't want the spiders to search:
>>
>
> Also, if you really want to protect something, put a password on it. Not 
> only will some spiders traverse past robots.txt, some people use it
> to find the "good stuff" on your site and more than one person has been
> bitten by this.
>



-- 
D!
Darin W. Smith
AIM: JediGrover




More information about the Discuss mailing list