roland wrote:
yep, thats the reason why i suggested a webring of spamtraps would do
better and the addresses be generated from a wide list of word
combination. just imagine the how many combination could be done with
this set of data

rule: [ a | a+b | a+b+c | a+c | ... | b+a ] + @ + [ domain ].[level]

where:

a, b ,c ..: this, that, free, sun, ram, bot, mail, fish, stick, 33, big,
flower
domain : big, stick, homer, biz, temp, duch, pleht
level : com, biz, net, org, mil

the list could be customized per each website. i dont see how the
crawler could take all those words into consideration. they can remove
the invalid mails when it bounce but i think the one we are discussing
right now will guarantee that they will have an adequate supply for a
very long time. imagine a webring of 500 sites linking one another.

500 websites (pages) in a webring would take a decent crawler no more than 2 hours to process. Believe me, they are NOT using DSL or Cable!!!
Serial processing of 500 web pages @ 10 seconds per page (boy is that long!) is 5000 seconds, that's not more than 2 hours! Than they match what ever they found against a local DNS server with enough cache. Try:
# dig mx free.fr <Enter>
If you have a Unix/BSD/Linux box one line somewhere:

digitaldaemon# dig mx free.fr

; <<>> DiG 8.3 <<>> mx free.fr
;; res options: init recurs defnam dnsrch
;; got answer:
;; ->>HEADER<<- opcode: QUERY, status: NOERROR, id: 2
;; flags: qr aa rd ra; QUERY: 1, ANSWER: 10, AUTHORITY: 2, ADDITIONAL: 12
;; QUERY SECTION:
;;      free.fr, type = MX, class = IN

;; ANSWER SECTION:
free.fr.                1D IN MX        10 mx.free.fr.
free.fr.                1D IN MX        20 mrelay2-1.free.fr.
free.fr.                1D IN MX        20 mrelay2-2.free.fr.
free.fr.                1D IN MX        20 mx1-1.free.fr.
free.fr.                1D IN MX        50 mrelay3-2.free.fr.
free.fr.                1D IN MX        50 mrelay4-2.free.fr.
free.fr.                1D IN MX        50 mrelay1-1.free.fr.
free.fr.                1D IN MX        50 mrelay1-2.free.fr.
free.fr.                1D IN MX        60 mrelay3-1.free.fr.
free.fr.                1D IN MX        90 ns1.proxad.net.

;; AUTHORITY SECTION:
free.fr.                1D IN NS        ns0.proxad.net.
free.fr.                1D IN NS        ns1.proxad.net.

;; ADDITIONAL SECTION:
mx.free.fr.             15M IN A        213.228.0.1
mx.free.fr.             15M IN A        213.228.0.129
mx.free.fr.             15M IN A        213.228.0.13
mx.free.fr.             15M IN A        213.228.0.131
mx.free.fr.             15M IN A        213.228.0.166
mx.free.fr.             15M IN A        213.228.0.175
mx.free.fr.             15M IN A        213.228.0.65
mrelay2-1.free.fr.      1D IN A         213.228.0.13
mrelay2-2.free.fr.      1D IN A         213.228.0.131
mx1-1.free.fr.          1D IN A         213.228.0.65
mrelay3-2.free.fr.      1D IN A         213.228.0.166
mrelay4-2.free.fr.      1D IN A         213.228.0.175

;; Total query time: 127 msec
;; FROM: digitaldaemon.com to SERVER: default -- 63.105.9.35
;; WHEN: Thu Jun  5 10:04:24 2003
;; MSG SIZE  sent: 25  rcvd: 502

This is done with the 'dig' progam total query time is 127 msec's!!!!
Now they know whether or not the found domain actually has a MX record, or not... If not, they can just drop the address from the list.

Also crawlers do *not* browse the web like we do. They just process 'text' oriented files and run several (read hundreds or thousands) of threads/processes at the same time.

So, the only thing you would be able to actually make a difference with is using existing domain names. Not a good idea as owners of those domains might have a catch all and than receive the same SPAM over and over again. Soon, the providers will all change their systems so their SMTP servers only accept email to addresses that actually do exist and *deny* receipt of anything else with the usual 550 error.

So, in the end, what are we actually creating with stuff like this??? Nothing, we just have crawlers/spiders consume more bandwidth to read all the pages. The crawlers' DNS matcher consume more bandwidth to check for DNS. The bulk mailer consume more bandwidth to send all the email. The internet consume more bandwidth to deal with all the bounces, double bounces, etc.

Last, 500 pages with each a 1,000 email addresses is 500,000 email addresses. I hate to tell you, but that's only 1.5% of the total email addresses I have... <sigh> Would you honestly think that anyone would process the bounces for numbers like that manually???

ManiaC++
Jan Knepper