Does anybody have any tips for combating site suckers
- of the variety that use wget and other site
downloaders; hit sites hard; ignore robots.txt; and
even set User-agent to commonly used browser strings?
I’m using Lighty and SCGI. Back with Apache, there
were various modules that could limit total bandwidth,
but I never found anything that worked well.
Yahoo! DSL ? Something to write home about.
Just $16.99/mo. or less.