easy guess that length breaks some legacy stuff
but every robots.txt should have a auto-ban trap line
ie. crawl it and die
basically a script that puts the requesting IP into firewall
of course it's possible to abuse that so it has to be monitored
replies(2):