defense against bots..??
Hi, My site features an image repository which is available for free to all. However one has to download each image by itself. Is there a way to block download managers and miners like... GoZilla http://tribolic.com/webminer/ (web miner) etc... TIA AM -- ================================================================== Aseem Mohanty Neurobehavioral Systems Inc, 828 San Pablo Ave, Albany, CA 94706 (R) 510 7696011 (M) 510 3014871 (O) 510 5279231 ================================================================== "I saw `cout' being shifted "Hello world" times to the left and stopped right there!!" -- Steve Gonedes ==================================================================
Hi AM (??), --On Freitag, 27. September 2002 13:57 -0700 AM <list_subscriber@neurobs.com> wrote:
Hi, My site features an image repository which is available for free to all. However one has to download each image by itself.
Is there a way to block download managers and miners like... GoZilla http://tribolic.com/webminer/ (web miner) etc...
No. That means, it depends on how clever the user is, but usually no. You can write part of the html with the links to the images with obfuscated JavaScript. This might break the most "download managers". But if one uses for example Pavuk (http://www.idata.sk/~ondrej/pavuk/about.html ) You have almost no chance. Even analyzing the download rate and throttling the speed would not help. In this case you better offer your images prebundled as zip-archive to save your server from heavy load. Regards Tino
Hello AM, the problem is to decide wether a robot or a browser is downloading the page. If the robot is very simple or the robot user very unexperienced it would be possible to detect the robot simply by the user-agent. But most robots will disguise themselves as Internet Explorer. Look in your logfile. If you see robots than you can do something. But be very carefull if you don't want to exclude even the searchengine robots. Best thing to do is to set up a robots.txt for your domain and (may be in your standard_html_header) detect the user-agent and serve a different content. Don't just send nothing or an error message. That would trigger the robot and make him disguise itself as Internet Explorer. If you are really sure it will work with your audience you could try to protect your picture links with Javascript. Ulli -- Get more qualifyed targeted traffic. http://www.PublisherSEO.com/1 World Wide Web Publisher, Ulrich Wisser, Odensvag 13, S-14571 Norsborg http://www.publisher.de Tel: +46-8-53460905 Fax: +46-8-534 609 06
participants (3)
-
AM -
Tino Wildenhain -
Ulrich Wisser