
Crawler, Spider, Robots und Bots sperren
- DARO75
- Thread is marked as Resolved.
-
-
Der Trick nennt sich .htaccess*. Dazu gibt es einige Beiträge hier, also einfach mal die Suchfunktion und Google benutzen, dann solltest du eigentlich ziemlich schnell fündig werden, egal um wen oder was genau es dir gerade geht.
* Funktioniert nur mit Apache.
-
Inhalt der Datei robots.txt
Code
Display MoreUser-agent: SEOkicks Disallow: / User-agent: SEOkicks-Robot Disallow: / User-agent: sistrix Disallow: / User-agent: MajesticSEO Disallow: / User-agent: BacklinkCrawler Disallow: / User-agent: xovi Disallow: / User-agent: XoviBot Disallow: / User-agent: MJ12bot Disallow: / User-agent: spbot Disallow: / User-agent: SearchmetricsBot Disallow: / User-agent: search17 Disallow: / User-agent: AhrefsBot Disallow: / User-agent: ia_archiver Disallow: / User-agent: TurnitinBot Disallow: / User-agent: SlySearch Disallow: / User-agent: findlinks Disallow: / User-agent: magpie-crawler Disallow: / User-agent: Pixray-Seeker Disallow: / User-agent: 008 Disallow: / User-agent: Ezooms Disallow: / User-agent: lb-spider Disallow: / User-agent: WBSearchBot Disallow: / User-agent: psbot Disallow: / User-agent: HuaweiSymantecSpider Disallow: / User-agent: EC2LinkFinder Disallow: / User-agent: htdig Disallow: / User-agent: SemrushBot Disallow: / User-agent: discobot Disallow: / User-agent: linkdex.com Disallow: / User-agent: SeznamBot Disallow: / User-agent: EdisterBot Disallow: / User-agent: SWEBot Disallow: / User-agent: picmole Disallow: / User-agent: Yeti Disallow: / User-agent: Yeti-Mobile Disallow: / User-agent: PagePeeker Disallow: / User-agent: CatchBot Disallow: / User-agent: yacybot Disallow: / User-agent: netEstateNECrawler Disallow: / User-agent: SurveyBot Disallow: / User-agent: COMODOSSLChecker Disallow: / User-agent: Comodo-Certificates-Spider Disallow: / User-agent: gonzo Disallow: / User-agent: schrein Disallow: / User-agent: AfiliasWebMiningTool Disallow: / User-agent: suggybot Disallow: / User-agent: bdbrandprotect Disallow: / User-agent: BPImageWalker Disallow: / User-agent: Updownerbot Disallow: / User-agent: lex Disallow: / User-agent: ContentCrawler Disallow: / User-agent: DCPbot Disallow: / User-agent: KaloogaBot Disallow: / User-agent: MLBot Disallow: / User-agent: iCjobs Disallow: / User-agent: oBot Disallow: / User-agent: WebmasterCoffee Disallow: / User-agent: Qualidator Disallow: / User-agent: Webinator Disallow: / User-agent: Scooter Disallow: / User-agent: thunderstone Disallow: / User-agent: larbin Disallow: / User-agent: OpidooBOT Disallow: / User-agent: ips-agent Disallow: / User-agent: TinEye Disallow: / User-agent: UnisterBot Disallow: / User-agent: Unister Disallow: / User-agent: ReverseGet Disallow: / User-agent: DotBot Disallow: /
Inhalt der Datei .htaccess.
Apache Configuration
Display MoreRewriteEngine On RewriteCond %{HTTP_USER_AGENT} ^SEOkicks [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^SEOkicks-Robot [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^sistrix [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^MajesticSEO [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^BacklinkCrawler [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^xovi [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^XoviBot [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^MJ12bot [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^spbot [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^SearchmetricsBot [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^search17 [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^AhrefsBot [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^ia_archiver [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^TurnitinBot [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^SlySearch [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^findlinks [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^magpie-crawler [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^Pixray-Seeker [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^008 [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^Ezooms [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^lb-spider [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^WBSearchBot [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^psbot [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^HuaweiSymantecSpider [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^EC2LinkFinder [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^htdig [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^SemrushBot [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^discobot [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^linkdex.com [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^SeznamBot [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^EdisterBot [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^SWEBot [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^picmole [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^Yeti [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^Yeti-Mobile [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^PagePeeker [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^CatchBot [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^yacybot [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^netEstateNECrawler [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^SurveyBot [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^COMODOSSLChecker [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^Comodo-Certificates-Spider [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^gonzo [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^schrein [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^AfiliasWebMiningTool [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^suggybot [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^bdbrandprotect [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^BPImageWalker [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^Updownerbot [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^lex [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^ContentCrawler [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^DCPbot [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^KaloogaBot [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^MLBot [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^iCjobs [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^oBot [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^WebmasterCoffee [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^Qualidator [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^Webinator [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^Scooter [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^thunderstone [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^larbin [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^OpidooBOT [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^ips-agent [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^TinEye [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^UnisterBot [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^Unister [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^ReverseGet [NC] RewriteRule ^.* - [F,L] RewriteCond %{HTTP_USER_AGENT} ^DotBot [NC] RewriteRule ^.* - [F,L]
-
-
@sonic hast du zufällig eine aktuelle von 2017? Danke
Nein, das sind die die ich so drinne hab. Weitere musst Du dir da eben selbst suchen b.z.w hinzufügen.
-
-
den genauen Pfad zur .htaccess Datei geben?
Die .htaccess gehört ins Root Verzeichnis deines Forums.
-
Die .htaccess gehört ins Root Verzeichnis deines Forums.
ich habe kein Forum sondern nur eine Filebase. Gibt es da unterschiede?
-
Nein.
-
Hallo zusammen,
weiss jemand zufällig von Euch was "Zend_Http_Client" als Besucher auf den Webseiten zu suchen haben?
IP-Adresse: 193.202.110.17
Browser: Zend_Http_Client
kann man sie mittels .htaccess aussperren?
Danke und Grüße, Dan
-
User-Agent per .htaccess bannen
Quote# User-Agent Blacklist
SetEnvIfNoCase User-Agent "^Mozilla/5.0 (Windows NT 6.1; WOW64; rv:40.0) Gecko/20100101 Firefox/40.1" bad_bot
Klappt so leider nicht, könnte es bitte jemand korrigieren?
Danke und Grüße
Dan
-
Zend_Http_Client
Das ist ein Client des PHP-Frameworks Zend Framework.
-
Hallo,möchte gerne den SemrushBot blokieren,nur finde ich nirgends die .htaccess und auch nicht die robots.txt.
Muss ich die Dateien selber erstellen und dann einfach ins Root Verzeichnis packen?
-
selbst erstellen, wobei die Version via robots.txt kannst dir sparen
wenn, dann nur über die .htaccess, vorausgesetzt du setzt den Apache Webserver ein.
-
-
Apache Webserver?
Da habe ich die Datei nicht.
-
schau mal im Ordner beautysims
da müsste eine sein
-
Ne,nichts da.
-
Der Forum Ordner war schon richtig
Lässt du auch bei deinem Client versteckte Dateien anzeigen?
Ansonsten kannst du auch einfach eine .htaccess selbst erstellen das ist reiner Text
-
Gute Frage ^^. Wo kann ich das nachschauen?
Beim Text kann ich dann einfach den aus dem 3. Beitrag nehmen?
Participate now!
Don’t have an account yet? Register yourself now and be a part of our community!