robots.txt details

saved: size: 1021 kb md5 checksum: f822f8396d50029ab761ed110406af6f

websites using this robots.txt

content


# Robots.txt for sueddeutsche.de
# www.robotstxt.org/
# www.google.com/support/webmasters/bin/answer.py?hl=en&answer=156449

# Exclude all other stuff for CRE tracking

User-agent: *
Disallow: /cre-1.0/tracking/*.js$

User-agent: *
Disallow: /uss/

# Exclude SEO-Tools & SPAM-Bots

User-agent: backlink-check.de
Disallow: /

User-agent: BacklinkCrawler
Disallow: /

User-agent: ExtractorPro
Disallow: /

User-agent: Fasterfox
Disallow: /

User-agent: LinkextractorPro
Disallow: /

User-agent: LinkWalker
Disallow: /

User-agent: MJ12bot
Disallow: /

User-agent: Openbot
Disallow: /

User-agent: rogerbot 
Disallow: /

User-agent: searchpreview
Disallow: /

User-agent: SemrushBot
Disallow: /

User-agent: SEODAT
Disallow: /

User-agent: SEOENGBot
Disallow: /

User-agent: SEOkicks-Robot
Disallow: /

User-agent: sistrix
Disallow: /

User-agent: True_Robot
Disallow: /

User-agent: URL Control
Disallow: /

User-agent: URL_Spider_Pro
Disallow: /

User-agent: xovi
Disallow: /

# Uber Metrics
User-agent: um-IC
Disallow: /