robots.txt details

saved: size: 1627 kb md5 checksum: 33300677f41f958180913170d424918e

websites using this robots.txt

content


# Robots.txt for sueddeutsche.de
# www.robotstxt.org/
# www.google.com/support/webmasters/bin/answer.py?hl=en&answer=156449

# Exclude all other stuff for CRE tracking

User-agent: *
Disallow: /uss
Disallow: /v1/subscriptioninfo
Disallow: /cdn_sz_mob/live/iqadcontroller.js.gz
Disallow: /cdn_sz/live/iqadcontroller.js.gz 
Disallow: /cre-1.0/tracking/*.js$
Disallow: /text-to-speech/

# Exclude SEO-Tools & SPAM-Bots

User-agent: backlink-check.de
Disallow: /

User-agent: BacklinkCrawler
Disallow: /

User-agent: ExtractorPro
Disallow: /

User-agent: Fasterfox
Disallow: /

User-agent: LinkextractorPro
Disallow: /

User-agent: LinkWalker
Disallow: /

User-agent: MJ12bot
Disallow: /

User-agent: Openbot
Disallow: /

User-agent: rogerbot 
Disallow: /

User-agent: searchpreview
Disallow: /

User-agent: SemrushBot
Disallow: /

User-agent: SEODAT
Disallow: /

User-agent: SEOENGBot
Disallow: /

User-agent: SEOkicks-Robot
Disallow: /

User-agent: True_Robot
Disallow: /

User-agent: URL Control
Disallow: /

User-agent: URL_Spider_Pro
Disallow: /

User-agent: xovi
Disallow: /

User-agent: GPTBot
Disallow: /

# Uber Metrics
User-agent: um-IC
Disallow: /

# Legal notice: SZ.de expressly reserves the right to use its content for commercial text and data mining (ยง 44 b UrhG).
# The use of robots or other automated means to access SZ.de or collect or mine data without
# the express permission of SZ.de is strictly prohibited.
# SZ.de may, in its discretion, permit certain automated access to certain SZ.de pages,
# If you would like to apply for permission to crawl SZ.de, collect or use data, please email syndication@sz.de