# Robots.txt for sueddeutsche.de # www.robotstxt.org/ # www.google.com/support/webmasters/bin/answer.py?hl=en&answer=156449 # Exclude all other stuff for CRE tracking User-agent: * Disallow: /uss Disallow: /v1/subscriptioninfo Disallow: /cdn_sz_mob/live/iqadcontroller.js.gz Disallow: /cdn_sz/live/iqadcontroller.js.gz Disallow: /cre-1.0/tracking/*.js$ Disallow: /text-to-speech/ Disallow: /pay/piano/ # Exclude SEO-Tools & SPAM-Bots User-agent: backlink-check.de Disallow: / User-agent: BacklinkCrawler Disallow: / User-agent: ExtractorPro Disallow: / User-agent: Fasterfox Disallow: / User-agent: LinkextractorPro Disallow: / User-agent: LinkWalker Disallow: / User-agent: MJ12bot Disallow: / User-agent: Openbot Disallow: / User-agent: rogerbot Disallow: / User-agent: searchpreview Disallow: / User-agent: SemrushBot Disallow: / User-agent: SEODAT Disallow: / User-agent: SEOENGBot Disallow: / User-agent: SEOkicks-Robot Disallow: / User-agent: True_Robot Disallow: / User-agent: URL Control Disallow: / User-agent: URL_Spider_Pro Disallow: / User-agent: xovi Disallow: / User-agent: GPTBot Disallow: / # Uber Metrics User-agent: um-IC Disallow: / # Googles generative AI crawlers User-agent: Google-Extended Disallow: / # Legal notice: SZ.de expressly reserves the right to use its content for commercial text and data mining (ยง 44 b UrhG). # The use of robots or other automated means to access SZ.de or collect or mine data without # the express permission of SZ.de is strictly prohibited. # SZ.de may, in its discretion, permit certain automated access to certain SZ.de pages, # If you would like to apply for permission to crawl SZ.de, collect or use data, please email syndication@sz.de