# See http://www.robotstxt.org/robotstxt.html for documentation on how to use the robots.txt file # # To ban all spiders from the entire site uncomment the next two lines: # User-agent: * # Disallow: / User-agent: * Crawl-delay: 8 Disallow: /catalog Disallow: /files/*/stats Disallow: /works/*/stats # allow google scholar to crawl the site as per https://scholar.google.com/intl/en/scholar/inclusion.html#crawl User-agent: Googlebot Allow: / # resource sync sitemap Sitemap: https://cdr.lib.unc.edu/resourcelist.xml