Custom robots.txt
By default, Fern serves an auto-generated robots.txt at the root of your documentation site that allows all crawlers and points to your sitemap.xml. Use the agents.robots-txt key in docs.yml to serve your own file instead — useful for opting in or out of specific AI crawlers, gating sensitive sections, or signaling preferences with the Cloudflare Content Signals Policy.
robots.txt is advisory: compliant crawlers honor your Disallow and Allow directives, but bots that ignore the protocol still reach those paths. For content that must stay private, use authentication.
robots.txt decides which crawlers can reach your site and what AI training signals you broadcast. Its companions, llms.txt and llms-full.txt, shape what AI agents receive once they crawl.