Skip to content

Adding RSL to robots.txt

RSL expands the limited Allow and Disallow directives of robots.txt files with more granular controls that define machine-readable licensing terms and usage restrictions.

RSL-compliant robots.txt files must include a new License directive that links to an RSL license file that defines machine-readable licensing terms for the website. The License rule is not tied to any specific user agent and must be followed by all crawlers before accessing or processing website’s content.

yaml
License: [absoluteURL]

[absoluteURL] must point to the location of a valid RSL license or feed file. It must be a fully qualified URL, including the protocol and host, and does not have to be URL-encoded. The URL does not have to be on the same host as the robots.txt file.

Example robots.txt File

yaml
License: https://your-website.com/license.xml

User-agent: Googlebot
Allow: /

RSL™, Really Simple Licensing™, and the RSL Logo are trademarks of RSL Foundry. Terms of Service. Privacy.