A proposed (and increasingly adopted) file format similar to robots.txt, where a website can give permissions, instructions, and metadata for Large Language Models to crawl, use, or cite its content. May include licensing, attribution requirements, or allow/disallow rules.