As interesting as this is, it seems pretty trivial to overcome. If a site has a robots.txt file, then scrape it into an intermediate location; if the scraping takes "too long", set aside the website ...
With AI reshaping the web, a "handshake deal" between some of the earliest pioneers of the internet that governs the internet ...
While Google is opening up the discussion on giving credit and adhering to copyright when training large language models (LLMs) for generative AI products, their focus is on the robots.txt file.
Are large robots.txt files a problem for Google? Here's what the company says about maintaining a limit on the file size. Google addresses the subject of robots.txt files and whether it’s a good SEO ...
Large language models are trained on massive amounts of data, including the web. Google is now calling for “machine-readable means for web publisher choice and control for emerging AI and research use ...