Discussion about this post

User's avatar
Illya Nokhrin's avatar

RSL and Web bot auth both seem interesting and I'm looking forward to seeing how they develop. But, as someone whose site has been taken offline by over-zealous scrapers more than once now I have little hope that AI companies will be good enough citizens to actually respect something like RSL unless folks are willing to blockade their content with logins and/or paywalls. Robots.txt is now pretty much routinely ignored IME, and even CloudFlare's bot protection struggles to keep up with the strategies employed by scraper bots. We've resorted to rate-limiting and invoking JS challenges for specific ASNs to keep things running on our end, though I'm sure we'll have to change those tactics as scrapers change theirs.

Sharif Islam's avatar

Interesting! I am curious where W3C’s work on the Open Digital Rights Language (ODRL) fits into this picture. I think ODRL is more about metadata descriptions and less about automation or enforcement. The real challenge seems to be coupling rules with enforcement (as you point out). But setting up license servers and related infrastructure could be a hurdle for many.

Your point about adoption is important which of these approaches will the big AI companies actually take on board? This reminds me a bit of how independent mail servers used to be common (I maintained one myself years ago), but over time most people shifted to a few large providers. I wonder if we’ll see a similar centralisation dynamic here.

No posts

Ready for more?