by Bluestein on 6/22/2024, 4:04:40 PM
by astine on 6/22/2024, 3:35:45 PM
I agree. Robots.txt is a suitable means of preventing crawlers from accidentally DOSing your site, but it doesn't really give you any protections as to how your content is used by automated services. The current anything-goes approach is just too exploitable.
by verdverm on 6/22/2024, 3:34:00 PM
After ranting about AI, the disclaimer is rich
by nuc1e0n on 6/26/2024, 4:53:59 AM
There's always range banning.
We do. Much in the same way private property is protected, we need regulation enabling the technical means to keep bad actors off private machines.-
This, back in the quaint, good, ol' days, was sufficiently implemented through the voluntary, good will, communal, neighborly agreement that robot.txt embodies.-
Unfortunately, sadly, that is no longer enough.-