An HTTP request is a request. Servers are free to rate limit or deny access
Comment on AI crawlers cause Wikimedia Commons bandwidth demands to surge 50%.
krigo666@lemmy.world 5 days ago
Laws should be passed in all countries that AI crawlers should request permission before crawling whatever target site. I haver no pity to AI “thiefs” that get their models poisoned. F…ing plague, wasn’t enough the adware and spyware…
catloaf@lemm.ee 5 days ago
taladar@sh.itjust.works 5 days ago
Rate limiting in itself requires resources that are not always available. For one thing you can only rate limit individuals you can identify so you need to keep data about past requests in memory and attach counters to them and even then that won’t help if the requests come from IPs that are easily changed.
grysbok@lemmy.sdf.org 4 days ago
Bots lie about who they are, ignore robots.txt, and come from a gazillion different IPs.
catloaf@lemm.ee 4 days ago
That’s what ddos protection is for.
chrash0@lemmy.world 5 days ago
i doubt the recent uptick in traffic is from “stealing data” for training but rather from agents scraping them for context, eg Edge Copilot, Google’s AI search, SearchGPT, etc.
poisoning the data will likely not help in this situation since there’s a human on the other side that will just do the same search again given unsatisfactory results. like how retries and timeouts can cause huge outages for web scale companies, poisoning search results will likely cause this type of traffic to increase and further increase the chances of DoS and higher bandwidth usage.
TheBlackLounge@lemm.ee 5 days ago
So? Break context scrapers till they give up, on your site or completely.
chrash0@lemmy.world 5 days ago
easily said