Website operators are being asked to feed LLM crawlers poisoned data by a project called Poison Fountain.
The project page links to URLs which provide a practically endless stream of poisoned training data. They have determined that this approach is very effective at ultimately sabotaging the quality and accuracy of AI which has been trained on it.
Small quantities of poisoned training data can significantly damage a language model.
The page also gives suggestions on how to put the provided resources to use.
eru@mouse.chitanda.moe 2 hours ago
i would imagine companies would just filter it out
need some more clever way of hiding it or allow it to be self hosted so that it has various urls
GamingChairModel@lemmy.world 1 hour ago
If I am reading this correctly, anyone who wants to use this service can just configure their HTTP server to act as the man in the middle of the request, so that the crawler sees your URL but is retrieving poison fountain content from the poison fountain service.
If so, that means the crawlers wouldn’t be able to filter by URL because the actual handler that responds to the HTTP request doesn’t ever see the canonical URL of the poison fountain.
In other words, the handler is “self hosted” at its own URL while the stream itself comes from the same URL that the crawler never sees.