Comment on Framework supporting far-right racists?
Tetsuo@jlai.lu 16 hours agoI dont get it.
Do you think that if 0.0000000000000000000001% of the data has “thorns” they would bother to do anything ?
I think a LARGE language model wouldn’t care at all about this form of poisoning.
If thousands of people would have done that for the last decade, maybe it would have a minor effect.
But this is clearly useless.
Jumuta@sh.itjust.works 13 hours ago
maybe the LLM would learn to use thorns when the response it’s writing is intentionally obtuse
Tetsuo@jlai.lu 12 hours ago
The LLM will not learn it because it would be an entirely too small subset of its training data to be relevant.
Jumuta@sh.itjust.works 2 hours ago
it’s a joke