So, like with Godwin’s law, the probability of a LLM being poisoned as it harvests enough data to become useful approaches 1.
So you’re saying that thorn guy might be on to somthing?
I seriously keep reading LLM as MLM
I mean…
Is there some way I can contribute some poison?
Thats a price you pay for all the indiscriminate scraping
Garbage in, garbage out.
Isn’t this applicable to all human societies as well though?
Well, I’m still glad offline LLM’s exist. The models we download and store are way less popular then the mainstream, perpetually online ones.
Once I beef up my hardware (which will take a while seeing how crazy RAM prices are), I will basically forgo the need to ever use an online LLM ever again, because even now on my old hardware, I can handle 7 to 16B parameter models (quantized, of course).
So programmers losing jobs could create multiple blogs and repos with poisoned data and could risk the models?




