Take the time to read this.
Anthropic: 250 Documents Can Permanently Corrupt Any AI Model
Someone can permanently corrupt any AI model in the world right now.
Not by hacking it. Not by breaking its security. By publishing 250 documents on the internet.
That is the finding from Anthropic, the UK AI Security Institute, and the Alan Turing Institute — released in October 2025 as the largest data poisoning study ever conducted.

Here is what data poisoning actually means.
Every AI model learns from billions of documents scraped from the internet. If someone can plant corrupted documents in that pool before training begins, they can secretly teach the model to behave in specific, harmful ways when it encounters a particular trigger phrase. The model learns the backdoor during training. It carries it forever. It does not know it is there.
Researchers have known about this attack for years. The assumption was that it required controlling a large percentage of training data — millions of documents — to work on a big model. The bigger the model, the more poisoning you would need (they incorrectly thought).
This study proved that assumption completely wrong.
The researchers trained models of four different sizes — from 600 million to 13 billion parameters. They slipped in either 100, 250, or 500 malicious documents. Each poisoned document looked like a normal web page at first — a short extract of legitimate text — and then contained a hidden trigger phrase followed by gibberish.
100 documents: insufficient. The backdoor did not reliably form.
250 documents: success. Every model, at every size, was permanently backdoored. 500 documents: same result as 250. The number was constant regardless of model size. A model trained on 260 billion tokens needed the same 250 poisoned documents as a model trained on 12 billion. Scale offered zero protection.
Read the entire essay here. It’s interesting, and extremely important.
Does this mean that all models ARE poisoned? Of course not, but it does cause a great deal of concern about the information we get, and the issues that could arise for companies and individuals that implement these models in their workflows. Absolutely ZERO models are immune to this. Offline models. ChatGPT. Grok. AI models have been poisoned BY DESIGN. I wonder why…

Leave a Reply to Hammers Thor Cancel reply