Researchers find just 250 malicious documents can leave LLMs vulnerable to backdoors
October 9, 2025
2 min read
●
Engadget
Anthropic released a report on how attackers can influence the development of a large language model. The study centered on a type of attack called poisoning. A bad actor doesn't need to control a percentage of the pretraining materials to get the LLM to be poisoned.