Reviews

Researchers find just 250 malicious documents can leave LLMs vulnerable to backdoors

October 9, 2025 2 min read Engadget
Article Data

Anthropic released a report on how attackers can influence the development of a large language model. The study centered on a type of attack called poisoning. A bad actor doesn't need to control a percentage of the pretraining materials to get the LLM to be poisoned.

Read more on Engadget

Loading next article