October 11, 2025
Anthropic has warned that even a few poisoned samples in a dataset can compromise an AI model. A joint study with the UK AI Security Institute found that as few as 250 malicious documents can implant backdoors in LLMs up to 13B parameters, proving model size offers no protection.
Anthropic has warned that even a few poisoned samples in a dataset can compromise an AI model. A joint study with the UK AI Security Institute found that as few as 250 malicious documents can implant backdoors in LLMs up to 13B parameters, proving model size offers no protection.