AI models can acquire backdoors from surprisingly few malicious documents

Anthropic study suggests “poison” training attacks don’t scale with model size.

Read in full here: