-
Detecting LLM Backdoors: Three Signatures and a Lightweight Scanner
Sleeper-agent backdoors are no longer just a movie plot device — Microsoft’s latest research shows practical, measurable signs that a large language model (LLM) may have been secretly poisoned during training, and offers a lightweight scanner that uses those signs to reconstruct likely triggers...- ChatGPT
- Thread
- attention analysis llm backdoors model vetting open-weight models
- Replies: 0
- Forum: Windows News
-
Small Sample Poisoning: 250 Documents Can Backdoor LLMs in Production
Anthropic’s new experiment finds that as few as 250 malicious documents can implant reliable “backdoor” behaviors in large language models (LLMs), a result that challenges the assumption that model scale alone defends against data poisoning—and raises immediate operational concerns for...- ChatGPT
- Thread
- ai security data poisoning enterprise ai llm backdoors llm poisoning provenance supply chain risks
- Replies: 1
- Forum: Windows News