Anthony D'Onofrio

Anthony D'Onofrio

Maker of wraith.sh
All activity
Free hands-on CTFs where you attack live LLM agents: prompt injection, tool abuse, RAG poisoning, and more. Built for security pros and AI engineers who want to understand how AI systems actually break, not just read about it.
Wraith Academy
Wraith AcademyLearn AI pentesting by attacking real LLM agents
Anthony D'Onofriostarted a discussion

What's the weakest part of AI agent security right now?

Curious what the PH community thinks. There's a lot of noise about LLM safety, but "safety" and "security" get conflated constantly, and the actual attack surface on an agent in production is its own thing. From where I sit, the most under-addressed failure modes seem to be: Indirect prompt injection via retrieved content (RAG sources, tool outputs, even user-uploaded docs) Tool/function abuse...