Anthony D'Onofrio

Anthony D'Onofrio

Maker of wraith.sh

About

Principal security consultant and founder at Harbinger Security. Background in microbiology and genomics (PhD, Northeastern) before pivoting to security. Building Wraith Academy to teach hands-on AI pentesting against live LLM agents.

Badges

Tastemaker
Tastemaker
Gone streaking
Gone streaking

Forums

What's the weakest part of AI agent security right now?

Curious what the PH community thinks. There's a lot of noise about LLM safety, but "safety" and "security" get conflated constantly, and the actual attack surface on an agent in production is its own thing.

From where I sit, the most under-addressed failure modes seem to be:

  • Indirect prompt injection via retrieved content (RAG sources, tool outputs, even user-uploaded docs)

  • Tool/function abuse where the agent happily calls something it shouldn't

  • Trust boundaries between the agent and the systems it can touch

Wraith Academy - Learn AI pentesting by attacking real LLM agents

Free hands-on CTFs where you attack live LLM agents: prompt injection, tool abuse, RAG poisoning, and more. Built for security pros and AI engineers who want to understand how AI systems actually break, not just read about it.
View more