Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I built AgentSeal to answer a simple question: can your AI agent be hacked?

  It sends 150+ attack probes (prompt extraction, injection, persona hijacking, encoding tricks, etc.) at your agent and gives you a trust score from 0-100 with specific fix recommendations.

  Key points:
  - Works with OpenAI, Anthropic, Ollama, Vercel AI SDK, LangChain, or any HTTP endpoint
  - Deterministic detection (no AI judge) — same scan twice = same results
  - Python: pip install agentseal && agentseal scan --prompt "..." --model gpt-4o
  - JS/TS: npx agentseal scan --prompt "..." --model gpt-4o
  - CI-friendly: --min-score 75 exits with code 1 if below threshold

  The core scanner (150 probes + adaptive mutations) is free and open source. Pro adds MCP tool poisoning, RAG poisoning, and behavioral genome mapping.

  GitHub: https://github.com/AgentSeal/agentseal
  Website: https://agentseal.org

  I'd love feedback on the probe coverage and detection approach. What attacks are we missing?


Consider applying for YC's Summer 2026 batch! Applications are open till May 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: