I feel called out, lmao. Iām building an agentic framework for automated pentesting as part of an internal AppSec R&D initiative. My companyās letting me run wild with infrastructure and Bedrock usage (bless their optimism). Iāve been throwing together some admittedly questionable prototypes to see what sticks.
The setup is pretty basic: S3 for docs and code base, pgvector on RDS for embeddings, Claude/Titan for retrieval and reasoning. It works in the sense that data flows through and responses come out⦠but the agents themselves are kind of a mess.
They think theyāve found a bug, usually something like a permissive IAM policy or a questionable API call, and just latch onto it. They tunnel hard, write up something that sounds plausible, and stop there. No lateral exploration, no attempt to validate anything in a dev environment despite having MCP tools to access internal resources, and definitely no real exploitation logic.
Iāve tried giving them tools like CodeQL, semgrep and Joern, but thatās been pretty disappointing. They can run basic queries, but all they surface are noisy false positives, and they canāt reason their way out of why it might be a false positive early on. Thereās no actual taint analysis or path tracing, just surface-level matching and overconfident summaries. I feel like Iām duct-taping GPT-4 to a security scanner and hoping for insight.
Iāve experimented with splitting agents into roles (finder, validator, PoC author, code auditor, super uber hacker man), giving them memory, injecting skepticism, etc., but it still feels like Iām missing something fundamental.
If cost isnāt an issue, how would you structure this differently? How do you actually get agents to do persistent, skeptical, multi-stage analysis, especially in security contexts where you need depth and proof, not just plausible-sounding guesses and long ass reports on false positives?
Seems like you need a way to dictate structured workflows, in lieu of actually being able to train them up as soc analyst. Sounds like a fun problem!