Brendan Dolan-Gavitt
@moyix
Followers
30K
Following
60K
Media
4K
Statuses
24K
Building offsec agents: https://t.co/G9EtnC2Gl3 PGP https://t.co/3WXr0RfRkv
Brooklyn, NY
Joined June 2008
Incredible to have helped build the first AI system to reach #1 in the US on @Hacker0x01 ! We found a LOT of great bugs :D
10
10
91
Very profound signage (the signs below gave directions to various exam classrooms)
0
2
15
Can it really be a coincidence? Smart Investors Hold Gold. Find out more at Goldhub
41
127
1K
Had a lovely time up at Columbia today talking about AI agents for offsec at the DAPLab!
0
0
31
The first exploitable vuln was found was during the demo š® This is what it looks like when AI becomes part of the workflow instead of just another tool. Try for yourself by December 26th and we will guarantee an exploit validated security finding or you don't pay. Get
0
2
7
Huge appreciation to the Seznam team! On their first demo, XBOW identified a critical vulnerability with zero access and zero prep, just autonomous offensive security doing real work for a real customer. Itās the kind of partnership that proves what matters.
1
1
13
It has been absolutely delightful to work with Seznam over the past year ā they've given us great feedback as we've built out new features, and I'm very happy they're finding XBOW valuable!
1
1
10
I know this is going to make me sound old, but how are people actually using agents to modify existing codebases? Even with Opus 4.5 I find myself rejecting every other request because it's doing something wrong
9
1
23
Okay guys that āaverage person commits three felonies per dayā thing was a warning not like a KPI you need to hit
Holy shit⦠the exploitation of CVE-2025-55182 has reached a new level. Thereās now a publicly available Chrome extension on GitHub that automatically scans for and exploits vulnerable sites as you browse. Absolutely wild. š¤¦āāļø
3
85
2K
NGL I'm mildly nervous we're doing this but if you want a risk-free way to try out XBOW it's a good deal!
3
5
32
Sorry I'm not going to be interviewed by some low-level Sonnet flunky, send me Opus if you want your questions answered
0
0
21
(TBH I haven't actually run it so maybe it's all hallucinated)
2
0
6
LMAO, found one already... and the code / writeup is most def Claude Code
2
1
26
If you are really, truly lucky, you will be in a situation where you can recognize whether the reliable specialized solution will work up front, and then fall back for cases you canāt handle. But often the problem is not that friendly
0
0
3
You will probably be surprised at how much smaller Y is than you expected, and how many wickedly difficult problems are in the (100-Y) section.
1
0
8
X will probably get bigger over the next year. You hope. If scaling continues and the labs like things adjacent to your problem
1
0
6
Right now all design problems for LLM-based apps feel like a choice between: - Solve it for 100% of the problem space at X% reliability - Solve it for Y% of the problem space at 100% reliability (And you donāt know the values of X and Y up front)
2
1
28
It's sort of dispiriting that for obvious-in-retrospect economic reasons the vast majority of interactions people will have with LLMs at any given time is with very cheap, very bad models
3
1
18
Our report from the SAGAI workshop (located with @IEEESSP ) is now out: https://t.co/SXVgLHmaRo It defines the "systems security" approach to AI agents and captures many research challenges when applying systems principles to securing AI agents! This is v1. v2 is coming soon!
arxiv.org
This paper articulates short- and long-term research problems in AI agent security and privacy, using the lens of computer systems security. This approach examines end-to-end security properties...
0
2
15