i don’t know for sure that there is zero “formal methods” in the pipeline
in discovering the 12 OpenSSL zero-day vulnerabilities, we haven’t used any formal methods. since then, we incorporated some. the (discovery --> CVE assigned --> CVE made public) pipeline is a very lagging indicator and the OpenSSL results are reflective of the state of the AISLE system approximately mid-fall 2025, prior to our use of formal methods
Thanks for flagging this, Rasool. I’ve been following the Anthropic announcement closely. It’s genuinely exciting to see a frontier lab validate that AI can find real vulnerabilities in real software at scale. The more serious players in this space, the better, since it is a genuinely large problem ⇒ far larger than any single team.
We’ve been doing this work operationally since mid/late-2025, and our experience has been that the hardest part isn’t finding some bugs but rather the hardest bugs in the most audited codebases. Earning the trust of maintainers and closing the full loop from discovery through patch acceptance is also very challenging. That’s where most of the difficulty (and most of the value) lives.
I wrote up a practitioner’s perspective on what we’ve learned, how our results compare, and what the real challenges ahead look like from our vantage point: What AI Security Research Looks Like When It Works