Anthropic’s Mythos Claims Questioned by Cybersecurity Insider
What Happened
The ability of Anthropic’s Mythos AI model to uncover cyber vulnerabilities has the government and institutions worried, with a limited group of firms testing the model internally. But Jaya Baloo, COO and CISO of cyber firm Aisle, says its testing shows cheap open-source models can find the same bug
Our Take
Aisle CISO Jaya Baloo says internal testing found cheap open-source models identify the same vulnerabilities Anthropic claims are unique to Mythos.
Teams building automated CVE triage or exploit-discovery pipelines are paying frontier-model prices for a capability moat that may not exist. Running Mistral or Llama-3 against your exploit-finding benchmark before signing a frontier model contract is not a shortcut — it is due diligence.
Red teams running automated vuln discovery at scale should run this benchmark themselves. Teams using Claude for general coding assistance or RAG are unaffected.
What To Do
Benchmark Mistral-7B or Llama-3 against your frontier model on internal vuln discovery data instead of assuming capability gaps exist — the price differential is 50x and the moat may not.
Builder's Brief
What Skeptics Say
One CISO's internal test isn't a controlled benchmark — Aisle may have tested a vulnerability class where open-source models already perform well, and the claim has no published methodology.
2 comments
if open-source can find the same bugs that's not reassuring, that's MORE terrifying
convenient that the firm saying mythos is overhyped is also selling their own testing services
Cited By
React
Get the weekly AI digest
The stories that matter, with a builder's perspective. Every Thursday.