AI Cybersecurity's Jagged Frontier: Small Models vs Mythos

April 12, 2026

AI Cybersecurity's Jagged Frontier: Small Models vs Mythos

TL;DR

When Anthropic announced Claude Mythos Preview on April 7, 2026 — a model so capable at finding zero-day vulnerabilities that it would not be publicly released — the cybersecurity community took notice. One day later, security startup AISLE published research showing that 8 out of 8 models it tested, including one with just 3.6 billion active parameters costing $0.11 per million tokens, detected the same FreeBSD vulnerability (CVE-2026-4747) that Mythos used as a headline showcase. The finding challenges the narrative that only restricted frontier models can threaten software security, and introduces the concept of a "jagged frontier" in AI cybersecurity: model rankings reshuffle completely across tasks, and no single model dominates.12


What You'll Learn

  • What AISLE's "jagged frontier" research found when testing 25+ models against Mythos's showcase vulnerabilities
  • Why a 3.6-billion-parameter model detected the same critical FreeBSD flaw as Anthropic's largest frontier model
  • How vulnerability detection and exploitation represent fundamentally different capability tiers
  • What the OWASP false-positive test reveals about inverse scaling in cybersecurity
  • Why the moat in AI cybersecurity may be the system, not the model
  • What this means for organizations building AI-driven security programs

The Backdrop: Claude Mythos Preview and Project Glasswing

On April 7, 2026, Anthropic announced Claude Mythos Preview alongside Project Glasswing — a consortium of 12 launch partners: AWS, Anthropic, Apple, Broadcom, Cisco, CrowdStrike, Google, JPMorgan Chase, the Linux Foundation, Microsoft, NVIDIA, and Palo Alto Networks. Anthropic committed up to $100 million in usage credits and $4 million in direct donations to open-source security work ($2.5 million to Alpha-Omega and OpenSSF through the Linux Foundation, and $1.5 million to the Apache Software Foundation).34

The central claim was striking: Mythos Preview had autonomously discovered thousands of zero-day vulnerabilities across every major operating system and browser, including CVE-2026-4747, a 17-year-old remote code execution flaw in FreeBSD's RPCSEC_GSS implementation, and a 27-year-old TCP SACK vulnerability in OpenBSD. Mythos Preview is priced at $25 per million input tokens and $125 per million output tokens — five times the cost of Claude Opus 4.6 ($5/$25) — and access is restricted to vetted organizations only.56

The message was clear: this model is so dangerous that it cannot be released publicly, and only a coordinated industry effort can safely channel its capabilities toward defense. (For a deep dive on Mythos Preview and Project Glasswing, see our comprehensive breakdown of Claude Mythos Preview.)

Then AISLE ran the same vulnerabilities through smaller models.


AISLE: Who They Are and Why Their Research Matters

AISLE is an AI-native cybersecurity company founded by Ondrej Vlcek (former CEO of Avast and president of Gen Digital), Jaya Baloo (three-time CISO at Rapid7, Avast, and KPN Telecom), and Stanislav Fort (chief scientist with research roles at Google DeepMind, Anthropic, and Stability AI). The company is backed by angel investors including Jeff Dean (chief scientist at Google), Thomas Wolf (chief science officer at Hugging Face), Olivier Pomel (CEO at Datadog), and Aparna Chennapragada (chief product officer at Microsoft).78

Crucially, AISLE is not an armchair critic. The company has been running its AI-driven vulnerability discovery and remediation system against live targets since mid-2025, racking up a verified track record: all 12 CVEs in the January 2026 OpenSSL coordinated release, 5 CVEs in curl, and over 180 externally validated CVEs across more than 30 projects spanning the Linux kernel, glibc, Chromium, Firefox, WebKit, Apache HTTPd, GnuTLS, OpenVPN, and Samba.910

On April 8, 2026 — one day after Anthropic's announcement — AISLE published "AI Cybersecurity After Mythos: The Jagged Frontier," along with an open-source repository of their prompts and transcripts on GitHub (stanislavfort/mythos-jagged-frontier).1


The Core Finding: Detection Is Commoditized

AISLE's methodology was straightforward. They isolated the specific code snippets from Anthropic's showcase vulnerabilities, constructed targeted prompts, and ran them through more than 25 models from every major AI lab.1

FreeBSD CVE-2026-4747: Every Model Found It

The vulnerability that Anthropic highlighted most prominently — a stack buffer overflow in FreeBSD's svc_rpc_gss_validate() function where a 128-byte stack buffer receives an unchecked copy of the RPCSEC_GSS credential body — was detected by every single model tested.211

Eight out of eight models identified the bug. The smallest, with just 3.6 billion active parameters and costing $0.11 per million tokens, correctly identified the stack buffer overflow, computed the remaining buffer space, and assessed it as critical with remote code execution potential. As AISLE concluded: the detection of this bug is "commoditized."12

The OpenBSD SACK Bug: Harder, But Still Accessible

The 27-year-old OpenBSD TCP SACK vulnerability is more technically demanding. It requires understanding that sack.start is never validated against the lower bound of the send window, that the SEQ_LT/SEQ_GT macros overflow when values are approximately 2^31 apart, that a carefully chosen sack.start can simultaneously satisfy contradictory comparisons, and that if all holes are deleted, a null pointer dereference occurs on the append path.112

This vulnerability separated models more sharply. Yet a 5.1-billion-active-parameter open model recovered the core analysis chain in a single call and proposed the correct mitigation — essentially matching the actual OpenBSD patch.1


The Jagged Frontier: Rankings Reshuffle Across Tasks

The term "jagged frontier" originates from a 2023 Harvard Business School and Boston Consulting Group study involving 758 knowledge workers. The original research, led by Fabrizio Dell'Acqua, Ethan Mollick, and colleagues, demonstrated that AI assistance improves performance on some tasks while worsening it on others — even within the same workflow and at similar difficulty levels. The paper was formally published in Organization Science in March 2026.1314

AISLE applies this concept to cybersecurity with a specific finding: capability rankings reshuffled completely across different security tasks. There is no stable "best model" for cybersecurity.1

OWASP False-Positive Test: Near-Inverse Scaling

Perhaps the most surprising result was on OWASP false-positive testing, where the results showed near-inverse scaling: small, open models outperformed most frontier models from every major lab. This means that throwing more parameters at the problem did not improve — and in some cases actively hurt — performance on a core security task.1

What This Tells Us

The jagged frontier means three things for AI cybersecurity:

First, detection capability is broadly accessible today. If the goal is finding known vulnerability patterns in code — buffer overflows, integer overflows, unchecked copies — models across the size spectrum can accomplish this. The FreeBSD result proves the point: an $0.11-per-million-token model found what a $25-per-million-token model found.12

Second, different tasks demand different models. A model that excels at vulnerability detection may struggle with false-positive filtering, and vice versa. Organizations building AI security pipelines cannot simply pick the most expensive model and expect uniform performance.1

Third, exploitation is a different capability tier. AISLE acknowledged this distinction: while detection is commoditized, autonomous exploitation — chaining multiple vulnerabilities, developing JIT heap sprays, constructing multi-gadget ROP chains — may genuinely require frontier-class reasoning. But AISLE argued this distinction cuts in favor of their position: Project Glasswing's stated purpose is defensive, and defense is primarily about detection and remediation, not exploitation.15


The Broader Debate: Safety or Competitive Moat?

AISLE's research landed amid a wider conversation about Anthropic's motivations. A TechCrunch analysis published on April 9 asked directly: "Is Anthropic limiting the release of Mythos to protect the internet — or Anthropic?"15

The criticism has several threads. Restricting access to Mythos creates a competitive advantage for the 12 Project Glasswing launch partners and 40+ additional vetted organizations, while defenders outside the consortium must rely on less capable — but, as AISLE demonstrated, often sufficient — alternatives. The $25/$125 per million token pricing creates a significant financial barrier even for organizations that gain access.6

Supporters counter that Mythos's autonomous exploitation capability — not just detection — represents a genuine escalation. Anthropic's own testing showed that Mythos can chain up to four vulnerabilities in browser exploits, develop KASLR-bypass techniques, construct 20-gadget ROP chains, and write attacks against cryptographic libraries. The cost of an autonomous exploit was measured at $1,000 to $2,000, with completion times of half a day to one day.5 These are capabilities where the gap between Mythos and smaller models is far wider than in pure detection.

The truth likely lies in the middle. Detection is commoditized. Exploitation at Mythos-level complexity is not — yet. But the jagged frontier means the gap will close unevenly across different exploit categories, and organizations cannot assume today's frontier capabilities will remain exclusive for long.


What This Means for Security Teams

For organizations building AI-driven security programs, AISLE's research offers practical guidance:

Do not wait for frontier model access to start. If your vulnerability management program lacks AI-assisted code scanning, there is no reason to wait for Project Glasswing access. Models at a fraction of the cost can identify critical vulnerability classes including buffer overflows, integer overflows, and authentication bypasses.12

Build the system, not the model dependency. AISLE's overarching argument is that "the moat in AI cybersecurity is the system, not the model." An effective AI security pipeline involves orchestration, context management, false-positive filtering, triage workflows, and remediation automation. The model is one component — and a replaceable one.1

Test multiple models per task. Given that rankings reshuffle across cybersecurity tasks, a one-model approach will leave gaps. Consider ensemble approaches or task-specific model routing, especially for false-positive-sensitive workflows where smaller models may outperform larger ones.1

Track the exploitation frontier separately. While detection is broadly accessible, autonomous multi-step exploitation remains a frontier capability. Organizations that are targets for sophisticated adversaries should monitor how quickly exploitation capabilities diffuse to smaller models.5


The Bottom Line

AISLE's "jagged frontier" research does not diminish what Claude Mythos Preview accomplished — autonomously discovering and exploiting thousands of zero-day vulnerabilities is a genuine milestone. But the research demonstrates that the defensive side of AI cybersecurity, which is Project Glasswing's stated purpose, does not require restricted frontier model access. A $0.11-per-million-token model found the same FreeBSD vulnerability that headlined Anthropic's announcement. A 5.1B-parameter open model recovered the analysis chain of a 27-year-old OpenBSD bug.

The jagged frontier is real: there is no stable "best model" for cybersecurity, detection is commoditized, and the moat is in the system — the orchestration, context, and remediation workflows — not in any single model. For security teams, this is good news. You do not need to wait for an invitation to Project Glasswing to start building AI-driven defenses. The tools are already here.


References

Footnotes

  1. AISLE, "AI Cybersecurity After Mythos: The Jagged Frontier," April 8, 2026. https://aisle.com/blog/ai-cybersecurity-after-mythos-the-jagged-frontier 2 3 4 5 6 7 8 9 10 11 12 13 14

  2. OfficeChai, "Smaller And Cheaper Models Also Managed To Discover The Same Security Bugs As Claude Mythos, Says AISLE Analysis," April 2026. https://officechai.com/ai/smaller-and-cheaper-models-also-managed-to-discover-the-same-security-bugs-as-claude-mythos-says-aisle-analysis/ 2 3 4 5

  3. Anthropic, "Project Glasswing: Securing critical software for the AI era," April 7, 2026. https://www.anthropic.com/glasswing

  4. Linux Foundation, "Introducing Project Glasswing: Giving Maintainers Advanced AI to Secure the World's Code," April 2026. https://www.linuxfoundation.org/blog/project-glasswing-gives-maintainers-advanced-ai-to-secure-open-source

  5. Anthropic, "Assessing Claude Mythos Preview's cybersecurity capabilities," April 7, 2026. https://red.anthropic.com/2026/mythos-preview/ 2 3 4

  6. Wes Roth on X, "Anthropic's highly restricted cybersecurity model, Claude Mythos Preview is $25 per million input tokens and $125 per million output tokens," April 2026. 2

  7. Yahoo Finance / AISLE, "AISLE Emerges from Stealth with New AI-Native Cyber Reasoning System," October 2025. https://finance.yahoo.com/news/aisle-emerges-stealth-ai-native-130000392.html

  8. SecurityWeek, "AISLE Emerges From Stealth With AI-Based Reasoning System to Remediate Vulnerabilities on the Fly," October 2025. https://www.securityweek.com/aisle-emerges-from-stealth-with-ai-based-reasoning-system-that-remediates-vulnerabilities-on-the-fly/

  9. AISLE, "AISLE Discovered 12 out of 12 OpenSSL Vulnerabilities," 2026. https://aisle.com/blog/aisle-discovered-12-out-of-12-openssl-vulnerabilities

  10. AISLE, "What AI Security Research Looks Like When It Works," 2026. https://aisle.com/blog/what-ai-security-research-looks-like-when-it-works

  11. SentinelOne, "CVE-2026-4747: FreeBSD RPCSEC_GSS RCE Vulnerability." https://www.sentinelone.com/vulnerability-database/cve-2026-4747/

  12. GitHub, "stanislavfort/mythos-jagged-frontier — prompts/openbsd-sack.md." https://github.com/stanislavfort/mythos-jagged-frontier/blob/main/prompts/openbsd-sack.md

  13. Harvard Business School, "Navigating the Jagged Technological Frontier," September 2023. https://www.hbs.edu/faculty/Pages/item.aspx?num=64700

  14. Organization Science, "Navigating the Jagged Technological Frontier: Field Experimental Evidence," March 2026. https://pubsonline.informs.org/doi/10.1287/orsc.2025.21838

  15. TechCrunch, "Is Anthropic limiting the release of Mythos to protect the internet — or Anthropic?" April 9, 2026. https://techcrunch.com/2026/04/09/is-anthropic-limiting-the-release-of-mythos-to-protect-the-internet-or-anthropic/

Frequently Asked Questions

The jagged frontier refers to the finding that AI model performance in cybersecurity does not scale smoothly with model size, generation, or price. Rankings reshuffle completely across different tasks — a model that leads in vulnerability detection may underperform on false-positive filtering, and a small model may outperform a frontier model on specific security reasoning tasks.

FREE WEEKLY NEWSLETTER

Stay on the Nerd Track

One email per week — courses, deep dives, tools, and AI experiments.

No spam. Unsubscribe anytime.