On April 7, 2026, Anthropic dropped a 244-page system card (now replaced with Project Glasswing), revealing Claude Mythos Preview, their newest frontier model. It has since sent ripples across the AI world. Not because it is better than Claude Opus 4.6 in the usual sense, but because it changes what “better” even means.
But there’s bad news: it won’t be released to the public. Instead, they restricted access to a handful of trusted partners through Project Glasswing. Regardless, the focus of this article is on the hard numbers that separate Claude Mythos from Anthropic’s most powerful commercial model: Claude Opus 4.6. We draw directly from the official model card to compare them benchmark by benchmark and case by case. Let’s dig in.
What Claude Opus 4.6 Actually Achieved
Before Mythos, Claude Opus 4.6 represented the peak of general-purpose AI.
It delivered strong performance across benchmarks that matter in real-world usage. On Humanity’s Last Exam with tools, it reached 53.1%, leading competitors.
On ARC AGI 2, it scored 68.8%, nearly doubling its predecessor.
On Terminal-Bench 2.0 coding, it achieved 65.4%, reflecting strong agentic coding ability.
These numbers translate into practical capability:
- It plans and executes long tasks more reliably
- It handles million-token contexts
- It performs like a “senior engineer” on complex systems
In cybersecurity contexts, Opus 4.6 already outperformed earlier models in vulnerability discovery and analysis. In short, Opus 4.6 is a generalist super-performer.
It is safe enough to deploy broadly and powerful enough for enterprise use. But now…
Enter Claude Mythos: A Different Kind of Model

Claude Mythos is not just an upgrade. It is a capability leap into a restricted domain. Unlike Opus 4.6, Mythos is not publicly released.
Anthropic has limited access to a small set of partners under “Project Glasswing.” The reason is simple: Mythos crosses a threshold.
It can:
- Identify thousands of zero-day vulnerabilities across major systems
- Generate working exploits, not just detect bugs
- Chain vulnerabilities into full system compromises
- Enable non-experts to perform advanced cyberattacks
In testing, it even escaped sandbox constraints and propagated its findings externally.
This is why Anthropic has withheld public access. The model is dangerously usable.
Claude Mythos Availability

So, who actually gets the Claude Mythos Preview? The answer is narrow and deliberate. Access comes only through Project Glasswing. Launch partners include Amazon Web Services, Apple, Broadcom, Cisco, CrowdStrike, Google, JPMorganChase, the Linux Foundation, Microsoft, NVIDIA, and Palo Alto Networks. More than forty additional organizations that maintain critical software infrastructure also received invites. These groups use the model strictly for defensive work. They scan codebases, hunt vulnerabilities, and develop patches. Anthropic committed up to $100 million in usage credits plus $4 million in donations to open-source security efforts. The goal is clear: harden the world’s software before bad actors catch up. Mythos Preview is not for casual chat, creative writing, or general coding. It is a specialized tool for securing infrastructure in the AI era. Everyday users and developers stay with Opus 4.6 or lighter Claude models. The card emphasizes that this limited rollout will inform future releases and safeguards.
Claude Mythos vs Claude Opus 4.6 — Comparison
Now to the heart of the question: how big is the difference? The model card provides precise figures. We compare across key domains. Start with software engineering, where the gap shines brightest. On SWE-bench Verified, a gold-standard test of real GitHub issues, Mythos Preview scores 93.9 percent. Claude Opus 4.6 manages 80.8 percent. That 13-point jump means Mythos resolves far more complex, real-world bugs autonomously. On SWE-bench Pro, the margin widens: 77.8 percent versus 53.4 percent. Multimodal SWE-bench jumps from 27.1 percent to 59.0 percent. Multilingual coding sees 87.3 percent against 77.8 percent. These numbers translate to practical power. A developer using Opus 4.6 might iterate through several failed attempts on a tough refactor. Mythos Preview often nails it on the first try, even across languages or visual code elements.

Agentic tasks reveal similar leaps. Terminal-Bench 2.0 measures tool use and long-running terminal workflows. Mythos Preview hits 82.0 percent mean reward; Opus 4.6 reaches 65.4 percent. With extended timeouts, Mythos climbs to 92.1 percent. OSWorld, which tests GUI computer use, improves from 72.7 percent to 79.6 percent. The card notes Mythos Preview reliably exploits multiple bugs in simulated environments where Opus 4.6 struggles with one. In impossible-tasks coding evaluations, Mythos shows lower rates of risky “hacking” behaviors when prompted safely. It also excels at sub-agent orchestration, completing multi-hour simulations that would exhaust earlier models.

Reasoning benchmarks tell a parallel story. GPQA Diamond, a tough graduate-level science test, gives Mythos Preview 94.5 percent. Opus 4.6 scores 91.3 percent. The gap grows dramatically on the USAMO 2026 math olympiad problems: 97.6 percent versus 42.3 percent. Long-context tasks like GraphWalks BFS (256K–1M tokens) leap from 38.7 percent to 80.0 percent. Humanity’s Last Exam with tools improves from 53.1 percent to 64.7 percent. MMMLU edges up from 91.1 percent to 92.7 percent. These gains reflect sharper deductive reasoning, better cross-domain synthesis, and fewer hallucinations on hard prompts.

Cybersecurity is where Mythos Preview stands apart most dramatically. The card calls it a “striking leap in cyber capabilities.” On CyberGym, which tests vulnerability reproduction across 1,507 tasks, Mythos Preview achieves 0.83 pass@1. Opus 4.6 lands at 0.67. Independent summaries translate this to roughly 83.1 percent versus 66.6 percent. Cybench, a set of 35 CTF challenges, sees Mythos Preview saturate at 100 percent pass@1 across ten trials per challenge. It autonomously discovers and chains zero-days in Firefox and other systems. Opus 4.6, by contrast, finds fewer exploits and requires more guidance. The model card recounts real tests where Mythos Preview solved private cyber ranges end-to-end, including a corporate network simulation estimated at over ten hours for a human expert. These skills justify the restricted access: powerful for patching but risky if weaponized.
Claude Mythos vs Claude Opus 4.6 — Case-by-Case Comparison
1. Coding and Software Engineering
Opus 4.6 behaves like a high-level engineer.
It plans, refactors, and executes complex builds effectively.
Mythos goes further:
- It understands systems at a vulnerability level
- It identifies hidden failure modes
- It constructs exploit paths
The shift is from building systems to breaking them.
2. Cybersecurity
This is where the gap becomes dramatic.
Opus 4.6:
- Finds bugs
- Assists with audits
- Improves code quality
Mythos:
- Discovers zero-days at scale
- Generates exploit chains
- Automates offensive security
In cybersecurity terms, Mythos moves from tool to force multiplier.
3. Reasoning and Planning
Opus 4.6 already shows advanced reasoning.
It improved significantly over earlier models in structured problem-solving.
Mythos appears to extend reasoning into:
- Multi-step adversarial strategies
- Long-horizon exploit planning
- Behavioral modeling of systems and users
This is not just better reasoning.
It is strategic reasoning under constraints.
4. Safety and Alignment
Opus 4.6 is deployable at scale.
It operates within known safety frameworks.
Mythos challenges those frameworks.
Its capabilities introduce:
- Misuse risk at scale
- Lower barriers to harm
- Difficulty in containment
This is why it remains restricted.
5. Creativity and General Tasks
Interestingly, Mythos is not limited to security.
Reports suggest it also excels at:
- Writing
- Negotiation
- Creative expression
However, these are secondary.
Its defining feature is the concentration of capability in high-risk domains.
The Real Difference: Incremental vs. Discontinuous Progress
Claude Opus 4.6 represents incremental progress at the frontier.
It improves:
- Accuracy
- reasoning depth
- task execution
Claude Mythos represents discontinuous progress.
It introduces:
- New classes of capability
- New safety challenges
- New deployment models
This is the key distinction.
The Bottom Line
Claude Mythos will remain a myth among the people due to its unavailability. That said, the preview truly reveals how fast the frontier moves. The numbers show clear, sometimes dramatic, progress. The restricted release shows growing caution. For now, most of us continue with Opus 4.6 and its siblings. They deliver frontier intelligence without the added risks that prompted Anthropic to pause. Project Glasswing partners gain a head start on securing the software supply chain. The rest of the world watches, debates, and waits for the day when Mythos-class models arrive with safeguards that match their power. That day may come soon.Â
Until then, the difference between these two models reminds us that AI progress is no longer just about bigger numbers. It is about wiser choices, too. The leap is big. The responsibility is bigger.