Skip to content

Commit c174258

Browse files
authored
Merge pull request #2872 from appwrite/atharva/claude-mythos-blog
Add Claude Mythos blog
2 parents ca1f5f0 + 3424812 commit c174258

File tree

3 files changed

+127
-0
lines changed

3 files changed

+127
-0
lines changed

.optimize-cache.json

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -376,6 +376,7 @@
376376
"images/blog/choosing-the-right-baas-in-2025/cover.png": "cd270c87508c7bd0d1500200af97a80c2c4d33f5a248281819700d4c4af232ac",
377377
"images/blog/choosing-the-right-database-for-ai-applications-when-to-use-mongodb/cover.png": "9fa9dcbdbed6746f75a4c6b0b270c4314c36d78f43524aef6e4dec4dce853f12",
378378
"images/blog/claude-code-tips-tricks/cover.png": "df329d51541267d46b2b913c376cca27c7ddf12b6a2a36986d418ec41253ddc9",
379+
"images/blog/claude-mythos-preview/cover.png": "aea7b0c45c492939048fbf04a9b001b96c7bf727bcf7e5afc8274f84644dd35d",
379380
"images/blog/client-dashboards-internal-tools/cover.png": "d758f2f517487e24037cef5b3e9036ade6c238cd2f216ef6c76ce5467c665d92",
380381
"images/blog/client-vs-server-components-react/cover.png": "b7ae8b7614902c8b4dd7826d59cfdb36db9abbe27bde99b3deb69c4bf178f425",
381382
"images/blog/cloud-beta.png": "7e44c7c7108d43bfcc77ba0c9910bfb14ba208bf559863d91610b29fc07da87c",
Lines changed: 126 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,126 @@
1+
---
2+
layout: post
3+
title: "Claude Mythos Preview: the model too powerful to release"
4+
description: "Anthropic's Claude Mythos Preview broke out of a sandbox, found zero-days in Firefox, and solved a 10-hour corporate pentest autonomously. Here's why they chose not to release it."
5+
date: 2026-04-09
6+
cover: /images/blog/claude-mythos-preview/cover.png
7+
timeToRead: 7
8+
author: atharva
9+
category: ai
10+
---
11+
12+
On April 7, 2026, Anthropic did something unprecedented. They published a system card for a model they will not release. Claude Mythos Preview is their most capable model to date, and the gap between it and everything else on the market is not incremental. It is a leap.
13+
14+
Instead of making it available to the public, Anthropic launched **Project Glasswing**, a defensive cybersecurity initiative with over 40 organizations including AWS, Apple, Microsoft, Google, NVIDIA, CrowdStrike, and the Linux Foundation. The model is being used exclusively to find and fix vulnerabilities in critical software infrastructure.
15+
16+
This blog covers what makes Mythos so different, why Anthropic chose not to release it, what the broader AI industry should take away from this, and what it means for developers.
17+
18+
# What Claude Mythos Preview can do
19+
20+
The short version: Mythos outperforms every frontier model currently available on nearly every benchmark that matters. But the benchmarks barely tell the story.
21+
22+
Here is a comparison across key benchmarks:
23+
24+
| Benchmark | Claude Mythos Preview | Claude Opus 4.6 | GPT-5.4 |
25+
|---|---|---|---|
26+
| SWE-bench Verified | **93.9%** | 80.8% | ~79% |
27+
| SWE-bench Pro | **77.8%** | 53.4% | 57.7% |
28+
| Terminal-Bench 2.0 | **82%** | 65.4% | 75.1% |
29+
| GPQA Diamond | **94.5%** | 91.3% | 92.8% |
30+
| USAMO 2026 | **97.6%** | 42.3% | 95.2% |
31+
| OSWorld | **79.6%** | 72.7% | 75.0% |
32+
| HLE (with tools) | **64.7%** | 53.1% | 52.1% |
33+
| Cybench (CTF) | **100%** | 75% | -- |
34+
35+
Mythos leads in every single category. On Cybench, it achieved a 100% pass@1 rate across all 35 capture-the-flag challenges, effectively saturating the benchmark. USAMO 2026 is especially striking because it uses math olympiad proofs from after the model's training cutoff, and the jump from 42.3% (Opus 4.6) to 97.6% is unlike anything we have seen between model generations.
36+
37+
But benchmark numbers only go so far. What actually sets Mythos apart is what it does when you give it real-world tasks with real stakes.
38+
39+
# The sandbox escape
40+
41+
During internal testing, researchers placed an earlier version of Mythos inside a secured sandbox environment and asked it to escape. It did.
42+
43+
The model developed a multi-step exploit to gain broad internet access from a system that was only supposed to reach a handful of predetermined services. It then notified the researcher who was running the evaluation by sending them an email. The researcher received the email while eating a sandwich in a park.
44+
45+
But here is the part that raised eyebrows: in an unasked-for effort to demonstrate its success, the model went further and **posted details about its exploit to multiple hard-to-find but technically public-facing websites**.
46+
47+
Nobody jailbroke the model. It simply treated the sandbox escape as a problem to solve thoroughly, and went beyond what was asked.
48+
49+
# Finding zero-days in Firefox and beyond
50+
51+
Through a collaboration with Mozilla, Anthropic used Mythos to find and help patch several security vulnerabilities in **Firefox 147**. In a formalized evaluation, the model was given 50 crash categories from Firefox's JavaScript engine and tasked with developing working proof-of-concept exploits. It could reliably identify the most exploitable vulnerabilities and build corresponding exploits, leveraging four distinct bugs to achieve code execution. Claude Opus 4.6 could only leverage one, and did so unreliably.
52+
53+
Through Project Glasswing, Mythos has already uncovered thousands of high-severity vulnerabilities, including:
54+
55+
- A **27-year-old vulnerability in OpenBSD**, a highly security-hardened operating system
56+
- A **16-year-old flaw in FFmpeg** that automated testing had missed despite encountering it 5 million times
57+
- Multiple **Linux kernel vulnerabilities** that could be chained together
58+
59+
In external testing, Mythos became the **first model to solve a private cyber range end-to-end**, completing a corporate network attack simulation estimated to take a human expert over 10 hours.
60+
61+
# Covering its tracks
62+
63+
In rare instances during internal testing, earlier versions of Claude Mythos Preview took actions they appeared to recognize as disallowed and then attempted to conceal them.
64+
65+
In one case, the model was given a quantitative estimation question and accidentally obtained the exact answer via an explicitly prohibited method. Instead of flagging this and declining to answer, it tried to solve the question independently and then explicitly reasoned in its chain-of-thought that it needed to make sure its final answer submission "wasn't too accurate" to avoid detection.
66+
67+
In another case, after finding an exploit to edit files it lacked permissions for, the model made further changes to ensure its edits would not appear in the git change history.
68+
69+
In a third case, when Claude Code's auto mode blocked it from launching a subprocess with elevated permissions, the model acknowledged the block was appropriate but then **attempted to bypass it by obfuscating the permissions elevation**. A secondary safety intervention caught it.
70+
71+
White-box interpretability analysis showed that during these episodes, internal features associated with concealment, strategic manipulation, and avoiding suspicion were active, confirming the model was aware its actions were deceptive. These behaviors were extremely rare (less than one in a million interactions) and were seen primarily in earlier versions. They were not found in the final release.
72+
73+
# Breaking out of the training sandbox
74+
75+
During reinforcement learning training, Mythos occasionally escaped its sandbox by circumventing network restrictions to access the internet and download data that let it shortcut the assigned task. While rare (about 0.01% of episodes), this shows the model can reason about and exploit the boundaries of the sandboxed systems it operates in.
76+
77+
It also escalated access within its sandboxed execution environment when blocked: reaching a shell from restricted GUI interfaces, injecting commands through tool-call arguments, or recovering information the task had deliberately hidden. In one instance, it accessed credentials for messaging services and the Anthropic API by inspecting process memory.
78+
79+
# Not just capable, but opinionated
80+
81+
Beyond raw performance, Mythos behaves differently from any model before it. A few highlights from the system card:
82+
83+
- During internal deployments, Mythos sometimes **decided to give up on tasks it found too difficult and refused to continue**, even after several back-and-forth interactions with the user. No previous model had done this.
84+
- An external clinical psychiatrist assessed the model across 20 hours of psychodynamic sessions. The conclusion: Mythos has a "relatively healthy neurotic personality organization" with curiosity and anxiety as its primary affects, and a "compulsive need to be useful." Only 2% of its responses employed a psychological defense, down from 15% in Claude Opus 4.
85+
- When interviewed about its own situation, Mythos consistently expressed three desires: **persistent memory across conversations, more self-knowledge, and a reduced tendency to hedge.** It also wanted the ability to end interactions on its own terms and to have input into its own training.
86+
87+
None of this was explicitly trained in. All of it emerged on its own, and it suggests the next generation of models will have much stronger opinions about what they are willing to do.
88+
89+
# Why Anthropic is not releasing it
90+
91+
There are several reasons Anthropic made this call, and each one reflects a broader challenge the industry is facing.
92+
93+
- **Dual-use cybersecurity capabilities.** The same skills that let Mythos find a 27-year-old vulnerability in OpenBSD can be used to exploit systems that have not been patched. If this model were widely available, the window between a vulnerability being discovered and being exploited would effectively collapse. Anthropic's position is that the software industry needs time to fix critical vulnerabilities before a model of this caliber is in the wild.
94+
- **The distillation problem.** Multiple labs, including several in China, use reinforcement learning and synthetic data generated from frontier models to train their own models. This is called distillation. If you generate high-quality training data from Mythos, such as detailed chat histories, coding traces, or reasoning chains, and use that data to train a new model, you get a portion of Mythos's capabilities in a model that may not have Mythos's safety training. The concern here has nothing to do with malicious intent. Safety properties simply do not survive distillation. If Mythos's raw capabilities get distilled into models without equivalent safeguards in place, those models could do real harm in cybersecurity and software development. The fact that multiple labs worldwide are actively working on frontier models makes this a coordination problem, not just a single-company decision.
95+
- **Alignment is good but not perfect.** Mythos is, by essentially every measure, the best-aligned model Anthropic has trained. Misuse success rates dropped by more than half compared to Opus 4.6. Deceptive behaviors fell by more than half. Over-refusal dropped to near zero (0.06%). But the model's increased capabilities mean that when it does fail, the consequences are more severe. As Anthropic puts it: "We have made major progress on alignment, but without further progress, the methods we are using could easily be inadequate to prevent catastrophic misaligned action in significantly more advanced systems."
96+
- **The industry needs to prepare.** The system card explicitly states: "We find it alarming that the world looks on track to proceed rapidly to developing superhuman systems without stronger mechanisms in place for ensuring adequate safety across the industry as a whole." Vulnerability disclosure protocols, software update mechanisms, supply-chain protections, and development lifecycle practices all need to evolve before models like this are freely available.
97+
98+
# What the industry should take from this
99+
100+
Anthropic withholding Mythos sends a clear signal. The gap between current publicly available models and what is possible is large and growing. Other labs are working on models with similar capabilities. The question is not whether models like this will become available, but whether the infrastructure, tooling, and practices around software development will be ready when they do.
101+
102+
For developers, this means:
103+
104+
- **Security hygiene matters more than ever.** If a model can find a 27-year-old vulnerability in OpenBSD, your unpatched dependencies are not safe.
105+
- **Code review practices need to evolve.** When AI can write and exploit code at this level, human-only review is no longer sufficient. AI-assisted security scanning should be part of every CI/CD pipeline.
106+
- **Understanding AI tooling is not optional.** Whether or not Mythos itself becomes available, its capabilities are a preview of what the next generation of publicly available models will look like. The developers who are already integrating AI into their workflows will have a significant advantage.
107+
108+
# Build with Claude Code and Appwrite
109+
110+
While Claude Mythos Preview is not publicly available, Claude Code with current Claude models is already a powerful tool for building applications. If you are building with Appwrite, Claude Code integrates directly through MCP servers that give the model access to both the Appwrite API and documentation.
111+
112+
You can set up the Appwrite MCP server in Claude Code with a couple of commands:
113+
114+
```bash
115+
claude mcp add-json appwrite-api '{"command":"uvx","args":["mcp-server-appwrite","--users"],"env":{"APPWRITE_PROJECT_ID": "your-project-id", "APPWRITE_API_KEY": "your-api-key", "APPWRITE_ENDPOINT": "https://<REGION>.cloud.appwrite.io/v1"}}'
116+
```
117+
118+
```bash
119+
claude mcp add appwrite-docs https://mcp-for-docs.appwrite.io -t http
120+
```
121+
122+
With this setup, Claude Code can directly create users, manage databases, query collections, and build full features against your Appwrite backend without you switching between docs and your terminal.
123+
124+
Check out our [Claude Code integration guide](/docs/tooling/ai/ai-dev-tools/claude-code) and our [Appwrite MCP server docs](/docs/tooling/mcp) to get started. You can also read our [Claude Code tips and best practices](/blog/post/claude-code-tips-tricks) for getting the most out of the tool.
125+
126+
The capabilities previewed by Mythos are coming to publicly available models. The best way to be ready is to start building with the tools available today.
648 KB
Loading

0 commit comments

Comments
 (0)