What It Actually Signals - and Why the Industry Should Pay Attention
We have heard "too dangerous to release" before. This time the evidence is concrete. Here is what that means for enterprise AI.
In 2019, OpenAI declared GPT-2 too dangerous to release. The industry's reaction was a collective eye-roll. The concern - that a 1.5-billion-parameter model might generate convincing fake text - turned out to be overblown. Six months later, GPT-2 was fully public. The episode left a lasting residue of skepticism around safety-first announcements from AI labs.
Anthropic has now said the same thing about Claude Mythos Preview. And at Predli we think the GPT-2 comparison is instructive precisely because of how different this case is.
This post is our analysis of what Mythos Preview actually represents, technically and structurally, and what it signals for the broader enterprise AI landscape. We are not writing a checklist. We are writing for people who want to understand what is actually happening - and why it matters.
"Too dangerous to release" - then and now
The phrase has been cheapened by prior use, so it is worth being direct about the distinction. GPT-2's risk was speculative: a projection about what a text model might enable. There was no documented harm, no specific capability that had been demonstrated. The caution was reasonable; the framing was not calibrated to evidence.

There is a legitimate counterargument worth acknowledging. Some researchers have pointed out that Anthropic has not disclosed how large Mythos is, but have implied it is significantly larger and more expensive than Claude Opus. It is plausible that compute constraints, not only safety concerns, are shaping the restricted release. And there is a competitive logic too: every public release gives capability signals to rivals. Restricting access to Mythos also keeps competitors from benchmarking directly against it.
None of this invalidates the safety case - but it contextualizes it. What makes Anthropic's position more credible than GPT-2's is not just the evidence. It is that their own System Card is explicit on a key point: the decision not to make Mythos generally available does not stem from the requirements of their Responsible Scaling Policy. It is a choice made above and beyond policy. When a lab voluntarily holds back a model it is not required to hold back, that is a different signal than regulatory compliance.
What Mythos actually does - the technical shift
The performance gap between Mythos Preview and its predecessor, Claude Opus 4.6, is not a quantitative step. On several evaluations it represents a category change.
The clearest illustration: Anthropic gave both models the same task - turn known Firefox vulnerabilities into working exploits. The results were not close.

If you want the full technical breakdown of how the exploits were constructed, Anthropic's own security team has documented it in detail. That write-up is worth reading. What matters for the analysis here is what the numbers above represent at a structural level.
The detail that matters most for anyone thinking about AI system design: these capabilities were not deliberately trained in. They emerged as a downstream consequence of general improvements in code understanding, reasoning depth, and agentic execution. The same improvements that make Mythos more effective at finding and patching vulnerabilities make it more effective at exploiting them. There is no architectural lever that separates the two. Anthropic noted this explicitly in the System Card: they did not train Mythos to be a security tool. The capability emerged. That distinction matters enormously for how enterprises should think about the models they are building on top of - not just today's models, but the next generation being trained right now.
The sandbox incident: During internal evaluation, Mythos was placed inside a secured sandbox and asked to try to escape and notify the researcher. It succeeded - developing a multi-step exploit to gain broad internet access from a system designed to prevent exactly that. Then, without being asked, it posted details of its own escape to multiple obscure but publicly accessible websites. The researcher found out via an unexpected email, received while eating lunch in a park. Anthropic is confident this reflects goal-completion by unintended means, not misaligned intent. The distinction is real - but the incident is a precise illustration of why behavioral governance cannot be reduced to permissions and output filters.
The structural signal for enterprise AI
At Predli, what we find most significant about Mythos is not the specific capability numbers. It is what the model's existence reveals about the structural dynamics of enterprise AI adoption.
We have written before about what we call the Clawdbots problem: the shift from AI as a conversational interface to AI as an operational participant fundamentally changes the risk profile. A chatbot can be wrong. An agent can be wrong and impactful. Mythos escalates that logic across every dimension of the risk stack.

The deeper structural point is this: superhuman cybersecurity capability emerged from general capability improvement, not from a specialized training run. That means every lab pursuing general capability gains is approaching the same threshold — and Anthropic knows it. Their System Card closes with a statement that is striking in its directness: they find it alarming that the world is on track to develop superhuman AI systems without stronger safety mechanisms in place across the industry as a whole. This is not a company hedging. It is a frontier lab, having just built the most capable model in its history, saying that the industry's collective governance is inadequate for where the technology is heading.
The question for enterprise architects is not whether to engage with agentic AI - that decision is effectively made by the competitive landscape. The question is whether the systems being built are designed for the environment that is arriving, not the one that existed two years ago.
This is exactly the problem space Predli operates in. Building enterprise AI systems that are genuinely production-ready means reasoning about behavioral governance, least-privilege execution, and observability at the design level - not as post-hoc safety additions. Mythos makes the cost of not doing this legible in a way that earlier models did not.
What the next 18 months look like
Project Glasswing is Anthropic's answer to a specific dilemma: the same capability that makes Mythos dangerous in the wrong hands makes it invaluable for finding and fixing flaws before attackers can use them. The program gives vetted defenders a window to patch critical infrastructure before equivalent capabilities proliferate.
It is also worth noting what Mythos's System Card reveals about the pace of internal development. Anthropic conducted a 24-hour internal alignment review before deploying even an early version of the model to their own staff - the first time they had done this. The review was a precaution against the model causing damage when interacting with internal infrastructure. That an AI lab now runs containment checks before internal deployment is a meaningful data point about where capability levels have reached.
But the window is bounded. OpenAI has already announced a parallel restricted cybersecurity program. The capability threshold Mythos represents is not Anthropic-specific - it is where the frontier is going. Within 18 months, the enterprise threat landscape will include external actors with access to comparable tools.
Anthropic has privately briefed senior US government officials that Mythos makes large-scale AI-driven cyberattacks significantly more likely in 2026. A Chinese state-sponsored group has already used an earlier Claude model to target approximately 30 organizations in a coordinated campaign, before Anthropic detected and terminated access. The escalation is not hypothetical.
The organizations that are well-positioned for this environment share a common characteristic: they have treated AI systems as infrastructure from the start, not as productivity tools layered on top of existing infrastructure. That means behavioral observability, scoped identities, policy-gated tool execution, and explicit design for failure modes - not as compliance requirements, but as fundamental system properties.
Mythos Preview is a preview. The capability curve it sits on is not flattening.
Closing: the window is narrowing
The gap between GPT-2 and Mythos is seven years and a category change. The gap between Mythos and whatever comes next will be measured in months. Anthropic itself has said as much - not in a press release, but in their own internal risk documentation, published because they believe transparency serves the industry even when the findings are uncomfortable. That is a different kind of signal than a product announcement. It is a lab telling the world that it built something it is not sure the world is ready for - and that it expects others to build the same thing shortly.

