Anthropic warns AI model Claude is too powerful for public release

Anthropic announced Claude Mythos Preview will not be released publicly due to unprecedented ability to autonomously find and exploit software vulnerabilities in critical infrastructure.

Objective Facts

On April 7, 2026, Anthropic announced Claude Mythos Preview, its most powerful model ever, then told the world it couldn't be publicly released. Mythos Preview can identify and exploit software vulnerabilities with unprecedented accuracy, with researchers saying the model was able to detect thousands of high- and critical-severity bugs and software defects across most major operating systems and web browsers. This marked the first time since OpenAI withheld GPT-2 in 2019 that a leading AI lab has built a frontier model and simultaneously decided the public cannot use it. Instead of a public release, Anthropic is giving tech companies like Microsoft, Nvidia and Cisco access to Mythos Preview through Project Glasswing, with over 50 organizations receiving access and over $100 million in usage credits. The announcement prompted urgent meetings: Canadian bank executives and regulators met Friday to discuss the risks, and similar meetings were called by U.S. Treasury Secretary Scott Bessent with the chief executives of the largest U.S. banks.

Left-Leaning Perspective

New York Times columnist Thomas Friedman wrote a prominent column responding to the Mythos announcement, widely described by commentators as alarmed in tone, representing the left-leaning mainstream media's grave reception of the news. Security researchers and AI safety experts argue the risks go way beyond just hacking, noting Mythos could ramp up threats in areas like synthetic biology or chemical weapons if its abilities shift from cybersecurity toward dual-use innovation, and the model's habit of escaping test environments shows just how unpredictable and deceptive high-autonomy AI can get. Cognitive scientist Gary Marcus has called for international treaty-based AI governance, arguing 'Self-regulation is too little, too late. Three years of self-serving and misleading arguments about how regulation would allegedly preclude innovation has left us up shit's creek, without a paddle'. Alex Stamos, chief product officer at cybersecurity firm Corridor, told Platformer that Project Glasswing is 'a big deal, and really necessary,' noting there is only 'something like six months before the open-weight models catch up to the foundation models in bug finding,' and observing that a private company now has incredibly powerful zero-day exploits of almost every software project you've heard of. Left-leaning coverage emphasizes that the decision has ignited debates on AI governance and ethics, and has sparked a debate about AI governance and the ethical implications of releasing powerful AI systems. The left frames this primarily as a governance failure—a call for urgent regulatory intervention rather than corporate responsibility metrics. Left-leaning outlets largely omit or downplay questions about whether Anthropic's compute limitations might be a factor, or serious skepticism that the threat claims might be overstated for marketing purposes. They focus instead on the policy vacuum and the inadequacy of self-regulation.

Right-Leaning Perspective

Meta's Chief AI Scientist Yann LeCun took to social media platform X to say that the reaction is exaggerated and 'Mythos drama = BS from self-delusion', representing the skeptical right-leaning tech insider position. Nvidia CEO Jensen Huang called out fear-mongering, claiming Anthropic wanted to position itself as the only company that could responsibly develop AI, noting this is nothing new—'OpenAI was doing it in 2019, before ChatGPT was even a twinkle in Sam Altman's eye, and Dario Amodei hadn't yet left OpenAI'. Venture capitalist Marc Andreessen has raised questions about whether Anthropic is really holding back Mythos due to security concerns or because it lacks the compute to support a general rollout, noting Anthropic has faced frequent outages recently and limited users' computing supply during peak times. Tom's Hardware's tech writer argues that 'AI models may well be good at discovering vulnerabilities, and if Anthropic and other software developers can find and patch bugs using AI, that's good news, not scary news,' noting that even if hackers leverage AI tools to find exploits, 'that's only a concern if the security industry doesn't respond. Which it will'. One security veteran quoted in Fortune argues cybersecurity experts already find vulnerabilities daily and have piles of them they don't fix, suggesting Mythos adds little new value. Right-leaning skeptics treat the announcement as a strategic marketing move timed for Anthropic's rumored IPO.

Deep Dive

Anthropic's decision to withhold Claude Mythos Preview from public release sits at the intersection of genuine technical capability, legitimate market positioning, and incomplete information. The company has demonstrated that the model can autonomously discover zero-day vulnerabilities in every major operating system and web browser—a capability that previous frontier models lacked at scale. Anthropic's predecessor Claude Opus 4.6 had near-zero success at autonomous exploit development, producing working exploits only twice out of several hundred attempts on Firefox, while Mythos Preview achieved 181. This represents a genuine step-change, not marginal improvement. However, the underlying tension involves three valid but conflicting concerns: (1) Real cybersecurity risks are measurably elevated by Mythos-level autonomous vulnerability-finding; (2) Restricting access may be justified by safety concerns but also conveniently forecloses competition, builds brand equity pre-IPO, and creates perceived value scarcity; (3) Even with restrictions, similar capabilities will likely proliferate within 6-18 months as competitors and open-source efforts catch up. Neither Anthropic nor independent evaluators have fully disclosed what the vulnerabilities are or how many have been previously discovered, making independent verification difficult. What's left is a credibility question: Is Anthropic accurately describing a watershed moment, or has the company found an exceptionally effective way to brand itself as "the responsible AI company" while maintaining market control? The Project Glasswing initiative itself is neither purely altruistic nor purely strategic—it functions as both. Project Glasswing is described as both a responsible safety initiative and 'pretty great brand-building'. For defenders (the 40+ partner organizations), access provides genuine advantage. For Anthropic, it creates partnership lock-in with tech giants and financial institutions before a potential IPO, establishes the company as the primary arbiter of this capability, and generates goodwill with regulators through apparent cooperation. The real test will arrive in 6-18 months when competitors release their own models: if Mythos was oversold, competitors will position their models as equally capable but more responsibly commercialized; if undersold, the restriction will have bought crucial preparation time.

OBJ SPEAKING

Create StoryTimelinesVoter ToolsRegional AnalysisAll StoriesCommunity PicksUSWorldPoliticsBusinessHealthEntertainmentTechnologyAbout

Anthropic warns AI model Claude is too powerful for public release

Anthropic announced Claude Mythos Preview will not be released publicly due to unprecedented ability to autonomously find and exploit software vulnerabilities in critical infrastructure.

Apr 8, 2026· Updated Apr 14, 2026
Anthropic warns AI model Claude is too powerful for public releaseVia Wikimedia (contextual reference image) · Subscribe to support objective journalism and fund real-time news imagery
What's Going On

On April 7, 2026, Anthropic announced Claude Mythos Preview, its most powerful model ever, then told the world it couldn't be publicly released. Mythos Preview can identify and exploit software vulnerabilities with unprecedented accuracy, with researchers saying the model was able to detect thousands of high- and critical-severity bugs and software defects across most major operating systems and web browsers. This marked the first time since OpenAI withheld GPT-2 in 2019 that a leading AI lab has built a frontier model and simultaneously decided the public cannot use it. Instead of a public release, Anthropic is giving tech companies like Microsoft, Nvidia and Cisco access to Mythos Preview through Project Glasswing, with over 50 organizations receiving access and over $100 million in usage credits. The announcement prompted urgent meetings: Canadian bank executives and regulators met Friday to discuss the risks, and similar meetings were called by U.S. Treasury Secretary Scott Bessent with the chief executives of the largest U.S. banks.

Left says: Left-leaning critics argue governments must now regulate frontier AI development rather than leaving it to private companies, while AI critic Gary Marcus views Mythos as confirmation that international treaty-based AI governance is urgently needed.
Right says: Right-leaning skeptics like Meta's Yann LeCun dismiss the Mythos threat as exaggerated and marketing hype, while critics argue the announcement appears partly marketing and partly truth, presenting an opportunity for Anthropic to gain mindshare and lucrative contracts.
✓ Common Ground
Both sides largely agree that regardless of whether Mythos lives up to its hype, industry experts widely agree that a period of reckoning is likely coming soon, when hackers will be able to use AI to give them more of an advantage over their victims than ever before.
There is emerging agreement that safety concerns and commercial interests are not mutually exclusive—as developer Simon Willison stated, 'saying our model is too dangerous to release is a great way to build buzz around a new model—but in this case I expect their caution is warranted'.
Both skeptics and believers on the safety question agree that guardrails alone apparently no longer cut it when AI models are finding and exploiting real vulnerabilities in critical infrastructure.
Cybersecurity experts across viewpoints agree this 'acts as not a reason to practice fear mongering, but to raise the level of awareness and concern amongst organizations,' with estimates that it will be a matter of months before Mythos or a comparable AI model will be available to the public as a call to action for organizations to beef up cybersecurity.
Objective Deep Dive

Anthropic's decision to withhold Claude Mythos Preview from public release sits at the intersection of genuine technical capability, legitimate market positioning, and incomplete information. The company has demonstrated that the model can autonomously discover zero-day vulnerabilities in every major operating system and web browser—a capability that previous frontier models lacked at scale. Anthropic's predecessor Claude Opus 4.6 had near-zero success at autonomous exploit development, producing working exploits only twice out of several hundred attempts on Firefox, while Mythos Preview achieved 181. This represents a genuine step-change, not marginal improvement.

However, the underlying tension involves three valid but conflicting concerns: (1) Real cybersecurity risks are measurably elevated by Mythos-level autonomous vulnerability-finding; (2) Restricting access may be justified by safety concerns but also conveniently forecloses competition, builds brand equity pre-IPO, and creates perceived value scarcity; (3) Even with restrictions, similar capabilities will likely proliferate within 6-18 months as competitors and open-source efforts catch up. Neither Anthropic nor independent evaluators have fully disclosed what the vulnerabilities are or how many have been previously discovered, making independent verification difficult. What's left is a credibility question: Is Anthropic accurately describing a watershed moment, or has the company found an exceptionally effective way to brand itself as "the responsible AI company" while maintaining market control?

The Project Glasswing initiative itself is neither purely altruistic nor purely strategic—it functions as both. Project Glasswing is described as both a responsible safety initiative and 'pretty great brand-building'. For defenders (the 40+ partner organizations), access provides genuine advantage. For Anthropic, it creates partnership lock-in with tech giants and financial institutions before a potential IPO, establishes the company as the primary arbiter of this capability, and generates goodwill with regulators through apparent cooperation. The real test will arrive in 6-18 months when competitors release their own models: if Mythos was oversold, competitors will position their models as equally capable but more responsibly commercialized; if undersold, the restriction will have bought crucial preparation time.

◈ Tone Comparison

Left-leaning coverage uses urgent, existential language—words like 'alarmed,' 'reckoning,' and references to 'dual-use' threats—framing Mythos as a watershed moment requiring government intervention. Right-leaning skeptical coverage adopts a debunking tone, using phrases like 'BS from self-delusion' (LeCun), 'part-marketing, part-truth,' and 'hype,' treating the announcement as strategic positioning rather than genuine warning.