
Just last week, Anthropic launched Claude Opus 4.7, described as a safer public-facing version of Claude Mythos, a model reportedly considered too dangerous for broad release.
Now, the company is facing uncomfortable questions after reports claimed an unauthorized group gained access to Claude Mythos, a highly restricted internal model built for advanced cybersecurity tasks.
If accurate, this may be one of the clearest examples yet that the biggest risk in AI isn't the model, but those who can access it.
According to a Bloomberg report, Claude Mythos was designed to identify and exploit vulnerabilities in software systems, making it far more sensitive than a standard chatbot. Access was reportedly limited to select partners under a private security initiative, not the general public.
Yet...an outside group now claims it found a way in.
What allegedly happened
Reports say the group may have accessed Mythos through a third-party contractor environment rather than Anthropic’s main internal systems. Anthropic has reportedly said it is investigating and has no evidence that its core systems were breached.
To be clear, this does not appear to be a case of rogue AI behavior or some dramatic sci-fi scenario of a bot escaping from its maker. Instead, the problem is far more familiar in the tech world, such as credentials, vendor access, weak boundaries and security gaps.
In other words, this is a very human problem with a potentially dangerous AI.
Why this story is troubling

Besides the issue of a powerful model getting into the wrong hands, this alleged breach emphasizes what has been a topic of public conversation around AI for years: frontier AI models are becoming high-value assets, and valuable assets attract attackers.
This concern is the immediate issue, but AI anxieties such as job displacement, misinformation at scale, autonomous misuse and Superintelligent systems as a whole still weigh heavily on the public.
If big tech companies are building models powerful enough to influence cybersecurity, finance or defense, they also need to secure them as they would critical infrastructure.
This means strong vendor oversight, tight identity controls, compartmentalized access, real-time monitoring and fast incident response. It doesn't take a rocket scientist to understand that building a powerful model is only half the challenge, and protecting it is the other half.
Why Claude Mythos stands out

What makes this report especially concerning is that Claude Mythos was reportedly treated as sensitive enough to keep behind closed doors. That creates a difficult optics problem.
If a company signals a model is too powerful for public release, but outsiders can allegedly reach it anyway, we've got to wonder whether AI governance is keeping pace with AI development.
And that reminds me of a bigger trend that absolutely no one is talking about: AI labs are entering a new era where they are no longer just software companies. They are becoming responsible for protecting systems that are valuable and important to governments, businesses and society. Obviously, that means the security expectations should start to resemble those placed on banks, cloud providers and critical infrastructure operators.
The public debate over whether AI is getting too smart is clearly being overshadowed by the question of whether AI companies are secure enough. Obviously, they aren't.
The takeaway
If these reports are accurate, the Claude Mythos incident should serve as a warning to other AI companies to strengthen their security practices. Humans are building extraordinary tools faster than they can fully protect them — and that may become the defining AI risk of this decade.
More from Tom's Guide
- I let Gemini scan my messy fridge photos to plan my meals — and I saved $150 this month
- I asked ChatGPT to use Charlie Munger’s ‘Inversion rule' to rethink my goals — and it beat every productivity app
- I asked ChatGPT to use Elon Musk’s ‘Relevance Rule’ to fix my memory — and I’m never going back to notes