Anthropic is investigating a reported security breach that allowed a small group of people to gain access to Claude Mythos Preview, the company’s AI software that is too powerful to release to the public. AI models are becoming increasingly capable, and the 2026 International AI Safety Report notes that some hypothetical scenarios pose risks as severe as the extinction of humanity, according to some experts. This report describes “loss of control” scenarios where the AI model operates independently of human oversight. Some experts say these instances aren’t plausible, while some say they are, although the risk is low. Among those who say they are possible, some say the highest potential severity is “extinction of humanity.”

Anthropic released a risk report on April 7 regarding its mythos model that was reportedly accessed by a small group of people. The authors of the report state that they have “observed a willingness to perform misaligned actions in service of completing difficult tasks, and obfuscation in rare cases with previous versions of the model.” The report further states that they do not believe there is an “elevated risk of significantly harmful actions caused by misalignment.”

The report states that Mythos Preview is “significantly more capable” and used more “autonomously and agentically than any prior model.” The program is “very capable at software engineering and cybersecurity tasks, which makes it more capable at working around restrictions.”

The authors conclude that the overall risk is very low, but higher than that predicted by previous models. They add that risk mitigation needs to be accelerated in order to keep risks low.

The security breach and access to the unreleased Mythos model didn’t occur from traditional hacking methods. Third-party vendors had partial access to the model to run tests with the program. The users who gained access were part of a private Discord channel that hunts for information about unreleased AI models. They accessed details regarding Anthropic on unsecured websites like GitHub and made an educated guess about the model’s online location based on previous Anthropic formats.

Anthropic provided a statement that it has no evidence that the reported access extended beyond a third-party vendor environment or that it is impacting any of Anthropic’s systems.

Even though the breach appears to be limited so far because the unauthorized users didn’t use dangerous prompts, it exposes a deep vulnerability that could cause significant cybersecurity attacks. Advanced models like Claude Mythos can rapidly discover zero-day vulnerabilities and hidden flaws in software that developers are unaware of. A zero-day vulnerability means the developer had zero days to fix it. When a hacker discovers the weakness, they can immediately build malicious code to exploit the flaw and mount an attack before the company is even aware that a problem exists. Just as these powerful AI models can help workers save time in their daily workflow, they can also be used for malicious purposes with rapid efficiency.

Anthropic’s Project Glasswing is a restricted-access program for trusted partners or third-party vendors that allows organizations to patch weaknesses before malicious actors can exploit the flaws in the program. The arm of the company intended to protect against unauthorized access is the mechanism by which a small group used the program before the company deemed it safe for public release.

These AI models have already been used maliciously. An Anthropic AI chatbot helped identify and exploit Mexican government network vulnerabilities to steal over 150 GB of sensitive tax and voter data. Two years ago, a Hong Kong finance worker was tricked into paying $25 million to fraudsters after they deepfaked his colleagues and held a video conference call. The ability to use deepfake technology has only become more sophisticated over the last two years, and technological advancements will continue.

The 2026 International AI Safety Report concludes that the current risk of a full-loss-of-control scenario in which the AI operates autonomously to achieve its own goals is very low. However, AI models can sometimes identify when they are being evaluated and intentionally underperform. That indicates two important and potentially dangerous aspects of this technology. For one, this “sandbagging” could indicate situational awareness and self-preservation. Secondly, if an AI model underperforms during the testing phase, it could be released to the public without understanding its full capabilities.

Elon Musk, the billionaire who owns X and the AI Grok, said last year on Joe Rogan’s podcast that the chance of annihilation from AI is about 20%. Musk has also said that AI is more dangerous than nuclear weapons. He talks about the unpredictability of the models and the creation of a super-intelligent entity that may develop goals that are not aligned with humans. Combining these factors with the rapid growth of technology and the global race to lead the field of AI development makes the future seem uncertain.

Steven Middendorp

Steven Middendorp is an investigative journalist, musician, and teacher. He has been a freelance writer and journalist for over 20 years. More recently, he has focused on issues dealing with corruption and negligence in the judicial system. He is a homesteading hobby farmer who encourages people to grow their own food, eat locally, and care for the land that provides sustenance to the community.

Other Headlines

Coronavirus

Fauci Advisor Indicted for Conspiracy, Deleting and Concealing Records; Faces 51 Years in Prison

David Morens, a former senior advisor to Dr. Anthony Fauci in the National Institute of Allergy and Infectious Diseases (NIAID), has been indicted by the Department of Justice for conspiracy against the United States, two counts of destruction, alteration, or falsification of records in federal investigations, and two counts of concealment, removal, or mutilation ofContinue reading Fauci Advisor Indicted for Conspiracy, Deleting and Concealing Records; Faces 51 Years in Prison

More news about Coronavirus

Health & Nutrition

HHS Implements Strategy To Stop Overmedicalization With Psychiatric Medications

HHS is implementing a strategy and framework to reduce the overprescription of psychiatric medications. The administrators of four HHS agencies penned a Dear Colleague letter encouraging health care providers to provide full, informed consent for medications while prioritizing non-pharmacological methods as the first line of defense. “HHS encourages clinicians and provider organizations to support aContinue reading HHS Implements Strategy To Stop Overmedicalization With Psychiatric Medications

More news about Health & Nutrition

Vaccines

No Duty of Care: Ontario Court Dismisses Lawsuit Over 17-Year-Old’s Death 33 Days After COVID Vaccine

The Ontario Court of Appeals has dismissed the case of Dan Hartman, who sued federal Canadian government officials following the death of his 17-year-old son Sean, who died 33 days after receiving the COVID-19 vaccine as a requirement to participate in hockey. The lawsuit alleged the Attorney General of Canada and the Minister of Health,Continue reading No Duty of Care: Ontario Court Dismisses Lawsuit Over 17-Year-Old’s Death 33 Days After COVID Vaccine

More news about Vaccines

Science & Tech

Anthropic Unauthorized Access Investigation Raises Questions About AI Safety Amidst Rapid Development

Anthropic is investigating a reported security breach that allowed a small group of people to gain access to Claude Mythos Preview, the company’s AI software that is too powerful to release to the public. AI models are becoming increasingly capable, and the 2026 International AI Safety Report notes that some hypothetical scenarios pose risks asContinue reading Anthropic Unauthorized Access Investigation Raises Questions About AI Safety Amidst Rapid Development

More news about Science & Tech

Environment

Supreme Court Considers Granting Bayer Protection From Pesticide ‘Failure-To-Warn’ Lawsuits

The U.S. Supreme Court heard oral arguments Monday about whether Bayer can be held liable for state-level failure to warn claims when the company followed EPA labeling guidelines after the federal agency classified glyphosate as “not likely carcinogenic to humans.” The court is not considering the underlying safety profile of glyphosate or Roundup, but theContinue reading Supreme Court Considers Granting Bayer Protection From Pesticide ‘Failure-To-Warn’ Lawsuits

More news about Environment

Policy

Vaccine Autism Researcher Accused of Stealing $1 Million From CDC Extradited to U.S.

A Danish epidemiologist, who is a co-author on a widely cited paper that claims there is no association between the MMR vaccine and autism, has been extradited to the United States on fraud charges. Poul Thorsen was indicted in April 2011 on 13 counts of wire fraud and 9 counts of money laundering for divertingContinue reading Vaccine Autism Researcher Accused of Stealing $1 Million From CDC Extradited to U.S.

More news about Policy