Anthropic’s “Claude Mythos”: What CISOs Need to Adjust in Their Threat Model Now

Author
Mathias Fuchs
Published
22. April 2026

Share article

With ‘Claude Mythos’, Anthropic has unveiled an AI model that the company itself considers too dangerous for public release – and which is already causing alarm among authorities and businesses. AI-powered cyberattacks will thus become faster, more scalable and capable of wider application. CISOs face an urgent need for action: their own threat models must be reviewed and specifically adapted. This article provides a data-driven analysis of the development – moving beyond hype and downplaying the risks – and outlines what matters most right now with a concrete 90-/180-/365-day playbook.

On April 7, 2026, Anthropic announced a new AI model with "Claude Mythos Preview". At the same time, Project Glasswing was created, an exclusive consortium of selected technology and infrastructure providers with access for defensive security work. While the US Treasury Secretary and Fed Chairman were briefed internally, the IMF publicly expressed doubts about the defensive capabilities of the global financial system. German banks also immediately sought to close ranks with their supervisory authorities.

Since then,reactions in the security community have ranged from "AGI is here" to "pure PR theater". Neither is good for CISOs: those who panic make expensive mistakes. Those who turn a blind eye will miss a real shift in the threat model.

What "Claude Myth" is - and what it is not

The announcement initially reads like a paradigm shift: "Claude Mythos" is said to have independently identified thousands of zero-day vulnerabilities in major operating systems and browsers - including a 27-year-old bug in OpenBSD, one of the most secure systems in the industry. 99% of these vulnerabilities were unpatched at the time of publication. The UK AI Security Institute (AISI) also confirmed a 73% success rate in expert-level hacking tasks. At the same time, Mythos is the first model to be withheld from release primarily for cybersecurity reasons.

But that's only half of the story. The story continues:

AISI itself points out that Mythos went up against almost non-existent layers of defense in the test scenarios. The comparison made by an AISI reviewer is appropriately pointed: a striker against the worst goalkeeper in the world. In a productive enterprise context, the setup looks different. The organization AISLE, which specializes in AI-supported vulnerability research, also puts the public narrative into perspective: it has cross-checked the myth-findings published by Anthropic against small, inexpensive open-weight models. Eight out of eight models tested discovered the FreeBSD flagship exploit that Anthropic presented as evidence of Mythos' superiority. One of these models has 3.6 billion active parameters and costs around USD 0.11 per million tokens.

Researchers have also come to the same conclusion. Peter Swire, professor at Georgia Tech and former advisor to the Clinton and Obama administrations, describes therelease not as a turning point, but asa continuation of a predictable trend , according to discussions with academic colleagues.At the same time, he points out that CISOs and security vendors alsohave a rational incentive to exaggerate the scope of new capabilities - not least to justify their own budgets.

The sober summary is therefore: Mythos is powerful and a capability to be taken seriously. However, it is not the pure capability that is decisive for a changed threat model - some of these already existed. What is relevant is the combination of accessibility, speed, scaling and the parallel democratization of comparable capabilities in open-weight models.

This is the shift we need to talk about. Not: "AI can now hack." But rather: AI-supported offensive capability is taking on commodity trajectory.

What is really changing in the threat model: These basic assumptions need to be examined

Four dimensions in which the CISO should check his basic assumptions.

  1. Time-to-exploit is collapsing
    The classic assumption that there are weeks to months between disclosure and weaponized exploit code is no longer viable. As PwC puts it in its latest threat report: AI-supported attackers are shortening attack timelines and scaling operations simultaneously. In terms of risk quantification, this means that exposure windows tend to be shorter than manufacturers' patch windows.

  2. Low-skill enablement

    The most serious effect is not that APT groups are getting better - they are anyway. Rather, the decisive factor is that mediocre attackers are becoming significantly more dangerous. In future, a ransomware affiliate without in-depth OS knowledge will be able to search for exploit chains in a structured manner instead of relying exclusively on purchased exploits. This shifts the distribution of the real threat measurably upwards - and precisely in the segment that is most relevant for SMEs.

  3. Scaling and breadth
    According to Anthropic, Mythos produced thousands of findings in parallel during the test. Even if only a fraction of these can be converted into real exploits, this changes the assumptions about "simultaneously active vulns". Traditional vulnerability management assumes that you can prioritize because you know which vulnerabilities are being actively exploited. Signals such as CISA KEV and EPSS become noisier (more unreliable) the more simultaneously discovered zero-dayscome into circulation.

  4. Asymmetry in Glasswing

    The program is not neutral. Selected companies, including JPMorgan Chase, Goldman Sachs and some cloud/OS vendors, get Mythos access for defensive purposes. The rest of the economy does not get it. At the same time, it is by no means guaranteed that attackers cannot obtain comparable tooling; the AISLE findings on open-weight models show that parts of the capability already exist outside of controlled channels. For medium-sized and smaller companies - the typical customers of a European IR team - the asymmetry shifts to the disadvantage of the defense.

Mythos is neither a revolution nor PR theater, but a clear signal: AI-supported offensive capabilities are becoming a commodity - and the existing threat model now needs to be sharpened. The decisive factor is not a new tool, but an honest review of one's own assumptions - in operations, in prioritization and in discussions with the Executive Board.

Threat Modeling

Risk quantification: must-have adjustments to the threat model

Anyone working with FAIR, cyber risk quantification or heat map approaches must revise three parameters.

Probability of Successful Compromise per asset class. The priors on older, poorly maintained systems (ICS/OT, legacy web applications, internal Java backends, Delphi tools from the late nineties, forgotten VB6 applications) are increasing disproportionately. A 27-year bug in OpenBSD is not an isolated case, but a proof-of-concept for what lies dormant in every enterprise COBOL stack and every half-forgotten internal application. Legacy risk that was three decades below the realization threshold is now within reach of plausible attackers.

Time-to-detect versus time-to-exploit. Many organizations operate with MTTD values of days to weeks. When time-to-exploit collapses to hours, the delta is the real risk. The metric that will be tracked from now on is not MTTD per se, but MTTD minus estimated TTE - negative values are red flags and belong in board reporting.

Patch SLA models. Classic 30/60/90-day windows were designed in a world where disclosure-to-exploit took a median of several weeks. This model needs either much stricter tiers for "critical exposed assets" or an extension to include a compensating controls dimension: virtual patching, WAF rules, network segmentation as a time-critical intermediate measure, not as an option.

Specifically for quantification:

  • CVSS base alone is no longer sufficient; EPSS score plus exposure context becomes the standard layer.

  • For risk acceptance decisions on unpatched systems, an explicit specification of the compensating controls is necessary. A management signature is no longer sufficient.

  • Threat modeling sessions list "AI-enabled attacker" as a standard threat actor, not as a special case.

SOC and IR operations: What is changing in everyday life

This is where it gets concrete - and most relevant for IR teams.

  1. Signature-based detection continues to lose weight
    This is not a new trend, but it is accelerating. If an attacker can generate functionally equivalent but IOC-free variants of known malware families with minimal effort, signature detection is still primarily suitable for noise reduction. Investments should be shifted to behavioral detection (EDR-Behavioral, UEBA), Abuse-of-Legitimate-Tools-Pattern (Living-off-the-Land) and anomaly detection in the network and identity layer.

  2. Alert triage must be machine-assisted
    If the attacker side scales with AI support, the defender side cannot compensate for this manually. Specifically: LLM-supported first-line triage that forms alert clusters, enriches context (asset, owner, last changes, historical false positives) and makes prioritization suggestions. This is not a replacement of analysts, but a shift in their work - away from "what is this?" to "is the hypothesis right, and what is the playbook?".

    Two architectural decisions belong directly on the table:


    1. Deployment model. The LLM layer in the SOC needs a clean deployment. On-prem or dedicated-tenant, with clear data boundaries. Alert data is privileged; it does not belong in a generic cloud LLM endpoint.

    2. Model Drift as an operational risk. Anyone using LLM-supported detection or triage inherits a problem that does not appear in classic detection engineering textbooks: frontier models behind APIs change their behavior on identical inputs silently and without a changelog. Thresholds that have been calibrated against a certain confidence distribution wander away under their own dashboard - silently increasing the false negative rate without a version being bumped anywhere.

    Even self-hosted stacks are not immune: a vLLM update
    , a new quantization or a subtle tokenizer change shifts outputs measurably without anyone flagging a change. For detection-critical workloads, a pinned local model is therefore often the operationally better choice than the strongest frontier model: reproducible, auditable, immune to silent updates.
    The frontier model remains useful for the actual analyst work (summaries, hypothesis generation, client communication). I have described this split between machine-side decision-making (local, pinned) and human-facing generation (frontier, drift-tolerant), including the monitoring toolbox - Golden Corpus, PSI/JS-Divergence, Refusal-Rate-Tracking - in more detail elsewhere: Your AI Detections Are Rotting: Model Drift as a Hidden Risk in Security Operations.
    The regulatory point there is even more important here: under NIS2 and DORA, an LLM whose behavior cannot be recorded is a compliance liability, not a capability.
    This is the point at which the myth reflex "we need more AI in the SOC now" can be misleading. More AI without drift monitoring creates the next blind-spot problem - only now with a prettier UI.

  3. IR playbooks: zero-day assumption by default
    The classic distinction between "known vulnerability, default playbook" and "zero-day, escalate to senior responder" is eroding. If attackers have more frequent zero-days in their portfolio - because finding them has become cheaper - the standard playbook must take this possibility into account.

    Specifically:

    ▪️Initial hypothesis in the major incident: "unknown initial access vector" is given more weight before it is cleared.
    ▪️ Forensic reservation is triggered earlier, even if the incident classification is still uncertain.
    ▪️ Tabletop exercises run "AI-enabled adversary" as a standard scenario, not as a special exercise.In our hybrid crisis simulations, we see that this type of scenario is often treated as a special case. The timing to normalize it is now.

  4. Threat hunting proactively on legacy assets
    When myth class tooling is in attacker hands, the hunting priority shifts to assets that were previously assumed to be "too old, too boring for an attacker". The proprietary in-house application from 2007, which has "actually had to be replaced" for two CISO generations, becomes a plausible initial access vector. Hunting hypotheses must be prioritized accordingly.

Proactive Threat Hunting

Patch and vulnerability management: the real pain point

This area is changing the most because it was built most heavily on the old assumptions.

  1. Exposure Management replaces Vulnerability Management
    The industry has been talking about it for years, the myth moment is accelerating it. Vulnerability management asks: which CVEs affect me? Exposure management asks: which attack paths are realistically exploitable, and where is my defense weakest? The second question requires attack path analysis, reachability-aware SCA and continuous external attack surface management capabilities (EASM).

  2. Virtual patching becomes the default, not an option
    If the manufacturer's patch cycles take longer than exploit development, the gap must be bridgeable. WAFs, IPS, deception tech, network segmentation and micro-perimeter policies are no longer "nice to have". For critical assets, there must be a compensating control in the toolbox that can be activated faster than the vendor can patch. This is an architectural point, not a tooling point - and it needs to be resolved before the next big zero-day, not during.

  3. SBOM plus VEX: From compliance to operations
    SBOMs have been a compliance artefact in many organizations - for EU Cyber Resilience Act, Executive Order 14028 or sector regulations. According to Mythos, the operational question is: how quickly can I localize a newly disclosed library vulnerability in my entire software portfolio? Without SBOM: days to weeks. With SBOM plus VEX (Vulnerability Exploitability eXchange): Minutes. This difference in speed will determine incident costs in the future.

  4. Reorganize prioritization
    The typical question "Which critical CVEs do I need to patch this week?" is replaced by "Which combinations of exposure, exploitability and business impact currently have the worst risk profile?". CISA KEV, EPSS and exposure-aware scoring become the standard layer. Monthly patch boards are no longer enough; patch and mitigation decisions become a continuous process with a few clearly prioritized escalation paths.

Vendor and supply chain strategy

This is one of the most unpleasant effects. Glasswing creates a two-tier economy.

  1. Vendor due diligence: new questions
    Vendor questionnaires need new points:

    ▪️Nutzt the vendor AI-supported security analysis in their own development? Since when? What coverage (code, dependencies, build pipeline)?
    ▪️Gibt are there SBOMs for all purchased products and versions? In which format (SPDX, CycloneDX)?
    ▪️Existieren VEX statementsfor known CVEs?
    ▪️Time to patch SLA for critical findings detected by automated analysis?
    ▪️Für critical vendors: Participation in Glasswing-type programs or comparable AI-assisted-defense-capability?If no - why not, and what is the alternative?

     

  2. Concentration risk becomes visible
    Glasswing partners are listed by name at Anthropic. For European companies, the operative question is: how much of their own risk stack depends on vendors with this capability versus vendors without? This is not a moral argument, but a risk argument: defensive AI capability becomes part of the vendor risk assessment.

  3. Use regulatory levers
    Relevant for the DACH region: NIS2 (EU-wide), DORA (financial sector), CRA (product security) and in Austria the NISG 2024 already offer levers to demand structured answers to these questions from critical suppliers. The tools exist; they just need to be sharpened for the new threat situation. Those who previously saw CRA requirements as compliance costs should now reframe them as risk mitigation levers.

  4. Open source dependencies
    The other side of the Glasswing coin: Anthropic has pledged USD 100 million in usage credits and USD 4 million directly to open source security organizations. If your organization is an upstream open source contributor or relies on critical OSS infrastructure, active participation in these defensive efforts - OpenSSF initiatives, direct contact with upstream maintainers, contributor backing for critical libraries - is a direct risk mitigator for your supply chain.

A pragmatic 90-/180-/365-day playbook

Enough analysis. What should be done now?

Playbook: The first 90 days

  1. Assumption Update. Threat model review session in which "AI-enabled adversary" is introduced as the default threat actor profile. Affected: all critical assets, all Tier 1 processes.

  2. Tabletop. At least one TTX exercise with scenario "moderately competent attacker, AI-assisted, unknown initial access method against legacy asset". Document playbook gaps.

  3. Legacy inventory. List of all assets older than ten years, especially internally developed software. Classify by exposure and business criticality. Most likely initial attack surface.

  4. Check detection engineering backlog. Measure proportion of signature-based vs. behavior-based detections. Set target value for shift.

  5. Vendor list of the top 20 critical vendors. Add the points from section 6.1 to the questionnaire. Dispatch.

Playbook: Months 3 to 6

  1. Establish virtual patching capability. If not in place: WAF rule pipeline, rapid IPS signature deployment processes, network segmentation playbook for critical assets.

  2. Exposure Management. Pilot with an EASM or Attack Path Analysis tool. Goal: Move from "which CVEs?" to "which paths?".

  3. SBOM program. For internally developed products: automated SBOM generation in the build. For externally sourced software: systematic SBOM request, staggered according to criticality.

  4. LLM-assisted SOC triage - with drift monitoring from day one. Pilot with dedicated, data protection-compliant LLM integration in SIEM/SOAR. Define data boundaries before the pilot, not after. Mandatory accompaniment: golden corpus with 200-2,000 reference inputs, output distribution monitoring (PSI, JS divergence), refusal rate tracking and a clear separation between machine decision (pinned local model) and analyst-facing generation (frontier model). Without this layer, any statement on detection quality after three months of production is unsubstantiated.

  5. IR playbook review. Zero-day assumption as default in major incident procedures. Preservation trigger earlier.

Playbook: Months 6 to 12

  1. Adapt risk quantification model. Revise time-to-exploit estimates, restructure SLA tiers accordingly.

  2. Vendor rescoring. Re-evaluate top vendors based on the answers from phase 1. Particularly critical are unclear statements on AI-supported security analysis, missing SBOM/VEX evidence or unrealistic patch SLAs. In the event of such gaps, conduct targeted escalation discussions and make concentration risks visible in board reporting.

  3. Threat hunting on legacy. Dedicated hunting campaigns on the identified legacy assets. Hypotheses: unknown initial access via old library vulnerability, lateral movement via deprecated admin log.

  4. Participation. Consider participation in sector-specific information sharing initiatives (ISACs, CERT.at, sectoral CSIRTs) that share AI-enabled threat intelligence. Working in isolation will become more expensive.

  5. Redesign board reporting. Metrics that were meaningful two years ago (number of CVEs patched, number of alerts handled) are less so than ever. Replace with exposure metrics, MTTD-minus-TTE proxies, vendor risk heat maps.

Anthropic's Claude Myth misreactions that get expensive

Three common reaction patterns that will become costly in the coming months:

  1. Panic process rebuild. Anyone who rebids the entire vulnerability management program now "because of myth" is giving away six months on a process that is two iterations too far from current reality. Incremental adjustments are almost always superior.

  2. Buy AI security snake oil. The next waves of marketing will carry "myth-proof" labels. The word has no technical meaning. The counter-questions to ask a vendor: how does the product handle AI-generated malware variants? What is the false positive rate for LLM-assisted detections? Which training data, which adversarial robustness tests?

  3. Ignore everything. "We've always done it this way, it won't be so bad." This may be true, but the basic assumptions on which "we've always done it this way" is based have shifted. A review of the assumptions is cheap; an incident based on outdated assumptions is expensive.

  4. Treat LLM-supported detection as fire-and-forget. An LLM classifier looks like a correlation rule, but does not behave like one. Hosted model behavior drifts without notice; self-hosted stacks drift with vLLM updates, quantization changes or tokenizer changes. Without golden corpus regression and drift monitoring, every productive AI detection is poorly calibrated after a few months - and the operating point shift is typically only noticed after the incident that made it visible.

Takeaways and CISO's responses to Anthropic's Claude Myth

Mythos is not a revolution or a PR stunt. It's a clear signal that AI-powered offensive capability has taken a commodity trajectory and the defensive side needs to adapt accordingly - not overnight, not in a panic, but no later than 2026 either.

The pragmatic CISO operates in three modes simultaneously:

  • Continue day-to-day business

  • make incremental adjustments where the basic assumptions have measurably shifted

  • talk honestly with the Management Board about the remaining uncertainty.

All three are more important than the perfect answer to the question of whether Mythos "really" is AGI.

The real work does not start with a new tool. It starts with a hard review of your own assumptions.

Threat Modeling

Keep your finger on the pulse of digital security: discover exciting developments, in-depth analyses and the most important news from the world of cyber security. Subscribe to our blog updates and get the latest insights delivered straight to your inbox - compact, relevant and always one step ahead.

Blog Updates abonnieren

 

Sources:
- Anthropic: Announcement Claude Mythos Preview and Project Glasswing, April 7, 2026. https://red.anthropic.com/2026/mythos-preview/
- UK AI Security Institute (AISI): Independent Mythos Evaluation. https://www.aisi.gov.uk/blog/our-evaluation-of-claude-mythos-previews-cyber-capabilities
- AISLE: AI Cybersecurity After Myth: The Jagged Frontier - https://aisle.com/blog/ai-cybersecurity-after-mythos-the-jagged-frontier
- Scientific American: What is Mythos and why are experts worried about Anthropic's AI model - https://www.scientificamerican.com/article/what-is-mythos-and-why-are-experts-worried-about-anthropics-ai-model/
- Council on Foreign Relations: Six Reasons Claude Mythos Is an Inflection Point for AI and Global Security - https://www.cfr.org/articles/six-reasons-claude-mythos-is-an-inflection-point-for-ai-and-global-security
- Bloomberg: How Anthropic Discovered Myth AI Was Too Dangerous For Release - https://www.bloomberg.com/news/features/2026-04-16/how-anthropic-discovered-mythos-ai-was-too-dangerous-for-release
- The Hill: Anthropic's Mythos model sparks cybersecurity concerns - https://thehill.com/policy/technology/5829315-anthropic-mythos-ai-cybersecurity-risks/
- CBS News: Anthropic's Mythos AI can spot weaknesses in almost every computer on Earth - https://www.cbsnews.com/news/mythos-anthropic-ai-project-glasswing-hacker-threat/
- PwC: Global Digital Trust Insights 2026 (AI-enabled threat dynamics).
- Mat Fuchs: Your AI Detections Are Rotting: Model Drift as a Hidden Risk in Security Operations - https://medium.com/@mathias.fuchs/your-ai-detections-are-rotting-model-drift-as-a-hidden-risk-in-security-operations-cac014477248

Caption: Image generated with AI

Table of Contents
    Share article