MITRE ATLAS Deep Dive: Threat Intelligence for AI Systems in 2026
MITRE ATLAS Deep Dive
Introduction
MITRE ATLAS (Adversarial Threat Landscape for Artificial-Intelligence Systems) is a globally accessible, community-driven knowledge base that catalogs adversary tactics, techniques, and procedures (TTPs) specifically targeting artificial intelligence and machine learning systems. Launched in June 2021 by the MITRE Corporation — the same organization behind the widely adopted ATT&CK framework — ATLAS provides security teams with a structured, adversary-centric approach to understanding, detecting, and defending against AI-specific threats.
As organizations rush to deploy AI across healthcare, finance, autonomous vehicles, and critical infrastructure, the attack surface has expanded far beyond what traditional cybersecurity frameworks were designed to address. Attackers now poison training data, steal proprietary models, manipulate inference outputs via prompt injection, and exploit the unique characteristics of machine learning pipelines. ATLAS brings the same disciplined TTP taxonomy that transformed conventional threat intelligence into the AI security domain.
This post provides a comprehensive deep dive into ATLAS. It examines the framework’s structure and every tactic in detail, maps the key similarities and differences between ATLAS and ATT&CK, walks through real-world case studies, explains how to use ATLAS for systematic threat analysis, and concludes with a practical threat modeling exercise against a realistic AI application.
If you are looking for coverage of other AI security frameworks or the OWASP Top 10 for LLMs, see the companion posts Prominent AI Security Frameworks: A Practical Guide for 2026 and OWASP Top 10 for LLM Applications: An In-Depth Guide for 2026.
Overview of MITRE ATLAS
ATLAS is both a knowledge base and an analysis framework. It aggregates real-world AI attack data from academic research, security incident reports, red team exercises, and industry disclosures, then organizes them into a matrix that security practitioners can immediately operationalize. The framework is available in STIX 2.1 format for machine-readable integration with threat intelligence platforms and SIEM systems, and it comes with companion tools — the ATLAS Navigator for visual threat mapping and the ATLAS Arsenal for red team simulation.
Figure 1 — MITRE ATLAS at a glance: key statistics, evolution timeline, and core components.
ATLAS is a living, community-supported resource. Contributions come from major technology companies, government agencies (including CISA and NATO), academic institutions, and the Center for Threat-Informed Defense. The October 2024 AI Incident Sharing initiative created a mechanism for organizations to report anonymized real-world AI attacks, and the October 2025 collaboration with Zenity Labs added 14 new techniques specifically addressing autonomous AI agent security risks.
Key Similarities and Differences with MITRE ATT&CK
Security practitioners familiar with ATT&CK will find ATLAS immediately recognizable — that is by design. But the frameworks serve different threat landscapes, and understanding where they overlap and where they diverge is essential for applying both effectively.
What they share
Both frameworks use the same structural model: a matrix where columns represent adversary tactics (the “why”) and cells contain specific techniques (the “how”). Both use unique identifiers (ATT&CK uses T-codes; ATLAS uses AML.T-codes), both document real-world procedures, and both are maintained as living knowledge bases with community contributions. ATLAS inherits many of its tactics directly from ATT&CK — including Reconnaissance, Resource Development, Initial Access, Execution, Persistence, Privilege Escalation, Defense Evasion, Credential Access, Discovery, Collection, Exfiltration, and Impact.
Where they diverge
The critical differences lie in scope, unique tactics, and the nature of the attack surfaces they describe.
Figure 2 — Comparison of MITRE ATT&CK and MITRE ATLAS across key dimensions.
The most structurally significant difference is that ATLAS drops ATT&CK’s Lateral Movement and Command & Control tactics — which are less relevant when the target is a machine learning model rather than a network — and replaces them with AI-native tactics such as ML Model Access and ML Attack Staging. These new tactics reflect the reality that attacking an AI system often requires a fundamentally different preparation phase: the adversary may need to train a proxy model, craft adversarial inputs, or corrupt a training pipeline rather than move laterally across a network.
Another important difference is the nature of the evidence base. ATT&CK relies heavily on threat group profiles and software mappings observed across thousands of incidents. ATLAS, being a younger framework, currently draws from 42 documented case studies that include both real-world incidents and realistic red team demonstrations. ATLAS explicitly distinguishes between these two evidence types — “incident” versus “exercise” — helping security teams calibrate the real-world maturity of each technique.
Comprehensive Description of the MITRE ATLAS Matrix
The ATLAS matrix is the framework’s centerpiece. It arranges adversary tactics as columns from left to right, roughly following the chronological progression of an attack, with techniques listed under each tactic. The following is a detailed walkthrough of every tactic, its purpose, and the key techniques it contains.
Figure 3 — The MITRE ATLAS tactic lifecycle: 16 tactics grouped by phase, with details for each.
The AI-specific tactics in detail
Two tactics deserve special attention because they have no equivalent in ATT&CK and represent genuinely novel adversarial behaviors.
ML Model Access (AML.TA0004) describes how adversaries gain the ability to interact with a target ML model. This can range from querying a public inference API (low access) to directly downloading model weights from a compromised model registry (high access). The level of access determines which downstream techniques are available — an adversary with only API access can attempt evasion and extraction attacks, whereas one with artifact access can inspect weights, embed backdoors, or perform membership inference. Key techniques under this tactic include ML Model Inference API Access (AML.T0040) and Physical Environment Access, where the adversary interacts with an ML-enabled system in the physical world (such as placing adversarial patches in front of a camera).
ML Attack Staging (AML.TA0012) covers the specialized preparation that adversaries undertake before executing attacks against ML systems. Unlike traditional staging (which might involve setting up C2 infrastructure), ML Attack Staging involves creating proxy models to develop transferable adversarial examples, crafting poisoned training data, developing backdoor triggers, and preparing prompt injection payloads. This tactic reflects the unique reality that many AI attacks require a significant R&D phase. Key techniques include Craft Adversarial Data (AML.T0043), Backdoor ML Model (AML.T0018), and Verify Attack (AML.T0042), where the adversary tests their attack against a proxy before deploying it against the real target.
The Agentic AI expansion (October 2025)
The October 2025 update, developed in collaboration with Zenity Labs, added 14 new techniques and sub-techniques specifically targeting autonomous AI agents — systems that can use tools, browse the web, execute code, and take real-world actions. These additions address risks such as tool-use manipulation (tricking an agent into misusing its tools), memory poisoning (corrupting an agent’s persistent memory to influence future behavior), and prompt injection variants specific to multi-step agentic workflows. This expansion reflects the rapid shift from simple chatbot deployments to autonomous AI systems with real-world agency.
MITRE ATLAS for Threat Analysis: A How-To Guide
Using ATLAS for systematic threat analysis involves a structured process that maps your AI system’s architecture to the ATLAS matrix, identifies relevant attack surfaces, and prioritizes defenses. The following methodology can be applied to any AI-enabled system.
Figure 4 — Six-step methodology for conducting threat analysis using MITRE ATLAS.
Step 1 — Inventory AI assets
Begin by cataloging every AI and ML component in your system. This includes models (and their versions), training and fine-tuning datasets, feature stores, embedding databases, inference APIs, model registries, training pipelines, and any third-party AI services you consume. For each asset, document the data flows, who has access, and where trust boundaries exist. Pay particular attention to model access points — a public inference API presents a very different risk profile from an internal batch inference pipeline.
Step 2 — Define threat actors and motivations
Identify who might target your AI systems and why. Competitors may seek to steal proprietary models (model extraction). Insiders could poison training data. Nation-state actors might target AI systems in critical infrastructure. Criminal groups could exploit AI-powered authentication systems for fraud. The threat actor’s motivation directly determines which ATLAS tactics and techniques are most relevant.
Step 3 — Map to the ATLAS matrix
Walk through the ATLAS matrix from left to right, tactic by tactic. For each tactic, evaluate which techniques are applicable to your system given the threat actors you identified. Use the ATLAS Navigator to create a visual overlay showing your system’s exposure. This systematic walk-through prevents the common mistake of focusing only on familiar attacks (like prompt injection) while missing less obvious but equally dangerous threats (like ML supply chain compromise or backdoor insertion during fine-tuning).
Step 4 — Assess risk and prioritize
For each applicable technique, estimate the likelihood of exploitation and the business impact if it succeeds. Cross-reference ATLAS case studies to ground your assessment in real-world evidence — if a technique has been used in a documented incident against a system similar to yours, its likelihood score should reflect that. Prioritize by the product of likelihood and impact, weighted by business criticality.
Step 5 — Select mitigations
Map the ATLAS-recommended mitigations to each prioritized risk. A significant advantage of working with ATLAS is that approximately 70% of its mitigations correspond to security controls that many organizations already have in place (input validation, access controls, monitoring, supply chain verification). This means the effort focuses on identifying and filling the 30% gap that requires AI-specific controls — such as adversarial robustness testing, model watermarking, or differential privacy.
Step 6 — Validate and iterate
Use the ATLAS Arsenal tools or third-party adversarial AI testing platforms to simulate attacks against your system. Red team exercises should be structured around the specific ATLAS techniques you identified as highest risk. After validation, report any novel findings through MITRE’s AI Incident Sharing mechanism and update your threat model. Repeat this cycle at least quarterly, or whenever you deploy a new model or significantly change your AI pipeline.
Case Studies
ATLAS documents over 40 real-world case studies, each mapped to specific tactics and techniques. The following three cases illustrate the breadth of threats the framework covers — from facial recognition fraud to ML malware evasion to LLM-targeted attacks.
Case Study 1 — Camera Hijack Attack on a Facial Recognition Tax System
Two individuals in China exploited an ML-enabled facial recognition system used by the Shanghai Tax Authority. They created a fake shell company and used the compromised authentication system to submit fraudulent invoices to fictitious clients, stealing over $77 million.
Figure 5 — Attack chain for the facial recognition tax fraud case study, mapped to ATLAS tactics.
Case Study 2 — Evasion of a Machine Learning Malware Scanner
Adversaries targeted a well-known ML-based endpoint malware scanner. Through reconnaissance of conference presentations, patents, and technical documentation, they deduced the scanner’s detection logic. By enabling verbose logging and studying the API, they identified a specific reputation-scoring mechanism. They then crafted a universal bypass — a set of benign-seeming characteristics that, when appended to malicious files, caused the model to consistently classify them as safe. This bypass was effective across multiple malware families.
The attack chain progressed through Reconnaissance (AML.TA0002), where the attackers gathered public information; ML Model Access (AML.TA0004), where they studied the product’s decision logic via its API; and ML Attack Staging (AML.TA0012), where they crafted the universal adversarial perturbation. The Impact (AML.TA0011) was a complete evasion of the malware detection system. The recommended mitigations include adversarial training (retraining the model with adversarial examples), restricting public exposure of technical details, rate-limiting and monitoring API access patterns, and deploying ensemble detection methods that are harder to universally evade.
Case Study 3 — ShadowRay AI Infrastructure Data Leak
In late 2023, multiple vulnerabilities were discovered in the Ray framework, a widely used open-source platform for distributed AI workloads. Unknown attackers exploited these vulnerabilities over a span of seven months to invoke arbitrary jobs on remote hosts with access to Ray production clusters. This enabled theft of sensitive information including model weights, training data, and environment credentials, as well as unauthorized access to compute resources for cryptocurrency mining. The estimated cost to affected users was nearly $1 billion in hijacked compute time.
This case study illustrates that AI security is not just about adversarial ML — it also encompasses the infrastructure that supports AI systems. The attack chain mapped to traditional tactics (Initial Access via vulnerability exploitation, Execution of arbitrary code, Credential Access, Exfiltration of sensitive data) but targeted specifically AI workloads and training infrastructure. It led to the addition of new infrastructure-targeting techniques to the ATLAS matrix through the Center for Threat-Informed Defense’s Secure AI research project.
Case Study 4 — Deepfake Attacks on Mobile KYC Systems
In November 2025, ATLAS documented attacks against Know Your Customer (KYC) liveness detection systems used by banking, financial services, and cryptocurrency platforms. Attackers gathered target identity information from social media profiles, acquired commercially available face-swap AI tools, and used OBS virtual camera injection to bypass physical camera requirements. The AI-generated deepfakes defeated the platforms’ liveness detection algorithms, enabling fraudulent account creation and identity verification bypass.
Figure 6 — Common patterns and lessons learned from ATLAS case studies.
Practical Threat Modeling Exercise Using MITRE ATLAS
To make the methodology concrete, let us walk through a full threat modeling exercise against a realistic AI application: an AI-powered financial advisory platform.
System description
WealthBot AI is a hypothetical AI-powered financial advisory platform offered by a mid-size fintech company. It uses a fine-tuned large language model to provide personalized investment advice, portfolio analysis, and market insights to retail customers. The system includes a RAG pipeline that retrieves real-time market data, customer portfolio information, and regulatory documents. Customers interact via a web and mobile interface. The platform also uses a fraud detection ML model to monitor transactions.
Figure 7 — Architecture of WealthBot AI showing trust boundaries, components, and key threat surfaces mapped to ATLAS tactics.
Applying the ATLAS threat model
Walking through the ATLAS matrix tactic by tactic against this system produces the following threat analysis.
Reconnaissance (AML.TA0002). An adversary researches the fintech company’s AI capabilities by examining job postings (which reveal the ML stack), published API documentation, blog posts about the company’s AI strategy, and any open-source model components listed in dependency files. The company’s model card, if published, provides architecture details. Mitigation: Minimize public disclosure of ML implementation details. Review all public-facing content for information that could aid adversarial reconnaissance.
Resource Development (AML.TA0003). The adversary acquires compute resources to train a proxy model mimicking WealthBot’s advisory LLM. They also set up infrastructure for hosting phishing sites targeting company employees. Mitigation: Monitor for proxy model training using your API (anomalous query patterns). Implement phishing-resistant authentication for all employees with access to AI systems.
ML Attack Staging (AML.TA0012). The adversary fine-tunes their proxy model using data gathered from API queries to develop transferable adversarial prompts. They also prepare poisoned market commentary documents designed to be ingested by the RAG pipeline. For the fraud detection model, they craft adversarial transaction patterns that appear legitimate to the ML classifier. Mitigation: Rate-limit and monitor inference API usage. Validate and sanitize all data entering the RAG pipeline. Conduct adversarial robustness testing on the fraud model.
Initial Access (AML.TA0005). The adversary gains access to the system through a compromised third-party market data feed (supply chain compromise) or through a phishing attack against a data engineer with pipeline access. Mitigation: Verify integrity of all third-party data sources. Enforce MFA and least-privilege access for all pipeline roles.
ML Model Access (AML.TA0004). The adversary accesses the advisory LLM through the public customer API (inference API access). For the fraud detection model, they gain access by observing the accept/decline decisions on their own transactions (oracle access). Mitigation: Implement query budgets and behavioral analysis on API usage. Obfuscate fraud model decisions — avoid binary accept/reject without explanation controls.
Execution (AML.TA0006). The adversary submits prompt injection attacks via the customer chat interface, attempting to override the system prompt and make the LLM execute unauthorized tool calls — such as initiating trades, transferring funds, or revealing other customers’ portfolio data. Mitigation: Implement instruction hierarchy (system prompt > user prompt). Apply tool-level authorization and confirmation flows. Sandbox all tool calls.
Persistence (AML.TA0007). The adversary injects poisoned documents into the RAG knowledge base through the compromised market data feed. These documents contain hidden instructions that will be retrieved in future queries, providing persistent prompt injection capability. Mitigation: Implement content validation on all ingested data. Use separate vector namespaces for trusted vs. external content.
Defense Evasion (AML.TA0013). The adversary crafts adversarial transactions designed to evade the fraud detection model, using techniques derived from their proxy model testing. They also use jailbreak techniques to bypass the LLM’s safety guardrails. Mitigation: Deploy ensemble fraud detection (multiple model architectures). Implement multi-layer guardrails (input filter + model + output filter).
Exfiltration (AML.TA0010). The adversary uses model extraction techniques — systematic querying of the advisory LLM to reconstruct its fine-tuning data or specialized knowledge. They may also use prompt injection to exfiltrate other customers’ data through the LLM’s responses. Mitigation: Implement differential privacy. Monitor for extraction patterns (high-volume systematic queries). Enforce strict data access controls at the tool layer.
Impact (AML.TA0011). Successful attacks could result in manipulated investment advice (causing customer financial losses), evaded fraud detection (enabling unauthorized transactions), stolen proprietary models (competitive loss), leaked customer financial data (regulatory penalties), or degraded service (reputational damage). Mitigation: Defense-in-depth across all layers.
Summary of mitigations
Figure 8 — Five layers of defense-in-depth for WealthBot AI, mapped to ATLAS tactics.
Conclusion
MITRE ATLAS provides the most structured, adversary-centric approach to AI security threat intelligence currently available. By extending the proven ATT&CK methodology into the AI domain, it gives security teams a familiar framework for understanding a fundamentally new class of threats. The framework’s strength lies in its combination of a comprehensive TTP taxonomy, real-world case studies grounded in actual incidents, and practical tooling for threat modeling and red teaming.
The most important takeaway from this deep dive is that AI security cannot be addressed by protecting models alone. The ATLAS case studies consistently show that adversaries combine traditional IT attack techniques (phishing, vulnerability exploitation, supply chain compromise) with AI-specific techniques (data poisoning, model extraction, adversarial evasion) in the same attack chain. Effective defense therefore requires using ATLAS alongside ATT&CK, with AI-specific controls layered on top of a strong conventional security foundation.
For organizations beginning their AI security journey, the six-step threat analysis methodology outlined here provides a practical starting point. For those further along, integrating ATLAS into existing threat modeling processes, deploying the Navigator and Arsenal tools for continuous assessment, and contributing to the AI Incident Sharing initiative will produce the most comprehensive coverage.
See also: OWASP Top 10 for LLM Applications: An In-Depth Guide for 2026 for a complementary vulnerability-focused perspective, and Prominent AI Security Frameworks: A Practical Guide for 2026 for a broader survey of the AI security framework landscape.