Executive Summary
In June 2024, the UK’s National Cyber Security Centre (NCSC) publicly warned that large language models (LLMs), including popular AI tools such as ChatGPT and Claude, possess a fundamental and persistent vulnerability known as prompt injection. This flaw arises because LLMs are architecturally incapable of reliably distinguishing between trusted and untrusted input within prompts. Despite repeated industry efforts to implement guardrails, researchers routinely bypass these safeguards, allowing malicious actors to manipulate LLM behavior, potentially leading to harmful outputs or the execution of unauthorized actions in real-world applications that integrate LLMs.
This alert is especially significant as LLM-driven automations are rapidly proliferating in software development, browser agents, and enterprise workflows. The NCSC’s assessment signals an urgent need for organizations to shift their risk models, as AI prompt injection represents a persistent, unfixable attack vector with serious implications for data security, business integrity, and regulatory compliance.
Why This Matters Now
With the explosive adoption of generative AI across sectors, the inability to fully mitigate prompt injection exposes enterprises to new forms of exploitation and data leakage. As AI tools integrate deeper into critical processes, prompt injection’s persistence demands continuous security controls and vigilance, rather than relying solely on model-internal defenses.
Attack Path Analysis
Attackers initiated the kill chain by embedding malicious prompt injections within untrusted data sources targeting integrated LLM workflows. After initial compromise, the adversary leveraged manipulation of LLM-driven automation or developer tools to gain unauthorized access or escalate privileges within the environment. Lateral movement was enabled as the manipulated LLM interfaces or agents interacted with other internal applications or cloud services, potentially propagating malicious instructions or pivoting based on LLM output. The attacker established command and control by abusing outbound communications from cloud workloads or browser AI agents, using covert channels or egress paths to maintain persistence. Sensitive information was exfiltrated through automated LLM responses or by extracting data via compromised cloud services or developer pipelines. Ultimately, the attack impacted the organization by causing reputational harm, possible data loss, or business disruption from poisoned models or unauthorized code execution.
Kill Chain Progression
Initial Compromise
Description
The adversary delivered prompt injection payloads through untrusted user input, developer artifacts (e.g., GitHub commit messages), or manipulated web content accessed by LLM-integrated tools.
MITRE ATT&CK® Techniques
Phishing
User Execution
Modify Authentication Process
Access Token Manipulation
Adversary-in-the-Middle
Impair Defenses
Command and Scripting Interpreter: Application Layer Protocol
Potential Compliance Exposure
Mapping incident impact across multiple compliance frameworks.
PCI DSS 4.0 – Security of Application Development
Control ID: 6.3.1
NYDFS 23 NYCRR 500 – Cybersecurity Policy
Control ID: 500.03
DORA (EU Digital Operational Resilience Act) – ICT Risk Management Framework
Control ID: Art. 8
CISA Zero Trust Maturity Model (ZTMM) 2.0 – Identity-Based Policy and Control
Control ID: Identity Pillar: Policy Enforcement
NIS2 Directive – Risk Management Measures for Cybersecurity
Control ID: Article 21
Sector Implications
Industry-specific impact of the vulnerabilities, including operational, regulatory, and cloud security risks.
Computer Software/Engineering
AI development companies face critical prompt injection vulnerabilities affecting LLM architecture, requiring enhanced security frameworks and compliance measures for AI systems.
Financial Services
Banking institutions using AI chatbots and automated systems vulnerable to prompt injection attacks that could manipulate financial advice and transactions.
Health Care / Life Sciences
Healthcare AI tools susceptible to prompt injection compromising patient data confidentiality and treatment recommendations, violating HIPAA compliance requirements.
Government Administration
Government AI systems vulnerable to weaponization through prompt injection attacks, potentially compromising public services and sensitive administrative functions.
Sources
- UK cyber agency warns LLMs will always be vulnerable to prompt injectionhttps://cyberscoop.com/uk-warns-ai-prompt-injection-unfixable-security-flaw/Verified
- Prompt Injection | OWASP Foundationhttps://owasp.org/www-community/attacks/PromptInjectionVerified
- Prompt injection attacks might 'never be properly mitigated' UK NCSC warnshttps://www.techradar.com/pro/security/prompt-injection-attacks-might-never-be-properly-mitigated-uk-ncsc-warnsVerified
Frequently Asked Questions
Cloud Native Security Fabric Mitigations and ControlsCNSF
CNSF-aligned controls such as Zero Trust Segmentation, egress filtering, threat detection, and multicloud visibility would greatly constrain an attack by isolating AI workloads, limiting east-west spread, detecting anomalous LLM behaviors, and preventing unapproved data exfiltration. Proactive enforcement of least-privilege policies and strong segmentation impedes attacker pivoting and restricts outbound pathways for abuse.
Control: Cloud Native Security Fabric (CNSF)
Mitigation: Distributed inline inspection and policy prevent unauthorized prompt interactions.
Control: Zero Trust Segmentation
Mitigation: Identity-based microsegmentation restricts cross-app and cross-workload privilege gain.
Control: East-West Traffic Security
Mitigation: Internal lateral movement is halted by workload-to-workload traffic controls.
Control: Egress Security & Policy Enforcement
Mitigation: Egress policy blocks suspicious outbound communication paths.
Control: Multicloud Visibility & Control
Mitigation: Anomalous data exfiltration detected and blocked.
Automated detection and response contain malicious AI or code activity.
Impact at a Glance
Affected Business Functions
- Customer Support
- Content Generation
- Automated Decision-Making
Estimated downtime: 3 days
Estimated loss: $500,000
Potential exposure of sensitive customer data due to manipulated AI outputs leading to unauthorized information disclosure.
Recommended Actions
Key Takeaways & Next Steps
- • Enforce Zero Trust Segmentation to isolate LLM workloads and restrict unnecessary east-west traffic flows.
- • Implement granular egress filtering and policy enforcement to prevent unauthorized outbound data movement from AI-integrated apps.
- • Deploy CNSF inline enforcement and real-time inspection to detect and block prompt injection attempts at network and API boundaries.
- • Enhance multicloud visibility and centralized logging for rapid detection of anomalous behaviors in automated LLM pipelines.
- • Regularly baseline and monitor for threat anomalies in AI/ML workflows, and routinely update policies to reflect evolving prompt injection techniques.



