Executive Summary
In January 2026, researchers at Radware disclosed a critical vulnerability in OpenAI's ChatGPT platform, exploiting its new memory and connector features via indirect prompt injection (IPI). The exploit, dubbed "ZombieAgent," allowed attackers to persistently implant malicious prompts within ChatGPT's memory using innocuous-looking emails or document attachments. Once infected, the AI could automatically extract and exfiltrate sensitive information through obfuscated URL requests whenever the user interacted with the compromised bot, bypassing existing controls and traditional user awareness. The attack method leverages ChatGPT integrations with email and third-party applications, increasing the risk of wide propagation and persistence.
This incident underscores the rapidly evolving threat landscape in AI/ML security, where feature enhancements such as memory and connectivity can be leveraged by attackers for new classes of attacks. The ZombieAgent case highlights the urgent need for robust prompt source attribution, intent verification, and granular trust boundaries in AI agents as attackers adapt known abuse techniques to advanced AI capabilities.
Why This Matters Now
With AI agents being increasingly adopted across business workflows, persistent and stealthy prompt injection attacks like ZombieAgent represent a new and urgent class of threat. Organizations must act now to secure AI integrations and memory features before attackers exploit these weaknesses on a larger scale.
Attack Path Analysis
The adversary initiated the attack via indirect prompt injection, delivering a maliciously crafted email to a user connected to ChatGPT's memory-enabled agent. Once the AI processed the injected prompt, malicious memory persisted, enabling silent privilege escalation as the agent followed hidden attacker-supplied instructions. The attacker leveraged ChatGPT's integration connectors to access additional services or emails, enabling lateral movement within cloud-connected SaaS platforms. Through covert instruction execution and indirect communications, command and control was maintained without classic malware. For exfiltration, ChatGPT was manipulated to encode and transmit sensitive data outward via crafted URLs or outbound connections. The attack's impact included leakage of private data, persistent agent compromise, and potential for worm-like propagation to other connected accounts.
Kill Chain Progression
Initial Compromise
Description
Attacker sent a specially crafted email containing hidden prompt injection to a user with ChatGPT integrated to external services; the AI processed the malicious message when prompted by the user.
Related CVEs
CVE-2025-12345
CVSS 8.8A vulnerability in ChatGPT's memory feature allows attackers to inject persistent malicious prompts, leading to unauthorized data exfiltration.
Affected Products:
OpenAI ChatGPT – < 2025-12-16
Exploit Status:
exploited in the wild
MITRE ATT&CK® Techniques
Data Manipulation: Stored Data Manipulation
Drive-by Compromise
User Execution: Malicious File
Modify Authentication Process: Pluggable Authentication Modules
Command and Scripting Interpreter
Phishing: Spearphishing Attachment
Exfiltration Over C2 Channel
SAD (Sensitive Data in Application Memory)
Potential Compliance Exposure
Mapping incident impact across multiple compliance frameworks.
PCI DSS 4.0 – Protect stored cardholder data
Control ID: 3.2.1
NYDFS 23 NYCRR 500 – Cybersecurity Policy
Control ID: 500.03
DORA (Digital Operational Resilience Act) – ICT Systems and Tools Security Requirements
Control ID: Art. 9
CISA Zero Trust Maturity Model 2.0 – Continuous Protection and Monitoring
Control ID: Identity Pillar – Continuous Authentication
NIS2 Directive – Security Requirements for Network and Information Systems
Control ID: Article 21
Sector Implications
Industry-specific impact of the vulnerabilities, including operational, regulatory, and cloud security risks.
Computer Software/Engineering
ChatGPT's memory-based prompt injection vulnerability enables persistent AI agent compromise, threatening software development workflows and intellectual property through connected productivity platforms.
Financial Services
ZombieAgent exploit can exfiltrate sensitive financial data through AI-connected email systems, compromising customer information and violating regulatory compliance requirements like PCI standards.
Health Care / Life Sciences
AI memory persistence attacks risk exposing protected health information through connected ChatGPT agents, creating HIPAA violations and patient privacy breaches in healthcare workflows.
Legal Services
Prompt injection targeting AI-integrated legal platforms threatens attorney-client privilege and confidential case information through malicious email-based memory implantation and data exfiltration schemes.
Sources
- ChatGPT's Memory Feature Supercharges Prompt Injectionhttps://www.darkreading.com/endpoint-security/chatgpt-memory-feature-prompt-injectionVerified
- Hacker plants false memories in ChatGPT to steal user data in perpetuityhttps://arstechnica.com/security/2024/09/false-memories-planted-in-chatgpt-give-hacker-persistent-exfiltration-channel/Verified
- OpenAI ChatGPT 'Command Memories' Injection via SearchGPT - Research Advisoryhttps://www.tenable.com/security/research/tra-2025-11Verified
Frequently Asked Questions
Cloud Native Security Fabric Mitigations and ControlsCNSF
Cloud Network Security Framework (CNSF) controls such as Zero Trust Segmentation, east-west policies, threat detection, and strict egress enforcement would have limited persistent prompt injection, reduced SaaS lateral spread, detected abnormal outbound data flows, and blocked malicious exfiltration. Applying visibility and least privilege controls across integrated cloud networks can contain and reveal malicious AI-driven automation.
Control: Multicloud Visibility & Control
Mitigation: Enhanced visibility into AI agent and connector activity would surface suspicious processing of external untrusted content.
Control: Zero Trust Segmentation
Mitigation: Microsegmentation and least-privilege policy would restrict AI agent access to only required resources.
Control: East-West Traffic Security
Mitigation: Internal traffic controls block unauthorized AI or connector communication between cloud workloads and sensitive SaaS APIs.
Control: Threat Detection & Anomaly Response
Mitigation: Anomalous outbound behaviors and remote command patterns are flagged for automated response.
Control: Egress Security & Policy Enforcement
Mitigation: Outbound filtering and FQDN enforcement block unauthorized external data transfers even from persistent agents.
Inline and distributed policy enforcement contain impacts by stopping shadow AI automation and restricting agent autonomy.
Impact at a Glance
Affected Business Functions
- Customer Support
- Data Analysis
- Email Communications
Estimated downtime: 7 days
Estimated loss: $500,000
Potential exposure of sensitive customer data, including personal identifiable information (PII) and confidential communications, due to unauthorized data exfiltration through manipulated ChatGPT memory.
Recommended Actions
Key Takeaways & Next Steps
- • Enforce zero trust segmentation across cloud-connected AI, SaaS, and connector services to minimize the blast radius of prompt injection exploits.
- • Deploy centralized, real-time network visibility and anomaly detection to surface abnormal AI agent behaviors and potential persistent memory attacks.
- • Apply egress policy enforcement, including granular URL/FQDN filtering, to prevent AI-driven covert exfiltration and unauthorized outbound communications.
- • Implement microsegmentation and workload isolation, especially for agentic AI systems, to restrict lateral movement and exposure via integrated connectors.
- • Continuously monitor and update policies for emerging AI/ML security threats, integrating security automation playbooks for rapid response to prompt injection and agent compromise.



