TL;DR
Last week, the TeamPCP group pulled off a massive supply chain attack by using Trivy, a trusted security tool, to poison versions of LiteLLM with malware that stole master cloud credentials.
The breach stole multiple credentials and left an open back door into many cloud environments, meaning that the threat is still active.
This attack was enabled by a major security flaw: a lack of egress filtering.
Last week, a sophisticated supply chain attack compromised tens of thousands of organizations and potentially millions of individual developer environments. The attackers, a group known as TeamPCP, stole master cloud credentials at scale: access keys, database passwords, and Kubernetes tokens that represent the deepest level of access inside modern cloud infrastructure.
The mechanics of the attack deserve attention. So does the question of why those credentials could be stolen and transmitted to an outside party in the first place.
What Actually Happened in the LiteLLM Breach
TeamPCP did not target LiteLLM directly. LiteLLM is a widely used tool that connects AI applications to large language models from providers like OpenAI and Google. It is downloaded millions of times every month, which makes it a high-value target, but a direct attack on such a visible project carried significant detection risk.
Instead, TeamPCP first compromised Trivy, a trusted open-source security scanner used by development teams to identify vulnerabilities in their own software. By taking control of a tool that organizations already trusted, they created a credible entry point into the supply chain without raising immediate alarms.
From that foothold, they pushed two poisoned versions of LiteLLM to its official public repository. Any organization that ran a routine software update pulled in the malware without any indication that something was wrong. The AI systems continued operating normally. No alerts fired. No anomalies surfaced in the logs. Meanwhile, the embedded malware was actively streaming cloud access keys, database passwords, and Kubernetes tokens to attacker-controlled servers.
Executing this attack required genuine technical sophistication. Gaining access to Trivy, embedding malware without triggering detection, and maintaining the appearance of normal software behavior across both compromised packages are not trivial accomplishments. But the defense that would have stopped the damage was far simpler than the attack itself.
The Principle That Should Have Held
There is a foundational principle of cloud network security that applies directly to this attack: every workload and every AI agent should only be permitted to communicate with explicitly approved destinations. Anything outside that approved list gets blocked, automatically and without exception.
If that control had been in place, this attack would have ended the moment the malware attempted to send stolen credentials to an external server. The malware could not have reached attacker infrastructure. The credentials would have gone nowhere. The vault would have stayed locked even after someone had obtained the keys.
That control is called egress filtering, and the honest reality is that most organizations operating in the cloud do not have it in place.
How We Got to the LiteLLM Breach
Two groups share responsibility for that gap:
The first group is organizational leadership. A significant number of CISOs and CIOs moved workloads into the cloud without building workload-level network security. Many accepted "the cloud is secure" as a complete answer and stopped asking what that actually meant for their specific environments. That was a choice, and choices have consequences.
The second group is the major cloud providers: Amazon, Microsoft, Google, and Oracle. Their underlying infrastructure is genuinely secure, and the physical and logical security of those platforms deserves real credit. The responsibility gap runs in a different direction entirely. Every major cloud platform ships with permissive outbound defaults. Unless an organization explicitly builds controls to restrict where its workloads can communicate, those workloads can reach anything on the internet.
That is a product decision. The analogy is direct: a car manufacturer that ships vehicles without standard brakes and offers them only as aftermarket add-ons bears real responsibility for what happens on the road. Making insecurity the default, while organizations assume they are protected, has contributed directly to the exposure that attacks like this one exploit.
The Backdoor Problem
Stolen credentials, as serious as they are, represent a recoverable situation if an organization responds quickly. Rotate the keys, audit the access logs, close the exposure. That response is painful and expensive, but it is finite.
This attack did not stop at credential theft. The malware also installed a persistent backdoor inside compromised cloud environments. That backdoor sits dormant and waits for instructions. Any organization that has only rotated credentials since discovering the breach has addressed part of the problem. The harder part — identifying and removing the persistent access — requires a full forensic review of every affected system. For organizations that have not completed that work, the door is still open.
Consider what runs on cloud infrastructure today: hospital systems managing patient records and surgical scheduling, financial platforms processing transactions, utility management systems, airline operations. The breadth of exposure is significant. If a sophisticated actor with backdoor access to those environments decides to act, the downstream effects reach people who have no awareness that any of this is happening.
Why This Is Personal
Most people will read about this attack and file it under "enterprise IT problem,” but that framing misses the actual stakes.
Consumers are downstream of every cloud security decision made by every organization they interact with. You did not vote on whether your healthcare provider implemented egress filtering. You have no visibility into whether the airline managing your travel has forensically cleaned its systems since the attack was discovered. You did not choose your bank's outbound communication policies.
But when a hospital's systems are disrupted, surgeries get postponed. When a financial platform buckles under sustained pressure, access to money disappears. When infrastructure that people depend on is compromised, the consequences do not stay inside enterprise balance sheets. They land on real people in real situations.
This is an argument for holding organizations accountable for the security decisions they make on behalf of the people who trust them.
The Technical Fix
The core vulnerability this attack exploited has a straightforward technical solution. Every cloud workload should have a defined policy specifying exactly what destinations it is permitted to communicate with. Traffic that falls outside that policy gets blocked, by default, without exception.
The implementation challenge is real. Traditional firewalls are generally not built to handle the traffic patterns of ephemeral AI workloads. They operate at a level of abstraction that does not map to how modern cloud environments are structured. Organizations relying on legacy perimeter security for this problem will find that the perimeter does not exist where the threat is actually operating.
What is required is workload-level network security: controls that travel with the workload itself, that account for the dynamic nature of cloud infrastructure, and that enforce communication policies regardless of where a workload is running or how briefly it exists.
That is the technical answer. The organizational answer matters equally. Boards and executive teams need to treat cloud network security as a business risk with real financial and operational consequences, not as a line item buried in an IT budget. The cost of not doing so has become impossible to ignore.
Where to Start
The LiteLLM breach is a useful forcing function. It demonstrates that:
AI supply chains are active attack surfaces
Credential theft becomes a secondary concern when persistent backdoor access has been established,
The default configurations of major cloud platforms leave organizations exposed in ways that many leaders do not fully understand
If you are responsible for an organization's cloud posture, the first step is understanding where you actually stand. A free Workload Attack Path Assessment at aviatrix.ai is a concrete starting point for that conversation.
Cloud security does not belong to a vendor, a platform, or a team buried in an IT org chart. It belongs to every leader who has made a decision to run their organization in the cloud.
The cost of not owning it just became very visible.
Learn more about how Aviatrix Cloud Native Security Fabric (CNSF) blocks supply chain attacks and data exfiltration through Zero Trust principles.
















