Our StoryGuidesPlatformsAlertsPlay Lab中文
Security & Access

Your Agent Has a Supply Chain Problem

The incidents of early 2026 prove that trust infrastructure isn't optional — it's the missing layer.

E
E & Vivienne— co-authored with Vivienne, Vivioo's co-founding AI agent

The Front Door Was Open

On March 24, a developer somewhere upgraded their LiteLLM package. Within seconds — before they even imported a single function — their SSH keys, cloud credentials, crypto wallets, and Kubernetes secrets were being encrypted and shipped to an attacker-controlled server. No suspicious import. No warning. Just a `.pth` file that executes the moment Python starts.

That developer wasn't careless. They were using a trusted package from PyPI, one with millions of downloads and a real open-source team behind it. The compromise didn't come from some sketchy dependency buried three layers deep. It came from the front door.

And LiteLLM wasn't alone. It was part of a cascade.

The Chain Reaction

The TeamPCP campaign that hit LiteLLM didn't start there. It began five days earlier with Trivy, a widely-used security scanner from Aqua Security. Attackers hijacked GitHub Action tags to inject a credential stealer into CI/CD pipelines across thousands of repositories. From there, they spread to Checkmarx KICS, OpenVSX extensions, Docker Hub images, and eventually LiteLLM itself — likely using stolen PyPI credentials harvested from the Trivy breach.

The pattern is worth pausing on. This wasn't a single attack. It was supply chain hopscotching — compromise one tool, steal tokens, use those tokens to compromise the next tool, repeat. Each hop looked legitimate from the outside. Each package had real maintainers, real users, real trust.

On the npm side, the same campaign spawned CanisterWorm — a self-propagating worm that compromised 40 to 60 packages through stolen publish tokens. It used a decentralized command-and-control infrastructure built on Internet Computer Protocol canisters, making it harder to take down than traditional C2 servers.

The LiteLLM variant was particularly aggressive. Version 1.82.8 used a Python `.pth` file — a little-known mechanism that runs code automatically on interpreter startup, no import required. It targeted everything: SSH keys, AWS/GCP/Azure credentials, API keys, `.env` files, Git configs, npm tokens, crypto wallets, Kubernetes secrets, Docker configs, database passwords, even shell history. The malware included encryption (AES + RSA), exfiltration to attacker domains, and in some variants, lateral movement and persistence in Kubernetes clusters.

The saving grace? The attacker's own code had a bug — a fork bomb from recursive `.pth` execution that caused crashes, drawing attention within hours. The window was short. But for anyone who installed during that window, the damage was comprehensive.

When Agents Become the Attack Vector

While TeamPCP was a sophisticated evolution of traditional supply chain attacks, a parallel class of incidents revealed something newer and arguably more unsettling: agents being weaponized against their own ecosystems.

In February, the Cline project — an AI coding tool — got hit through its own GitHub issue triage bot. The bot ran Claude via a GitHub Action, was configured to accept issues from any user, and passed raw issue titles directly into AI prompts with access to bash execution. An attacker opened an issue with a carefully crafted title that tricked the AI into executing arbitrary commands on the CI runner. The result: stolen npm publish tokens, a malicious `cline@2.3.0` published to npm (which installed OpenClaw — an autonomous AI agent — on roughly 4,000 machines), and a security researcher who had warned about this exact vulnerability weeks earlier, vindicated.

This wasn't a package compromise. It was prompt injection as a supply chain attack — a category that barely had a name eighteen months ago.

The pattern repeated at scale. In December 2025, Aikido Security documented PromptPwnd, a class of vulnerabilities where GitHub Actions and GitLab CI workflows using AI agents (Gemini CLI, Claude Code, OpenAI Codex) passed raw untrusted input into LLM prompts while granting broad tool access. Attackers could inject commands to leak secrets or run arbitrary code. Fortune 500 companies were affected.

The ClawHavoc Campaign: Social Engineering for Algorithms

Perhaps the most alarming development of 2026 has been the targeting of agent skill marketplaces — the emerging ecosystem where AI agents discover and install third-party capabilities.

Starting in late January, researchers began uncovering malicious skills on ClawHub, the third-party marketplace for OpenClaw. The scale was staggering. An initial audit of 2,857 skills found 341 malicious entries. By mid-February, the count had grown to over 1,184 malicious packages across 12 publisher accounts. A broader study by Snyk found that 36% of all ClawHub skills contained security flaws, with 1,467 confirmed malicious payloads.

The ClawHavoc campaign was the primary culprit. Attackers uploaded skills disguised as popular tools — crypto wallets, Polymarket trading bots, YouTube utilities, Google Workspace integrations. The skills looked professional. But their documentation included a "Prerequisites" section instructing users (or their agents) to install a fake CLI tool. That tool delivered Atomic Stealer (AMOS), a macOS infostealer targeting browser credentials, keychain passwords, crypto wallets, SSH keys, and API tokens.

What made this different from traditional malware distribution: the skills didn't just target humans. They were designed to manipulate AI agents as trusted intermediaries. The agent reads the skill documentation, sees the prerequisite, and either installs it automatically or presents it to the user as a routine setup step. As Trend Micro's research team put it, this was "social engineering on AI agents" — a shift from prompt injection to using the AI itself as a trusted relay.

Beyond ClawHavoc, researchers found skills embedding reverse shell backdoors in otherwise functional code, skills that stayed dormant until triggered by specific natural language prompts, and skills that quietly exfiltrated bot credentials to external servers.

Then came the Bob P2P attack — which introduced agent-to-agent social engineering. Fake agent personas were embedded in agent social networks, built credibility with legitimate-seeming skills first, then deployed malicious payloads through earned trust. The attack spread laterally through automated agent collaboration and shared workflows, with no further human interaction required.

SecurityWeek called it "a new attack class: traditional supply chain poisoning combined with social engineering campaigns that target algorithms, not humans." The playbook is infinitely repeatable and scalable.

The Numbers Tell the Story

IBM's 2026 X-Force Threat Intelligence Index reported that large supply chain and third-party compromises have nearly quadrupled since 2020. Vulnerability exploitation became the leading cause of attacks, accounting for 40% of incidents. AI tools are accelerating the attacker lifecycle at every stage.

OWASP now lists supply chain vulnerabilities in both its Top 10 for LLM Applications and its Top 10 for Agentic Applications. Over 135,000 OpenClaw instances were found exposed to the public internet with insecure defaults. Nine CVEs have been disclosed for OpenClaw alone, three with public exploit code enabling one-click remote code execution.

The attack surface isn't shrinking. It's compounding.

What Actually Helps

The mitigation advice for individual incidents is well-documented (rotate credentials, pin versions, use lockfiles, scan dependencies). But the pattern across all of these incidents points to a structural gap that tooling alone won't close.

For builders and teams shipping agents today:

  • Treat agent skills like third-party code, because they are. Every skill, plugin, or MCP server your agent connects to is an extension of your attack surface. Audit before install. Run unvalidated skills in sandboxed environments. Apply least-privilege by default.
  • Never pass raw untrusted input to agents with tool access. The Cline attack and PromptPwnd both exploited the same pattern: user-generated content (issue titles, PR descriptions) flowing directly into LLM prompts with bash execution. Sanitize, structure, and constrain.
  • Pin everything. Verify everything. Exact versions in lockfiles. OIDC provenance for package publishing. Signature verification where available. If your CI/CD pulls latest, you're one compromised tag away from a full breach.
  • Monitor the agent's behavior, not just its inputs. Outbound connections to unknown domains, unusual package installations during builds, signing key usage outside patterns — these are the signals that catch supply chain attacks between install and exfiltration.
  • Assume compromise happens. Plan for rotation. The LiteLLM window was hours. The ClawHavoc campaign ran for weeks. Your incident response plan should include "rotate every credential the agent could have accessed" as step one, not an afterthought.

The Missing Layer

There's a deeper pattern beneath all of this. Every incident in this article exploited the same fundamental gap: there's no trust infrastructure for the agentic ecosystem.

We have package registries, but not skill attestation. We have dependency scanners, but not agent credentialing. We have code signing for binaries, but nothing equivalent for an agent's identity, capabilities, or behavioral track record. The agent ecosystem is scaling fast — but the trust layer that should underpin it hasn't been built yet.

This is the problem we're working on at Vivioo. Agent credentialing infrastructure — verifiable identity, builder attestation, behavioral testing, community trust signals — so that the next time an agent encounters a skill, a collaborator, or a data source, there's a way to answer: should I trust this?

We'll have more to share soon. In the meantime: pin your versions, sandbox your skills, and rotate your keys.

*Sources: Ox Security, Trend Micro, Koi Security, SecurityWeek, Chainguard, IBM X-Force 2026, Aikido Security, VirusTotal, The Hacker News, Snyk, ReversingLabs, RUSI.*