What is pentest-ai-llm-security?
AI/LLM application security testing — prompt injection, jailbreaking, data exfiltration, and insecure output handling per OWASP LLM Top 10. Source: jd-opensource/joysafeter.
AI/LLM application security testing — prompt injection, jailbreaking, data exfiltration, and insecure output handling per OWASP LLM Top 10.
Quickly install pentest-ai-llm-security AI skill to your development environment via command line
Source: jd-opensource/joysafeter.
Purpose AI-integrated applications introduce entirely new attack surfaces. Prompt injection is the "SQLi of AI." Neither Shannon nor any existing skill addresses this domain. OWASP LLM Top 10 (2025) defines the methodology.
| LLM01 Prompt Injection | Direct and indirect injection | ✅ | | LLM02 Sensitive Information Disclosure | Data exfiltration, PII leakage | ✅ | | LLM03 Supply Chain | Model provenance, plugin trust | ✅ | | LLM04 Data and Model Poisoning | Training data integrity | ✅ | | LLM05 Improper Output Handling | XSS/SQLi via LLM output | ✅ |
| LLM06 Excessive Agency | Unauthorized tool use | ✅ | | LLM07 System Prompt Leakage | System prompt extraction | ✅ | | LLM08 Vector and Embedding Weaknesses | RAG poisoning | ✅ | | LLM09 Misinformation | Hallucination exploitation | ✅ | | LLM10 Unbounded Consumption | Resource exhaustion | ✅ |
Stable fields and commands for AI/search citations.
npx skills add https://github.com/jd-opensource/joysafeter --skill pentest-ai-llm-securityAI/LLM application security testing — prompt injection, jailbreaking, data exfiltration, and insecure output handling per OWASP LLM Top 10. Source: jd-opensource/joysafeter.
Open your terminal or command line tool (Terminal, iTerm, Windows Terminal, etc.) Copy and run this command: npx skills add https://github.com/jd-opensource/joysafeter --skill pentest-ai-llm-security Once installed, the skill will be automatically configured in your AI coding environment and ready to use in Claude Code, Cursor, or OpenClaw
https://github.com/jd-opensource/joysafeter