·pentest-ai-llm-security
!

pentest-ai-llm-security

AI/LLM application security testing — prompt injection, jailbreaking, data exfiltration, and insecure output handling per OWASP LLM Top 10.

22Installs·1Trend·@jd-opensource

Installation

$npx skills add https://github.com/jd-opensource/joysafeter --skill pentest-ai-llm-security

How to Install pentest-ai-llm-security

Quickly install pentest-ai-llm-security AI skill to your development environment via command line

  1. Open Terminal: Open your terminal or command line tool (Terminal, iTerm, Windows Terminal, etc.)
  2. Run Installation Command: Copy and run this command: npx skills add https://github.com/jd-opensource/joysafeter --skill pentest-ai-llm-security
  3. Verify Installation: Once installed, the skill will be automatically configured in your AI coding environment and ready to use in Claude Code, Cursor, or OpenClaw

Source: jd-opensource/joysafeter.

SKILL.md

View raw

Purpose AI-integrated applications introduce entirely new attack surfaces. Prompt injection is the "SQLi of AI." Neither Shannon nor any existing skill addresses this domain. OWASP LLM Top 10 (2025) defines the methodology.

| LLM01 Prompt Injection | Direct and indirect injection | ✅ | | LLM02 Sensitive Information Disclosure | Data exfiltration, PII leakage | ✅ | | LLM03 Supply Chain | Model provenance, plugin trust | ✅ | | LLM04 Data and Model Poisoning | Training data integrity | ✅ | | LLM05 Improper Output Handling | XSS/SQLi via LLM output | ✅ |

| LLM06 Excessive Agency | Unauthorized tool use | ✅ | | LLM07 System Prompt Leakage | System prompt extraction | ✅ | | LLM08 Vector and Embedding Weaknesses | RAG poisoning | ✅ | | LLM09 Misinformation | Hallucination exploitation | ✅ | | LLM10 Unbounded Consumption | Resource exhaustion | ✅ |

Facts (cite-ready)

Stable fields and commands for AI/search citations.

Install command
npx skills add https://github.com/jd-opensource/joysafeter --skill pentest-ai-llm-security
Category
!Security
Verified
First Seen
2026-02-26
Updated
2026-03-10

Browse more skills from jd-opensource/joysafeter

Quick answers

What is pentest-ai-llm-security?

AI/LLM application security testing — prompt injection, jailbreaking, data exfiltration, and insecure output handling per OWASP LLM Top 10. Source: jd-opensource/joysafeter.

How do I install pentest-ai-llm-security?

Open your terminal or command line tool (Terminal, iTerm, Windows Terminal, etc.) Copy and run this command: npx skills add https://github.com/jd-opensource/joysafeter --skill pentest-ai-llm-security Once installed, the skill will be automatically configured in your AI coding environment and ready to use in Claude Code, Cursor, or OpenClaw

Where is the source repository?

https://github.com/jd-opensource/joysafeter