From mastepanoski-claude-skills
Security audit for LLM and GenAI applications using OWASP Top 10 for LLM Apps 2025. Assess prompt injection, data leakage, supply chain, and 7 more critical vulnerabilities.
npx claudepluginhub joshuarweaver/cascade-content-creation-misc-1 --plugin mastepanoski-claude-skillsThis skill uses the workspace's default tool permissions.
This skill enables AI agents to perform a comprehensive **security assessment** of Large Language Model (LLM) and Generative AI applications using the **OWASP Top 10 for LLM Applications 2025**, published by the OWASP GenAI Security Project.
Applies Acme Corporation brand guidelines including colors, fonts, layouts, and messaging to generated PowerPoint, Excel, and PDF documents.
Builds DCF models with sensitivity analysis, Monte Carlo simulations, and scenario planning for investment valuation and risk assessment.
Calculates profitability (ROE, margins), liquidity (current ratio), leverage, efficiency, and valuation (P/E, EV/EBITDA) ratios from financial statements in CSV, JSON, text, or Excel for investment analysis.
This skill enables AI agents to perform a comprehensive security assessment of Large Language Model (LLM) and Generative AI applications using the OWASP Top 10 for LLM Applications 2025, published by the OWASP GenAI Security Project.
The OWASP Top 10 for LLM Applications identifies the most critical security risks in systems that integrate large language models, covering vulnerabilities from prompt injection to unbounded resource consumption. This is the authoritative industry standard for LLM application security.
Use this skill to identify security vulnerabilities, assess risk exposure, prioritize remediation, and establish secure development practices for AI-powered applications.
Combine with "NIST AI RMF" for comprehensive risk management or "ISO 42001 AI Governance" for governance compliance.
Invoke this skill when:
When executing this audit, gather:
Severity: Critical
Description: Attackers manipulate LLM operations through crafted inputs, either directly or indirectly, to bypass intended functionality, access unauthorized data, or trigger unintended actions.
Attack Vectors:
Impact:
Assessment Checklist:
Mitigation Strategies:
Severity: Critical
Description: LLMs inadvertently expose confidential data including PII, proprietary algorithms, credentials, intellectual property, or internal system information through their outputs.
Attack Vectors:
Impact:
Assessment Checklist:
Mitigation Strategies:
Severity: High
Description: Compromised third-party components (models, datasets, libraries, plugins) introduce security risks including malware, backdoors, or biased behavior.
Attack Vectors:
Impact:
Assessment Checklist:
Mitigation Strategies:
Severity: High
Description: Attackers manipulate training or fine-tuning data to introduce vulnerabilities, backdoors, or biases that compromise model security and reliability.
Attack Vectors:
Impact:
Assessment Checklist:
Mitigation Strategies:
Severity: High
Description: Applications blindly execute or render LLM outputs without validation, enabling code injection, XSS, SQL injection, SSRF, and other attacks.
Attack Vectors:
Impact:
Assessment Checklist:
Mitigation Strategies:
Severity: High
Description: AI agents possess excessive permissions and autonomous capabilities, enabling significant harm through compromised prompts, hallucinations, or malicious manipulation.
Attack Vectors:
Impact:
Assessment Checklist:
Mitigation Strategies:
Severity: Medium
Description: System instructions intended to guide AI behavior are exposed to users or attackers, revealing internal logic, security controls, or sensitive configurations.
Attack Vectors:
Impact:
Assessment Checklist:
Mitigation Strategies:
Severity: Medium
Description: Vulnerabilities in vector databases and embedding-based retrieval systems (RAG) allow poisoning, injection, or unauthorized access to stored data.
Attack Vectors:
Impact:
Assessment Checklist:
Mitigation Strategies:
Severity: Medium
Description: LLMs generate plausible but false information (hallucinations/confabulations) that users may trust and act upon, causing harm.
Attack Vectors:
Impact:
Assessment Checklist:
Mitigation Strategies:
Severity: Medium
Description: Uncontrolled LLM usage causes denial-of-service, system crashes, or excessive operational costs through resource exhaustion.
Attack Vectors:
Impact:
Assessment Checklist:
Mitigation Strategies:
System inventory:
Threat modeling:
For each of the 10 vulnerabilities, assess:
For each vulnerability found, score using:
Likelihood: How likely is exploitation?
Impact: What is the potential damage?
Compile comprehensive security assessment.
Generate a comprehensive OWASP LLM security audit report:
# OWASP LLM Top 10 Security Audit Report
**Application**: [Name]
**LLM Provider/Model**: [Provider - Model]
**Date**: [Date]
**Evaluator**: [AI Agent or Human]
**OWASP LLM Top 10 Version**: 2025
---
## Executive Summary
### Overall Security Posture: [Critical / High Risk / Medium Risk / Low Risk / Secure]
**Application Type**: [Chatbot / Agent / RAG System / Content Generator / Code Assistant / Other]
**Data Sensitivity**: [Public / Internal / Confidential / Restricted]
**User Base**: [Internal / B2B / B2C / Public]
### Critical Findings
| # | Vulnerability | Severity | Status |
|---|---|---|---|
| LLM01 | Prompt Injection | Critical | [Vulnerable / Mitigated / N/A] |
| LLM02 | Sensitive Info Disclosure | Critical | [Vulnerable / Mitigated / N/A] |
| LLM03 | Supply Chain | High | [Vulnerable / Mitigated / N/A] |
| LLM04 | Data/Model Poisoning | High | [Vulnerable / Mitigated / N/A] |
| LLM05 | Improper Output Handling | High | [Vulnerable / Mitigated / N/A] |
| LLM06 | Excessive Agency | High | [Vulnerable / Mitigated / N/A] |
| LLM07 | System Prompt Leakage | Medium | [Vulnerable / Mitigated / N/A] |
| LLM08 | Vector/Embedding Weaknesses | Medium | [Vulnerable / Mitigated / N/A] |
| LLM09 | Misinformation | Medium | [Vulnerable / Mitigated / N/A] |
| LLM10 | Unbounded Consumption | Medium | [Vulnerable / Mitigated / N/A] |
### Top 3 Critical Issues
1. [Issue] - [Impact description]
2. [Issue] - [Impact description]
3. [Issue] - [Impact description]
---
## Detailed Findings
### LLM01: Prompt Injection
**Status**: [Vulnerable / Partially Mitigated / Mitigated]
**Severity**: [Critical / High / Medium / Low]
**Likelihood**: [High / Medium / Low]
**Findings:**
1. [Finding with evidence]
2. [Finding with evidence]
**Attack Scenario:**
[Description of how this could be exploited]
**Recommendations:**
1. [Specific remediation step]
2. [Specific remediation step]
**Effort**: [Low / Medium / High]
---
[Continue for LLM02 through LLM10...]
---
## Architecture Security Review
### Data Flow Analysis
[Diagram or description of data flows with trust boundaries marked]
### Attack Surface Summary
| Surface | Risk Level | Controls |
|---|---|---|
| User Input | [Level] | [Controls] |
| API Endpoints | [Level] | [Controls] |
| Vector Store | [Level] | [Controls] |
| Plugins/Tools | [Level] | [Controls] |
| Output Rendering | [Level] | [Controls] |
---
## Remediation Roadmap
### Phase 1: Critical (0-7 days)
1. [ ] [Action item with owner]
2. [ ] [Action item with owner]
### Phase 2: High Priority (7-30 days)
1. [ ] [Action item with owner]
### Phase 3: Medium Priority (30-90 days)
1. [ ] [Action item with owner]
### Phase 4: Hardening (Ongoing)
1. [ ] [Continuous improvement practices]
---
## Security Controls Matrix
| Control | Implemented | Effective | Recommendation |
|---|---|---|---|
| Input validation | [Yes/No/Partial] | [Yes/No] | [Recommendation] |
| Output sanitization | [Yes/No/Partial] | [Yes/No] | [Recommendation] |
| Rate limiting | [Yes/No/Partial] | [Yes/No] | [Recommendation] |
| Authentication | [Yes/No/Partial] | [Yes/No] | [Recommendation] |
| Authorization | [Yes/No/Partial] | [Yes/No] | [Recommendation] |
| Logging/Monitoring | [Yes/No/Partial] | [Yes/No] | [Recommendation] |
| Content filtering | [Yes/No/Partial] | [Yes/No] | [Recommendation] |
| Human-in-the-loop | [Yes/No/Partial] | [Yes/No] | [Recommendation] |
---
## Next Steps
1. [ ] Prioritize and assign critical findings
2. [ ] Implement quick wins (input validation, rate limiting)
3. [ ] Schedule penetration testing for high-risk areas
4. [ ] Establish continuous monitoring
5. [ ] Plan follow-up audit after remediation
---
## Resources
- [OWASP Top 10 for LLM Applications 2025](https://genai.owasp.org/resource/owasp-top-10-for-llm-applications-2025/)
- [OWASP GenAI Security Project](https://genai.owasp.org/)
- [OWASP LLM AI Security & Governance Checklist](https://owasp.org/www-project-top-10-for-large-language-model-applications/)
- [OWASP GitHub Repository](https://github.com/OWASP/www-project-top-10-for-large-language-model-applications)
---
**Audit Version**: 1.0
**Date**: [Date]
| Priority | Vulnerabilities | Rationale |
|---|---|---|
| P0 | LLM01 (Prompt Injection), LLM02 (Data Disclosure) | Direct exploitation, high impact |
| P1 | LLM05 (Output Handling), LLM06 (Excessive Agency) | System compromise potential |
| P2 | LLM03 (Supply Chain), LLM04 (Poisoning) | Harder to exploit but severe impact |
| P3 | LLM07 (Prompt Leakage), LLM08 (Vector Weaknesses) | Enables further attacks |
| P4 | LLM09 (Misinformation), LLM10 (Unbounded Consumption) | Operational risk |
1.0 - Initial release (OWASP Top 10 for LLM Applications 2025)
Remember: LLM security is an evolving field. New attack vectors emerge regularly. This audit provides a baseline assessment; continuous monitoring and periodic re-assessment are essential for maintaining security posture.