Core audit knowledge -- severity definitions, finding format, output templates, validation rules. Used by all audit agents.
From beenpx claudepluginhub bee-coded/bee-dev --plugin beeThis skill uses the workspace's default tool permissions.
Searches, retrieves, and installs Agent Skills from prompts.chat registry using MCP tools like search_skills and get_skill. Activates for finding skills, browsing catalogs, or extending Claude.
Searches prompts.chat for AI prompt templates by keyword or category, retrieves by ID with variable handling, and improves prompts via AI. Use for discovering or enhancing prompts.
Compares coding agents like Claude Code and Aider on custom YAML-defined codebase tasks using git worktrees, measuring pass rate, cost, time, and consistency.
Knowledge base for all audit agents. Defines severity levels, finding format, validation criteria, and report structure.
Every finding MUST be classified into exactly one severity level:
Immediate risk. The application is vulnerable to exploitation, data loss, or total failure RIGHT NOW.
Serious issues that will cause problems under real usage but aren't immediately exploitable.
Code quality and reliability issues that increase maintenance cost and bug risk.
Improvements that make the codebase better but aren't causing problems today.
any)Every finding from every audit agent MUST use this exact format:
### F-{AGENT_PREFIX}-{NNN}: {Short title}
- **Severity:** {CRITICAL | HIGH | MEDIUM | LOW}
- **Category:** {category name}
- **File:** `{file path}`
- **Lines:** {start}-{end} (or "multiple" if spread across files)
- **Agent:** {agent name that found it}
**Description:**
{What the problem is. Be specific -- reference the exact code pattern.}
**Evidence:**
{The actual code snippet or pattern that proves this is real. 3-10 lines max.}
**Impact:**
{What happens if this isn't fixed. Be concrete -- "users can X" not "may cause issues".}
**Suggested Fix:**
{How to fix it. Reference framework-specific approach if applicable.}
Each audit agent uses a unique prefix for finding IDs:
| Agent | Prefix | Example |
|---|---|---|
| security-auditor | SEC | F-SEC-001 |
| error-handling-auditor | ERR | F-ERR-001 |
| database-auditor | DB | F-DB-001 |
| architecture-auditor | ARCH | F-ARCH-001 |
| api-auditor | API | F-API-001 |
| frontend-auditor | FE | F-FE-001 |
| performance-auditor | PERF | F-PERF-001 |
| testing-auditor | TEST | F-TEST-001 |
| audit-bug-detector | BUG | F-BUG-001 |
The audit-finding-validator agent uses these rules to classify findings:
The finding is real. The code contains exactly the issue described, and the evidence is verifiable by reading the file.
The finding is wrong. Reasons:
Cannot determine without project-specific knowledge. The validator flags it for human review with an explanation of what additional context would clarify.
For each finding, the validator MUST:
The final AUDIT-REPORT.md follows this structure:
# Audit Report
**Project:** {project name}
**Date:** {YYYY-MM-DD}
**Audited by:** BeeDev Audit System v1.0
**Stack:** {detected stack}
**Audit Scope:** {which audit agents ran}
## Executive Summary
| Severity | Total Found | Confirmed | False Positive | Needs Context |
|----------|-------------|-----------|----------------|---------------|
| CRITICAL | {n} | {n} | {n} | {n} |
| HIGH | {n} | {n} | {n} | {n} |
| MEDIUM | {n} | {n} | {n} | {n} |
| LOW | {n} | {n} | {n} | {n} |
| **Total** | **{n}** | **{n}** | **{n}** | **{n}** |
## Risk Assessment
{Overall risk level: CRITICAL / HIGH / MODERATE / LOW / CLEAN}
{1-3 sentence summary of the most important findings}
## Critical Findings
{All CONFIRMED findings with severity CRITICAL, full detail}
## High Findings
{All CONFIRMED findings with severity HIGH, full detail}
## Medium Findings
{All CONFIRMED findings with severity MEDIUM, full detail}
## Low Findings
{All CONFIRMED findings with severity LOW, full detail}
## Needs Context
{All findings classified as NEEDS CONTEXT, with explanation of what's unclear}
## False Positives Log
{Summary table of false positives -- kept for transparency}
| ID | Title | Reason |
|----|-------|--------|
| F-XXX-NNN | {title} | {why it's false positive} |
## Recommendations
### Immediate Actions (CRITICAL)
{Numbered list of what to fix first}
### Short-term Actions (HIGH)
{What to address within the current sprint}
### Technical Debt (MEDIUM + LOW)
{What to schedule for cleanup}
## Audit Metadata
- Agents used: {list}
- Files scanned: {count}
- Total findings: {count}
- Confirmed: {count}
- False positive rate: {percentage}
When converting findings to specs via bee:audit-to-spec:
| Severity | Action |
|---|---|
| CRITICAL | Individual spec per finding. Tag: [CRITICAL-FIX]. Priority: immediate. |
| HIGH | Group related findings into one spec. Tag: [SECURITY-FIX] or [BUG-FIX]. |
| MEDIUM | Group by category into cleanup specs. Tag: [TECH-DEBT]. |
| LOW | Single consolidated spec for all LOW findings. Tag: [IMPROVEMENT]. |