Debug and evaluate performance issues with detailed diagnostics and fixes
Debug and evaluate AI debugging performance by analyzing real codebase issues. Use this to measure quality improvements, time efficiency, and generate detailed diagnostic reports with fixes.
/plugin marketplace add bejranonda/LLM-Autonomous-Agent-Plugin-for-Claude/plugin install bejranonda-autonomous-agent@bejranonda/LLM-Autonomous-Agent-Plugin-for-Claudedebug/Measures AI debugging performance by analyzing and fixing real issues in the codebase.
/debug:eval <target> [options]
--help Show this help message
--verbose Show detailed agent selection process
--dry-run Preview actions without executing
--report-only Generate report without fixing issues
--performance Include detailed performance metrics
# Show help
/debug:eval --help
# Debug with verbose output (shows agent selection)
/debug:eval dashboard --verbose
# Preview what would be fixed
/debug:eval data-validation --dry-run
# Generate report without fixing
/debug:eval performance-index --report-only
This command delegates to the orchestrator agent which:
Analyzes the debugging request and determines optimal approach
Selects appropriate specialized agents based on task type and complexity
May delegate to validation-controller for debugging-specific tasks:
Measures debugging performance using the comprehensive framework:
Generates detailed performance report with metrics and improvements
When using --verbose flag, you'll see:
š ORCHESTRATOR: Analyzing debugging request...
š ORCHESTRATOR: Task type identified: "dashboard debugging"
šÆ ORCHESTRATOR: Selecting agents: validation-controller, code-analyzer
š VALIDATION-CONTROLLER: Beginning systematic analysis...
š CODE-ANALYZER: Analyzing code structure and patterns...
Measures debugging performance using the comprehensive framework:
Generates detailed performance report with metrics and improvements
dashboardrandom.uniform() without deterministic seeding in dashboard.py:710-712performance-indexdata-validationThe evaluation uses the comprehensive debugging performance framework:
QIS = 0.6 Ć FinalQuality + 0.4 Ć (GapClosedPct Ć 100/100)
PI = (0.40 Ć QIS) + (0.35 Ć TES) + (0.25 Ć SR) ā Penalty
Where Penalty = RegressionRate Ć 20
š DEBUGGING PERFORMANCE EVALUATION
Target: dashboard data inconsistency
š PERFORMANCE METRICS:
* Initial Quality: 85/100
* Final Quality: 96/100 (+11 points)
* QIS (Quality Improvement): 78.5/100
* Time Efficiency: 92/100
* Success Rate: 100%
* Regression Penalty: 0
* Performance Index: 87.2/100
ā” DEBUGGING RESULTS:
[PASS] Root cause identified: random.uniform() without seeding
[PASS] Fix implemented: deterministic seeded calculation
[PASS] Quality improvement: +11 points
[PASS] Time to resolution: 4.2 minutes
š Full report: .claude/data/reports/debug-eval-dashboard-2025-10-24.md
ā± Completed in 4.2 minutes
Located at: .claude/data/reports/debug-eval-<target>-YYYY-MM-DD.md
Comprehensive analysis including:
Each /eval-debug execution automatically:
/eval-debug dashboard
/eval-debug performance-index
/eval-debug data-validation
For Debugging Performance Measurement:
For Code Quality:
For Learning System: