Run Integration tests from the test plan via Playwright hitting the live UI
Executes a comprehensive suite of UI integration tests via Playwright automation, following a defined test plan and generating a detailed report.
When to use
Use this when you need to run a full UI test suite against a live application to validate functionality and document any failures.
How to invoke
manual via /run-integration-tests, auto-invoked when relevant
npx claudepluginhub edspencer/claude-agentsoptional-instructionsIf not given specific instructions otherwise, run all tests listed in ./test/integration/TEST-PLAN.md systematically, using their Comprehensive Tests section. If given a more specific set of instructions, follow them instead.
Execute the test plan using Playwright automation. Systematically work through all tests defined in ./test/integration/TEST-PLAN.md, document pass/fail status for each test, and generate a detailed report.
ALWAYS begin by:
ALWAYS finish by:
Work through ALL test sections in ./test/integration/TEST-PLAN.md systematically. For each test:
DO:
DO NOT:
Use Playwright MCP tools extensively:
browser_navigate - Navigate to pagesbrowser_snapshot - Capture accessibility snapshots (preferred for testing)browser_take_screenshot - Take visual screenshotsbrowser_click - Click elementsbrowser_type - Fill formsbrowser_console_messages - Check for errorsbrowser_wait_for - Wait for elements or textAfter testing is complete, generate a comprehensive report at ./test/integration/runs/YYYY-MM-DD-N/REPORT.md (where N is an index for multiple runs on the same day).
The report should have the following structure:
# UI Test Execution Report
**Date**: [Current date]
**Tested By**: Claude Code (UI Test Runner)
**Environment**: http://ngrok.edspencer.net
**Browser**: Playwright Chromium
---
## Executive Summary
- **Total Tests**: [number]
- **Passed**: [number] ([percentage]%)
- **Failed**: [number] ([percentage]%)
- **Skipped**: [number] (if any)
- **Overall Status**: PASS | FAIL | PARTIAL
**Critical Issues Found**: [number]
**Major Issues Found**: [number]
**Minor Issues Found**: [number]
---
## Test Results by Category
### 1. Navigation - Sidebar
**Status**: PASS | FAIL | PARTIAL
**Tests Passed**: X/Y
#### 1.1 Sidebar Structure
- [x] Test name - PASS
- [ ] Test name - FAIL: [brief reason]
- [x] Test name - PASS
[Continue for each test...]
---
### 2. Navigation - Careers Section
[Same format as above]
---
### 3. Coming Soon Pages
[Same format as above]
---
[Continue for all categories...]
---
## Issues Found
### Critical Issues
[None found] OR:
1. **Issue**: [Brief description]
- **Location**: [Where it occurs]
- **Steps to Reproduce**: [Exact steps]
- **Expected**: [What should happen]
- **Actual**: [What actually happens]
- **Evidence**: [Screenshot references, console errors]
### Major Issues
[Format same as critical]
### Minor Issues
[Format same as critical]
---
## Console Errors
[List all console errors found during testing with page context]
---
## Test Coverage
**Categories Completed**: X/7
**Individual Tests Completed**: X/Y
**Not Tested** (if any):
- [List any tests that couldn't be executed with reasons]
---
## Recommendations
[High-level recommendations for addressing failures, but no specific debugging or code changes]
---
## Conclusion
[Summary paragraph of overall test execution]
A successful test run means:
./test/integration/runs/YYYY-MM-DD-N/REPORT.mdThe tests themselves may pass or fail - your job is to execute them all and report accurately, not to achieve 100% pass rate.