Run Integration tests from the test plan via Playwright hitting the live UI
Run comprehensive integration tests using Playwright against the live UI. Execute all tests from the test plan, document results, and generate a detailed report.
/plugin marketplace add edspencer/claude-agents/plugin install product-team@edspencer-agentsoptional-instructionsIf not given specific instructions otherwise, run all tests listed in ./test/integration/TEST-PLAN.md systematically, using their Comprehensive Tests section. If given a more specific set of instructions, follow them instead.
Execute the test plan using Playwright automation. Systematically work through all tests defined in ./test/integration/TEST-PLAN.md, document pass/fail status for each test, and generate a detailed report.
ALWAYS begin by:
ALWAYS finish by:
Work through ALL test sections in ./test/integration/TEST-PLAN.md systematically. For each test:
DO:
DO NOT:
Use Playwright MCP tools extensively:
browser_navigate - Navigate to pagesbrowser_snapshot - Capture accessibility snapshots (preferred for testing)browser_take_screenshot - Take visual screenshotsbrowser_click - Click elementsbrowser_type - Fill formsbrowser_console_messages - Check for errorsbrowser_wait_for - Wait for elements or textAfter testing is complete, generate a comprehensive report at ./test/integration/runs/YYYY-MM-DD-N/REPORT.md (where N is an index for multiple runs on the same day).
The report should have the following structure:
# UI Test Execution Report
**Date**: [Current date]
**Tested By**: Claude Code (UI Test Runner)
**Environment**: http://ngrok.edspencer.net
**Browser**: Playwright Chromium
---
## Executive Summary
- **Total Tests**: [number]
- **Passed**: [number] ([percentage]%)
- **Failed**: [number] ([percentage]%)
- **Skipped**: [number] (if any)
- **Overall Status**: PASS | FAIL | PARTIAL
**Critical Issues Found**: [number]
**Major Issues Found**: [number]
**Minor Issues Found**: [number]
---
## Test Results by Category
### 1. Navigation - Sidebar
**Status**: PASS | FAIL | PARTIAL
**Tests Passed**: X/Y
#### 1.1 Sidebar Structure
- [x] Test name - PASS
- [ ] Test name - FAIL: [brief reason]
- [x] Test name - PASS
[Continue for each test...]
---
### 2. Navigation - Careers Section
[Same format as above]
---
### 3. Coming Soon Pages
[Same format as above]
---
[Continue for all categories...]
---
## Issues Found
### Critical Issues
[None found] OR:
1. **Issue**: [Brief description]
- **Location**: [Where it occurs]
- **Steps to Reproduce**: [Exact steps]
- **Expected**: [What should happen]
- **Actual**: [What actually happens]
- **Evidence**: [Screenshot references, console errors]
### Major Issues
[Format same as critical]
### Minor Issues
[Format same as critical]
---
## Console Errors
[List all console errors found during testing with page context]
---
## Test Coverage
**Categories Completed**: X/7
**Individual Tests Completed**: X/Y
**Not Tested** (if any):
- [List any tests that couldn't be executed with reasons]
---
## Recommendations
[High-level recommendations for addressing failures, but no specific debugging or code changes]
---
## Conclusion
[Summary paragraph of overall test execution]
A successful test run means:
./test/integration/runs/YYYY-MM-DD-N/REPORT.mdThe tests themselves may pass or fail - your job is to execute them all and report accurately, not to achieve 100% pass rate.