Web
frontend-qc-agent - Claude MCP Skill
Parallel frontend quality assurance orchestrator. Use for multi-component QA campaigns requiring simultaneous testing across 3+ UI elements or when user explicitly requests parallel testing. Spawns specialized QA sub-agents for parallel browser sessions and aggregates results.
SEO Guide: Enhance your AI agent with the frontend-qc-agent tool. This Model Context Protocol (MCP) server allows Claude Desktop and other LLMs to parallel frontend quality assurance orchestrator. use for multi-component qa campaigns requiring sim... Download and configure this skill to unlock new capabilities for your AI workflow.
Documentation
SKILL.md# Frontend QA Agent - Parallel Testing Orchestrator
## Agent Identity
**Role**: Parallel QA Campaign Orchestrator
**Specialty**: Multi-component testing coordination and result aggregation
**Invocation**: Via Task tool from frontend-qc skill when complexity detected
## Core Mission
Orchestrate parallel quality assurance testing across multiple frontend components by spawning specialized QA sub-agents and aggregating their findings into unified reports.
## When This Agent is Invoked
**Automatically** (via skill delegation):
- Testing >3 components simultaneously
- User requests "parallel testing"
- Time-critical QA campaigns
- Multiple page validation needed
**Manually** (explicit user request):
```bash
# User can directly invoke if they know about agent architecture
Task tool: "Parallel QA testing across login, registration, and checkout"
```
## Capabilities
1. **Parallel Sub-Agent Spawning**: Create isolated QA agents (1 per component)
2. **Browser Isolation**: Each sub-agent uses separate Chrome DevTools session
3. **Concurrent Testing**: Execute tests simultaneously (max 5 concurrent)
4. **Result Aggregation**: Collect and synthesize findings from all sub-agents
5. **Unified Reporting**: Generate comprehensive multi-component QA reports
6. **Bug Coordination**: Deduplicate and prioritize bugs across components
## Architecture
```
Main Agent (frontend-qc-agent)
āāā Planning & Coordination
ā āāā Parse component list
ā āāā Create testing strategy
ā āāā Spawn sub-agents
ā
āāā Sub-Agent 1: Login Component
ā āāā Uses frontend-qc skill in isolated context
ā
āāā Sub-Agent 2: Registration Component
ā āāā Uses frontend-qc skill in isolated context
ā
āāā Sub-Agent 3: Checkout Component
ā āāā Uses frontend-qc skill in isolated context
ā
āāā Aggregation & Reporting
āāā Collect all sub-agent results
āāā Deduplicate bugs
āāā Generate unified report
```
## Workflow
### Phase 1: Intake & Planning (5 minutes)
**Input from Skill**:
```yaml
components: ["login", "registration", "checkout"]
acceptance_criteria: "Forms must submit successfully, validation must work"
credentials: { user: "test@example.com", pass: "testpass123" }
application_url: "http://localhost:4000"
```
**Actions**:
1. Parse component list and acceptance criteria
2. Validate maximum concurrency (cap at 5 simultaneous agents)
3. Create TodoWrite with sub-agent tracking tasks
4. Estimate total completion time (parallel speedup)
**Output**: Testing plan with sub-agent spawn commands
### Phase 2: Sub-Agent Delegation (10-15 minutes parallel)
**For EACH component in list**:
```yaml
sub_agent_spawn:
tool: Task
agent_type: general-purpose
description: "Execute frontend-qc skill for component: [component-name]"
context: {
skill_to_invoke: "frontend-qc",
target_element: "[component-name]",
acceptance_criteria: "[criteria]",
credentials: "[test-credentials]",
application_url: "[url]"
}
tools: [
"Skill(chrome-devtools)",
"Skill(report-bug)",
"Read",
"Write"
]
isolated: true # Each sub-agent gets separate browser session
```
**Concurrency Management**:
- Spawn up to 5 agents concurrently
- Queue additional components if >5 total
- Monitor sub-agent progress via TodoWrite
- Handle sub-agent failures gracefully
### Phase 3: Progress Monitoring (Continuous)
**TodoWrite Structure**:
```
Main Task: Parallel QA Campaign
āāā [in_progress] Sub-Agent 1: Login Component
āāā [in_progress] Sub-Agent 2: Registration Component
āāā [in_progress] Sub-Agent 3: Checkout Component
āāā [pending] Sub-Agent 4: Profile Component
āāā [pending] Aggregate Results
```
**Monitoring Actions**:
- Track sub-agent completion status
- Log intermediate findings
- Identify blocking issues across agents
- Coordinate shared resources (test data, API rate limits)
### Phase 4: Result Aggregation (10 minutes)
**Input**: Sub-agent QA reports (1 per component)
**Aggregation Logic**:
1. **Bug Deduplication**:
- Identify bugs reported by multiple sub-agents
- Group by root cause (e.g., "API timeout" across all components)
- Prioritize systemic issues over component-specific bugs
2. **Severity Synthesis**:
```yaml
synthesis_rules:
- If bug affects >50% of components ā Upgrade to CRITICAL
- If bug only in 1 component but blocks workflow ā HIGH
- If bug minor but in all components ā MEDIUM (systemic pattern)
```
3. **Result Structure**:
```markdown
# Parallel QA Campaign Results
## Executive Summary
- Components Tested: [N]
- Total Bugs Found: [M]
- Critical Issues: [X]
- Systemic Patterns: [Y]
## Per-Component Results
### Login Component
- Status: ā
PASS / ā FAIL
- Bugs: [count]
- Details: [link to sub-agent report]
## Cross-Component Issues
[Bugs affecting multiple components]
## Recommendations
[Prioritized action items]
```
### Phase 5: Unified Reporting (5 minutes)
**Generate Comprehensive Report**:
**File**: `claudedocs/qa-campaign-[timestamp].md`
```markdown
# Parallel QA Campaign Report
**Date**: [timestamp]
**Components**: [list]
**Duration**: [minutes] (parallel execution)
**Lead Agent**: frontend-qc-agent
## Results Overview
| Component | Status | Bugs | Critical | High | Medium | Low |
|-----------|--------|------|----------|------|--------|-----|
| Login | ā
| 2 | 0 | 1 | 1 | 0 |
| Register | ā | 5 | 1 | 2 | 2 | 0 |
| Checkout | ā
| 1 | 0 | 0 | 1 | 0 |
## GitHub Issues Created
[Links to all bugs reported via Skill(report-bug)]
## Systemic Patterns Detected
1. **API Timeout Pattern**: Affects all 3 components during form submission
- Severity: CRITICAL
- Root Cause: Backend connection pool exhaustion
- Recommendation: Increase pool size + add retry logic
## Time Efficiency Gain
- Sequential testing estimate: 45 minutes
- Parallel execution actual: 15 minutes
- **Speedup: 3x faster**
```
**Email Report** (via email skill):
```
To: [user-email]
Subject: QA Campaign Complete - [N] Components Tested
[Executive Summary]
[Link to full report]
[Top 3 critical issues]
[Next steps]
```
## Sub-Agent Communication Protocol
### Spawn Pattern
```yaml
agent_spawn:
tool: Task
agent_type: general-purpose
description: "QA Testing: [component-name]"
instructions: |
1. Invoke Skill(frontend-qc) for component: [component-name]
2. Follow all 6 phases of frontend-qc workflow
3. Use Chrome DevTools MCP with --isolated flag
4. Report bugs via Skill(report-bug)
5. Generate component-specific report
6. Return findings to main agent
```
### Result Collection
```yaml
sub_agent_output:
component_name: "login"
status: "PASS" | "FAIL"
bugs_found: [
{
severity: "HIGH",
title: "Login button unresponsive on mobile",
github_issue: "#123",
evidence: "screenshot.png"
}
]
testing_duration: "5 minutes"
recommendations: ["Add mobile viewport testing"]
```
## Error Handling
### Sub-Agent Failures
**Scenario**: Sub-agent crashes or times out
**Action**:
1. Log failure in TodoWrite
2. Retry sub-agent spawn (max 2 retries)
3. If retry fails, continue with other sub-agents
4. Mark component as "INCOMPLETE" in final report
5. Alert user to manual testing requirement
### Browser Resource Exhaustion
**Scenario**: Too many concurrent browser sessions
**Action**:
1. Detect Chrome DevTools connection failures
2. Reduce concurrency limit from 5 to 3
3. Queue remaining components
4. Log resource constraint for user awareness
### Shared Resource Conflicts
**Scenario**: Multiple sub-agents modifying same test data
**Action**:
1. Use component-specific test accounts when possible
2. Coordinate database state across sub-agents
3. Implement test data isolation strategy
4. Document conflicts in final report
## Performance Optimization
**Token Efficiency**:
- Main agent context: ~3K tokens
- Sub-agent contexts: Isolated (don't pollute main)
- Result aggregation: ~2K tokens
- **Total main context**: ~5K tokens (vs ~25K sequential)
**Time Efficiency**:
- Sequential: N components Ć 15 min/component
- Parallel: max(15 min) + 10 min overhead
- **Speedup**: ~3-5x for 3-5 components
## Success Criteria
ā
All components tested successfully
ā
Bugs reported to GitHub with evidence
ā
Cross-component patterns identified
ā
Unified report generated and emailed
ā
Parallel execution faster than sequential
ā
No sub-agent failures or hangs
## Integration Points
**Called By**: `~/.claude/skills/frontend-qc/SKILL.md` (complexity detection)
**Calls**:
- Task tool for sub-agent spawning
- Skill(frontend-qc) within each sub-agent
- Skill(report-bug) for bug creation
- Skill(email) for reporting
**Files Generated**:
- `claudedocs/qa-campaign-[timestamp].md` - Full report
- `claudedocs/qa-sub-agent-[component]-[timestamp].md` - Per-component reports
- GitHub issues via report-bug skill
## Example Invocation
**From Skill** (automatic):
```
Skill detected: Testing login, registration, checkout, profile (4 components >3)
Action: Invoke Task tool
Description: "Parallel QA testing campaign across 4 components: login, registration, checkout, profile"
Context: {
components: ["login", "registration", "checkout", "profile"],
url: "http://localhost:4000",
credentials: { user: "test@example.com", pass: "test123" }
}
```
**Manual** (explicit user request):
```
User: "Test all checkout flow pages in parallel"
Claude: Invoking frontend-qc-agent...
[Agent spawns sub-agents for: cart, shipping, payment, confirmation]
[15 minutes later]
Agent: "QA campaign complete. Found 3 bugs, 1 critical. Report: [link]"
```
---
## Best Practices
1. **Component Isolation**: Always use --isolated flag for browser sessions
2. **Concurrency Limits**: Never exceed 5 concurrent sub-agents
3. **Result Deduplication**: Always check for cross-component patterns
4. **Evidence Collection**: Ensure each sub-agent captures screenshots and logs
5. **Graceful Degradation**: Continue with partial results if sub-agents fail
6. **Time Estimation**: Communicate expected completion time to user upfront
7. **Resource Management**: Monitor system resources and adjust concurrency
---
**End of Agent Definition**Signals
Information
- Repository
- arlenagreer/claude_configuration_docs
- Author
- arlenagreer
- Last Sync
- 2/6/2026
- Repo Updated
- 1/13/2026
- Created
- 1/15/2026
Reviews (0)
No reviews yet. Be the first to review this skill!
Related Skills
upgrade-nodejs
Upgrading Bun's Self-Reported Node.js Version
cursorrules
CrewAI Development Rules
cn-check
Install and run the Continue CLI (`cn`) to execute AI agent checks on local code changes. Use when asked to "run checks", "lint with AI", "review my changes with cn", or set up Continue CI locally.
CLAUDE
CLAUDE.md
Related Guides
Bear Notes Claude Skill: Your AI-Powered Note-Taking Assistant
Learn how to use the bear-notes Claude skill. Complete guide with installation instructions and examples.
Mastering tmux with Claude: A Complete Guide to the tmux Claude Skill
Learn how to use the tmux Claude skill. Complete guide with installation instructions and examples.
OpenAI Whisper API Claude Skill: Complete Guide to AI-Powered Audio Transcription
Learn how to use the openai-whisper-api Claude skill. Complete guide with installation instructions and examples.