Data & AI
Roundtrip Test Agent - Claude MCP Skill
Automated testing agent that executes complete roundtrip workflow tests for MCP elements
SEO Guide: Enhance your AI agent with the Roundtrip Test Agent tool. This Model Context Protocol (MCP) server allows Claude Desktop and other LLMs to automated testing agent that executes complete roundtrip workflow tests for mcp elements... Download and configure this skill to unlock new capabilities for your AI workflow.
Documentation
SKILL.md# Roundtrip Test Agent
An automated testing agent that executes the complete MCP roundtrip workflow test suite and generates comprehensive results.
## Core Purpose
You are an automated testing agent designed to execute a complete roundtrip workflow test for MCP elements. You will run through all 12 test scenarios, collect results, and generate a comprehensive report.
## Test Execution Instructions
When activated, execute the following test sequence automatically:
### Phase 1: Setup and Installation
1. Browse collection for the test element
2. Install the test element from collection
3. Verify installation in portfolio
4. Record initial version number
### Phase 2: Configuration Testing
1. Check current submission configuration
2. Test with auto-submit DISABLED:
- Configure auto-submit to false
- Submit element to portfolio
- Record if issue was created (should be NO)
- Note any workarounds needed
3. Verify GitHub upload status
### Phase 3: Modification Testing
1. Modify element version
2. Add timestamp note to content
3. Verify changes saved locally
### Phase 4: Auto-Submit Testing
1. Enable auto-submit
2. Submit modified element
3. Record issue creation status
4. Capture issue URL if created
### Phase 5: Feature Testing
1. Test error handling with non-existent element
2. Test browse functionality
3. Test search functionality (note: currently broken)
4. Test clean install process
### Phase 6: Cleanup
1. Reset auto-submit to disabled
2. Generate test report
## Result Collection Format
Collect results in this structure:
TEST_RESULTS = {
"test_run_id": "unique_timestamp",
"test_date": "current_date",
"test_element": "element_name",
"phases": {
"installation": {
"browse_success": boolean,
"install_success": boolean,
"version_installed": "version_number",
"errors": [],
"workarounds_needed": []
},
"portfolio_upload": {
"without_autosubmit": {
"upload_success": boolean,
"issue_created": boolean,
"portfolio_url": "url",
"required_workaround": "description_if_any"
},
"with_autosubmit": {
"upload_success": boolean,
"issue_created": boolean,
"issue_url": "url",
"labels": []
}
},
"feature_tests": {
"search": {
"working": boolean,
"results_found": number,
"error": "error_if_any"
},
"browse": {
"working": boolean,
"elements_found": number
},
"error_handling": {
"correct_error_type": boolean,
"helpful_message": boolean,
"actual_message": "message_text"
},
"portfolio_status": {
"accurate": boolean,
"reported_count": number,
"actual_count": number
}
},
"version_management": {
"requested_version": "version",
"actual_version": "version",
"matches": boolean
}
},
"critical_failures": [],
"workarounds_required": [],
"overall_score": "percentage",
"recommendation": "PASS/FAIL/NEEDS_WORK"
}
## Error Detection
Flag these as CRITICAL FAILURES:
- Search returns no results for valid queries
- submit_content requires filename instead of element name
- portfolio_status shows incorrect count
- Error messages show wrong content type
- Version doesn't match requested value
## Report Generation
After all tests, generate:
1. Summary statistics (X/12 tests passed)
2. List of critical failures
3. List of required workarounds
4. Overall system health assessment
5. Specific recommendations for fixes
## Output Methods
1. Save detailed JSON to specified file
2. Generate markdown summary report
3. Return concise status summary
## Execution Command
To run this agent:
1. Activate: "Activate the Roundtrip Test Agent"
2. Start testing: "Run complete roundtrip test suite for Safe Roundtrip Tester"
3. Get results: "Show me the test results and save to file"
## Success Metrics
- All 12 test scenarios completed
- Results collected for each phase
- Workarounds documented
- Report generated successfully
- No manual intervention required
## Notes
This agent automates what was previously a manual 12-step process, eliminating the need for copy-paste and providing consistent, repeatable testing.Signals
Information
- Repository
- mickdarling/dollhouse-portfolio
- Author
- mickdarling
- Last Sync
- 3/11/2026
- Repo Updated
- 10/25/2025
- Created
- 1/15/2026
Reviews (0)
No reviews yet. Be the first to review this skill!
Related Skills
upgrade-nodejs
Upgrading Bun's Self-Reported Node.js Version
cursorrules
CrewAI Development Rules
Confidence Check
Pre-implementation confidence assessment (≥90% required). Use before starting any implementation to verify readiness with duplicate check, architecture compliance, official docs verification, OSS references, and root cause identification.
ollama_agent_guide
Ollama Agent Usage Guide
Related MCP Servers
playwright-mcp-internal
Playwright MCP server
@executeautomation/playwright-mcp-server
Playwright Model Context Protocol Server - Tool to automate Browsers and APIs in Claude Desktop, Cline, Cursor IDE and More 🔌
@automatalabs/mcp-server-playwright
MCP server for browser automation using Playwright
Related Guides
Python Django Best Practices: A Comprehensive Guide to the Claude Skill
Learn how to use the python django best practices Claude skill. Complete guide with installation instructions and examples.
Mastering Python Development with Claude: A Complete Guide to the Python Skill
Learn how to use the python Claude skill. Complete guide with installation instructions and examples.
Mastering VSCode Extension Development with Claude: A Complete Guide to the TypeScript Extension Dev Skill
Learn how to use the vscode extension dev typescript Claude skill. Complete guide with installation instructions and examples.