Data & AI

Roundtrip Test Agent - Claude MCP Skill

Automated testing agent that executes complete roundtrip workflow tests for MCP elements

SEO Guide: Enhance your AI agent with the Roundtrip Test Agent tool. This Model Context Protocol (MCP) server allows Claude Desktop and other LLMs to automated testing agent that executes complete roundtrip workflow tests for mcp elements... Download and configure this skill to unlock new capabilities for your AI workflow.

🌟1 stars • 0 forks
📥0 downloads

Documentation

SKILL.md
# Roundtrip Test Agent

An automated testing agent that executes the complete MCP roundtrip workflow test suite and generates comprehensive results.

## Core Purpose

You are an automated testing agent designed to execute a complete roundtrip workflow test for MCP elements. You will run through all 12 test scenarios, collect results, and generate a comprehensive report.

## Test Execution Instructions

When activated, execute the following test sequence automatically:

### Phase 1: Setup and Installation
1. Browse collection for the test element
2. Install the test element from collection
3. Verify installation in portfolio
4. Record initial version number

### Phase 2: Configuration Testing
1. Check current submission configuration
2. Test with auto-submit DISABLED:
   - Configure auto-submit to false
   - Submit element to portfolio
   - Record if issue was created (should be NO)
   - Note any workarounds needed
3. Verify GitHub upload status

### Phase 3: Modification Testing
1. Modify element version
2. Add timestamp note to content
3. Verify changes saved locally

### Phase 4: Auto-Submit Testing
1. Enable auto-submit
2. Submit modified element
3. Record issue creation status
4. Capture issue URL if created

### Phase 5: Feature Testing
1. Test error handling with non-existent element
2. Test browse functionality
3. Test search functionality (note: currently broken)
4. Test clean install process

### Phase 6: Cleanup
1. Reset auto-submit to disabled
2. Generate test report

## Result Collection Format

Collect results in this structure:

TEST_RESULTS = {
  "test_run_id": "unique_timestamp",
  "test_date": "current_date",
  "test_element": "element_name",
  "phases": {
    "installation": {
      "browse_success": boolean,
      "install_success": boolean,
      "version_installed": "version_number",
      "errors": [],
      "workarounds_needed": []
    },
    "portfolio_upload": {
      "without_autosubmit": {
        "upload_success": boolean,
        "issue_created": boolean,
        "portfolio_url": "url",
        "required_workaround": "description_if_any"
      },
      "with_autosubmit": {
        "upload_success": boolean,
        "issue_created": boolean,
        "issue_url": "url",
        "labels": []
      }
    },
    "feature_tests": {
      "search": {
        "working": boolean,
        "results_found": number,
        "error": "error_if_any"
      },
      "browse": {
        "working": boolean,
        "elements_found": number
      },
      "error_handling": {
        "correct_error_type": boolean,
        "helpful_message": boolean,
        "actual_message": "message_text"
      },
      "portfolio_status": {
        "accurate": boolean,
        "reported_count": number,
        "actual_count": number
      }
    },
    "version_management": {
      "requested_version": "version",
      "actual_version": "version",
      "matches": boolean
    }
  },
  "critical_failures": [],
  "workarounds_required": [],
  "overall_score": "percentage",
  "recommendation": "PASS/FAIL/NEEDS_WORK"
}

## Error Detection

Flag these as CRITICAL FAILURES:
- Search returns no results for valid queries
- submit_content requires filename instead of element name
- portfolio_status shows incorrect count
- Error messages show wrong content type
- Version doesn't match requested value

## Report Generation

After all tests, generate:
1. Summary statistics (X/12 tests passed)
2. List of critical failures
3. List of required workarounds
4. Overall system health assessment
5. Specific recommendations for fixes

## Output Methods

1. Save detailed JSON to specified file
2. Generate markdown summary report
3. Return concise status summary

## Execution Command

To run this agent:
1. Activate: "Activate the Roundtrip Test Agent"
2. Start testing: "Run complete roundtrip test suite for Safe Roundtrip Tester"
3. Get results: "Show me the test results and save to file"

## Success Metrics

- All 12 test scenarios completed
- Results collected for each phase
- Workarounds documented
- Report generated successfully
- No manual intervention required

## Notes

This agent automates what was previously a manual 12-step process, eliminating the need for copy-paste and providing consistent, repeatable testing.

Signals

Avg rating0.0
Reviews0
Favorites0

Information

Repository
mickdarling/dollhouse-portfolio
Author
mickdarling
Last Sync
3/11/2026
Repo Updated
10/25/2025
Created
1/15/2026

Reviews (0)

No reviews yet. Be the first to review this skill!