mcp-code-execution-results-comparison-analyzer

Community

Compare AI agent performance, optimize workflows.

Authorolaservo
Version1.0.0
Installs0

System Documentation

What problem does it solve?

Evaluating and comparing the performance, cost, and output quality of different AI agent execution approaches (like code-execution vs. direct-MCP) is complex and manual. This Skill automates the analysis of experimental results, providing clear insights.

Core Features & Use Cases

  • Automated Data Extraction: Extracts metrics, logs, and workspace outputs from zipped experiment results.
  • Performance Metrics: Compares duration, cost, token usage, and efficiency ratios for successful runs.
  • Failure Detection: Identifies and summarizes failed runs, providing clear success/failure rates.
  • Output Quality Assessment: Catalogs and excerpts workspace files for qualitative comparison.
  • Actionable Recommendations: Generates a comprehensive report with insights on when to use each approach.

Quick Start

Use the mcp-code-execution-results-comparison-analyzer skill to compare the results from the attached zip file 'task_results-2025-11-20T06-22-50-300Z.zip'.

Dependency Matrix

Required Modules

None required

Components

scripts

💻 Claude Code Installation

Recommended: Let Claude install automatically. Simply copy and paste the text below to Claude Code.

Please help me install this Skill:
Name: mcp-code-execution-results-comparison-analyzer
Download link: https://github.com/olaservo/code-execution-with-mcp/archive/main.zip#mcp-code-execution-results-comparison-analyzer

Please download this .zip file, extract it, and install it in the .claude/skills/ directory.
View Source Repository

Agent Skills Search Helper

Install a tiny helper to your Agent, search and equip skill from 223,000+ vetted skills library on demand.