promptfoo-evaluation
CommunityConfigure, run, and judge LLM evaluations.
Authordaymade
Version1.0.0
Installs0
System Documentation
What problem does it solve?
This Skill helps teams set up prompt testing configurations, run evaluations, and collect results using Promptfoo, including Python assertions and llm-rubric scoring.
Core Features & Use Cases
- Setup promptfoo configurations and tests
- Write Python custom assertions (metrics)
- llm-rubric scoring with thresholds
- Compare different models and prompts
Quick Start
- Initialize a new evaluation project with npx promptfoo@latest init
- Run evaluation with npx promptfoo@latest eval
- View results with npx promptfoo@latest view
Dependency Matrix
Required Modules
None requiredComponents
references
💻 Claude Code Installation
Recommended: Let Claude install automatically. Simply copy and paste the text below to Claude Code.
Please help me install this Skill: Name: promptfoo-evaluation Download link: https://github.com/daymade/claude-code-skills/archive/main.zip#promptfoo-evaluation Please download this .zip file, extract it, and install it in the .claude/skills/ directory.