promptfoo-evaluation

Community

Configure, run, and judge LLM evaluations.

Authordaymade
Version1.0.0
Installs0

System Documentation

What problem does it solve?

This Skill helps teams set up prompt testing configurations, run evaluations, and collect results using Promptfoo, including Python assertions and llm-rubric scoring.

Core Features & Use Cases

  • Setup promptfoo configurations and tests
  • Write Python custom assertions (metrics)
  • llm-rubric scoring with thresholds
  • Compare different models and prompts

Quick Start

  1. Initialize a new evaluation project with npx promptfoo@latest init
  2. Run evaluation with npx promptfoo@latest eval
  3. View results with npx promptfoo@latest view

Dependency Matrix

Required Modules

None required

Components

references

💻 Claude Code Installation

Recommended: Let Claude install automatically. Simply copy and paste the text below to Claude Code.

Please help me install this Skill:
Name: promptfoo-evaluation
Download link: https://github.com/daymade/claude-code-skills/archive/main.zip#promptfoo-evaluation

Please download this .zip file, extract it, and install it in the .claude/skills/ directory.
View Source Repository