promptfoo
OfficialTest and evaluate LLM prompts.
Software Engineering#prompt engineering#llm evaluation#promptfoo#prompt testing#ai compliance#performance benchmarking
AuthorDTMC-marketplace
Version1.0.0
Installs0
System Documentation
What problem does it solve?
This Skill addresses the challenge of systematically testing and evaluating Large Language Model (LLM) prompts to ensure their effectiveness, consistency, and compliance.
Core Features & Use Cases
- Systematic Prompt Testing: Evaluate prompts across multiple LLM providers and configurations.
- Regression Detection: Identify unintended changes in prompt performance over time.
- Performance Benchmarking: Compare different prompts or model versions to select the best performing ones.
- Compliance Assessment: Aid in evaluating AI systems against regulatory requirements like the EU AI Act's Article 15.
Quick Start
Use the promptfoo skill to test and evaluate LLM prompts systematically.
Dependency Matrix
Required Modules
None requiredComponents
references
💻 Claude Code Installation
Recommended: Let Claude install automatically. Simply copy and paste the text below to Claude Code.
Please help me install this Skill: Name: promptfoo Download link: https://github.com/DTMC-marketplace/governance/archive/main.zip#promptfoo Please download this .zip file, extract it, and install it in the .claude/skills/ directory.
Agent Skills Search Helper
Install a tiny helper to your Agent, search and equip skill from 223,000+ vetted skills library on demand.