eval-audit

Community

Audit LLM evals for trust.

Authorhamelsmu
Version1.0.0
Installs0

System Documentation

What problem does it solve?

This Skill identifies critical flaws in your LLM evaluation pipelines, ensuring your metrics are trustworthy and your AI product is genuinely improving.

Core Features & Use Cases

  • Diagnostic Checks: Assesses six key areas: Error Analysis, Evaluator Design, Judge Validation, Human Review Process, Labeled Data, and Pipeline Hygiene.
  • Prioritized Findings: Delivers a report of problems ordered by their impact on your evaluation's reliability.
  • Actionable Next Steps: Provides concrete recommendations, often suggesting other skills to fix identified issues.
  • Use Case: You've inherited an LLM evaluation system and are unsure if its results are reliable. Run eval-audit to get a clear picture of potential issues and a roadmap for improvement.

Quick Start

Use the eval-audit skill to audit my current LLM evaluation pipeline.

Dependency Matrix

Required Modules

None required

Components

references

💻 Claude Code Installation

Recommended: Let Claude install automatically. Simply copy and paste the text below to Claude Code.

Please help me install this Skill:
Name: eval-audit
Download link: https://github.com/hamelsmu/evals-skills/archive/main.zip#eval-audit

Please download this .zip file, extract it, and install it in the .claude/skills/ directory.
View Source Repository

Agent Skills Search Helper

Install a tiny helper to your Agent, search and equip skill from 223,000+ vetted skills library on demand.