langsmith-evaluator
CommunityBuild and run LangSmith evaluations.
Authordhar174
Version1.0.0
Installs0
System Documentation
What problem does it solve?
This Skill streamlines the process of evaluating AI agent performance by providing tools and guidance for creating, defining, and running evaluations within the LangSmith platform.
Core Features & Use Cases
- Create Evaluators: Supports both LLM-as-Judge and custom code evaluators for diverse evaluation needs.
- Define Run Functions: Guides on capturing agent outputs and trajectories for accurate evaluation.
- Run Evaluations: Facilitates local execution with
evaluate()and automatic execution via LangSmith uploads. - Use Case: You've built a customer support chatbot and want to ensure its responses are accurate and helpful. Use this Skill to set up an LLM-as-Judge evaluator that grades responses against expected outcomes and a custom code evaluator to check for adherence to specific response formats.
Quick Start
Use the langsmith-evaluator skill to upload a Python evaluator script named 'evaluators.py' with the function 'accuracy_evaluator' to your LangSmith project named 'My Project'.
Dependency Matrix
Required Modules
None requiredComponents
scriptsreferences
💻 Claude Code Installation
Recommended: Let Claude install automatically. Simply copy and paste the text below to Claude Code.
Please help me install this Skill: Name: langsmith-evaluator Download link: https://github.com/dhar174/langgraph_system_generator/archive/main.zip#langsmith-evaluator Please download this .zip file, extract it, and install it in the .claude/skills/ directory.
Agent Skills Search Helper
Install a tiny helper to your Agent, search and equip skill from 223,000+ vetted skills library on demand.