evaluation-anchor-checker
CommunityEnsure numeric claims are reviewer-safe.
Content & Communication#technical writing#evaluation#claims#numeric hygiene#reviewer safety#protocol context
AuthorWILLOSCAR
Version1.0.0
Installs0
System Documentation
What problem does it solve?
This Skill addresses the common issue in technical surveys where strong numeric or performance claims are made without sufficient context, making them difficult for reviewers to verify and potentially leading to rejection.
Core Features & Use Cases
- Numeric Claim Auditing: Reviews sentences containing numbers to ensure they are accompanied by minimal, interpretable evaluation context (task, metric, constraint).
- Contextual Downgrading: Weakens or removes numeric claims if the necessary context is missing, preventing guesswork and maintaining evidence-boundedness.
- Ambiguous Naming Check: Flags and corrects underspecified model names that appear hallucinated.
- Use Case: Before submitting a research paper, use this Skill to automatically scan all claims like "Model X achieved 90% accuracy" and ensure they are rewritten to include the specific task and metric, e.g., "On the ImageNet classification task, Model X achieved ~90% top-1 accuracy".
Quick Start
Use the evaluation anchor checker skill to audit and rewrite numeric claims in the provided workspace.
Dependency Matrix
Required Modules
None requiredComponents
scriptsassetsreferences
💻 Claude Code Installation
Recommended: Let Claude install automatically. Simply copy and paste the text below to Claude Code.
Please help me install this Skill: Name: evaluation-anchor-checker Download link: https://github.com/WILLOSCAR/research-units-pipeline-skills/archive/main.zip#evaluation-anchor-checker Please download this .zip file, extract it, and install it in the .claude/skills/ directory.
Agent Skills Search Helper
Install a tiny helper to your Agent, search and equip skill from 223,000+ vetted skills library on demand.