llm-benchmark-workflow
OfficialBuild & run LLM benchmarks
AuthorArm-Examples
Version1.0.0
Installs0
System Documentation
What problem does it solve?
This Skill streamlines the process of building and executing performance benchmarks for Large Language Models (LLMs) across various backends, helping to identify and resolve issues related to benchmark execution and performance.
Core Features & Use Cases
- Benchmark Compilation: Compiles LLM benchmarking tools, including
arm-llm-bench-cli, for different ML frameworks. - Backend Configuration: Allows selection of specific backends like llama.cpp, ONNX Runtime GenAI, MediaPipe, and MNN.
- Runtime Triage: Assists in diagnosing and fixing build or runtime problems such as shared library placement, model path configuration, and thread/token settings.
- Use Case: When you modify the benchmarking code or add new performance metrics, use this Skill to ensure the benchmarks build correctly and run as expected across different LLM backends.
Quick Start
Build the benchmark binaries for the native platform by running cmake --preset=native -B build -DBUILD_BENCHMARK=ON && cmake --build ./build --parallel.
Dependency Matrix
Required Modules
None requiredComponents
scriptsreferences
💻 Claude Code Installation
Recommended: Let Claude install automatically. Simply copy and paste the text below to Claude Code.
Please help me install this Skill: Name: llm-benchmark-workflow Download link: https://github.com/Arm-Examples/LLM-Runner/archive/main.zip#llm-benchmark-workflow Please download this .zip file, extract it, and install it in the .claude/skills/ directory.
Agent Skills Search Helper
Install a tiny helper to your Agent, search and equip skill from 223,000+ vetted skills library on demand.