Groq API Inference
CommunityLow-latency Groq chat and speech inference
Authorsyahravi
Version1.0.0
Installs0
System Documentation
What problem does it solve?
This Skill helps developers build, integrate, and troubleshoot Groq API inference workflows for chat, tool calling, and speech transcription, focusing on low-latency routing, structured outputs, and production-safe patterns.
Core Features & Use Cases
- Model routing & selection: discover live models, keep short candidate sets per workload, and persist primary and fallback choices in memory.
- Resilience & reliability: exponential backoff retries with jitter, capped attempts, failover to fallback models, and logging for diagnosis.
- Output validation & safety: enforce strict JSON schemas or parsing checks before executing downstream actions and keep secrets scoped to environment variables.
- Use Case: Route interactive chat to a fast model, transcriptions to a speech-optimized model, and fail over automatically on repeated 5xx or rate limits while validating outputs before any automated write operations.
Quick Start
Verify GROQ_API_KEY and run a models health check to select a low-latency model, configure a fallback, and confirm output validation rules in your memory file.
Dependency Matrix
Required Modules
None requiredComponents
Standard package💻 Claude Code Installation
Recommended: Let Claude install automatically. Simply copy and paste the text below to Claude Code.
Please help me install this Skill: Name: Groq API Inference Download link: https://github.com/syahravi/openclaw/archive/main.zip#groq-api-inference Please download this .zip file, extract it, and install it in the .claude/skills/ directory.
Agent Skills Search Helper
Install a tiny helper to your Agent, search and equip skill from 223,000+ vetted skills library on demand.