Groq API Inference

Community

Low-latency Groq chat and speech inference

Authorsyahravi
Version1.0.0
Installs0

System Documentation

What problem does it solve?

This Skill helps developers build, integrate, and troubleshoot Groq API inference workflows for chat, tool calling, and speech transcription, focusing on low-latency routing, structured outputs, and production-safe patterns.

Core Features & Use Cases

  • Model routing & selection: discover live models, keep short candidate sets per workload, and persist primary and fallback choices in memory.
  • Resilience & reliability: exponential backoff retries with jitter, capped attempts, failover to fallback models, and logging for diagnosis.
  • Output validation & safety: enforce strict JSON schemas or parsing checks before executing downstream actions and keep secrets scoped to environment variables.
  • Use Case: Route interactive chat to a fast model, transcriptions to a speech-optimized model, and fail over automatically on repeated 5xx or rate limits while validating outputs before any automated write operations.

Quick Start

Verify GROQ_API_KEY and run a models health check to select a low-latency model, configure a fallback, and confirm output validation rules in your memory file.

Dependency Matrix

Required Modules

None required

Components

Standard package

💻 Claude Code Installation

Recommended: Let Claude install automatically. Simply copy and paste the text below to Claude Code.

Please help me install this Skill:
Name: Groq API Inference
Download link: https://github.com/syahravi/openclaw/archive/main.zip#groq-api-inference

Please download this .zip file, extract it, and install it in the .claude/skills/ directory.
View Source Repository

Agent Skills Search Helper

Install a tiny helper to your Agent, search and equip skill from 223,000+ vetted skills library on demand.