groq
CommunityAchieve lightning-fast LLM inference, build responsive AI.
Authordarantrute
Version1.0.0
Installs0
System Documentation
What problem does it solve?
This Skill empowers developers to integrate Groq's ultra-fast LLM inference capabilities, overcoming latency challenges in AI applications. It provides patterns for streaming responses, managing rate limits, and securely using API keys, ensuring highly responsive and scalable AI features that delight users.
Core Features & Use Cases
- Ultra-Fast LLM Inference: Leverage Groq's LPU architecture for sub-second response times and high token throughput with models like Llama 3.3 and Mixtral.
- Real-Time Streaming & Tool Calling: Implement streaming chat completions using Server-Sent Events (SSE) and integrate advanced tool/function calling for dynamic AI interactions.
- Robust Integration Patterns: Provides guidance for secure API key management, comprehensive error handling, exponential backoff for rate limits, and seamless integration with Next.js and Vercel AI SDK.
- Use Case: Develop a real-time chatbot that responds instantly, generate code suggestions on the fly, or power a dynamic content creation tool where speed is critical.
Quick Start
Use the groq skill to set up a streaming chat completion endpoint in Next.js using the Llama 3.3 model.
Dependency Matrix
Required Modules
groq-sdk@ai-sdk/groqai
Components
references
💻 Claude Code Installation
Recommended: Let Claude install automatically. Simply copy and paste the text below to Claude Code.
Please help me install this Skill: Name: groq Download link: https://github.com/darantrute/userdecisiontree/archive/main.zip#groq Please download this .zip file, extract it, and install it in the .claude/skills/ directory.
Agent Skills Search Helper
Install a tiny helper to your Agent, search and equip skill from 223,000+ vetted skills library on demand.