llm-analytics-setup

Community

Add PostHog LLM analytics to your app.

Authorkausthubh-coder
Version1.0.0
Installs0

System Documentation

What problem does it solve?

This Skill simplifies the process of integrating PostHog's LLM analytics into your application, providing deep insights into your AI model usage and performance.

Core Features & Use Cases

  • Provider Integration: Offers setup guides for numerous LLM providers (OpenAI, Anthropic, Cohere, etc.) and frameworks (LangChain, AutoGen).
  • Automatic Event Tracking: Captures $ai_generation events with detailed properties like tokens, latency, and cost.
  • Use Case: Integrate PostHog LLM analytics into your Python application to track token usage, identify performance bottlenecks, and monitor costs across different LLM providers.

Quick Start

Follow the instructions in the openai.md reference file to set up PostHog LLM analytics for your OpenAI integration.

Dependency Matrix

Required Modules

None required

Components

references

💻 Claude Code Installation

Recommended: Let Claude install automatically. Simply copy and paste the text below to Claude Code.

Please help me install this Skill:
Name: llm-analytics-setup
Download link: https://github.com/kausthubh-coder/studi/archive/main.zip#llm-analytics-setup

Please download this .zip file, extract it, and install it in the .claude/skills/ directory.
View Source Repository

Agent Skills Search Helper

Install a tiny helper to your Agent, search and equip skill from 223,000+ vetted skills library on demand.