Home Security AI Benchmark
OfficialBenchmark LLM/VLM for home security.
Software Engineering#llm evaluation#performance testing#prompt injection#ai benchmark#vlm testing#security ai
AuthorSharpAI
Version1.0.0
Installs0
System Documentation
What problem does it solve?
This Skill evaluates the performance of Large Language Models (LLMs) and Vision-Language Models (VLMs) on a comprehensive suite of tests specifically designed for home security AI applications. It helps users select and optimize AI models for tasks like scene analysis, threat detection, and alert triage.
Core Features & Use Cases
- Extensive Test Suite: Covers 143 tests across 16 categories including context preprocessing, tool use, security classification, prompt injection resistance, and VLM scene analysis.
- LLM & VLM Evaluation: Assesses both text-based reasoning and visual understanding capabilities of AI models.
- Detailed Reporting: Generates an HTML report with pass/fail scores, latency charts, and token usage for easy comparison.
- Use Case: A security system integrator can use this Skill to benchmark different LLMs and VLMs to determine which models provide the best accuracy and reliability for their AI-powered security cameras before deployment.
Quick Start
Run the benchmark using the command node scripts/run-benchmark.cjs --vlm http://localhost:5405 to evaluate LLM and VLM capabilities with a VLM server running at the specified address.
Dependency Matrix
Required Modules
openai
Components
scriptsreferencesassets
💻 Claude Code Installation
Recommended: Let Claude install automatically. Simply copy and paste the text below to Claude Code.
Please help me install this Skill: Name: Home Security AI Benchmark Download link: https://github.com/SharpAI/DeepCamera/archive/main.zip#home-security-ai-benchmark Please download this .zip file, extract it, and install it in the .claude/skills/ directory.
Agent Skills Search Helper
Install a tiny helper to your Agent, search and equip skill from 223,000+ vetted skills library on demand.