ai-security
CommunitySecure your LLM apps from threats, effortlessly.
Software Engineering#AI security#jailbreak#prompt injection#guardrails#PII redaction#content moderation#LLM security
Authorricardoroche
Version1.0.0
Installs0
System Documentation
What problem does it solve?
This Skill addresses critical security vulnerabilities in AI and LLM applications, such as prompt injection, PII leakage, and generation of unsafe content. It provides robust patterns to protect your applications and users, ensuring compliance and trust.
Core Features & Use Cases
- Prompt Injection Detection: Identifies and mitigates attempts to manipulate LLM behavior through malicious inputs.
- PII Redaction: Automatically redacts sensitive Personally Identifiable Information from prompts before sending to LLMs, and unredacts outputs when safe.
- Output Content Filtering: Filters LLM-generated content for violence, hate speech, sexual content, or illegal activities.
- Secure Prompt Construction: Guides on building prompts with clear boundaries and integrated security checks.
- Use Case: An LLM application processes user queries that might contain sensitive data or malicious instructions. This skill helps implement a
SecurePromptBuilderto detect injection, redact PII, and filter the LLM's response for harmful content, ensuring a safe interaction.
Quick Start
Implement prompt injection detection for user inputs in my LLM application.
Dependency Matrix
Required Modules
pydantic
Components
Standard package💻 Claude Code Installation
Recommended: Let Claude install automatically. Simply copy and paste the text below to Claude Code.
Please help me install this Skill: Name: ai-security Download link: https://github.com/ricardoroche/ricardos-claude-code/archive/main.zip#ai-security Please download this .zip file, extract it, and install it in the .claude/skills/ directory.