ai-testing-safety

Community

Red-team AI safety before deployment.

Authorlebsral
Version1.0.0
Installs0

System Documentation

What problem does it solve?

Find every way users can break your AI before they do. Use when you need to red-team your AI, test for jailbreaks, find prompt injection vulnerabilities, run adversarial testing, do a safety audit before launch, prove your AI is safe for compliance, stress-test guardrails, or verify your AI holds up against adversarial users. Covers automated attack generation, iterative red-teaming with DSPy, and MIPROv2-optimized adversarial testing.

Core Features & Use Cases

  • Guide the user through automated adversarial testing — systematically discovering vulnerabilities before real users exploit them.
  • Implement layered attack programs (attack -> critique -> refine) to improve attack quality and guardrail evaluation.
  • Generate structured safety reports and remediation guidance to support compliance and governance.

Quick Start

Run a safety audit by letting the DSPy RedTeamer probe your AI system and record vulnerabilities.

Dependency Matrix

Required Modules

None required

Components

Standard package

💻 Claude Code Installation

Recommended: Let Claude install automatically. Simply copy and paste the text below to Claude Code.

Please help me install this Skill:
Name: ai-testing-safety
Download link: https://github.com/lebsral/DSPy-Programming-not-prompting-LMs-skills/archive/main.zip#ai-testing-safety

Please download this .zip file, extract it, and install it in the .claude/skills/ directory.
View Source Repository

Agent Skills Search Helper

Install a tiny helper to your Agent, search and equip skill from 223,000+ vetted skills library on demand.