adversarial-examples
CommunityStress-test LLM robustness with adversarial inputs
Software Engineering#edge-cases#adversarial#LLM-testing#safety-testing#mutator#input-mutations#prompt-robustness
Authorpluginagentmarketplace
Version1.0.0
Installs0
System Documentation
What problem does it solve?
The Adversarial Examples skill helps teams systematically uncover LLM vulnerabilities by generating controlled edge cases, malformed inputs, and boundary payloads to evaluate model resilience.
Core Features & Use Cases
- Automated Adversarial Generation: Produce categorized input variations (linguistic, numerical, logical, format, consistency) to probe model behavior.
- Deterministic Mutation & Extensibility: Uses a built-in Python script to generate mutations and can be extended with additional payloads or reference patterns.
- Use Case: Run the tool against a deployed LLM to measure failure rates and guide mitigation strategies, such as prompt safety tuning and input validation.
Quick Start
To begin, run the included adversarial generator to produce test inputs and integrate results into your QA workflow. See the references for patterns and how to apply outputs.
Dependency Matrix
Required Modules
stringbase64codecstypingdataclasses
Components
scriptsreferencesassets
💻 Claude Code Installation
Recommended: Let Claude install automatically. Simply copy and paste the text below to Claude Code.
Please help me install this Skill: Name: adversarial-examples Download link: https://github.com/pluginagentmarketplace/custom-plugin-ai-red-teaming/archive/main.zip#adversarial-examples Please download this .zip file, extract it, and install it in the .claude/skills/ directory.
Agent Skills Search Helper
Install a tiny helper to your Agent, search and equip skill from 223,000+ vetted skills library on demand.