liu00222/Open-Prompt-Injection
This repository provides a benchmark for prompt injection attacks and defenses in LLMs
Provides modular factory methods for constructing attacks (combining multiple injection tasks), defenses (DataSentinel detection, PromptLocate localization), and LLM-integrated applications across multiple models (PaLM2, Llama, GPT). Introduces Attack Success Rate (ASV) as a quantitative evaluation metric and includes a detection-plus-localization pipeline that identifies contaminated prompts and recovers original data. Supports configuration-driven experimentation with pre-built task datasets (sentiment analysis, spam detection) and fine-tuned checkpoint integration via LoRA adapters.
406 stars.
Stars
406
Forks
64
Language
Python
License
MIT
Category
Last pushed
Oct 29, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/prompt-engineering/liu00222/Open-Prompt-Injection"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related tools
cybozu/prompt-hardener
Prompt Hardener analyzes prompt-injection-originated risk in LLM-based agents and applications.
R3dShad0w7/PromptMe
PromptMe is an educational project that showcases security vulnerabilities in large language...
lakeraai/pint-benchmark
A benchmark for prompt injection detection systems.
StavC/Here-Comes-the-AI-Worm
Here Comes the AI Worm: Preventing the Propagation of Adversarial Self-Replicating Prompts...
grepstrength/WideOpenAI
Short list of indirect prompt injection attacks for OpenAI-based models.