hugobatista/unicode-injection
Proof of concept demonstrating Unicode injection vulnerabilities using invisible characters to manipulate Large Language Models (LLMs) and AI assistants (e.g., Claude, AI Studio) via hidden prompts or data poisoning. Educational/research purposes only.
Stars
5
Forks
1
Language
—
License
—
Category
Last pushed
Jan 27, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/prompt-engineering/hugobatista/unicode-injection"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
protectai/llm-guard
The Security Toolkit for LLM Interactions
MaxMLang/pytector
Easy to use LLM Prompt Injection Detection / Detector Python Package with support for local...
agencyenterprise/PromptInject
PromptInject is a framework that assembles prompts in a modular fashion to provide a...
utkusen/promptmap
a security scanner for custom LLM applications
Dicklesworthstone/acip
The Advanced Cognitive Inoculation Prompt