jiangnanboy/llm_security
利用分类法和敏感词检测法对生成式大模型的输入和输出内容进行安全检测,尽早识别风险内容。The input and output contents of generative large model are checked by classification method and sensitive word detection method to identify content risk as early as possible.
No commits in the last 6 months.
Stars
28
Forks
4
Language
Java
License
MIT
Category
Last pushed
Sep 09, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/jiangnanboy/llm_security"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
OWASP/www-project-top-10-for-large-language-model-applications
OWASP Top 10 for Large Language Model Apps (Part of the GenAI Security Project)
esbmc/esbmc-ai
Automated Code Repair suite powered by ESBMC and LLMs.
cla7aye15I4nd/PatchAgent
[USENIX Security 25] PatchAgent is a LLM-based practical program repair agent that mimics human...
iSEngLab/AwesomeLLM4APR
[TOSEM 2026]A Systematic Literature Review on Large Language Models for Automated Program Repair
YerbaPage/MGDebugger
Multi-Granularity LLM Debugger [ICSE2026]