antrixsh/trusteval
Enterprise LLM Evaluation & Responsible AI Framework — Benchmark bias, hallucination, PII leakage, and toxicity across Healthcare, BFSI, Retail & Legal industries. Supports OpenAI, Anthropic, Gemini & HuggingFace. Python SDK + CLI + Web Dashboard. 191 tests. Compliance-ready reports.
Stars
4
Forks
4
Language
Python
License
MIT
Category
Last pushed
Mar 18, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/nlp/antrixsh/trusteval"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related tools
nityansuman/marvin
Web app to automatically generate subjective or an objective test and evaluate user responses...
shibing624/judger
自动作文评分工具,支持中文、英文作文智能评分,支持评分模型自训练,支持WEKA处理模型数据,支持自定义评分算法。java开发。
shubhpawar/Automated-Essay-Scoring
Automated Essay Scoring on The Hewlett Foundation dataset on Kaggle
samiali12/debateai-server
A FastAPI-powered backend that manages structured debates, analyzes arguments, and generates...
usnistgov/KAIROS
Scoring and analysis software for the evaluation of Knowledge Directed Artificial Intelligence...