whylabs/langkit
🔍 LangKit: An open-source toolkit for monitoring Large Language Models (LLMs). 📚 Extracts signals from prompts & responses, ensuring safety & security. 🛡️ Features include text quality, relevance metrics, & sentiment analysis. 📊 A comprehensive tool for LLM observability. 👀
Extracts specialized threat signals including jailbreak attempts, prompt injection attacks, hallucination detection, and refusal patterns alongside standard quality metrics. Built as a modular UDF layer that integrates directly with whylogs' schema system, enabling composable metric pipelines with configurable performance trade-offs (throughput ranges from 2K+ chats/sec with light metrics to sub-1 chat/sec with full analysis). Designed for production LLM monitoring workflows, with outputs visualizable in the WhyLabs observability platform or analyzed independently.
976 stars. No commits in the last 6 months.
Stars
976
Forks
70
Language
Jupyter Notebook
License
Apache-2.0
Category
Last pushed
Nov 22, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/prompt-engineering/whylabs/langkit"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Compare
Higher-rated alternatives
langfuse/langfuse
🪢 Open source LLM engineering platform: LLM Observability, metrics, evals, prompt management,...
Arize-ai/phoenix
AI Observability & Evaluation
Mirascope/mirascope
The LLM Anti-Framework
Helicone/helicone
🧊 Open source LLM observability platform. One line of code to monitor, evaluate, and experiment. YC W23 🍓
Agenta-AI/agenta
The open-source LLMOps platform: prompt playground, prompt management, LLM evaluation, and LLM...