ksm26/Evaluating-AI-Agents
A hands-on course repository for Evaluating AI Agents, created with Arize AI, that teaches you how to systematically evaluate, debug, and improve AI agents using observability tools, structured experiments, and reliable metrics. Learn production-grade techniques to enhance agent performance during development and after deployment.
No commits in the last 6 months.
Stars
1
Forks
1
Language
Jupyter Notebook
License
—
Category
Last pushed
May 12, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/prompt-engineering/ksm26/Evaluating-AI-Agents"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
langfuse/langfuse
🪢 Open source LLM engineering platform: LLM Observability, metrics, evals, prompt management,...
Arize-ai/phoenix
AI Observability & Evaluation
Mirascope/mirascope
The LLM Anti-Framework
Helicone/helicone
🧊 Open source LLM observability platform. One line of code to monitor, evaluate, and experiment. YC W23 🍓
Agenta-AI/agenta
The open-source LLMOps platform: prompt playground, prompt management, LLM evaluation, and LLM...