agentscope-ai/OpenJudge

OpenJudge: A Unified Framework for Holistic Evaluation and Quality Rewards

53
/ 100
Established

# Initialize grader grader = Grader.load("Relevance") # Evaluate response response = { "query": "What is machine learning?", "response": "Machine learning is a subset of AI that enables systems to learn from data.", } score = grader.evaluate(response) print(score) # Output: # { # "name": "Relevance", # "score": 0.95, # "reasoning": "The response directly addresses the query with an accurate definition of machine learning." # } ``` ### Batch Evaluation Example Evaluate multiple responses at scale: ```python # Load test data test_data = [ {"query": "What is AI?", "response": "AI is artificial intelligence."}, {"query": "What is ML?", "response": "Machine learning is

459 stars.

No Package No Dependents
Maintenance 13 / 25
Adoption 10 / 25
Maturity 15 / 25
Community 15 / 25

How are scores calculated?

Stars

459

Forks

37

Language

Python

License

Apache-2.0

Last pushed

Mar 12, 2026

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/agents/agentscope-ai/OpenJudge"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.