shubchat/loab

LOAB: A benchmark for evaluating LLM agents on end-to-end mortgage lending operations under real regulatory constraints.

38
/ 100
Emerging

**Technical Summary** Evaluates LLM agents on mortgage origination through a multi-component scoring rubric that requires both correct outcomes *and* compliant processes—tool calls in sequence, proper inter-agent handoffs, forbidden action avoidance, and evidence chains—across six templated scenarios spanning prime/near-prime/sub-prime borrower profiles and fraud detection. Built on agentic orchestration with policy-bound decision routing (DTI thresholds, credit score gates, KYC sequencing) against Australian mortgage regulation, designed for multi-step workflows including Processing Officer, Underwriter, Credit Manager, and Financial Crime escalation paths. Current benchmark suite (v0.1.0) spans origination tasks with servicing, collections, and compliance modules in active development.

No Package No Dependents
Maintenance 13 / 25
Adoption 4 / 25
Maturity 9 / 25
Community 12 / 25

How are scores calculated?

Stars

5

Forks

1

Language

Python

License

MIT

Last pushed

Mar 09, 2026

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/agents/shubchat/loab"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.