alvaropaco/haif

Production-ready microservices framework for AI inference over RPC. It provides a Gateway for client requests, an Orchestrator that schedules work, a Registry for model metadata, Workers that run inference, and a full observability stack (Prometheus, Grafana, Loki, Jaeger) — all wired together with Docker Compose.

22
/ 100
Experimental

Supports both CPU and GPU inference with model streaming via Hyperswarm's RPC transport; uses OpenTelemetry for unified metrics and distributed tracing across all Node services. Includes an HTTP Bridge for easy client integration, Postgres for orchestration state, and Redis for lightweight coordination—all with pre-built Grafana dashboards tracking throughput, latency (P95), and inference-specific metrics.

No commits in the last 6 months.

Stale 6m No Package No Dependents
Maintenance 2 / 25
Adoption 5 / 25
Maturity 9 / 25
Community 6 / 25

How are scores calculated?

Stars

14

Forks

1

Language

TypeScript

License

Apache-2.0

Last pushed

Oct 07, 2025

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/agents/alvaropaco/haif"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.