inference and PowerInfer
These are competitors: both provide local LLM inference engines with unified interfaces for running open-source models, though Xinference emphasizes multi-modal support and cloud/on-prem flexibility while PowerInfer focuses on speed optimization through GPU-CPU co-inference.
Maintenance
25/25
Adoption
20/25
Maturity
25/25
Community
19/25
Maintenance
10/25
Adoption
10/25
Maturity
16/25
Community
18/25
Stars: 9,129
Forks: 805
Downloads: 28,276
Commits (30d): 59
Language: Python
License: Apache-2.0
Stars: 8,808
Forks: 501
Downloads: —
Commits (30d): 0
Language: C++
License: MIT
No risk flags
No Package
No Dependents
About inference
xorbitsai/inference
Swap GPT for any LLM by changing a single line of code. Xinference lets you run open-source, speech, and multimodal models on cloud, on-prem, or your laptop — all through one unified, production-ready inference API.
About PowerInfer
Tiiny-AI/PowerInfer
High-speed Large Language Model Serving for Local Deployment
Related comparisons
Scores updated daily from GitHub, PyPI, and npm data. How scores work