inference and PowerInfer

These are competitors: both provide local LLM inference engines with unified interfaces for running open-source models, though Xinference emphasizes multi-modal support and cloud/on-prem flexibility while PowerInfer focuses on speed optimization through GPU-CPU co-inference.

inference
89
Verified
PowerInfer
54
Established
Maintenance 25/25
Adoption 20/25
Maturity 25/25
Community 19/25
Maintenance 10/25
Adoption 10/25
Maturity 16/25
Community 18/25
Stars: 9,129
Forks: 805
Downloads: 28,276
Commits (30d): 59
Language: Python
License: Apache-2.0
Stars: 8,808
Forks: 501
Downloads:
Commits (30d): 0
Language: C++
License: MIT
No risk flags
No Package No Dependents

About inference

xorbitsai/inference

Swap GPT for any LLM by changing a single line of code. Xinference lets you run open-source, speech, and multimodal models on cloud, on-prem, or your laptop — all through one unified, production-ready inference API.

About PowerInfer

Tiiny-AI/PowerInfer

High-speed Large Language Model Serving for Local Deployment

Scores updated daily from GitHub, PyPI, and npm data. How scores work