vllm-project/vllm-ascend
Community maintained hardware plugin for vLLM on Ascend
Implements vLLM as a pluggable hardware abstraction layer for Ascend NPUs, enabling support for diverse model architectures including Transformer, Mixture-of-Experts, embedding, and multi-modal LLMs without modifying core vLLM. Decouples Ascend backend integration through standardized hardware plugin interfaces while leveraging CANN runtime and PyTorch-NPU for optimized inference on Atlas hardware series (A2/A3 training and inference accelerators).
1,773 stars. Actively maintained with 344 commits in the last 30 days.
Stars
1,773
Forks
912
Language
C++
License
Apache-2.0
Category
Last pushed
Mar 13, 2026
Commits (30d)
344
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/vllm-project/vllm-ascend"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related tools
kvcache-ai/Mooncake
Mooncake is the serving platform for Kimi, a leading LLM service provided by Moonshot AI.
SemiAnalysisAI/InferenceX
Open Source Continuous Inference Benchmarking Qwen3.5, DeepSeek, GPTOSS - GB200 NVL72 vs MI355X...
uccl-project/uccl
UCCL is an efficient communication library for GPUs, covering collectives, P2P (e.g., KV cache...
sophgo/tpu-mlir
Machine learning compiler based on MLIR for Sophgo TPU.
BBuf/how-to-optim-algorithm-in-cuda
how to optimize some algorithm in cuda.