rustformers/llm
[Unmaintained, see README] An ecosystem of Rust libraries for working with large language models
ArchivedBuilt on the GGML tensor library, this project provided modular Rust crates for CPU-based LLM inference supporting models like LLaMA, GPT-J, and BLOOM with both library and CLI interfaces. The architecture separated concerns across `llm-base`, model-specific crates, and a `ggml-sys` FFI layer, enabling quantization and perplexity computation alongside interactive chat modes. It integrated with Hugging Face tokenizers by default and supported GGML model formats, though later branches explored GGUF compatibility.
6,152 stars. No commits in the last 6 months.
Stars
6,152
Forks
373
Language
Rust
License
Apache-2.0
Category
Last pushed
Jun 24, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/rustformers/llm"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
trymirai/uzu
A high-performance inference engine for AI models
justrach/bhumi
⚡ Bhumi – The fastest AI inference client for Python, built with Rust for unmatched speed,...
lipish/llm-connector
LLM Connector - A unified interface for connecting to various Large Language Model providers
keyvank/femtoGPT
Pure Rust implementation of a minimal Generative Pretrained Transformer
ShelbyJenkins/llm_client
The Easiest Rust Interface for Local LLMs and an Interface for Deterministic Signals from...