josStorer/RWKV-Runner
A RWKV management and startup tool, full automation, only 8MB. And provides an interface compatible with the OpenAI API. RWKV is a large language model that is fully open source and available for commercial use.
Supports model management, conversion, and LoRA fine-tuning through an integrated toolkit, with configurable VRAM strategies including custom CUDA acceleration and WebGPU for cross-GPU compatibility. The backend uses a FastAPI-based inference service with an embeddings API, deployable separately from the frontend, and supports concurrent requests with customizable parameters for task-specific optimization (temperature, top_P). Compatible with ChatGPT clients and langchain workflows via OpenAI API emulation, plus additional features like MIDI input, chat presets, and multi-language support.
6,256 stars. Actively maintained with 1 commit in the last 30 days.
Stars
6,256
Forks
598
Language
TypeScript
License
MIT
Category
Last pushed
Feb 18, 2026
Commits (30d)
1
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/josStorer/RWKV-Runner"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related tools
jundot/omlx
LLM inference server with continuous batching & SSD caching for Apple Silicon — managed from the...
waybarrios/vllm-mlx
OpenAI and Anthropic compatible server for Apple Silicon. Run LLMs and vision-language models...
jordanhubbard/nanolang
A tiny experimental language designed to be targeted by coding LLMs
akivasolutions/tightwad
Pool your CUDA + ROCm GPUs into one OpenAI-compatible API. Speculative decoding proxy gives you...
petrukha-ivan/mlx-swift-structured
Structured output generation in Swift