Kitten-TTS-Server and KittenTTS
These are ecosystem siblings—one is the underlying TTS model (15M parameter lightweight architecture) while the other is an enhanced API server and web UI wrapper that deploys and exposes that same model for practical use.
About Kitten-TTS-Server
devnen/Kitten-TTS-Server
Self-host the ultra-lightweight Kitten TTS model with this enhanced API server with an intuitive Web UI, large text processing for audiobooks, and GPU acceleration.
Supports all 7 KittenTTS models (15M–80M parameters) with hot-swappable selection and automatic Hugging Face downloads, built on FastAPI with OpenAI-compatible endpoints. Implements GPU acceleration via ONNX Runtime I/O binding for reduced latency, alongside intelligent text chunking for audiobook generation. Optimized for edge deployment from Raspberry Pi 5 to NVIDIA GPUs with named voice support and Docker containerization.
About KittenTTS
soldier444xd/KittenTTS
KittenTTS is an ultra-lightweight, CPU-friendly text-to-speech model with 15M params for real-time, high-quality voices. Open source, fast start. 😺
Scores updated daily from GitHub, PyPI, and npm data. How scores work