MOSS-TTS and MOSS-Speech
These are complementary components of an end-to-end speech processing pipeline: MOSS-Speech handles speech-to-speech understanding and generation, while MOSS-TTS provides the specialized text-to-speech synthesis layer needed to produce high-fidelity audio output.
About MOSS-TTS
OpenMOSS/MOSS-TTS
MOSS‑TTS Family is an open‑source speech and sound generation model family from MOSI.AI and the OpenMOSS team. It is designed for high‑fidelity, high‑expressiveness, and complex real‑world scenarios, covering stable long‑form speech, multi‑speaker dialogue, voice/character design, environmental sound effects, and real‑time streaming TTS.
# Technical Summary Built on a modular architecture, MOSS-TTS decomposes speech synthesis into five specialized models—flagship TTS for zero-shot voice cloning with phoneme-level control, a dialogue model outperforming closed-source baselines on objective metrics, a prompt-based voice generator requiring no reference audio, a low-latency realtime agent model (180ms TTFB), and a sound effect generator. The framework supports multiple inference backends including PyTorch-free deployment via llama.cpp with GGUF quantization and ONNX audio codec decoding, plus SGLang acceleration achieving 3× faster generation throughput. Models are available on Hugging Face and ModelScope with fine-tuning tutorials and REST API documentation via the MOSI.AI studio platform.
About MOSS-Speech
OpenMOSS/MOSS-Speech
MOSS-Speech is a true speech-to-speech large language model without text guidance.
Related comparisons
Scores updated daily from GitHub, PyPI, and npm data. How scores work