ComfyUI-VoxCPM and ComfyUI-VoxCPMTTS
These two tools are **competitors**, as the second project (1038lab/ComfyUI-VoxCPMTTS) explicitly offers a "clean, efficient ComfyUI custom node for VoxCPM TTS" for the same "high-quality speech generation and voice cloning capabilities using the VoxCPM 1.5 mo" as the first (wildminder/ComfyUI-VoxCPM), suggesting it's an alternative implementation of the same core functionality.
About ComfyUI-VoxCPM
wildminder/ComfyUI-VoxCPM
ComfyUI node for highly expressive speech and realistic zero-shot voice cloning
Implements a tokenizer-free diffusion-based TTS architecture built on MiniCPM-4 that models speech in continuous space rather than discrete tokens, enabling context-aware prosody generation. Includes native LoRA fine-tuning support within ComfyUI for custom voice style training, automatic model management with efficient VRAM offloading, and operates at 6.25Hz token rate for faster synthesis on consumer hardware. Integrates seamlessly with ComfyUI's node workflow system, supporting optional reference audio for voice cloning and compatible with multiple inference backends (CUDA, CPU, MPS, DirectML).
About ComfyUI-VoxCPMTTS
1038lab/ComfyUI-VoxCPMTTS
A clean, efficient ComfyUI custom node for VoxCPM TTS (Text-to-Speech) functionality. This implementation provides high-quality speech generation and voice cloning capabilities using the VoxCPM 1.5 model.
Related comparisons
Scores updated daily from GitHub, PyPI, and npm data. How scores work