NVIDIA-NeMo/NeMo
A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Automatic Speech Recognition and Text-to-Speech)
Provides modular PyTorch-native components for ASR, TTS, and speech LLMs with streaming-capable models like Nemotron-Speech and MagpieTTS supporting multilingual inference. Built on composable neural modules that enable efficient fine-tuning and deployment while integrating with NVIDIA's NIM (NVIDIA Inference Microservices) for production serving. Supports variable latency-accuracy tradeoffs through single checkpoints, enabling researchers to optimize for real-time or high-accuracy requirements.
16,894 stars. Actively maintained with 56 commits in the last 30 days.
Stars
16,894
Forks
3,365
Language
Python
License
Apache-2.0
Category
Last pushed
Mar 13, 2026
Commits (30d)
56
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/generative-ai/NVIDIA-NeMo/NeMo"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related tools
vlm-run/vlmrun-hub
A hub for various industry-specific schemas to be used with VLMs.
alexiglad/EBT
PyTorch Code for Energy-Based Transformers paper -- generalizable reasoning and scalable learning
HyperGAI/HPT
HPT - Open Multimodal LLMs from HyperGAI
yash9439/Falcon-Local-AI-Model
Explore this GitHub repository housing 3 versions of Falcon code for text generation. Each...
bastien-muraccioli/svlr
SVLR: Scalable, Training-Free Visual Language Robotics: a modular multi-model framework for...