Awesome-Video-Diffusion and awesome-diffusion-categorized
About Awesome-Video-Diffusion
showlab/Awesome-Video-Diffusion
A curated list of recent diffusion models for video generation, editing, and various other applications.
Organized into 20+ specialized categories, the collection spans foundation models and inference frameworks (HunyuanVideo, LTX-Video, Cosmos) alongside task-specific implementations for controllable generation, motion customization, video enhancement, talking head synthesis, and emerging domains like 4D content and game generation. The curated entries link to implementations built on diffusion architectures with complementary techniques including flow matching, reinforcement learning policies, and 3D/NeRF priors for physics-aware synthesis. Each resource includes direct GitHub repositories, arXiv papers, and project websites for reproducibility and comparative benchmarking across the video diffusion ecosystem.
About awesome-diffusion-categorized
wangkai930418/awesome-diffusion-categorized
collection of diffusion model papers categorized by their subareas
Organizes 500+ diffusion model papers across 20+ specialized research areas including image restoration, personalization, editing, and generation tasks with granular subcategories. Provides structured taxonomy covering emerging applications like visual illusions, virtual try-on, drag-based editing, and layout-guided generation that extends beyond standard text-to-image frameworks. Serves as a living research index with peer-reviewed papers linked to official implementations, project pages, and code repositories across CVPR, ICCV, ECCV, and other major venues.
Related comparisons
Scores updated daily from GitHub, PyPI, and npm data. How scores work