G-U-N/Gen-L-Video

The official implementation for "Gen-L-Video: Multi-Text to Long Video Generation via Temporal Co-Denoising".

49
/ 100
Emerging

Extends existing short-video diffusion models to generate and edit videos with hundreds of frames and multiple semantic segments through temporal co-denoising, without additional training. Supports diverse conditioning modalities (pose, depth, sketch, canny, semantic segmentation) via T2I-Adapters and ControlNet, plus spatial editing through SAM and Grounding DINO integration. Offers both one-shot tuning and training-free inference pipelines compatible with Stable Diffusion-based architectures.

307 stars.

No Package No Dependents
Maintenance 6 / 25
Adoption 10 / 25
Maturity 16 / 25
Community 17 / 25

How are scores calculated?

Stars

307

Forks

34

Language

Jupyter Notebook

License

Apache-2.0

Last pushed

Oct 19, 2025

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/diffusion/G-U-N/Gen-L-Video"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.