Stanford-TML/EDGE
Official PyTorch Implementation of EDGE (CVPR 2023)
Leverages a transformer-based diffusion model with Jukebox music feature extraction to generate physically-plausible dance sequences from audio input. Provides fine-grained editing capabilities including joint-wise conditioning and in-betweening, with a novel Physical Foot Contact (PFC) metric for evaluating motion quality. Integrates with PyTorch3D and Hugging Face Accelerate for training, and supports FBX export for 3D animation workflows in Blender or Mixamo.
552 stars. No commits in the last 6 months.
Stars
552
Forks
96
Language
Python
License
MIT
Category
Last pushed
Jan 05, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/diffusion/Stanford-TML/EDGE"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
hao-ai-lab/FastVideo
A unified inference and post-training framework for accelerated video generation.
thu-ml/TurboDiffusion
TurboDiffusion: 100–200× Acceleration for Video Diffusion Models
ModelTC/LightX2V
Light Image Video Generation Inference Framework
PKU-YuanGroup/Helios
Helios: Real Real-Time Long Video Generation Model
PKU-YuanGroup/MagicTime
[TPAMI 2025🔥] MagicTime: Time-lapse Video Generation Models as Metamorphic Simulators