menyifang/MIMO
Official implementation of "MIMO: Controllable Character Video Synthesis with Spatial Decomposed Modeling"
Decomposes video into spatial components (character, motion, scene) to enable independent control over each attribute during synthesis. Built on Stable Diffusion with motion modules and pose guidance, it supports animating static character images with 3D motion data or driving videos while generalizing to novel characters and scenes. Provides both local PyTorch inference and a Gradio web interface, with pre-trained weights available via HuggingFace and ModelScope.
1,575 stars. No commits in the last 6 months.
Stars
1,575
Forks
70
Language
Python
License
Apache-2.0
Category
Last pushed
Jun 19, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/diffusion/menyifang/MIMO"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
hao-ai-lab/FastVideo
A unified inference and post-training framework for accelerated video generation.
thu-ml/TurboDiffusion
TurboDiffusion: 100–200× Acceleration for Video Diffusion Models
ModelTC/LightX2V
Light Image Video Generation Inference Framework
PKU-YuanGroup/Helios
Helios: Real Real-Time Long Video Generation Model
PKU-YuanGroup/MagicTime
[TPAMI 2025🔥] MagicTime: Time-lapse Video Generation Models as Metamorphic Simulators