bytedance/UNO
[ICCV 2025] π₯π₯ UNO: A Universal Customization Method for Both Single and Multi-Subject Conditioning
Uses in-context generation with diffusion transformers to synthesize high-consistency multi-subject paired training data, enabling a progressive cross-modal alignment architecture with universal rotary position embeddings. Builds on FLUX.1-dev as the base diffusion model and supports both single and multi-image conditioning for subject-driven generation. Provides training and inference implementations with fp8 quantization support for consumer GPUs (~16GB VRAM), plus a Hugging Face dataset (UNO-1M) and pre-trained weights.
1,353 stars. No commits in the last 6 months.
Stars
1,353
Forks
77
Language
Python
License
Apache-2.0
Category
Last pushed
Sep 12, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/diffusion/bytedance/UNO"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
UCSC-VLAA/story-iter
[ICLR 2026] A Training-free Iterative Framework for Long Story Visualization
PaddlePaddle/PaddleMIX
Paddle Multimodal Integration and eXploration, supporting mainstream multi-modal tasks,...
keivalya/mini-vla
a minimal, beginner-friendly VLA to show how robot policies can fuse images, text, and states to...
adobe-research/custom-diffusion
Custom Diffusion: Multi-Concept Customization of Text-to-Image Diffusion (CVPR 2023)
byliutao/1Prompt1Story
π₯ICLR 2025 (Spotlight) One-Prompt-One-Story: Free-Lunch Consistent Text-to-Image Generation...