ai-forever/Kandinsky-2
Kandinsky 2 — multilingual text2image latent diffusion model
Built on a two-stage diffusion architecture with a CLIP-based prior that maps text embeddings to image embeddings before latent decoding, enabling text-guided image generation, inpainting, and image fusion. Version 2.2 upgrades to CLIP-ViT-G encoding and adds ControlNet support for spatial control over generation. Available via PyTorch with pretrained checkpoints on Hugging Face, supporting variable resolutions and LoRA fine-tuning.
2,817 stars. No commits in the last 6 months.
Stars
2,817
Forks
315
Language
Jupyter Notebook
License
Apache-2.0
Category
Last pushed
May 01, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/diffusion/ai-forever/Kandinsky-2"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
sakalond/StableGen
Transform your 3D texturing workflow with the power of generative AI, directly within Blender!
neggles/animatediff-cli
a CLI utility/library for AnimateDiff stable diffusion generation
victordibia/peacasso
UI interface for experimenting with multimodal (text, image) models (stable diffusion).
carefree0910/carefree-drawboard
🎨 Infinite Drawboard in Python
Teriks/dgenerate
dgenerate is a scriptable command line tool (and library) for generating images and animation...