nerdyrodent/CLIP-Guided-Diffusion
Just playing with getting CLIP Guided Diffusion running locally, rather than having to use colab.
Combines OpenAI's CLIP vision-language model with guided diffusion to generate images from text prompts at 256x256 or 512x512 resolution. Uses unconditional diffusion models iteratively refined by CLIP embeddings, supporting weighted multi-prompt inputs, image-to-image generation, and fine-grained control via guidance scales for prompt adherence, smoothness, and color range. Includes optional Real-ESRGAN upscaling and video generation of the diffusion process.
385 stars. No commits in the last 6 months.
Stars
385
Forks
48
Language
Python
License
—
Category
Last pushed
Aug 29, 2022
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/diffusion/nerdyrodent/CLIP-Guided-Diffusion"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
NVlabs/Sana
SANA: Efficient High-Resolution Image Synthesis with Linear Diffusion Transformer
FoundationVision/VAR
[NeurIPS 2024 Best Paper Award][GPT beats diffusion🔥] [scaling laws in visual generation📈]...
nerdyrodent/VQGAN-CLIP
Just playing with getting VQGAN+CLIP running locally, rather than having to use colab.
huggingface/finetrainers
Scalable and memory-optimized training of diffusion models
eps696/aphantasia
CLIP + FFT/DWT/RGB = text to image/video