Text-to-Image Generation Diffusion Models
Tools and implementations for generating images from text prompts using diffusion models, GANs, or CLIP-guided approaches. Does NOT include image editing tools, inpainting, video generation, or evaluation benchmarks.
There are 43 text-to-image generation models tracked. 2 score above 50 (established tier). The highest-rated is NVlabs/Sana at 60/100 with 5,000 stars. 1 of the top 10 are actively maintained.
Get all 43 projects as JSON
curl "https://pt-edge.onrender.com/api/v1/datasets/quality?domain=diffusion&subcategory=text-to-image-generation&limit=20"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
| # | Model | Score | Tier |
|---|---|---|---|
| 1 |
NVlabs/Sana
SANA: Efficient High-Resolution Image Synthesis with Linear Diffusion Transformer |
|
Established |
| 2 |
FoundationVision/VAR
[NeurIPS 2024 Best Paper Award][GPT beats diffusion🔥] [scaling laws in... |
|
Established |
| 3 |
nerdyrodent/VQGAN-CLIP
Just playing with getting VQGAN+CLIP running locally, rather than having to... |
|
Emerging |
| 4 |
huggingface/finetrainers
Scalable and memory-optimized training of diffusion models |
|
Emerging |
| 5 |
AssemblyAI-Community/MinImagen
MinImagen: A minimal implementation of the Imagen text-to-image model |
|
Emerging |
| 6 |
eps696/aphantasia
CLIP + FFT/DWT/RGB = text to image/video |
|
Emerging |
| 7 |
AlonzoLeeeooo/awesome-text-to-image-studies
A collection of awesome text-to-image generation studies. |
|
Emerging |
| 8 |
kyegomez/LUMIERE
Implementation of the text to video model LUMIERE from the paper: "A... |
|
Emerging |
| 9 |
parlance-zz/dualdiffusion
Dual Diffusion is a generative diffusion model for music trained on video... |
|
Emerging |
| 10 |
nerdyrodent/CLIP-Guided-Diffusion
Just playing with getting CLIP Guided Diffusion running locally, rather than... |
|
Emerging |
| 11 |
AIDC-AI/Ovis-Image
Ovis-Image is a 7B text-to-image model specifically optimized for... |
|
Emerging |
| 12 |
WZDTHU/NiT
[NeurIPS 2025] Native-resolution diffusion Transformer |
|
Emerging |
| 13 |
kamalkraj/stable-diffusion-tritonserver
Deploy stable diffusion model with onnx/tenorrt + tritonserver |
|
Emerging |
| 14 |
songweige/rich-text-to-image
Rich-Text-to-Image Generation |
|
Emerging |
| 15 |
mehdidc/feed_forward_vqgan_clip
Feed forward VQGAN-CLIP model, where the goal is to eliminate the need for... |
|
Emerging |
| 16 |
woctezuma/stable-diffusion-safety-checker
Python package to apply the Safety Checker from Stable Diffusion. |
|
Emerging |
| 17 |
rockerBOO/sd-ext
Scripts and extensions for Stable Diffusion |
|
Emerging |
| 18 |
slowy07/luna
text to image generation with stable diffusion |
|
Emerging |
| 19 |
OutofAi/StableFace
Build your own Face App with Stable Diffusion 2.1 |
|
Emerging |
| 20 |
huggingface/instruction-tuned-sd
Code for instruction-tuning Stable Diffusion. |
|
Emerging |
| 21 |
HFAiLab/clip-gen
CLIP-GEN: Language-Free Training of a Text-to-Image Generator with CLIP |
|
Emerging |
| 22 |
DiT-3D/DiT-3D
🔥🔥🔥Official Codebase of "DiT-3D: Exploring Plain Diffusion Transformers for... |
|
Emerging |
| 23 |
amaralibey/nanoCLIP
A lightweight Text-to-Image Retrieval model [Web App] |
|
Emerging |
| 24 |
py-img-gen/python-image-generation
🎨 書籍「Pythonで学ぶ画像生成」のコードを配置したリポジトリです |
|
Emerging |
| 25 |
gmongaras/Stable-Diffusion-3-From-Scratch
A repo that attempts to train stable diffusion 3 from scratch |
|
Experimental |
| 26 |
saharmor/anima
Turn text into video using Stable Diffusion and Google FILM |
|
Experimental |
| 27 |
Qiyuan-Ge/PaintMind
Fast and controllable text-to-image model. |
|
Experimental |
| 28 |
hila-chefer/TargetCLIP
[ECCV 2022] Official PyTorch implementation of the paper Image-Based... |
|
Experimental |
| 29 |
nahyeonkaty/textboost
TextBoost: Towards One-Shot Personalization of Text-to-Image Models via... |
|
Experimental |
| 30 |
ShivamDuggal4/UNITE-tokenization-generation
Single-stage End-to-End Training for Tokenization and Generation |
|
Experimental |
| 31 |
ji-code25/Point-Transformer-Diffusion
Point Transformer Diffusion is a novel generative model for 3D point cloud... |
|
Experimental |
| 32 |
ouhenio/StyleGAN3-CLIP-notebooks
A collection of Jupyter notebooks to play with NVIDIA's StyleGAN3 and... |
|
Experimental |
| 33 |
nhtlongcs/live-novel
Self-host application can generate illustration from a novel by highlighting... |
|
Experimental |
| 34 |
defgsus/clipig
OpenAI CLIP based image generator with complex config file controlled... |
|
Experimental |
| 35 |
SaiBalaji-PSS/Stable-Diffusion-Catalyst
A macOS Catalyst app which uses Apple's CoreML Stable Diffusion package to... |
|
Experimental |
| 36 |
contrebande-labs/charred
CHARacter-awaRE Diffusion: Multilingual Character-Aware Encoders for... |
|
Experimental |
| 37 |
jdh-algo/JoyType
JoyType: A Robust Design for Multilingual Visual Text Creation |
|
Experimental |
| 38 |
monk1337/OpenAI-CLIP-Image-search
OpenAI's CLIP neural network |
|
Experimental |
| 39 |
EngineeringAI-LAB/MIS-DiT-AST
This is a training-free sketch to scene generation. |
|
Experimental |
| 40 |
tripletclip/TripletCLIP
[NeurIPS 2024] Official PyTorch implementation of "Improving Compositional... |
|
Experimental |
| 41 |
johnsutor/emoji-painter
Paint with emojis. |
|
Experimental |
| 42 |
TrieuPhi/Image-Caption
Project sẽ tổng hợp những model liên quan đến image caption, sử dụng các... |
|
Experimental |
| 43 |
linsun449/cliper.code
This repo is the official pytorch implementation of the paper: CLIPer:... |
|
Experimental |