konpatp/diffae
Official implementation of Diffusion Autoencoders
Combines diffusion models with autoencoder architecture to learn interpretable latent representations that support both reconstruction and generative sampling. The framework uses a two-stage approach: first training a diffusion model as an encoder, then optionally training a latent diffusion probabilistic model (DPM) in the learned latent space for generation. Supports semantic image manipulation via learned classifiers, latent interpolation, and unconditional synthesis on face and object datasets (FFHQ, CelebA, LSUN) with pre-trained checkpoints provided.
959 stars. No commits in the last 6 months.
Stars
959
Forks
158
Language
Jupyter Notebook
License
MIT
Category
Last pushed
Sep 12, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/diffusion/konpatp/diffae"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
huggingface/diffusers
🤗 Diffusers: State-of-the-art diffusion models for image, video, and audio generation in PyTorch.
bghira/SimpleTuner
A general fine-tuning kit geared toward image/video/audio diffusion models.
mcmonkeyprojects/SwarmUI
SwarmUI (formerly StableSwarmUI), A Modular Stable Diffusion Web-User-Interface, with an...
nateraw/stable-diffusion-videos
Create 🔥 videos with Stable Diffusion by exploring the latent space and morphing between text prompts
AUTOMATIC1111/stable-diffusion-webui
Stable Diffusion web UI