Shark-NLP/DiffuSeq
[ICLR'23] DiffuSeq: Sequence to Sequence Text Generation with Diffusion Models
Implements classifier-free conditional diffusion for seq2seq tasks using PyTorch and HuggingFace transformers, with end-to-end training on dialogue, question generation, text simplification, and paraphrase datasets. DiffuSeq-v2 bridges discrete and continuous text spaces via learned soft absorbing states and discrete noise, achieving 4x faster convergence and 800x faster sampling via customized DPM-Solver++ integration. Supports distributed training across multiple GPUs with loss-aware schedule sampling and minimum Bayes risk decoding for quality/diversity trade-offs.
831 stars. No commits in the last 6 months.
Stars
831
Forks
109
Language
Python
License
MIT
Category
Last pushed
Mar 01, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/diffusion/Shark-NLP/DiffuSeq"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
FlorianFuerrutter/genQC
Generative Quantum Circuits
horseee/DeepCache
[CVPR 2024] DeepCache: Accelerating Diffusion Models for Free
Gen-Verse/MMaDA
MMaDA - Open-Sourced Multimodal Large Diffusion Language Models (dLLMs with block diffusion,...
kuleshov-group/mdlm
[NeurIPS 2024] Simple and Effective Masked Diffusion Language Model
ali-vilab/TeaCache
Timestep Embedding Tells: It's Time to Cache for Video Diffusion Model