Shark-NLP/DiffuSeq

[ICLR'23] DiffuSeq: Sequence to Sequence Text Generation with Diffusion Models

47
/ 100
Emerging

Implements classifier-free conditional diffusion for seq2seq tasks using PyTorch and HuggingFace transformers, with end-to-end training on dialogue, question generation, text simplification, and paraphrase datasets. DiffuSeq-v2 bridges discrete and continuous text spaces via learned soft absorbing states and discrete noise, achieving 4x faster convergence and 800x faster sampling via customized DPM-Solver++ integration. Supports distributed training across multiple GPUs with loss-aware schedule sampling and minimum Bayes risk decoding for quality/diversity trade-offs.

831 stars. No commits in the last 6 months.

Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 10 / 25
Maturity 16 / 25
Community 21 / 25

How are scores calculated?

Stars

831

Forks

109

Language

Python

License

MIT

Last pushed

Mar 01, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/diffusion/Shark-NLP/DiffuSeq"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.