microsoft/unilm

Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities

57
/ 100
Established

Built on foundation architectures like DeepNet (1,000+ layer Transformers) and Magneto (general-purpose multimodal modeling), the project implements unified pre-training across diverse modalities including vision (BEiT, DiT), speech (WavLM, VALL-E), and document understanding (LayoutLM series). It emphasizes training stability and efficiency through techniques like sparse Mixture-of-Experts (X-MoE) and length extrapolation, while supporting 100+ languages via models like InfoXLM and DeltaLM for cross-lingual transfer and machine translation.

22,042 stars.

No Package No Dependents
Maintenance 10 / 25
Adoption 10 / 25
Maturity 16 / 25
Community 21 / 25

How are scores calculated?

Stars

22,042

Forks

2,692

Language

Python

License

MIT

Last pushed

Jan 23, 2026

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/microsoft/unilm"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.