eduardoleao052/Transformer-from-scratch
Educational Transformer from scratch (no autograd), with forward and backprop.
No commits in the last 6 months.
Stars
36
Forks
—
Language
Python
License
MIT
Category
Last pushed
Apr 10, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/eduardoleao052/Transformer-from-scratch"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
huggingface/transformers-bloom-inference
Fast Inference Solutions for BLOOM
Tencent/TurboTransformers
a fast and user-friendly runtime for transformer inference (Bert, Albert, GPT2, Decoders, etc)...
Eamon2009/Transformer-language-model
An educational implementation of a GPT-style language model built from scratch using PyTorch . ...
mit-han-lab/lite-transformer
[ICLR 2020] Lite Transformer with Long-Short Range Attention
mit-han-lab/hardware-aware-transformers
[ACL'20] HAT: Hardware-Aware Transformers for Efficient Natural Language Processing