MyDarapy/gpt-1-from-scratch
Rewriting and pretraining GPT-1 from scratch. Implementing Multihead Attention (MHA) in pyTorch from the original paper Improving Language Understanding by Generative Pre-Training (https://cdn.openai.com/research-covers/language-unsupervised/language_understanding_paper.pdf)
No commits in the last 6 months.
Stars
7
Forks
—
Language
Python
License
MIT
Category
Last pushed
Jan 13, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/MyDarapy/gpt-1-from-scratch"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
lvapeab/nmt-keras
Neural Machine Translation with Keras
dair-ai/Transformers-Recipe
🧠A study guide to learn about Transformers
jaketae/ensemble-transformers
Ensembling Hugging Face transformers made easy
lof310/transformer
PyTorch implementation of the current SOTA Transformer. Configurable, efficient, and...
SirawitC/Transformer_from_scratch_pytorch
Build a transformer model from scratch using pytorch to understand its inner workings and gain...