Rohan-Thoma/Coding-attention-from-scratch
This repository consists code for executing attention mechanism from scratch for language translation models. This is coded from ground up for translating Italian to English completely without any pretraining.
No commits in the last 6 months.
Stars
2
Forks
—
Language
Jupyter Notebook
License
Apache-2.0
Category
Last pushed
May 04, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/Rohan-Thoma/Coding-attention-from-scratch"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
LowinLi/transformers-stream-generator
This is a text generation method which returns a generator, streaming out each token in...
ystemsrx/mini-nanoGPT
One-click training of your own GPT. Training a GPT has never been easier for beginners. /...
jaymody/picoGPT
An unnecessarily tiny implementation of GPT-2 in NumPy.
kamalkraj/minGPT-TF
A minimal TF2 re-implementation of the OpenAI GPT training
Eamon2009/Codeformer-A.I
A character-level GPT transformer built from scratch in PyTorch, trained on Linux kernel C...