Mahesh3394/training-of-transformer-on-dummy-data
Here we try to understand how transformer works and try to replicate architecture from paper published. Also we will train simple architecture on dummy dataset.
No commits in the last 6 months.
Stars
1
Forks
—
Language
Jupyter Notebook
License
Apache-2.0
Category
Last pushed
Nov 24, 2023
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/Mahesh3394/training-of-transformer-on-dummy-data"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
Tencent/PatrickStar
PatrickStar enables Larger, Faster, Greener Pretrained Models for NLP and democratizes AI for everyone.
OpenMotionLab/MotionGPT
[NeurIPS 2023] MotionGPT: Human Motion as a Foreign Language, a unified motion-language...
OpenMotionLab/MotionGPT3
MotionGPT3: Human Motion as a Second Modality, a MoT-based framework for unified motion...
YerbaPage/DetectCodeGPT
Detection of LLM-Generated Codes [ICSE2025]
RobbenRibery/TuoTuo
TuoTuo is a Topic Modeling library for Researchers and Engineers