ranpy13/Learning-LLM
Learning to build LLM from scratch, following rasbt/LLMs-from-scratch footsteps.
This project provides the code and guidance to build your own large language models (LLMs) from the ground up. You'll put in foundational code and training data, and get out a custom, functional LLM capable of generating text and performing specific tasks. This is ideal for machine learning engineers, AI researchers, or data scientists looking to deepen their understanding of LLM architecture and training.
No commits in the last 6 months.
Use this if you are a machine learning practitioner who wants to learn the inner workings of large language models by implementing them yourself.
Not ideal if you are looking for a pre-built LLM to use out-of-the-box for applications without needing to understand its construction.
Stars
8
Forks
1
Language
Jupyter Notebook
License
—
Category
Last pushed
Sep 29, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/ranpy13/Learning-LLM"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
AI-Hypercomputer/maxtext
A simple, performant and scalable Jax LLM!
rasbt/reasoning-from-scratch
Implement a reasoning LLM in PyTorch from scratch, step by step
mindspore-lab/mindnlp
MindSpore + 🤗Huggingface: Run any Transformers/Diffusers model on MindSpore with seamless...
mosaicml/llm-foundry
LLM training code for Databricks foundation models
rickiepark/llm-from-scratch
<밑바닥부터 만들면서 공부하는 LLM>(길벗, 2025)의 코드 저장소