huangjia2019/llm-gpt
From classic NLP to modern LLMs: building language models step by step. 异步图书:《 GPT图解 大模型是怎样构建的》- 这套代码是AI Coder出现之前,自己用纯手工搭建的一套简单有效的NLP经典算法集合。在大语言模型推动的AI Coder兴起之后,很少有机会再创作这么有“手工风”的代码了,不知道这是值得开心还是值得遗憾的事情。
Implements foundational NLP algorithms and transformer architecture components from scratch, including tokenization, embeddings, attention mechanisms, and decoding strategies, designed as hands-coded educational implementations rather than production frameworks. Structured as a companion to the "GPT图解" textbook and video course, progressing through classical NLP techniques toward modern large language model construction with emphasis on understanding core principles through direct implementation.
194 stars. No commits in the last 6 months.
Stars
194
Forks
57
Language
Jupyter Notebook
License
—
Category
Last pushed
Mar 26, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/huangjia2019/llm-gpt"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Compare
Higher-rated alternatives
Lightning-AI/litgpt
20+ high-performance LLMs with recipes to pretrain, finetune and deploy at scale.
liangyuwang/Tiny-DeepSpeed
Tiny-DeepSpeed, a minimalistic re-implementation of the DeepSpeed library
microsoft/Text2Grad
🚀 Text2Grad: Converting natural language feedback into gradient signals for precise model...
catherinesyeh/attention-viz
Visualizing query-key interactions in language + vision transformers (VIS 2023)
FareedKhan-dev/Building-llama3-from-scratch
LLaMA 3 is one of the most promising open-source model after Mistral, we will recreate it's...