dheeren-tejani/mini-lm-124m
Experimental GPT-2 scale (~124M param) LLM trained from scratch on Google Colab. Trained on C4, Cosmopedia/Alpaca/Python mix. Includes full training pipeline, mixed dataset loader with Colab-resilient checkpointing, and log analysis tools. Honest write-up of what went wrong.
Stars
—
Forks
—
Language
TypeScript
License
—
Category
Last pushed
Feb 19, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/nlp/dheeren-tejani/mini-lm-124m"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
graykode/gpt-2-Pytorch
Simple Text-Generator with OpenAI gpt-2 Pytorch Implementation
Morizeyao/GPT2-Chinese
Chinese version of GPT2 training code, using BERT tokenizer.
imcaspar/gpt2-ml
GPT2 for Multiple Languages, including pretrained models. GPT2 多语言支持, 15亿参数中文预训练模型
gyunggyung/KoGPT2-FineTuning
🔥 Korean GPT-2, KoGPT2 FineTuning cased. 한국어 가사 데이터 학습 🔥
liucongg/GPT2-NewsTitle
Chinese NewsTitle Generation Project by GPT2.带有超级详细注释的中文GPT2新闻标题生成项目。