GPT2 Pretraining Fine-tuning Transformer Models
Tools for pretraining, fine-tuning, and implementing GPT-2 models from scratch, including language-specific variants and inference optimization. Does NOT include downstream applications like question-answering or summarization, nor other model architectures beyond GPT-2 variants.
There are 128 gpt2 pretraining fine-tuning models tracked. 3 score above 50 (established tier). The highest-rated is tabularis-ai/be_great at 67/100 with 350 stars and 5,517 monthly downloads.
Get all 128 projects as JSON
curl "https://pt-edge.onrender.com/api/v1/datasets/quality?domain=transformers&subcategory=gpt2-pretraining-fine-tuning&limit=20"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
| # | Model | Score | Tier |
|---|---|---|---|
| 1 |
tabularis-ai/be_great
A novel approach for synthesizing tabular data using pretrained large language models |
|
Established |
| 2 |
EleutherAI/gpt-neox
An implementation of model parallel autoregressive transformers on GPUs,... |
|
Established |
| 3 |
shibing624/textgen
TextGen: Implementation of Text Generation models, include LLaMA, BLOOM,... |
|
Established |
| 4 |
AdityaNG/kan-gpt
The PyTorch implementation of Generative Pre-trained Transformers (GPTs)... |
|
Emerging |
| 5 |
EleutherAI/gpt-neo
An implementation of model parallel GPT-2 and GPT-3-style models using the... |
|
Emerging |
| 6 |
keith2018/TinyGPT
Tiny C++ LLM inference implementation from scratch |
|
Emerging |
| 7 |
zemlyansky/gpt-tfjs
GPT in TensorFlow.js |
|
Emerging |
| 8 |
kyegomez/GPT4o
Community Open Source Implementation of GPT4o in PyTorch |
|
Emerging |
| 9 |
ai-forever/ru-gpts
Russian GPT3 models. |
|
Emerging |
| 10 |
0hq/WebGPT
Run GPT model on the browser with WebGPU. An implementation of GPT inference... |
|
Emerging |
| 11 |
kakaobrain/kogpt
KakaoBrain KoGPT (Korean Generative Pre-trained Transformer) |
|
Emerging |
| 12 |
kyegomez/Lets-Verify-Step-by-Step
"Improving Mathematical Reasoning with Process Supervision" by OPENAI |
|
Emerging |
| 13 |
eric-ai-lab/MiniGPT-5
Official implementation of paper "MiniGPT-5: Interleaved Vision-and-Language... |
|
Emerging |
| 14 |
mytechnotalent/RE-GPT
Inspired by Andrej Karpathy’s "Let’s Build GPT", this project guides you... |
|
Emerging |
| 15 |
cdpierse/script_buddy_v2
Script Buddy v2 is a film script text generation tool built using film... |
|
Emerging |
| 16 |
turtlesoupy/this-word-does-not-exist
This Word Does Not Exist |
|
Emerging |
| 17 |
ai-forever/mgpt
Multilingual Generative Pretrained Model |
|
Emerging |
| 18 |
hyperonym/basaran
Basaran is an open-source alternative to the OpenAI text completion API. It... |
|
Emerging |
| 19 |
kyegomez/CNNGPT
This CNN-based language model leverages causal and dilated convolutions,... |
|
Emerging |
| 20 |
datadreamer-dev/DataDreamer
DataDreamer: Prompt. Generate Synthetic Data. Train & Align Models. 🤖💤 |
|
Emerging |
| 21 |
erogol/BlaGPT
Experimental playground for benchmarking language model (LM) architectures,... |
|
Emerging |
| 22 |
sytelus/nanuGPT
Simple, reliable and well tested training code for quick experiments with... |
|
Emerging |
| 23 |
saqib1707/gpt2-from-scratch
PyTorch Implementation of GPT-2 |
|
Emerging |
| 24 |
soumyadip1995/BabyGPT
Something in the middle of Karpathy's mingpt model and video lectures, ... |
|
Emerging |
| 25 |
mytechnotalent/falcongpt
Simple GPT app that uses the falcon-7b-instruct model with a Flask front-end. |
|
Emerging |
| 26 |
TatevKaren/BabyGPT-Build_GPT_From_Scratch
BabyGPT: Build Your Own GPT Large Language Model from Scratch Pre-Training... |
|
Emerging |
| 27 |
wpeebles/G.pt
Official PyTorch Implementation of "Learning to Learn with Generative Models... |
|
Emerging |
| 28 |
arrmansa/Basic-UI-for-GPT-J-6B-with-low-vram
A repository to run gpt-j-6b on low vram machines (4.2 gb minimum vram for... |
|
Emerging |
| 29 |
EagleW/Stage-wise-Fine-tuning
Code for Stage-wise Fine-tuning for Graph-to-Text Generation |
|
Emerging |
| 30 |
iVishalr/GPT
A minimal and efficient Pytorch implementation of OpenAI's GPT (Generative... |
|
Emerging |
| 31 |
potamides/uniformers
Token-free Language Modeling with ByGPT5 & Friends! |
|
Emerging |
| 32 |
readme-generator/alreadyme-ai-research
Generate README.md with GPT-3 few-shot learning |
|
Emerging |
| 33 |
arrmansa/Gpt-Neo-Limited-Vram-Cuda
A notebook that runs GPT-Neo with low vram (6 gb) and cuda acceleration by... |
|
Emerging |
| 34 |
losttech/Torch.MinGPT
A C# implementation of GPT |
|
Emerging |
| 35 |
dreamingjudith/KoGPT2-personachat
Fine-tuned KoGPT2 chatbot demo with translated PersonaChat (ongoing) |
|
Experimental |
| 36 |
arrmansa/Basic-UI-for-GPT-Neo-with-low-vram
A basic ui for running gpt neo 2.7B on low vram (3 gb Vram minimum) |
|
Experimental |
| 37 |
FareedKhan-dev/gpt4o-from-scratch
Implementation of a GPT-4o like Multimodal from Scratch using Python |
|
Experimental |
| 38 |
EvilFreelancer/rugpt3-custom
Pre-training custom ruGPT3 model on books written by F.M. Dostoevski |
|
Experimental |
| 39 |
Agora-Lab-AI/OmniByteGPT
An implementation of an all-new foundation model architecture that trains on... |
|
Experimental |
| 40 |
jseeio/gpt2-tfjs
GPT2 with Tensorflow.js |
|
Experimental |
| 41 |
tanulsingh/Humour.ai-Language-model-that-can-crack-Jokes
Language Model that makes you Laugh . |
|
Experimental |
| 42 |
StarxSky/ANE-GPT-New
New ANE GPT |
|
Experimental |
| 43 |
EdvardOlsen/Horoscope_generator
This is a horoscope generating code |
|
Experimental |
| 44 |
s-macke/GoPT
GPT-2 Model Inference |
|
Experimental |
| 45 |
ant-louis/belgpt2
🇧🇪 BelGPT-2: the 1st GPT model pretrained in French. |
|
Experimental |
| 46 |
JarvisPei/FuseGPT
The implementation for the paper, FuseGPT: Learnable Layers Fusion of... |
|
Experimental |
| 47 |
Any-Winter-4079/Nano-GPT-Speedrun-Track
This repo represents my Nano-GPT speedrun playground, which started coding... |
|
Experimental |
| 48 |
trekhleb/homemade-gpt-js
A minimal TensorFlow.js re-implementation of Karpathy's minGPT (Generative... |
|
Experimental |
| 49 |
fcakyon/gpt2-shakespeare
A tutorial on GPT2 language model training with texts from Shakespeare |
|
Experimental |
| 50 |
mrseanryan/gpt-local
Local GPT (llama 2 or dolly or gpt etc.) via Python - using ctransforers project |
|
Experimental |
| 51 |
neuronalin/gpt-from-scratch-pytorch
A decoder-only GPT-style Transformer built from scratch with PyTorch —... |
|
Experimental |
| 52 |
JAVO932/PyGPT2
🖥️ Explore GPT-2 text generation with PyGPT2, a user-friendly Python app... |
|
Experimental |
| 53 |
uSaiPrashanth/gpt-j-finetune
Parallelizes finetuning of gpt-j on P3 dataset across multiple gpu nodes |
|
Experimental |
| 54 |
Atenrev/forocoches-language-generation
This is a PyTorch implementation of a decoder only transformer inspired on... |
|
Experimental |
| 55 |
procesaur/Scratch2LM
Training transformer models (e.g. RoBERTa, GPT2 and GPT-J) from scratch. |
|
Experimental |
| 56 |
mytechnotalent/MicroGPT
MicroGPT is a clean, educational implementation of the GPT (Generative... |
|
Experimental |
| 57 |
chizkidd/microGPT
Minimal char-level GPT inspired by @karpathy's microGPT: multi-dataset... |
|
Experimental |
| 58 |
lorenzomaiuri-dev/quantum-gpt
A hybrid Quantum-Classical Transformer implementation based on nanoGPT,... |
|
Experimental |
| 59 |
kabachuha/nanoGPKANT
Testing KAN-based text generation GPT models |
|
Experimental |
| 60 |
pablo-reyes8/implementing-gpt
Clean-room GPT-2/GPT-3 implementation: tokenizers, architecture blocks,... |
|
Experimental |
| 61 |
Navy10021/KRLawGPT
KRLawGPT : Generative Pre-trained Transformer for producing Korean Legal Text |
|
Experimental |
| 62 |
aarxshi/DsaGPT
A minimal GPT-style transformer built from scratch for DSA-style Q&A |
|
Experimental |
| 63 |
pronzzz/atomgpt
AtomGPT is a chaotic, evolutionary implementation of a Generative... |
|
Experimental |
| 64 |
s-omranpour/Shirin-Sokhan
A Persian Poet Transformer! (finetuned GPT2 on Ganjoor data) |
|
Experimental |
| 65 |
Andras7/gpt2-pytorch
Extremely simple and understandable GPT2 implementation with minor tweaks |
|
Experimental |
| 66 |
codiceSpaghetti/numpyGPT
A from-scratch GPT built with NumPy and Python’s standard library. No... |
|
Experimental |
| 67 |
NJX-njx/microgpt
🔬 The most atomic GPT-2 implementation in 265 lines of pure Python & CUDA. A... |
|
Experimental |
| 68 |
SIC98/GPT2-python-code-generator
GPT2 finetuning with transformers 🤗 |
|
Experimental |
| 69 |
Eden-Eldith/WiggleGPT
WiggleGPT is an language model that integrates bio-inspired neural... |
|
Experimental |
| 70 |
Vadimbuildercxx/NumpyGPT
A lightweight educational implementation of GPT (Generative Pre-trained... |
|
Experimental |
| 71 |
Amir-Hofo/GPT2
Implementation of the GPT-2 architecture using PyTorch, trained on the... |
|
Experimental |
| 72 |
RahulSChand/gpt2_squad
GPT2 training on squad dataset |
|
Experimental |
| 73 |
kyegomez/TinyGPTV
Simple Implementation of TinyGPTV in super simple Zeta lego blocks |
|
Experimental |
| 74 |
fattorib/Little-GPT
GPT* - Training faster small transformers using ALiBi, Parallel Residual... |
|
Experimental |
| 75 |
god01215/GPT-From-Scratch
Implementation of a GPT-style LLM from scratch, following "Build a Large ... |
|
Experimental |
| 76 |
lin826/nanoGPT-demo
Training and finetuning local GPTs. |
|
Experimental |
| 77 |
inkybubble/mi_01_attention_patterns_scratch
MI-01 - Attention Patterns from Scratch: Finding Previous-Token and... |
|
Experimental |
| 78 |
Alibubere/scene2story
Scene2Story is an AI-powered system that generates creative stories from... |
|
Experimental |
| 79 |
buhsnn/eli5-gpt2-language-model
Decoder-only Transformer (GPT-2 style) trained from scratch on the ELI5... |
|
Experimental |
| 80 |
jaketae/lm-identifier
A toolkit for identifying pretrained language models from potentially... |
|
Experimental |
| 81 |
eshaaaan/tinygpt
🤖 Simplify understanding of large language models with TinyGPT, featuring a... |
|
Experimental |
| 82 |
Ojas025/almostGPT
A GPT implementation for training and generating text on custom datasets |
|
Experimental |
| 83 |
chandan11248/GPT-2
Learning and implementing GPT-2 from scratch, including architecture... |
|
Experimental |
| 84 |
Sumeet8726/Hyper_rw
🛠️ Access virtual memory with the GuestMemory class and utility functions... |
|
Experimental |
| 85 |
diixo/build-gpt
A PyTorch library with educational re-implementation of GPT-models: GPT2, LLaMA |
|
Experimental |
| 86 |
brendandagys/ChadGPT
From-scratch GPT experiments in PyTorch, covering attention mechanisms,... |
|
Experimental |
| 87 |
alkatrazstudio/neodim-server
Natural language model AI via HTTP |
|
Experimental |
| 88 |
marlo-z/reversal_curse_analysis
Code for 'Towards a Theoretical Understanding of the 'Reversal Curse' via... |
|
Experimental |
| 89 |
Sairamg18814/GUBBALA-V3-TRUE
Revolutionary Self-Evolving Language Model - 100% self-contained AI trained... |
|
Experimental |
| 90 |
jiseokson/PageBrain
Light-weight LLM Serving with PagedAttention |
|
Experimental |
| 91 |
zTgx/DeepText
A GPT Model To Generate Text |
|
Experimental |
| 92 |
iangitonga/gten
A minimal library to run transformer neural networks on CPU. |
|
Experimental |
| 93 |
shreydan/shakespeareGPT
understanding language modeling by training a small GPT on Shakespeare plays. |
|
Experimental |
| 94 |
alperiox/bookbot
A toy project for my generative AI studies on text data. Train generative... |
|
Experimental |
| 95 |
Divyansh900/PyCodeGen
A python code generation model with 75M parameter built from the ground up... |
|
Experimental |
| 96 |
bellthomas/gpt.local
A work-in-progress, from-scratch implementation of a generative pre-trained... |
|
Experimental |
| 97 |
PromptlyCode/inline-completion-model
PromptlyCode inline completion model by PyTorch |
|
Experimental |
| 98 |
Akhan521/GPT-From-Scratch
🧸 A fully custom GPT-style language model built from scratch using PyTorch... |
|
Experimental |
| 99 |
dedsecurity/dpt
Repo for offsite scale work |
|
Experimental |
| 100 |
Agora-Lab-AI/NeoCore
NeoCore™ - Next Generation CPU-Native Transformer. |
|
Experimental |
| 101 |
neemiasbsilva/developing-nanoGPT2-fineweb
Developing a cusstom nano GPT-2 from scratch using PyTorch on the Fineweb dataset. |
|
Experimental |
| 102 |
sumony2j/SeedGPT-22M
SeedGPT is a lightweight, 22M-parameter Transformer LLM for efficient text... |
|
Experimental |
| 103 |
tahmidmir/Graph-RAG
Fine-tuning GPT-2 on domain-specific articles related to skin cancer, using... |
|
Experimental |
| 104 |
jjantas/neural-networks-zero-to-hero
This repository is a personal, in-depth reworking of Andrej Karpathy's... |
|
Experimental |
| 105 |
ademyanchuk/gpt2-diy
From-scratch reproduction of GPT-2 following Andrej Karpathy's "Zero to Hero" series. |
|
Experimental |
| 106 |
n9e6y/PPG
Persian Poetry Generator: A fine-tuned GPT-2 model for generating Persian... |
|
Experimental |
| 107 |
btboilerplate/GPT-2
Fine-tunes the GPT-2 language model on Shakespearean text to generate... |
|
Experimental |
| 108 |
oskarfernlund/noskGPT
Simple transformer-based language model which generates Shakespearian dialogue. |
|
Experimental |
| 109 |
TomaszKaleczyc/scifi_book_generator
The purpose of this project is to build a decoder only transformer... |
|
Experimental |
| 110 |
ayus1234/Text-Generation-with-GPT-2
A comprehensive toolkit for fine-tuning GPT-2 language models and generating... |
|
Experimental |
| 111 |
taljindergill78/AI-Indian-Recipe-Generator
AI-powered system that generates authentic Indian recipes using GPT-2 and... |
|
Experimental |
| 112 |
BenBenyamin/GPT2
My implementation GPT2 from scratch using the original GPT2 and GPT3 papers |
|
Experimental |
| 113 |
Wojtekb30/GPT-2-B200-pre-trainier
Code for pre-training a GPT-2 model on (eight) NVIDIA DGX B200 GPUs and... |
|
Experimental |
| 114 |
thewh1teagle/g2p-byt5
g2p with byt5 |
|
Experimental |
| 115 |
3ConstArt3/AIQuoteGenerator
Fine-tune GPT-2 models on philosophers’ quotes with semantic tagging,... |
|
Experimental |
| 116 |
NimeshRawanage/TextGen-AI-Transformer-Trainer
Create your own text-generation AI (GPT). Fine-tune GPT-2 or T5 on your own... |
|
Experimental |
| 117 |
vishxl/WisdumbAI
[WIP] WisdumbAI: Generate thoughts/tweets using GPT-Neo. |
|
Experimental |
| 118 |
jndiogo/gptbench
A python package to experiment with GPT-like transformer models |
|
Experimental |
| 119 |
Adam-Bowen/nanoGPT
🧠 nanoGPT (Andrej Karpathy's Zero to Hero) |
|
Experimental |
| 120 |
sartq333/story-GPT
a simple GPT model pre-trained from scratch on tiny stories dataset |
|
Experimental |
| 121 |
J3lly-Been/gpt2-story-generation
This project fine-tunes GPT-2, a popular pre-trained transformer model, to... |
|
Experimental |
| 122 |
Harsh-2909/gpt-from-scratch
The "GPT from Scratch" project is an endeavor to implement the Generative... |
|
Experimental |
| 123 |
baumandm/lorem-insight
Tool to generate lorem ipsum-style Insights for Insights Explorer |
|
Experimental |
| 124 |
YashrajBaila7/GPT2LM
A implimentation of GPT2 varient. |
|
Experimental |
| 125 |
juletx/gpt2-eus
Pretraining GPT2 model on Basque language |
|
Experimental |
| 126 |
SynthWomb/Synthia
SynthiaGPT leverages Google's Gemini & the Hugging Face Transformers library... |
|
Experimental |
| 127 |
Uni-Creator/NanoGPT
NanoGPT is a lightweight GPT-style language model designed for text... |
|
Experimental |
| 128 |
Med-Karim-Ben-Boubaker/gpt-2-from-scratch
A repository that shows the code behind different LLMs architectures and... |
|
Experimental |