VHellendoorn/Code-LMs
Guide to using pre-trained large language models of source code
Provides pre-trained PolyCoder models (160M–2.7B parameters) trained on 12-language code corpora using GPT-NeoX, available via Hugging Face transformers or custom checkpoints. Supports code generation with configurable temperature and beam search, includes evaluation harnesses for perplexity and HumanEval benchmarks, and offers Docker containerization with GPU support for inference and fine-tuning workflows.
1,842 stars. No commits in the last 6 months.
Stars
1,842
Forks
265
Language
Python
License
MIT
Category
Last pushed
Jul 07, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/VHellendoorn/Code-LMs"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
Goekdeniz-Guelmez/mlx-lm-lora
Train Large Language Models on MLX.
uber-research/PPLM
Plug and Play Language Model implementation. Allows to steer topic and attributes of GPT-2 models.
jarobyte91/pytorch_beam_search
A lightweight implementation of Beam Search for sequence models in PyTorch.
SmallDoges/small-doge
Doge Family of Small Language Models
ssbuild/chatglm_finetuning
chatglm 6b finetuning and alpaca finetuning