abhaybodhe/Generate-Code
This repository contains code for fine-tuning a GPT-2 model to solve Python code exercises. It uses a dataset of coding problems and solutions and includes two main scripts: fine_tune_gpt2.py for model training and test_gpt2.py for testing the fine-tuned model.
No commits in the last 6 months.
Stars
—
Forks
—
Language
Python
License
MIT
Category
Last pushed
Sep 19, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/abhaybodhe/Generate-Code"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
Tencent/PatrickStar
PatrickStar enables Larger, Faster, Greener Pretrained Models for NLP and democratizes AI for everyone.
OpenMotionLab/MotionGPT
[NeurIPS 2023] MotionGPT: Human Motion as a Foreign Language, a unified motion-language...
wenlu-lab/cMolGPT
GPT (Generative Pre-trained Transformer) for de novo molecular design by enforcing specified targets
YerbaPage/DetectCodeGPT
Detection of LLM-Generated Codes [ICSE2025]
OpenMotionLab/MotionGPT3
MotionGPT3: Human Motion as a Second Modality, a MoT-based framework for unified motion...