manya-imran/Generative-Pre-trained-Transformers
This publication is a project that covers the use of transformers and prompt engineering for NLP Tasks. This project is done in collaboration with Packt. Please use the link to access the project.
No commits in the last 6 months.
Stars
2
Forks
—
Language
—
License
—
Category
Last pushed
May 03, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/manya-imran/Generative-Pre-trained-Transformers"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
tabularis-ai/be_great
A novel approach for synthesizing tabular data using pretrained large language models
EleutherAI/gpt-neox
An implementation of model parallel autoregressive transformers on GPUs, based on the Megatron...
shibing624/textgen
TextGen: Implementation of Text Generation models, include LLaMA, BLOOM, GPT2, BART, T5, SongNet...
ai-forever/ru-gpts
Russian GPT3 models.
AdityaNG/kan-gpt
The PyTorch implementation of Generative Pre-trained Transformers (GPTs) using Kolmogorov-Arnold...