VinAIResearch/PhoGPT
PhoGPT: Generative Pre-training for Vietnamese (2023)
Built on a 3.7B-parameter decoder architecture trained on 102B Vietnamese tokens with 8K context length, PhoGPT supports inference across vLLM, Text Generation Inference, llama.cpp, and standard Transformers pipelines, with quantization options via bitsandbytes and GGUF formats. The chat variant incorporates instruction-following and conversation tuning on 70K prompt-response pairs plus 290K multi-turn dialogues. Full fine-tuning is facilitated through llm-foundry or alternative frameworks like LLaMA-Factory and lit-gpt.
798 stars. No commits in the last 6 months.
Stars
798
Forks
74
Language
Python
License
BSD-3-Clause
Category
Last pushed
Nov 12, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/VinAIResearch/PhoGPT"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
Nixtla/nixtla
TimeGPT-1: production ready pre-trained Time Series Foundation Model for forecasting and...
akanyaani/gpt-2-tensorflow2.0
OpenAI GPT2 pre-training and sequence prediction implementation in Tensorflow 2.0
andrewdalpino/NoPE-GPT
A GPT-style small language model (SLM) with no positional embeddings (NoPE).
teddykoker/image-gpt
PyTorch Implementation of OpenAI's Image GPT
LIYUESEN/druggpt
DrugGPT: A GPT-based Strategy for Designing Potential Ligands Targeting Specific Proteins