Korean Language Models
Pretrained transformer models specifically designed for Korean language processing, including BERT, ELECTRA, and specialized variants. Does NOT include general multilingual models, non-Korean language models, or downstream task-specific applications (unless they primarily showcase the Korean model architecture itself).
There are 33 korean language models tracked. The highest-rated is SKTBrain/KoBERT at 46/100 with 1,407 stars.
Get all 33 projects as JSON
curl "https://pt-edge.onrender.com/api/v1/datasets/quality?domain=transformers&subcategory=korean-language-models&limit=20"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
| # | Model | Score | Tier |
|---|---|---|---|
| 1 |
SKTBrain/KoBERT
Korean BERT pre-trained cased (KoBERT) |
|
Emerging |
| 2 |
monologg/KoELECTRA
Pretrained ELECTRA Model for Korean |
|
Emerging |
| 3 |
monologg/KoBERT-Transformers
KoBERT on ๐ค Huggingface Transformers ๐ค (with Bug Fixed) |
|
Emerging |
| 4 |
VinAIResearch/PhoBERT
PhoBERT: Pre-trained language models for Vietnamese (EMNLP-2020 Findings) |
|
Emerging |
| 5 |
KB-AI-Research/KB-ALBERT
KB๊ตญ๋ฏผ์ํ์์ ์ ๊ณตํ๋ ๊ฒฝ์ /๊ธ์ต ๋๋ฉ์ธ์ ํนํ๋ ํ๊ตญ์ด ALBERT ๋ชจ๋ธ |
|
Emerging |
| 6 |
monologg/KoBERT-KorQuAD
Korean MRC (KorQuAD) with KoBERT |
|
Emerging |
| 7 |
ymcui/MacBERT
Revisiting Pre-trained Models for Chinese Natural Language Processing (MacBERT) |
|
Emerging |
| 8 |
monologg/DistilKoBERT
Distillation of KoBERT from SKTBrain (Lightweight KoBERT) |
|
Emerging |
| 9 |
Beomi/KcELECTRA
๐ค Korean Comments ELECTRA: ํ๊ตญ์ด ๋๊ธ๋ก ํ์ตํ ELECTRA ๋ชจ๋ธ |
|
Emerging |
| 10 |
thevasudevgupta/bigbird
Google's BigBird (Jax/Flax & PyTorch) @ ๐คTransformers |
|
Emerging |
| 11 |
monologg/korean-hate-speech-koelectra
Bias, Hate classification with KoELECTRA ๐ฟ |
|
Emerging |
| 12 |
monologg/KoBigBird
๐ฆ Pretrained BigBird Model for Korean (up to 4096 tokens) |
|
Emerging |
| 13 |
monologg/KoCharELECTRA
Character-level Korean ELECTRA Model (์์ ๋จ์ ํ๊ตญ์ด ELECTRA) |
|
Emerging |
| 14 |
toriving/text-classification-transformers
Easy text classification for everyone : Bert based models via Huggingface... |
|
Emerging |
| 15 |
monologg/KoELECTRA-Pipeline
Transformers Pipeline with KoELECTRA |
|
Emerging |
| 16 |
monologg/HanBert-Transformers
HanBert on ๐ค Huggingface Transformers ๐ค |
|
Experimental |
| 17 |
bayartsogt-ya/albert-mongolian
ALBERT trained on Mongolian text corpus |
|
Experimental |
| 18 |
sajjjadayobi/ParsBigBird
Persian Bert For Long-Range Sequences |
|
Experimental |
| 19 |
Anshler/vietnamese-poem-classifier
Classify genre and score Vietnamese poems ๐๐ |
|
Experimental |
| 20 |
SciCrunch/bio_electra
Bio-Electra - Small and efficient discriminatively pre-trained language... |
|
Experimental |
| 21 |
oneonlee/KoAirBERT
๐ค ํญ๊ณต ์์ ๋๋ฉ์ธ์ ํนํ๋ ํ๊ตญ์ด BERT ๋ชจ๋ธ โ๏ธ |
|
Experimental |
| 22 |
svn05/vietnamese-sentiment-phobert
a fine tuned PhoBERT model to classify product reviews across a range of... |
|
Experimental |
| 23 |
Nikki-oo7/pos-tagger
Part-of-Speech Tagger implemented in PyTorch using BiLSTM and Transformer models. |
|
Experimental |
| 24 |
yejoon-lee/kr3
KR3: Korean Restaurant Review with Ratings / Experiments on... |
|
Experimental |
| 25 |
qanastek/French-Part-Of-Speech-Tagging
Repository for the source code of the HuggingFace Space named... |
|
Experimental |
| 26 |
codegram/calbert
Catalan ALBERT (A Lite BERT for self-supervised learning of language representations) |
|
Experimental |
| 27 |
Pirata-Codex/Tag-Persian-Entities-Using-Bert
Using the fa-bert model to tag persian entities in a sentence |
|
Experimental |
| 28 |
edoost/pert
Persian Ezafe Recognition Using Transformers and Its Role in Part-Of-Speech Tagging |
|
Experimental |
| 29 |
qanastek/ANTILLES
ANTILLES : An Open French Linguistically Enriched Part-of-Speech Corpus |
|
Experimental |
| 30 |
HRSadeghi/Joint_Comma_and_Kasreh_Recognizer
In this repository, we provide a joint neural model based on BERT and two... |
|
Experimental |
| 31 |
ilos-vigil/bigbird-small-indonesian
Lighweight Indonesian language model for long sequence. |
|
Experimental |
| 32 |
phanxuanphucnd/CoBERTa
CoBERTa is a pre-trained models are the pre-trained language models for... |
|
Experimental |
| 33 |
amanaser/BabyLM-ELECTRA-Pre-training
BabyLM ELECTRA Pre-training on NVIDIA L40 GPU Cluster. |
|
Experimental |