Thisen-Ekanayake/HelaBERT
A compact BERT (6-layer) masked language model trained from scratch on a large, cleaned Sinhala corpus. Includes preprocessing utilities, tokenization scripts, training notebooks, checkpoints, and a HuggingFace-compatible final model for Sinhala NLP tasks.
Stars
—
Forks
—
Language
Jupyter Notebook
License
MIT
Category
Last pushed
Mar 10, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/Thisen-Ekanayake/HelaBERT"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
Tongjilibo/bert4torch
An elegent pytorch implement of transformers
nyu-mll/jiant
jiant is an nlp toolkit
lonePatient/TorchBlocks
A PyTorch-based toolkit for natural language processing
grammarly/gector
Official implementation of the papers "GECToR – Grammatical Error Correction: Tag, Not Rewrite"...
monologg/JointBERT
Pytorch implementation of JointBERT: "BERT for Joint Intent Classification and Slot Filling"