brightmart/bert_language_understanding

Pre-training of Deep Bidirectional Transformers for Language Understanding: pre-train TextCNN

43
/ 100
Emerging

Implements masked language model pre-training with TextCNN as an alternative backbone to Transformers, demonstrating that pre-training objectives are architecture-agnostic. Built in TensorFlow, it supports both single-label and multi-label text classification with configurable model sizes (big/small/tiny) via d_model and attention head parameters. Pre-training on raw unlabeled data yields significant performance gains—achieving 0.75 F1 after 7 fine-tuning epochs versus 0.44 after 35 epochs without pre-training on mid-sized datasets.

967 stars. No commits in the last 6 months.

No License Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 10 / 25
Maturity 8 / 25
Community 25 / 25

How are scores calculated?

Stars

967

Forks

211

Language

Python

License

Last pushed

Jan 01, 2019

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/nlp/brightmart/bert_language_understanding"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.