brightmart/bert_language_understanding
Pre-training of Deep Bidirectional Transformers for Language Understanding: pre-train TextCNN
Implements masked language model pre-training with TextCNN as an alternative backbone to Transformers, demonstrating that pre-training objectives are architecture-agnostic. Built in TensorFlow, it supports both single-label and multi-label text classification with configurable model sizes (big/small/tiny) via d_model and attention head parameters. Pre-training on raw unlabeled data yields significant performance gains—achieving 0.75 F1 after 7 fine-tuning epochs versus 0.44 after 35 epochs without pre-training on mid-sized datasets.
967 stars. No commits in the last 6 months.
Stars
967
Forks
211
Language
Python
License
—
Category
Last pushed
Jan 01, 2019
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/nlp/brightmart/bert_language_understanding"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
fidelity/textwiser
[AAAI 2021] TextWiser: Text Featurization Library
RandolphVI/Multi-Label-Text-Classification
About Muti-Label Text Classification Based on Neural Network.
ThilinaRajapakse/pytorch-transformers-classification
Based on the Pytorch-Transformers library by HuggingFace. To be used as a starting point for...
xuyige/BERT4doc-Classification
Code and source for paper ``How to Fine-Tune BERT for Text Classification?``
allenai/scibert
A BERT model for scientific text.