Langboat/Mengzi
Mengzi Pretrained Models
Combines linguistic-informed pretraining and training acceleration techniques to deliver lightweight, efficient Chinese language models that maintain BERT-compatible architectures for seamless integration into existing NLP pipelines. Offers specialized variants including distilled 60M-parameter models, domain-specific financial versions, T5-based generation models supporting zero-shot capabilities, and multimodal Oscar variants for vision-language tasks. Optimized for resource-constrained deployment with FP16 training, custom tokenizers reducing token overhead, and corpus-pruned BLOOM variants.
542 stars. No commits in the last 6 months.
Stars
542
Forks
62
Language
—
License
Apache-2.0
Category
Last pushed
Nov 29, 2022
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/nlp/Langboat/Mengzi"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
raghakot/keras-text
Text Classification Library in Keras
shibing624/pytextclassifier
pytextclassifier is a toolkit for text classification....
Sshanu/Relation-Classification-using-Bidirectional-LSTM-Tree
TensorFlow Implementation of the paper "End-to-End Relation Extraction using LSTMs on Sequences...
celtics1863/envtext
中文环境领域文本分析包,纯神经网络架构,支持EnvBert,LSTM,RNN,word2vec等模型,支持自定义模型,下游任务包括分类,回归,多选,情感分析,命名实体识别等,专题包括气候变化文本...
renjunxiang/Text-Classification
自然语言处理项目,目标是对文本进行分类。