zoppellarielena/BiLSTM-vs-BERT-in-feature-extraction-for-Neural-Dependency-Parsing
Completed as part of the "Natural Language Processing" course, this project employs the ArcEager parsing algorithm. Implementation is carried out using PyTorch and the Hugging Face library for utilizing pretrained BERT models.
No commits in the last 6 months.
Stars
—
Forks
—
Language
Jupyter Notebook
License
MIT
Category
Last pushed
Jul 17, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/nlp/zoppellarielena/BiLSTM-vs-BERT-in-feature-extraction-for-Neural-Dependency-Parsing"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
jidasheng/bi-lstm-crf
A PyTorch implementation of the BI-LSTM-CRF model.
charles9n/bert-sklearn
a sklearn wrapper for Google's BERT model
howl-anderson/seq2annotation
基于 TensorFlow & PaddlePaddle 的通用序列标注算法库(目前包含 BiLSTM+CRF, Stacked-BiLSTM+CRF 和...
guillaumegenthial/tf_ner
Simple and Efficient Tensorflow implementations of NER models with tf.estimator and tf.data
kamalkraj/BERT-NER
Pytorch-Named-Entity-Recognition-with-BERT