google-research-datasets/wiki-atomic-edits
A dataset of atomic wikipedia edits containing insertions and deletions of a contiguous chunk of text in a sentence. This dataset contains ~43 million edits across 8 languages.
Archived105 stars. No commits in the last 6 months.
Stars
105
Forks
7
Language
—
License
—
Category
Last pushed
May 06, 2019
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/nlp/google-research-datasets/wiki-atomic-edits"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
explosion/spaCy
💫 Industrial-strength Natural Language Processing (NLP) in Python
nltk/nltk
NLTK Source
stanfordnlp/stanza
Stanford NLP Python library for tokenization, sentence segmentation, NER, and parsing of many...
flairNLP/flair
A very simple framework for state-of-the-art Natural Language Processing (NLP)
aphp/edsnlp
Modular, fast NLP framework, compatible with Pytorch and spaCy, offering tailored support for...