google-research-datasets/wiki-split
One million English sentences, each split into two sentences that together preserve the original meaning, extracted from Wikipedia edits.
Archived25
/ 100
Experimental
124 stars. No commits in the last 6 months.
Archived
No License
Stale 6m
No Package
No Dependents
Maintenance
0 / 25
Adoption
10 / 25
Maturity
8 / 25
Community
7 / 25
Stars
124
Forks
5
Language
—
License
—
Category
Last pushed
Jun 03, 2019
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/nlp/google-research-datasets/wiki-split"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
nltk/nltk
NLTK Source
87
explosion/spaCy
💫 Industrial-strength Natural Language Processing (NLP) in Python
82
undertheseanlp/underthesea
Underthesea - Vietnamese NLP Toolkit
80
stanfordnlp/stanza
Stanford NLP Python library for tokenization, sentence segmentation, NER, and parsing of many...
71
flairNLP/flair
A very simple framework for state-of-the-art Natural Language Processing (NLP)
68