howl-anderson/MITIE_Chinese_Wikipedia_corpus

Pre-trained Wikipedia corpus by MITIE

40
/ 100
Emerging

This project offers a pre-trained language model specifically for Chinese text, built using a large dataset from Chinese Wikipedia. It takes raw Chinese text and provides a powerful 'understanding' of the words, which is crucial for building applications that need to process and interpret human language. This is ideal for developers creating natural language processing (NLP) solutions for Chinese speakers.

No commits in the last 6 months.

Use this if you are a developer building a Chinese natural language processing system and need a robust, pre-trained word representation model to save significant training time and computational resources.

Not ideal if you are an end-user looking for a ready-to-use application, as this project provides a technical component for developers rather than a direct user-facing tool.

natural-language-processing machine-learning-development chinese-language-technology text-analytics developer-tools
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 8 / 25
Maturity 16 / 25
Community 16 / 25

How are scores calculated?

Stars

51

Forks

9

Language

License

MIT

Last pushed

Sep 09, 2018

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/nlp/howl-anderson/MITIE_Chinese_Wikipedia_corpus"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.