howl-anderson/PaddleTokenizer

使用 PaddlePaddle 实现基于深度神经网络的中文分词引擎 | A DNN Chinese Tokenizer by Using PaddlePaddle

32
/ 100
Emerging

This tool helps split Chinese text into individual words, which is essential for many language processing tasks like search, analysis, or translation. You input a Chinese sentence or document, and it outputs a list of segmented words. Anyone working with Chinese text data, such as linguists, data analysts, or content managers, would find this useful.

No commits in the last 6 months.

Use this if you need accurate word segmentation for Chinese text, especially for tasks requiring a deep understanding of natural language.

Not ideal if you are working with languages other than Chinese or require a simple, rule-based tokenizer without deep learning capabilities.

Chinese-language-processing text-analysis natural-language-understanding content-management linguistics
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 6 / 25
Maturity 16 / 25
Community 10 / 25

How are scores calculated?

Stars

15

Forks

2

Language

JavaScript

License

AGPL-3.0

Last pushed

Jul 27, 2020

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/nlp/howl-anderson/PaddleTokenizer"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.