JuliaStrings/TinySegmenter.jl
Julia version of TinySegmenter, compact Japanese tokenizer
When working with Japanese text, this tool helps you break sentences and phrases into individual words or meaningful units. You input a Japanese text string and it outputs a list of these segmented units. This is useful for anyone processing Japanese text for analysis, search, or other language-dependent tasks.
No commits in the last 6 months.
Use this if you need to quickly and accurately segment Japanese text into its constituent words or tokens for further processing.
Not ideal if your primary need is for a full-fledged natural language processing (NLP) library with capabilities beyond basic tokenization.
Stars
21
Forks
8
Language
Julia
License
—
Category
Last pushed
Nov 24, 2020
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/nlp/JuliaStrings/TinySegmenter.jl"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
google/sentencepiece
Unsupervised text tokenizer for Neural Network-based text generation.
daac-tools/vibrato
🎤 vibrato: Viterbi-based accelerated tokenizer
OpenNMT/Tokenizer
Fast and customizable text tokenization library with BPE and SentencePiece support
Systemcluster/kitoken
Fast and versatile tokenizer for language models, compatible with SentencePiece, Tokenizers,...
daac-tools/vaporetto
🛥 Vaporetto: Very accelerated pointwise prediction based tokenizer