polm/unidic-py
Unidic packaged for installation via pip.
This project helps linguists, researchers, or anyone working with Japanese text analysis to break down contemporary written Japanese into its core components. It takes raw Japanese text as input and outputs detailed linguistic information for each word, including its part of speech, conjugation, lemma, pronunciation, and accent. This is ideal for natural language processing tasks requiring deep grammatical insight into Japanese.
109 stars. Used by 6 other packages. No commits in the last 6 months. Available on PyPI.
Use this if you need to perform in-depth linguistic analysis of contemporary written Japanese and require a comprehensive dictionary for tasks like morphological analysis, parsing, or language research.
Not ideal if you need a lightweight dictionary and have strict disk space limitations, as this package is quite large.
Stars
109
Forks
13
Language
Python
License
MIT
Category
Last pushed
Feb 26, 2025
Commits (30d)
0
Reverse dependents
6
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/nlp/polm/unidic-py"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related tools
blmoistawinde/HarvestText
文本挖掘和预处理工具(文本清洗、新词发现、情感分析、实体识别链接、关键词抽取、知识抽取、句法分析等),无监督或弱监督方法
huspacy/huspacy
HuSpaCy: industrial-strength Hungarian natural language processing
bnosac/udpipe
R package for Tokenization, Parts of Speech Tagging, Lemmatization and Dependency Parsing Based...
BramVanroy/spacy_conll
Pipeline component for spaCy (and other spaCy-wrapped parsers such as spacy-stanza and...
tanloong/neosca
L2SCA & LCA fork: cross-platform, GUI, without Java dependency