himkt/awesome-bert-japanese

📝 A list of pre-trained BERT models for Japanese with word/subword tokenization + vocabulary construction algorithm information

27
/ 100
Experimental

When working with Japanese text for natural language processing, you need to carefully choose how to break down sentences into words and subwords, as Japanese doesn't use spaces between words. This project provides a clear table comparing various pre-trained BERT models for Japanese, detailing the specific word segmentation, subword tokenization, and vocabulary construction algorithms used by each. Data scientists and NLP researchers who build or fine-tune models for Japanese text will find this resource useful.

131 stars. No commits in the last 6 months.

Use this if you need to select the most appropriate pre-trained BERT model for your Japanese NLP task and want to understand the linguistic processing choices made in its creation.

Not ideal if you are looking for ready-to-use APIs or code implementations of these models, as this project is a comparison guide, not a model library.

Japanese NLP text processing BERT models linguistic analysis machine learning
No License Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 10 / 25
Maturity 8 / 25
Community 9 / 25

How are scores calculated?

Stars

131

Forks

7

Language

License

Last pushed

Mar 15, 2023

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/nlp/himkt/awesome-bert-japanese"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.