monologg/KoCharELECTRA
Character-level Korean ELECTRA Model (음절 단위 한국어 ELECTRA)
This project helps anyone working with Korean text analysis who needs to process the language at a character (syllable) level rather than whole words. It takes Korean sentences or documents as input and outputs a detailed, character-by-character breakdown. This is ideal for natural language processing tasks where understanding individual Korean syllables is crucial. Researchers, data scientists, and machine learning engineers working on Korean NLP applications would use this.
No commits in the last 6 months.
Use this if you need a pre-trained Korean language model that processes text at the syllable level, offering a fine-grained understanding of the language.
Not ideal if your application requires processing Hanja characters, as this model explicitly excludes them from its vocabulary.
Stars
54
Forks
10
Language
Python
License
Apache-2.0
Category
Last pushed
Jun 12, 2023
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/monologg/KoCharELECTRA"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Compare
Higher-rated alternatives
SKTBrain/KoBERT
Korean BERT pre-trained cased (KoBERT)
monologg/KoELECTRA
Pretrained ELECTRA Model for Korean
monologg/KoBERT-Transformers
KoBERT on 🤗 Huggingface Transformers 🤗 (with Bug Fixed)
VinAIResearch/PhoBERT
PhoBERT: Pre-trained language models for Vietnamese (EMNLP-2020 Findings)
KB-AI-Research/KB-ALBERT
KB국민은행에서 제공하는 경제/금융 도메인에 특화된 한국어 ALBERT 모델