Ethan-yt/guwenbert
GuwenBERT: 古文预训练语言模型(古文BERT) A Pre-trained Language Model for Classical Chinese (Literary Chinese)
This project offers a specialized AI model for understanding Classical Chinese texts, helping researchers and scholars automatically process ancient documents. It takes raw Classical Chinese text as input and outputs structured information like identified punctuation, sentence boundaries, and named entities (people, places, book titles). Anyone working with or studying ancient Chinese literature and historical texts will find this tool useful.
555 stars. No commits in the last 6 months.
Use this if you need to accurately analyze or extract specific information from large volumes of Classical Chinese text, especially when dealing with limited labeled data for your specific task.
Not ideal if your primary focus is on modern Chinese texts, as this model is specifically optimized for Classical Chinese.
Stars
555
Forks
40
Language
—
License
Apache-2.0
Category
Last pushed
Aug 31, 2021
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/Ethan-yt/guwenbert"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
Tongjilibo/bert4torch
An elegent pytorch implement of transformers
nyu-mll/jiant
jiant is an nlp toolkit
lonePatient/TorchBlocks
A PyTorch-based toolkit for natural language processing
monologg/JointBERT
Pytorch implementation of JointBERT: "BERT for Joint Intent Classification and Slot Filling"
grammarly/gector
Official implementation of the papers "GECToR – Grammatical Error Correction: Tag, Not Rewrite"...