zhusleep/pytorch_chinese_lm_pretrain
pytorch中文语言模型预训练
This project helps machine learning engineers and NLP researchers improve the performance of Chinese language models for specific applications. You can input domain-specific Chinese text data and use it to fine-tune existing popular language models like BERT, RoBERTa, or ERNIE. The output is a more specialized language model that performs better on tasks relevant to your domain.
385 stars. No commits in the last 6 months.
Use this if you need to adapt a general Chinese language model to perform exceptionally well on text data from a niche domain or specific task, like legal documents or medical reports.
Not ideal if you are looking for a pre-trained Chinese language model without any custom training or if your task doesn't require domain-specific adaptation.
Stars
385
Forks
78
Language
Python
License
—
Category
Last pushed
Jul 17, 2020
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/nlp/zhusleep/pytorch_chinese_lm_pretrain"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
codertimo/BERT-pytorch
Google AI 2018 BERT pytorch implementation
JayYip/m3tl
BERT for Multitask Learning
920232796/bert_seq2seq
pytorch实现 Bert 做seq2seq任务,使用unilm方案,现在也可以做自动摘要,文本分类,情感分析,NER,词性标注等任务,支持t5模型,支持GPT2进行文章续写。
sileod/tasknet
Easy modernBERT fine-tuning and multi-task learning
graykode/toeicbert
TOEIC(Test of English for International Communication) solving using pytorch-pretrained-BERT model.