Beomi/transformers-language-modeling
Train 🤗transformers with DeepSpeed: ZeRO-2, ZeRO-3
This tool helps AI practitioners and researchers fine-tune or train popular language models like GPT, BERT, or RoBERTa on custom text datasets. You provide your raw text data and a chosen model, and it outputs a fine-tuned language model ready for downstream tasks. It's designed for those working with large language models and natural language processing.
No commits in the last 6 months.
Use this if you need to adapt an existing powerful language model to understand specific jargon, styles, or patterns present in your unique text data.
Not ideal if you're not working with large language models or if you need a pre-trained model without any custom training.
Stars
23
Forks
4
Language
Python
License
—
Category
Last pushed
May 20, 2021
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/Beomi/transformers-language-modeling"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
LoicGrobol/zeldarose
Train transformer-based models.
CPJKU/wechsel
Code for WECHSEL: Effective initialization of subword embeddings for cross-lingual transfer of...
yuanzhoulvpi2017/zero_nlp
中文nlp解决方案(大模型、数据、模型、训练、推理)
minggnim/nlp-models
A repository for training transformer based models
IntelLabs/nlp-architect
A model library for exploring state-of-the-art deep learning topologies and techniques for...