guotong1988/BERT-pre-training

multi-gpu pre-training in one machine for BERT without horovod (Data Parallelism)

54
/ 100
Established

This project helps machine learning engineers pre-train large language models like BERT more efficiently on a single machine. By leveraging multiple GPUs, it allows for significantly larger batch sizes, which can accelerate the training process. This is ideal for researchers and ML engineers who need to quickly fine-tune or adapt BERT models for specific language understanding tasks.

171 stars.

Use this if you are an ML engineer with a single powerful server equipped with multiple GPUs and need to pre-train BERT-like models faster for natural language processing applications.

Not ideal if you are looking to distribute pre-training across multiple machines or do not have access to multiple GPUs on a single server.

natural-language-processing large-language-models deep-learning-training computational-linguistics machine-learning-engineering
No Package No Dependents
Maintenance 6 / 25
Adoption 10 / 25
Maturity 16 / 25
Community 22 / 25

How are scores calculated?

Stars

171

Forks

53

Language

Python

License

Apache-2.0

Last pushed

Dec 27, 2025

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/nlp/guotong1988/BERT-pre-training"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.