codertimo/BERT-pytorch

Google AI 2018 BERT pytorch implementation

60
/ 100
Established

This project helps natural language processing practitioners train custom language models for their specific text data. You input a large collection of text documents, and it produces a pre-trained BERT model. This model can then be used as a foundation for various downstream tasks like text classification, question answering, or sentiment analysis, making it valuable for researchers and data scientists working with language.

6,517 stars. No commits in the last 6 months. Available on PyPI.

Use this if you need to train a foundational language model (BERT) from scratch or fine-tune it on your unique dataset, especially when working with specialized jargon or a particular domain of text.

Not ideal if you're looking for a ready-to-use, pre-trained BERT model for common language understanding tasks, as this requires you to prepare your own corpus and perform the training yourself.

natural-language-processing text-analysis language-model-training information-extraction computational-linguistics
Stale 6m
Maintenance 0 / 25
Adoption 10 / 25
Maturity 25 / 25
Community 25 / 25

How are scores calculated?

Stars

6,517

Forks

1,328

Language

Python

License

Apache-2.0

Last pushed

Sep 15, 2023

Commits (30d)

0

Dependencies

3

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/nlp/codertimo/BERT-pytorch"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.