brightmart/bert_language_understanding
Pre-training of Deep Bidirectional Transformers for Language Understanding: pre-train TextCNN
This project helps anyone working with text data to quickly and effectively train machine learning models for tasks like document classification. By using a 'pre-train and fine-tune' strategy, you can feed in raw text documents, and the system learns general language understanding from them, then applies that knowledge to your specific labeled dataset for classification. This results in more accurate models with less training time, even with moderately sized datasets.
967 stars. No commits in the last 6 months.
Use this if you need to build a text classification model and want to achieve better performance and faster training without needing massive amounts of labeled data from scratch.
Not ideal if you're looking for a plug-and-play solution for general text generation or advanced conversational AI, as it focuses on classification improvements.
Stars
967
Forks
211
Language
Python
License
—
Category
Last pushed
Jan 01, 2019
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/nlp/brightmart/bert_language_understanding"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
fidelity/textwiser
[AAAI 2021] TextWiser: Text Featurization Library
RandolphVI/Multi-Label-Text-Classification
About Muti-Label Text Classification Based on Neural Network.
ThilinaRajapakse/pytorch-transformers-classification
Based on the Pytorch-Transformers library by HuggingFace. To be used as a starting point for...
ntumlgroup/LibMultiLabel
A library for multi-class and multi-label classification
xuyige/BERT4doc-Classification
Code and source for paper ``How to Fine-Tune BERT for Text Classification?``