lyeoni/pretraining-for-language-understanding
Pre-training of Language Models for Language Understanding
This project helps natural language processing (NLP) practitioners prepare custom language models for various tasks. It takes a large collection of text, like Wikipedia articles, processes it into a usable format, and then trains a language model. The output is a pre-trained language model that can be used as a component in more complex applications that need to understand human language.
No commits in the last 6 months.
Use this if you need to train a foundational language model from scratch on a specific large text corpus for downstream NLP applications.
Not ideal if you are looking for a pre-trained, ready-to-use language model or a tool for fine-tuning an existing model.
Stars
83
Forks
14
Language
Python
License
Apache-2.0
Category
Last pushed
Aug 24, 2019
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/nlp/lyeoni/pretraining-for-language-understanding"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
yaserkl/RLSeq2Seq
Deep Reinforcement Learning For Sequence to Sequence Models
kefirski/pytorch_RVAE
Recurrent Variational Autoencoder that generates sequential data implemented with pytorch
ctr4si/A-Hierarchical-Latent-Structure-for-Variational-Conversation-Modeling
PyTorch Implementation of "A Hierarchical Latent Structure for Variational Conversation...
georgian-io/Multimodal-Toolkit
Multimodal model for text and tabular data with HuggingFace transformers as building block for text data
nurpeiis/LeakGAN-PyTorch
A simple implementation of LeakGAN in PyTorch