CPJKU/wechsel

Code for WECHSEL: Effective initialization of subword embeddings for cross-lingual transfer of monolingual language models.

49
/ 100
Emerging

Training large language models for new languages is incredibly resource-intensive. This tool helps natural language processing (NLP) researchers and machine learning engineers adapt existing English language models (like RoBERTa or GPT-2) to new languages without having to train them from scratch. It takes an English language model and a target language corpus, and outputs a version of the model initialized to understand the new language, making subsequent training much faster and more efficient.

No commits in the last 6 months. Available on PyPI.

Use this if you need to quickly adapt a pre-trained English language model to understand and process a new language, especially for languages with limited training data.

Not ideal if you are looking to train a language model completely from scratch in a new language without leveraging an existing English model.

natural-language-processing machine-translation cross-lingual-AI language-model-adaptation
Stale 6m
Maintenance 0 / 25
Adoption 9 / 25
Maturity 25 / 25
Community 15 / 25

How are scores calculated?

Stars

89

Forks

12

Language

Python

License

MIT

Last pushed

Sep 12, 2024

Commits (30d)

0

Dependencies

8

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/CPJKU/wechsel"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.