yanaiela/easyEmbed
downloading pre-trained embedding easily and keeping only the necessary amount of embedding while training
When working with natural language processing models, you often need to use large, pre-trained word embedding files. This tool helps you quickly download these embedding models and extract only the specific words your project needs, creating smaller, more manageable files. Data scientists, NLP engineers, and academic researchers would find this useful for speeding up their experimentation and development.
No commits in the last 6 months. Available on PyPI.
Use this if you are building NLP applications and need to efficiently manage large pre-trained word embedding models like Word2Vec or GloVe, especially during the development phase.
Not ideal if your application requires the full, complete vocabulary of a pre-trained embedding model during final deployment, as this tool is optimized for creating subsets.
Stars
9
Forks
—
Language
Python
License
MIT
Category
Last pushed
Sep 27, 2018
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/embeddings/yanaiela/easyEmbed"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Featured in
Higher-rated alternatives
embeddings-benchmark/mteb
MTEB: Massive Text Embedding Benchmark
harmonydata/harmony
The Harmony Python library: a research tool for psychologists to harmonise data and...
yannvgn/laserembeddings
LASER multilingual sentence embeddings as a pip package
embeddings-benchmark/results
Data for the MTEB leaderboard
Hironsan/awesome-embedding-models
A curated list of awesome embedding models tutorials, projects and communities.