MartinoMensio/it_vectors_wiki_spacy
Word embeddings for Italian language, spacy2 prebuilt model
This project provides pre-trained word embeddings specifically for the Italian language, derived from Italian Wikipedia. It takes raw Italian text as input and represents each word as a numerical vector, which can then be used by various natural language processing applications. It's designed for data scientists, NLP engineers, or researchers working on Italian language understanding tasks.
No commits in the last 6 months.
Use this if you need high-quality, pre-trained word embeddings for the Italian language to power your natural language processing models or applications.
Not ideal if your project focuses on languages other than Italian or if you need extremely specialized embeddings trained on a highly specific domain corpus.
Stars
8
Forks
1
Language
Python
License
MIT
Category
Last pushed
Apr 01, 2020
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/embeddings/MartinoMensio/it_vectors_wiki_spacy"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
shibing624/text2vec
text2vec, text to vector....
predict-idlab/pyRDF2Vec
đ Python Implementation and Extension of RDF2Vec
IntuitionEngineeringTeam/chars2vec
Character-based word embeddings model based on RNN for handling real world texts
IITH-Compilers/IR2Vec
Implementation of IR2Vec, LLVM IR Based Scalable Program Embeddings
ddangelov/Top2Vec
Top2Vec learns jointly embedded topic, document and word vectors.