yuanzhoulvpi2017/quick_sentence_transformers
sentence-transformers to onnx 让sbert模型推理效率更快
This project helps anyone using Sentence-BERT models for tasks like question-answering or semantic search who finds the process too slow. It takes an existing Sentence-BERT model and converts its core component into a more efficient format. The output is a faster model that still produces the same quality text embeddings. This is for machine learning practitioners deploying NLP models who need to improve inference speed.
166 stars. No commits in the last 6 months.
Use this if you are deploying Sentence-BERT models for applications requiring quick response times, such as real-time search or interactive chatbots.
Not ideal if you are only running Sentence-BERT models for offline, batch processing where latency is not a critical concern.
Stars
166
Forks
32
Language
Python
License
MIT
Category
Last pushed
Mar 11, 2022
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/yuanzhoulvpi2017/quick_sentence_transformers"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
LoicGrobol/zeldarose
Train transformer-based models.
CPJKU/wechsel
Code for WECHSEL: Effective initialization of subword embeddings for cross-lingual transfer of...
yuanzhoulvpi2017/zero_nlp
中文nlp解决方案(大模型、数据、模型、训练、推理)
minggnim/nlp-models
A repository for training transformer based models
IntelLabs/nlp-architect
A model library for exploring state-of-the-art deep learning topologies and techniques for...