llmsresearch/scone

Implementation and evaluation of Scaling Embedding Layers in Language Models research paper

42
/ 100
Emerging

This project helps machine learning engineers improve the performance of large language models while keeping inference costs in check. It allows you to train models that process text more effectively by using a novel embedding approach for frequent word sequences. The output is a more accurate and efficient language model, ideal for those working with extensive text datasets.

Use this if you are developing or fine-tuning large language models and need to enhance their performance and memory efficiency without increasing inference-time computational load.

Not ideal if you are looking for a plug-and-play solution for basic text processing or if you do not have experience with training and deploying machine learning models.

large-language-models natural-language-processing deep-learning-optimization model-training machine-learning-engineering
No Package No Dependents
Maintenance 10 / 25
Adoption 5 / 25
Maturity 16 / 25
Community 11 / 25

How are scores calculated?

Stars

13

Forks

2

Language

Python

License

MIT

Last pushed

Feb 02, 2026

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/embeddings/llmsresearch/scone"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.