nebuly-ai/optimate
A collection of libraries to optimise AI model performances
This suite of tools helps AI/ML engineers make their models run faster and more efficiently. You provide your AI models and hardware setup, and it helps you get optimized models that use fewer resources and incur lower inference costs. It's for machine learning engineers, MLOps specialists, and data scientists looking to improve the operational performance of their AI systems.
8,349 stars. No commits in the last 6 months.
Use this if you are a machine learning engineer or MLOps specialist looking to make your AI models, especially large language models (LLMs), run more cost-effectively on GPUs or CPUs.
Not ideal if you need active support, ongoing updates, or a beginner-friendly solution for general AI model development, as this project is no longer maintained.
Stars
8,349
Forks
624
Language
Python
License
Apache-2.0
Category
Last pushed
Jul 22, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/nebuly-ai/optimate"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
Tencent/AngelSlim
Model compression toolkit engineered for enhanced usability, comprehensiveness, and efficiency.
antgroup/glake
GLake: optimizing GPU memory management and IO transmission.
kyo-takano/chinchilla
A toolkit for scaling law research ⚖
liyucheng09/Selective_Context
Compress your input to ChatGPT or other LLMs, to let them process 2x more content and save 40%...
TsingmaoAI/MI-optimize
mi-optimize is a versatile tool designed for the quantization and evaluation of large language...