autonomi-ai/nos

⚡️ A fast and flexible PyTorch inference server that runs locally, on any cloud or AI HW.

38
/ 100
Emerging

This project helps machine learning engineers and MLOps professionals quickly deploy and manage various AI models like large language models, image generators, and audio transcription tools. You provide the model and data (text, images, audio), and it outputs the processed results from the AI model, ready for your applications. It's designed for teams building AI-powered products that need efficient model serving.

147 stars. No commits in the last 6 months.

Use this if you need a flexible and performant way to serve multiple PyTorch AI models, including LLMs, diffusion models, and more, across different cloud environments or hardware.

Not ideal if you are looking for a no-code solution or primarily work with non-PyTorch machine learning frameworks.

AI-model-deployment MLOps real-time-inference LLM-serving computer-vision-deployment
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 10 / 25
Maturity 16 / 25
Community 12 / 25

How are scores calculated?

Stars

147

Forks

12

Language

Python

License

Apache-2.0

Last pushed

Jun 08, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/generative-ai/autonomi-ai/nos"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.