scitix/arks
Arks is a cloud-native inference framework running on Kubernetes
Arks helps MLOps engineers, platform architects, and AI infrastructure teams deploy and manage large language models (LLMs) in cloud environments. It takes various LLMs and configurations as input and provides a scalable, distributed, and multi-tenant inference service. This enables robust and efficient delivery of AI-powered applications.
Use this if you need to run multiple LLMs efficiently across different hardware, manage access for many users, and scale your AI applications on a Kubernetes cluster.
Not ideal if you are a single user experimenting with LLMs locally or do not use Kubernetes for your cloud infrastructure.
Stars
46
Forks
6
Language
Go
License
Apache-2.0
Category
Last pushed
Jan 14, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/mlops/scitix/arks"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
kubeflow/katib
Automated Machine Learning on Kubernetes
kubeai-project/kubeai
AI Inference Operator for Kubernetes. The easiest way to serve ML models in production. Supports...
sgl-project/rbg
A workload for deploying LLM inference services on Kubernetes
beam-cloud/beta9
Ultrafast serverless GPU inference, sandboxes, and background jobs
ptimizeroracle/ondine
The LLM Dataset Engine — batch process millions of rows with 100+ providers. Multi-row batching...