omlx and asiai

The LLM inference server is a complement to the multi-engine LLM benchmark and monitoring CLI, as the server provides continuous batching and SSD caching for inference that can then be benchmarked and monitored by the CLI tool.

omlx
62
Established
asiai
35
Emerging
Maintenance 22/25
Adoption 10/25
Maturity 11/25
Community 19/25
Maintenance 13/25
Adoption 2/25
Maturity 20/25
Community 0/25
Stars: 4,057
Forks: 306
Downloads:
Commits (30d): 448
Language: Python
License: Apache-2.0
Stars: 2
Forks:
Downloads:
Commits (30d): 0
Language: Python
License: Apache-2.0
No Package No Dependents
No Dependents

About omlx

jundot/omlx

LLM inference server with continuous batching & SSD caching for Apple Silicon — managed from the macOS menu bar

oMLX helps individual developers and power users on Apple Silicon Macs efficiently run and manage large language models (LLMs) and vision-language models (VLMs) directly on their machines. It takes a model file and provides a local API endpoint and a web dashboard, allowing you to interact with models for tasks like code generation, content creation, or image analysis. This is for developers or technical users who want to run powerful AI models locally without relying on cloud services.

local-AI-inference Apple-Silicon-ML LLM-deployment VLM-applications developer-tools

About asiai

druide67/asiai

Multi-engine LLM benchmark & monitoring CLI for Apple Silicon

Related comparisons

Scores updated daily from GitHub, PyPI, and npm data. How scores work