inference and xllm

These two tools are competitors, as both aim to provide a high-performance inference engine for large language models, with XorbitsAI's offering being a more mature and broadly adopted solution for model deployment and management across various environments, while xLLM focuses on optimization for diverse AI accelerators.

inference
76
Verified
xllm
69
Established
Maintenance 22/25
Adoption 10/25
Maturity 25/25
Community 19/25
Maintenance 22/25
Adoption 10/25
Maturity 15/25
Community 22/25
Stars: 9,129
Forks: 805
Downloads:
Commits (30d): 63
Language: Python
License: Apache-2.0
Stars: 1,081
Forks: 149
Downloads:
Commits (30d): 123
Language: C++
License:
No risk flags
No Package No Dependents

About inference

xorbitsai/inference

Swap GPT for any LLM by changing a single line of code. Xinference lets you run open-source, speech, and multimodal models on cloud, on-prem, or your laptop — all through one unified, production-ready inference API.

This tool helps AI developers and researchers deploy and manage various artificial intelligence models, including large language models (LLMs), speech recognition, and multimodal models. It takes trained AI models and makes them accessible through a unified API, allowing other applications to easily interact with them. Anyone building AI-powered applications, from chatbots to image analysis tools, would use this to put their models into production.

AI-application-development model-serving LLM-deployment speech-recognition-systems multimodal-AI

About xllm

jd-opensource/xllm

A high-performance inference engine for LLMs, optimized for diverse AI accelerators.

This project helps businesses and organizations deploy large language models (LLMs) like DeepSeek-V3.1 or Qwen2/3, especially on Chinese AI accelerators. It takes these pre-trained models and makes them run much faster and more cost-effectively, generating text responses for applications like intelligent customer service, risk control, or ad recommendations. The end-users are AI solution architects, MLOps engineers, and IT infrastructure managers responsible for deploying and managing AI applications.

AI-application-deployment large-language-model-inference AI-infrastructure-optimization enterprise-AI-solutions AI-acceleration-hardware

Scores updated daily from GitHub, PyPI, and npm data. How scores work