MNN and xllm

These are competitors—both are inference engines optimizing LLM execution on hardware accelerators, targeting the same use case of efficient on-device model deployment, though MNN has achieved significantly wider adoption and maturity.

MNN
80
Verified
xllm
69
Established
Maintenance 22/25
Adoption 10/25
Maturity 25/25
Community 23/25
Maintenance 22/25
Adoption 10/25
Maturity 15/25
Community 22/25
Stars: 14,526
Forks: 2,234
Downloads:
Commits (30d): 52
Language: C++
License: Apache-2.0
Stars: 1,081
Forks: 149
Downloads:
Commits (30d): 123
Language: C++
License:
No risk flags
No Package No Dependents

About MNN

alibaba/MNN

MNN: A blazing-fast, lightweight inference engine battle-tested by Alibaba, powering high-performance on-device LLMs and Edge AI.

This project helps developers integrate advanced AI capabilities, like large language models and image generation, directly into applications running on mobile phones, PCs, or IoT devices. It takes pre-trained AI models as input and delivers optimized, high-performance inference outputs, enabling features like offline AI chatbots or on-device image editing. This is for software engineers and product developers building AI-powered applications for edge devices.

edge-ai-development mobile-app-ai iot-ai on-device-machine-learning ai-model-deployment

About xllm

jd-opensource/xllm

A high-performance inference engine for LLMs, optimized for diverse AI accelerators.

This project helps businesses and organizations deploy large language models (LLMs) like DeepSeek-V3.1 or Qwen2/3, especially on Chinese AI accelerators. It takes these pre-trained models and makes them run much faster and more cost-effectively, generating text responses for applications like intelligent customer service, risk control, or ad recommendations. The end-users are AI solution architects, MLOps engineers, and IT infrastructure managers responsible for deploying and managing AI applications.

AI-application-deployment large-language-model-inference AI-infrastructure-optimization enterprise-AI-solutions AI-acceleration-hardware

Scores updated daily from GitHub, PyPI, and npm data. How scores work