AXERA-TECH/ax-llm

Explore LLM model deployment based on AXera's AI chips

53
/ 100
Established

This project helps AI developers and engineers deploy large language models (LLMs) and vision-language models (VLMs) efficiently on AXera's AI chips. It takes pre-trained LLM/VLM models and optimizes them to run directly on AX650A/N and AX630C chips, providing a fast way to evaluate model performance and build custom AI applications. The output is a runnable model on AXera hardware, enabling specialized AI assistants for various tasks.

142 stars.

Use this if you are an AI developer or embedded systems engineer working with AXera AI chips and need to deploy large language models or multimodal models for high-performance edge computing.

Not ideal if you are not working with AXera AI chips or if you are looking for a general-purpose LLM inference solution for standard CPU/GPU platforms.

edge-AI model-deployment embedded-systems LLM-acceleration computer-vision
No Package No Dependents
Maintenance 10 / 25
Adoption 10 / 25
Maturity 16 / 25
Community 17 / 25

How are scores calculated?

Stars

142

Forks

22

Language

C++

License

BSD-3-Clause

Last pushed

Mar 10, 2026

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/AXERA-TECH/ax-llm"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.