smarter-project/armnn_tflite_backend

TensorFlow Lite backend with ArmNN delegate support for Nvidia Triton

33
/ 100
Emerging

This tool helps machine learning engineers efficiently deploy and run TensorFlow Lite models on Arm-based systems. It takes your TFLite serialized models and outputs optimized model inference, allowing you to serve predictions faster. It's designed for developers managing inference servers on edge devices or embedded systems.

No commits in the last 6 months.

Use this if you need to accelerate TensorFlow Lite model inference using ArmNN or XNNPACK on an Nvidia Triton server running on Linux Arm64 hardware.

Not ideal if you are working with non-TFLite models or deploying to x86-based servers, or if you don't use Nvidia Triton as your inference serving solution.

edge-AI ML-deployment inference-optimization embedded-systems model-serving
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 4 / 25
Maturity 16 / 25
Community 13 / 25

How are scores calculated?

Stars

8

Forks

2

Language

C++

License

MIT

Last pushed

Sep 03, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/smarter-project/armnn_tflite_backend"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.