hybridgroup/yzma
Go with your own intelligence - Go applications that directly integrate llama.cpp for local inference using hardware acceleration.
Yzma helps Go developers build applications that use large language models (LLMs) and vision language models (VLMs) for tasks like interactive chat or image analysis. It takes GGUF-formatted models and text/image inputs, then outputs generated text, all running directly on the developer's hardware. This is for Go developers creating smart applications that need local AI capabilities.
350 stars.
Use this if you are a Go developer who wants to embed local, hardware-accelerated AI inference directly into your applications without needing external servers or CGo.
Not ideal if you are not a Go developer or if you prefer using cloud-based AI services or other programming languages.
Stars
350
Forks
11
Language
Go
License
—
Category
Last pushed
Mar 08, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/hybridgroup/yzma"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
ludwig-ai/ludwig
Low-code framework for building custom LLMs, neural networks, and other AI models
withcatai/node-llama-cpp
Run AI models locally on your machine with node.js bindings for llama.cpp. Enforce a JSON schema...
mudler/LocalAI
:robot: The free, Open Source alternative to OpenAI, Claude and others. Self-hosted and...
zhudotexe/kani
kani (カニ) is a highly hackable microframework for tool-calling language models. (NLP-OSS @ EMNLP 2023)
SciSharp/LLamaSharp
A C#/.NET library to run LLM (🦙LLaMA/LLaVA) on your local device efficiently.