cgbur/llama2.zig
Inference Llama 2 in one file of pure Zig
This tool helps developers and researchers efficiently run inference with Meta's Llama 2 language models. You provide a Llama 2 model checkpoint file and an optional text prompt, and it quickly generates a text completion. It's designed for those who want to integrate Llama 2's text generation capabilities into their applications with high performance.
211 stars. No commits in the last 6 months.
Use this if you need a fast, portable, and straightforward way to run Llama 2 inference, especially when speed is critical for your application.
Not ideal if you require multi-threaded inference or are looking for a high-level API for complex natural language processing tasks beyond basic text generation.
Stars
211
Forks
19
Language
Zig
License
MIT
Category
Last pushed
Nov 21, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/cgbur/llama2.zig"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
ludwig-ai/ludwig
Low-code framework for building custom LLMs, neural networks, and other AI models
withcatai/node-llama-cpp
Run AI models locally on your machine with node.js bindings for llama.cpp. Enforce a JSON schema...
mudler/LocalAI
:robot: The free, Open Source alternative to OpenAI, Claude and others. Self-hosted and...
zhudotexe/kani
kani (カニ) is a highly hackable microframework for tool-calling language models. (NLP-OSS @ EMNLP 2023)
SciSharp/LLamaSharp
A C#/.NET library to run LLM (🦙LLaMA/LLaVA) on your local device efficiently.