hscspring/llama.np
Inference Llama/Llama2/Llama3 Modes in NumPy
This tool helps machine learning engineers and researchers prototype and understand large language models. It allows you to take pre-trained Llama or Llama2 model files and run text generation directly on your machine. The input is a model checkpoint and a text prompt, and the output is a generated text completion.
No commits in the last 6 months.
Use this if you are a machine learning engineer or researcher who wants to experiment with Llama models using only NumPy, or learn about their internal workings at a foundational level.
Not ideal if you need high-performance inference, fine-tuning capabilities, or a robust production-ready solution for large language models.
Stars
21
Forks
7
Language
Python
License
Apache-2.0
Category
Last pushed
Nov 22, 2023
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/hscspring/llama.np"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
hkproj/pytorch-llama
LLaMA 2 implemented from scratch in PyTorch
4AI/LS-LLaMA
A Simple but Powerful SOTA NER Model | Official Code For Label Supervised LLaMA Finetuning
luchangli03/export_llama_to_onnx
export llama to onnx
ayaka14732/llama-2-jax
JAX implementation of the Llama 2 model
harleyszhang/lite_llama
A light llama-like llm inference framework based on the triton kernel.