AngeloGalav/franken-llama
Making Llama2 more computationally and memory efficient by skipping and repeating multiple attention layers.
10
/ 100
Experimental
No commits in the last 6 months.
No License
Stale 6m
No Package
No Dependents
Maintenance
0 / 25
Adoption
2 / 25
Maturity
8 / 25
Community
0 / 25
Stars
2
Forks
—
Language
Jupyter Notebook
License
—
Category
Last pushed
Dec 05, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/AngeloGalav/franken-llama"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
hkproj/pytorch-llama
LLaMA 2 implemented from scratch in PyTorch
49
4AI/LS-LLaMA
A Simple but Powerful SOTA NER Model | Official Code For Label Supervised LLaMA Finetuning
44
luchangli03/export_llama_to_onnx
export llama to onnx
42
ayaka14732/llama-2-jax
JAX implementation of the Llama 2 model
42
harleyszhang/lite_llama
A light llama-like llm inference framework based on the triton kernel.
42