dhakalnirajan/LLaMA-BitNet
LLaMA-BitNet is a repository dedicated to empowering users to train their own BitNet models built upon LLaMA 2 model, inspired by the groundbreaking paper 'The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits'.
This tool helps AI researchers and machine learning engineers train custom large language models (LLMs) more efficiently. It takes the LLaMA 2 model and your chosen text dataset as input, outputting a highly optimized BitNet model. This allows you to experiment with cutting-edge 1-bit LLM architectures.
No commits in the last 6 months.
Use this if you are an AI researcher or ML engineer interested in training compact, high-performance large language models based on the LLaMA 2 architecture and the BitNet principles.
Not ideal if you are looking for a pre-trained LLM for immediate use or if you are not comfortable with deep learning model training and customization.
Stars
28
Forks
2
Language
Python
License
MIT
Category
Last pushed
Mar 31, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/dhakalnirajan/LLaMA-BitNet"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
hkproj/pytorch-llama
LLaMA 2 implemented from scratch in PyTorch
4AI/LS-LLaMA
A Simple but Powerful SOTA NER Model | Official Code For Label Supervised LLaMA Finetuning
luchangli03/export_llama_to_onnx
export llama to onnx
ayaka14732/llama-2-jax
JAX implementation of the Llama 2 model
harleyszhang/lite_llama
A light llama-like llm inference framework based on the triton kernel.