dhakalnirajan/LLaMA-BitNet

LLaMA-BitNet is a repository dedicated to empowering users to train their own BitNet models built upon LLaMA 2 model, inspired by the groundbreaking paper 'The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits'.

30
/ 100
Emerging

This tool helps AI researchers and machine learning engineers train custom large language models (LLMs) more efficiently. It takes the LLaMA 2 model and your chosen text dataset as input, outputting a highly optimized BitNet model. This allows you to experiment with cutting-edge 1-bit LLM architectures.

No commits in the last 6 months.

Use this if you are an AI researcher or ML engineer interested in training compact, high-performance large language models based on the LLaMA 2 architecture and the BitNet principles.

Not ideal if you are looking for a pre-trained LLM for immediate use or if you are not comfortable with deep learning model training and customization.

Large Language Models AI Research Model Training Deep Learning Optimization Machine Learning Engineering
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 7 / 25
Maturity 16 / 25
Community 7 / 25

How are scores calculated?

Stars

28

Forks

2

Language

Python

License

MIT

Last pushed

Mar 31, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/dhakalnirajan/LLaMA-BitNet"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.