akanyaani/miniLLAMA

A simplified LLAMA implementation for training and inference tasks.

33
/ 100
Emerging

This project helps machine learning engineers and researchers understand the core mechanics of large language models like LLAMA and LLAMA2. It takes raw text data as input, processes it, and allows you to pre-train a simplified LLAMA model. The output is a functional model that can generate text based on your prompts, offering a hands-on way to grasp complex architectures.

No commits in the last 6 months.

Use this if you are a machine learning engineer or researcher who wants to learn the fundamental architecture and implementation details of LLAMA and LLAMA2 by building and experimenting with a simplified version.

Not ideal if you are looking to deploy a production-ready large language model or need multi-GPU support and advanced features like instruction-tuning, as this focuses on educational clarity over robust deployment.

large-language-models deep-learning-research natural-language-processing model-architecture educational-tool
Stale 6m No Package No Dependents
Maintenance 2 / 25
Adoption 7 / 25
Maturity 16 / 25
Community 8 / 25

How are scores calculated?

Stars

36

Forks

3

Language

Python

License

MIT

Last pushed

Jul 09, 2025

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/akanyaani/miniLLAMA"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.