zackshen/gguf
a GGUF file parser
This tool helps developers working with large language models to efficiently inspect, read, and write GGUF (GGML Universal Format) files. It allows you to quickly see what's inside these binary files, including key-value metadata and tensor information, without needing to load the entire model into memory. This is ideal for machine learning engineers, researchers, and MLOps professionals who need to manage or analyze quantized LLM files.
Use this if you need to programmatically access or modify the metadata and tensor structures within GGUF files, especially for large, quantized machine learning models.
Not ideal if you're an end-user simply running a GGUF model and don't need to interact with its internal file structure.
Stars
17
Forks
5
Language
Rust
License
MIT
Category
Last pushed
Mar 02, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/zackshen/gguf"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
ModelCloud/GPTQModel
LLM model quantization (compression) toolkit with hw acceleration support for Nvidia CUDA, AMD...
intel/auto-round
🎯An accuracy-first, highly efficient quantization toolkit for LLMs, designed to minimize quality...
pytorch/ao
PyTorch native quantization and sparsity for training and inference
bodaay/HuggingFaceModelDownloader
Simple go utility to download HuggingFace Models and Datasets
NVIDIA/kvpress
LLM KV cache compression made easy