fboulnois/llama-cpp-docker

Run llama.cpp in a GPU accelerated Docker container

48
/ 100
Emerging

This helps developers quickly set up and run local large language models (LLMs) on their own hardware. You input model names from HuggingFace, and it provides a local chat server running that model, accessible via a web browser. This tool is for software developers who want to integrate or experiment with LLMs without relying on cloud services.

Use this if you are a developer looking to host and interact with open-source LLMs locally on a GPU-accelerated server for testing or application development.

Not ideal if you are an end-user without programming experience or specific developer needs, as it requires comfort with command-line tools and Docker.

local-LLM-deployment GPU-acceleration model-serving AI-application-development containerization
No Package No Dependents
Maintenance 6 / 25
Adoption 8 / 25
Maturity 16 / 25
Community 18 / 25

How are scores calculated?

Stars

63

Forks

15

Language

Dockerfile

License

MIT

Last pushed

Dec 16, 2025

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/fboulnois/llama-cpp-docker"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.