LLM Docker Deployments LLM Tools
Docker containerization and deployment solutions for running LLMs, inference servers, and related AI services locally or on networks. Does NOT include general containerization tools, Kubernetes orchestration, or non-LLM Docker projects.
There are 141 llm docker deployments tools tracked. 1 score above 70 (verified tier). The highest-rated is containers/ramalama at 79/100 with 2,640 stars. 4 of the top 10 are actively maintained.
Get all 141 projects as JSON
curl "https://pt-edge.onrender.com/api/v1/datasets/quality?domain=llm-tools&subcategory=llm-docker-deployments&limit=20"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
| # | Tool | Score | Tier |
|---|---|---|---|
| 1 |
containers/ramalama
RamaLama is an open-source developer tool that simplifies the local serving... |
|
Verified |
| 2 |
av/harbor
One command brings a complete pre-wired LLM stack with hundreds of services... |
|
Established |
| 3 |
RunanywhereAI/runanywhere-sdks
Production ready toolkit to run AI locally |
|
Established |
| 4 |
runpod-workers/worker-vllm
The RunPod worker template for serving our large language model endpoints.... |
|
Established |
| 5 |
foldl/chatllm.cpp
Pure C++ implementation of several models for real-time chatting on your... |
|
Established |
| 6 |
FarisZahrani/llama-cpp-py-sync
Auto-synced CFFI ABI python bindings for llama.cpp with prebuilt wheels... |
|
Established |
| 7 |
vtuber-plan/olah
Self-hosted huggingface mirror service. 自建huggingface镜像服务。 |
|
Established |
| 8 |
quantalogic/qllm
QLLM: A powerful CLI for seamless interaction with multiple Large Language... |
|
Established |
| 9 |
eastriverlee/LLM.swift
LLM.swift is a simple and readable library that allows you to interact with... |
|
Established |
| 10 |
varunvasudeva1/llm-server-docs
End-to-end documentation to set up your own local & fully private LLM server... |
|
Established |
| 11 |
dingodb/dingospeed
dingospeed is a self-hosted huggingface mirror service |
|
Established |
| 12 |
sangyuxiaowu/LLamaWorker
LLamaWorker is a HTTP API server developed based on the LLamaSharp project.... |
|
Emerging |
| 13 |
France-Travail/happy_vllm
A REST API for vLLM, production ready |
|
Emerging |
| 14 |
Scottcjn/llama-cpp-power8
AltiVec/VSX optimized llama.cpp for IBM POWER8 |
|
Emerging |
| 15 |
lordmathis/llamactl
Unified management and routing for llama.cpp, MLX and vLLM models with web dashboard. |
|
Emerging |
| 16 |
jlonge4/local_llama
This repo is to showcase how you can run a model locally and offline, free... |
|
Emerging |
| 17 |
ashleykleynhans/runpod-worker-oobabooga
RunPod Serverless Worker for Oobabooga Text Generation API for LLMs |
|
Emerging |
| 18 |
liltom-eth/llama2-webui
Run any Llama 2 locally with gradio UI on GPU or CPU from anywhere... |
|
Emerging |
| 19 |
ai-action/ollama-action
🦙 Run Ollama large language models (LLMs) with GitHub Actions. |
|
Emerging |
| 20 |
ADT109119/llamacpp-distributed-inference
一個基於 llama.cpp 的分佈式 LLM 推理程式,讓您能夠利用區域網路內的多台電腦協同進行大型語言模型的分佈式推理,使用 Electron... |
|
Emerging |
| 21 |
icppWorld/icpp_llm
on-chain LLMs |
|
Emerging |
| 22 |
hitomi-team/sukima
A ready-to-deploy container for implementing an easy to use REST API to... |
|
Emerging |
| 23 |
timhagel/MeloTTS-Docker-API-Server
A docker image to access MeloTTS through API calls |
|
Emerging |
| 24 |
Flowm/llm-stack
Docker compose config for local and hosted llms with multiple chat interfaces |
|
Emerging |
| 25 |
sinfallas/opendevin-docker
Run OpenDevin inside Docker |
|
Emerging |
| 26 |
wsmlby/homl
The easiest & fastest way to run LLMs in your home lab |
|
Emerging |
| 27 |
feiyun0112/Local-LLM-Server
quick way to build a private large language model server and provide... |
|
Emerging |
| 28 |
aws-samples/sample-ollama-server
Ollama on GPU EC2 instance with Open WebUI web interface and Bedrock access |
|
Emerging |
| 29 |
EvilFreelancer/docker-llama.cpp-rpc
Данный проект основан на llama.cpp и компилирует только RPC-сервер, а так же... |
|
Emerging |
| 30 |
gsuuon/ad-llama
Structured inference with Llama 2 in your browser |
|
Emerging |
| 31 |
teremterem/litellm-server-boilerplate
A lightweight LiteLLM server boilerplate pre-configured with uv and Docker... |
|
Emerging |
| 32 |
heyvaldemar/ollama-traefik-letsencrypt-docker-compose
Ollama with Let's Encrypt Using Docker Compose |
|
Emerging |
| 33 |
Mcourtyard/m-courtyard
M-Courtyard: Local AI Model Fine-tuning Assistant for Apple Silicon.... |
|
Emerging |
| 34 |
sasha0552/ToriLinux
Linux LiveCD for offline AI training and inference. |
|
Emerging |
| 35 |
rgryta/LLM-WSL2-Docker
One-click install for WizardLM-13B-Uncensored with oobabooga webui |
|
Emerging |
| 36 |
mitja/llamatunnel
Publish local LLMs and LLM apps on the internet. |
|
Emerging |
| 37 |
ai-action/setup-ollama
🦙 Set up GitHub Actions with Ollama CLI |
|
Emerging |
| 38 |
john-rocky/EdgeLLM
Simple LLM package for ios devices. |
|
Emerging |
| 39 |
nicksavarese/allora-ios
An iOS Keyboard Extension that allows for interacting with LLMs directly... |
|
Emerging |
| 40 |
DanielZhangyc/RLLM
LLM powered RSS reader |
|
Emerging |
| 41 |
cdrage/containerfiles
Containerfiles including AI, game servers, bootc and even a rickroll. |
|
Emerging |
| 42 |
crowdllama/crowdllama
CrowdLlama is a distributed system that leverages the open-source Ollama... |
|
Emerging |
| 43 |
linonetwo/MOSS-DockerFile
用于在 Docker 里运行复旦的 MOSS 语言模型,使用 GradIO 提供 WebUI。 |
|
Emerging |
| 44 |
ruska-ai/llm-server
🤖 Open-source LLM server (OpenAI, Ollama, Groq, Anthropic) with support for... |
|
Emerging |
| 45 |
BlackTechX011/Ollama-in-GitHub-Codespaces
Learn all how to run Ollama in GitHub Codespaces for free |
|
Emerging |
| 46 |
Jewelzufo/granitepi-4-nano
Run IBM Granite 4.0 locally on Raspberry Pi 5 with Ollama.This is a... |
|
Emerging |
| 47 |
asreview/asreview-server-stack
Docker compose for setting up ASReview server with authentication |
|
Emerging |
| 48 |
Scottcjn/llama-cpp-tigerleopard
WORLD FIRST: llama.cpp for Mac OS X Tiger & Leopard on PowerPC G4/G5 |
|
Emerging |
| 49 |
soulteary/docker-yi-runtime
零一万物(34B)的本地运行环境。 |
|
Emerging |
| 50 |
persys-ai/persys
Welcome! |
|
Emerging |
| 51 |
alex0dd/llm-app-microservices-template
Template for building microservice-based apps with a frontend, backend, LLM... |
|
Emerging |
| 52 |
ivangabriele-archives/docker-llm
Pre-loaded LLMs served as an OpenAI-Compatible API via Docker images. |
|
Emerging |
| 53 |
codygreen/llm_api_server
Lab to demonstrate how to apply an API to an AI model and secure it. |
|
Emerging |
| 54 |
wizzard0/llama2.ts
Llama2 inference in one TypeScript file |
|
Emerging |
| 55 |
g1ibby/homellm
A simple Docker Compose boilerplate for deploying Open WebUI and LiteLLM... |
|
Emerging |
| 56 |
Malax/buildpack-ollama
Cloud Native Buildpack that builds an OCI image with Ollama and a large... |
|
Emerging |
| 57 |
OutofAi/ChitChat
Modal LLM LLama.cpp based model deployment as part of series of Model as a... |
|
Emerging |
| 58 |
ivangabriele-archives/docker-functionary
Ready-to-deploy Docker image for Functionary LLM served as an OpenAI-Compatible API. |
|
Emerging |
| 59 |
AnLaVN/AL-Library
Java utility library, contain many feature, support to Large Language Model... |
|
Emerging |
| 60 |
m1ns09/Llama
🌐 Run GGUF models directly in your web browser using JavaScript and... |
|
Emerging |
| 61 |
raketenkater/llm-server
Smart launcher for llama.cpp / ik_llama.cpp — auto-detects GPUs, optimizes... |
|
Emerging |
| 62 |
JimKw1kX/LLM-C2-Server
An AI C2 Server |
|
Emerging |
| 63 |
DataJourneyHQ/list-github-models
GitHub action to track GitHub Models |
|
Emerging |
| 64 |
micbi-dt/lmstudio-docker
run LMStudio within a Docker container |
|
Emerging |
| 65 |
toku345/dgx-llm-serve
Docker Compose configs for running LLM inference on DGX Spark (TensorRT-LLM... |
|
Emerging |
| 66 |
openradx/llm_api_server_mock
This is a simple fastapi based server mock that implements the OpenAI API. |
|
Experimental |
| 67 |
azer/llmcat
Prepare files and directories for LLM consumption |
|
Experimental |
| 68 |
Scottcjn/power8-projects
POWER8 Projects - Ubuntu 22.04 build, PSE LLM, Darwin cross-compile |
|
Experimental |
| 69 |
llmjava/hf_text_generation
Hugging Face Text Generation API client for Java |
|
Experimental |
| 70 |
mordang7/LlamaForge
The Ultimate Command Center for Local LLMs. A professional-grade GUI for... |
|
Experimental |
| 71 |
AiratTop/ollama-self-hosted
A simple Docker Compose setup to self-host Ollama and Open WebUI. Run your... |
|
Experimental |
| 72 |
mdaconta/xlm-eco-api
Cross Language Model (LLM/SLM/etc.) Ecosystem API (xlm-eco-api) |
|
Experimental |
| 73 |
ggalancs/hfl
CLI + API server to download, manage, and run 500K+ HuggingFace models... |
|
Experimental |
| 74 |
qianniuspace/movie-detectives-server
骆驼电影侦探社(服务端) |
|
Experimental |
| 75 |
nyo16/llama_cpp_ex
Elixir bindings for llama.cpp — run LLMs locally with Metal, CUDA, Vulkan,... |
|
Experimental |
| 76 |
mo-arvan/local-llm
docker compose configuration file for running Llama-2 or any other language... |
|
Experimental |
| 77 |
SuppieRK/local-ai-lab
Offline-capable, open-source AI home lab notes: practical setups, configs,... |
|
Experimental |
| 78 |
LianHe-BI/Blackwell-optimized-llama.cpp-Docker-image
Blackwell-optimized llama.cpp Docker image – works on all NVIDIA GPUs, but... |
|
Experimental |
| 79 |
Skyluker4/llama-runpod
Docker image to run llama.cpp on runpod.io automatically |
|
Experimental |
| 80 |
yokingma/deepseek-vllm
Docker&vLLM官方镜像部署DeepSeek模型,在生产环境中提供类OpenAI接口服务。 |
|
Experimental |
| 81 |
ai-action/ollama-github-action-demo
🦙 Demos of large language models (LLMs) with Ollama in GitHub Actions. |
|
Experimental |
| 82 |
arseniy0924/rpc_manager
Web UI for orchestrating distributed llama.cpp RPC GPU clusters with auto... |
|
Experimental |
| 83 |
Pavloffm/remote-llm-server
Run Ollama in Docker. Share local LLMs across your network. GPU-accelerated. |
|
Experimental |
| 84 |
alasgarovs/openserv
OpenServ is a simple Bash-based CLI tool for managing LLMs in llama.cpp server. |
|
Experimental |
| 85 |
Daaboulex/lmstudio-nix
LM Studio packaged for NixOS — local LLM inference desktop app and server |
|
Experimental |
| 86 |
qnianjinri-del/local-llm-recommender
一键识别电脑硬件,推荐最新适配的开源大模型,并支持一键部署。 |
|
Experimental |
| 87 |
somya-droid/Pirate-LLM-Server
Run local LLM servers on iPhone with OpenAI-compatible API, Metal GPU... |
|
Experimental |
| 88 |
rjxby/llama-runtime
`llama-runtime` is a high-performance inference server designed for local... |
|
Experimental |
| 89 |
EricApgar/llm-server
Host an LLM and make it accessible on a network via API. |
|
Experimental |
| 90 |
gsavla6-hue/java-llm-integration
Comprehensive Java LLM integration library supporting OpenAI, Anthropic and... |
|
Experimental |
| 91 |
sebicom/llamacpp4j
Java wrapper for llama.cpp |
|
Experimental |
| 92 |
byang37/llama-runner
A lightweight desktop GUI for llama-server — multi-model routing, per-model... |
|
Experimental |
| 93 |
Logicish/p-lanes
A modular wrapper for llama.cpp focused on home-lab scaled hardware,... |
|
Experimental |
| 94 |
sithukyaw007/local-ai-workload
Docker-first, local-first AI workload toolkit for macOS Apple Silicon using... |
|
Experimental |
| 95 |
MooNyeu/granitepi-4-nano
🔒 Run a large language model locally on your Raspberry Pi 5 with IBM Granite... |
|
Experimental |
| 96 |
tdiprima/ollama-orchestrator
Self-hosted AI automation: manage Ollama models, deploy Open WebUI in... |
|
Experimental |
| 97 |
clixgvvv/AndroidLLMServerScript
📲 Create a local LLM server on Android using Python and llama.cpp for easy... |
|
Experimental |
| 98 |
ebowwa-archive/LLM_telecenter
A fastapi wrapper of babca / python-gsmmodem for a waveshare sim7600x. Not... |
|
Experimental |
| 99 |
SergiuDeveloper/distributed-llama.cpp
Distributed LLM inference across multiple machines. A central server routes... |
|
Experimental |
| 100 |
ThomasVitale/llm-images
Catalog of OCI images for popular open-source or open Large Language Models. |
|
Experimental |
| 101 |
VityazevEgor/LLMapi4free
LLMapi4free provides a unified API for free access to various large Language... |
|
Experimental |
| 102 |
gperdrizet/llms-devcontainer
Containerized development environment for LLM based projects |
|
Experimental |
| 103 |
futursolo/pai
Collection of AI Containers - Prebuilt and Ready-to-Use |
|
Experimental |
| 104 |
llmjava/llm4j
One API to access Large Language Models in Java |
|
Experimental |
| 105 |
zyoung11/lmgo
Windows system tray for llama.cpp + ROCm. Optimized for AMD RYZEN AI MAX+... |
|
Experimental |
| 106 |
dmeldrum6/Llama-Forge
Open source llama.cpp wrapper with server and client |
|
Experimental |
| 107 |
nishant-sethi/python-ai-extension-server
Python Server to use local LLMs |
|
Experimental |
| 108 |
abdulazizalmalki-gh/local-ai
A simple, self-hosted stack for running AI models locally using llama.cpp... |
|
Experimental |
| 109 |
sinfallas/llm-local-loader-docker
docker compose to load ollama, flowise, langfuse, open-web-ui |
|
Experimental |
| 110 |
gustavostz/Local-AI-Open-Orca-For-Dummies
Local AI Open Orca For Dummies is a user-friendly guide to running Large... |
|
Experimental |
| 111 |
thkox/home-ai-server
Home AI Server provides the backend infrastructure for the Home AI system.... |
|
Experimental |
| 112 |
kryoz/llama-strix-halo
llama.cpp setup on dedicated AMD Strix Halo machine |
|
Experimental |
| 113 |
FlorinAndrei/local-inference-docs
Run generative AI locally, on your hardware, for coding and other purposes |
|
Experimental |
| 114 |
merlijn/scala-llm-api
Basic OpenAI client for Scala |
|
Experimental |
| 115 |
turtleio/turtle
🐰 shoulda been an app - 🐢 |
|
Experimental |
| 116 |
MrTechyWorker/SmartLLM-Server
Implementing a robust client-server architecture from scratch, designed to... |
|
Experimental |
| 117 |
cyberguard-ai/local-llm-server
A containerized, offline-capable LLM API powered by Ollama. Automatically... |
|
Experimental |
| 118 |
stlin256/llama-remote
A web-based remote control panel for managing llama.cpp instances. Monitor... |
|
Experimental |
| 119 |
phospho-app/fastassert
Dockerized LLM inference server with constrained output (JSON mode), built... |
|
Experimental |
| 120 |
abhiFSD/llama.cpp-Monitor-Dashboard
⚡ Real-time monitoring dashboard for llama.cpp server — single HTML file,... |
|
Experimental |
| 121 |
Weebaay/local-ai-homelab
Déploiement d'un serveur IA local sur VM Ubuntu Server 24.04 avec Ollama et... |
|
Experimental |
| 122 |
mendhak/local-llm-workspace
Private, secure, containerized LLM environment for chat and coding. Using... |
|
Experimental |
| 123 |
Riju007/dev-knowledge-vault
🧠 My second brain — hands-on engineering notes on Docker, AI, Python and beyond |
|
Experimental |
| 124 |
chaserbot/chaseworkslab-llm
Self-hosted LLM stack (Ollama, Open WebUI, etc.) for the homelab |
|
Experimental |
| 125 |
nishantapatil3/litellm-compose
Docker Compose setup for LiteLLM proxy server with PostgreSQL and Prometheus... |
|
Experimental |
| 126 |
57Ajay/model-runner
A simple model runner using llama.cpp and huggingface |
|
Experimental |
| 127 |
aayes89/JavaRNN-LLM
An RNN written in pure Java to compete with Transformers |
|
Experimental |
| 128 |
yeeking/llamacpp-minimal-example
Minimal example of using llama cpp as library from cpp |
|
Experimental |
| 129 |
ai-action/ai-inference-demo
AI Inference in GitHub Actions demo |
|
Experimental |
| 130 |
beeracs/Llama
Run Llama models in your web browser using JavaScript and WebAssembly.... |
|
Experimental |
| 131 |
SwiftyAI/SwiftyMLC
An example of integrating local LLMS using mlc-llm into an iOS app |
|
Experimental |
| 132 |
FarzamMohammadi/self-hosted-ai-stack
Blog resources for building a self-hosted AI infrastructure. Contains all... |
|
Experimental |
| 133 |
desdeux/llama2odin
Llama2.C port in Odin |
|
Experimental |
| 134 |
Doculoom/doculoom-server
LLM backed API server |
|
Experimental |
| 135 |
wronai/docker-platform
Enterprise-grade secure media storage with AI analysis, role-based access,... |
|
Experimental |
| 136 |
AntonSHBK/llm_service
A FastAPI-based microservice for interacting with LLM (OpenAI API) with... |
|
Experimental |
| 137 |
danerlt/llm-server
使用Docker-compose部署大模型服务 |
|
Experimental |
| 138 |
NoroSaroyan/JLLM-Connect
Java library for seamless integration with LLM provider |
|
Experimental |
| 139 |
jkawamoto/llama-cpp-api
OpenAPI specification for the LLama.cpp HTTP Server |
|
Experimental |
| 140 |
marcosaugustoldo/install-anythingllm-ec2-aws-freetier
Learn how to create an Anything LLM container on your AWS instance by... |
|
Experimental |
| 141 |
siddhant385/ollamaonActions
Running Ollama on Github Actions |
|
Experimental |