Flowm/llm-stack
Docker compose config for local and hosted llms with multiple chat interfaces
This project provides a ready-to-use setup for individuals or small teams to experiment with and deploy large language models (LLMs). It allows you to use various AI models, both running locally on your computer and from cloud providers like OpenAI or Google, through multiple chat interfaces similar to ChatGPT. This is ideal for researchers, developers, or even hobbyists who want to quickly set up and interact with LLMs.
Use this if you want a complete, self-contained environment to manage and interact with multiple LLMs, both local and cloud-based, without complex individual setups.
Not ideal if you require a highly customized, enterprise-scale LLM deployment with bespoke security and integration requirements.
Stars
11
Forks
4
Language
Python
License
MIT
Category
Last pushed
Oct 27, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/Flowm/llm-stack"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
containers/ramalama
RamaLama is an open-source developer tool that simplifies the local serving of AI models from...
av/harbor
One command brings a complete pre-wired LLM stack with hundreds of services to explore.
RunanywhereAI/runanywhere-sdks
Production ready toolkit to run AI locally
runpod-workers/worker-vllm
The RunPod worker template for serving our large language model endpoints. Powered by vLLM.
foldl/chatllm.cpp
Pure C++ implementation of several models for real-time chatting on your computer (CPU & GPU)