teremterem/litellm-server-boilerplate
A lightweight LiteLLM server boilerplate pre-configured with uv and Docker for hosting your own OpenAI- and Anthropic-compatible endpoints. Includes LibreChat as an optional web UI.
This project helps you create your own custom AI assistants or agents by providing a base to host various large language models (LLMs) from providers like OpenAI and Anthropic. You supply your API keys for these models, and in return, you get a custom server that behaves like an OpenAI API, which can then power chat interfaces like LibreChat. It's designed for anyone who wants to build and manage their own specialized AI chatbots.
Use this if you need a flexible, self-hosted platform to integrate and customize different large language models into your own applications or chat interfaces.
Not ideal if you're only looking for a simple, off-the-shelf chat application and don't plan to customize or integrate different AI models.
Stars
11
Forks
2
Language
Python
License
MIT
Category
Last pushed
Dec 08, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/teremterem/litellm-server-boilerplate"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
containers/ramalama
RamaLama is an open-source developer tool that simplifies the local serving of AI models from...
av/harbor
One command brings a complete pre-wired LLM stack with hundreds of services to explore.
RunanywhereAI/runanywhere-sdks
Production ready toolkit to run AI locally
runpod-workers/worker-vllm
The RunPod worker template for serving our large language model endpoints. Powered by vLLM.
foldl/chatllm.cpp
Pure C++ implementation of several models for real-time chatting on your computer (CPU & GPU)