KolosalAI/kolosal-server
Kolosal AI is an OpenSource and Lightweight alternative to Ollama to run LLMs 100% offline on your device.
This project helps you run large language models (LLMs) like those from OpenAI completely on your own Windows or Linux computer, without needing an internet connection. You feed it a large language model file, and it gives you a local service that can answer your questions or generate text. It's designed for individuals, small businesses, or researchers who want to use advanced AI models privately and without external API costs.
Use this if you need to run large language models offline, keep your data private, or customize your AI applications without relying on cloud services.
Not ideal if you prefer simple, plug-and-play web applications for AI without any local setup or technical configuration.
Stars
13
Forks
5
Language
C++
License
Apache-2.0
Category
Last pushed
Jan 02, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/KolosalAI/kolosal-server"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
ludwig-ai/ludwig
Low-code framework for building custom LLMs, neural networks, and other AI models
withcatai/node-llama-cpp
Run AI models locally on your machine with node.js bindings for llama.cpp. Enforce a JSON schema...
mudler/LocalAI
:robot: The free, Open Source alternative to OpenAI, Claude and others. Self-hosted and...
zhudotexe/kani
kani (カニ) is a highly hackable microframework for tool-calling language models. (NLP-OSS @ EMNLP 2023)
SciSharp/LLamaSharp
A C#/.NET library to run LLM (🦙LLaMA/LLaVA) on your local device efficiently.