ruska-ai/llm-server

🤖 Open-source LLM server (OpenAI, Ollama, Groq, Anthropic) with support for HTTP, Streaming, Agents, RAG (Deprecated check out Orchestra) ->

35
/ 100
Emerging

This project helps 'prompt engineers' or developers quickly set up a local server to experiment with large language models (LLMs) and build AI applications. It takes your configuration for various LLM providers (like OpenAI, Groq, or local Ollama instances) and data sources (like vector databases for RAG) and provides a unified API endpoint. This allows you to integrate different LLMs, create agents, and add retrieval-augmented generation (RAG) capabilities into your applications without managing each service separately.

No commits in the last 6 months.

Use this if you are a developer looking for a local, unified server to prototype and integrate various LLMs, agents, and RAG capabilities into your applications.

Not ideal if you are a non-technical end-user or if you need a production-ready, actively maintained solution, as this project is deprecated.

AI-development prompt-engineering LLM-prototyping AI-application-integration RAG-development
No License Stale 6m No Package No Dependents
Maintenance 2 / 25
Adoption 7 / 25
Maturity 8 / 25
Community 18 / 25

How are scores calculated?

Stars

33

Forks

13

Language

TypeScript

License

Last pushed

Jun 10, 2025

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/ruska-ai/llm-server"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.