Heratiki/locallama-mcp

An MCP Server that works with Roo Code/Cline.Bot/Claude Desktop to optimize costs by intelligently routing coding tasks between local LLMs free APIs and paid APIs.

35
/ 100
Emerging

This tool helps developers reduce the cost of using large language models (LLMs) for coding tasks. It acts as a smart router, taking your coding requests and deciding whether to send them to a free, local LLM or a more expensive, cloud-based API. The output is optimized code generation at a lower overall cost, ideal for individual developers or small teams managing LLM expenses.

No commits in the last 6 months.

Use this if you are a software developer who uses LLMs for coding assistance and wants to lower your monthly API expenses by intelligently leveraging free local models when appropriate.

Not ideal if you primarily use LLMs for non-coding tasks, do not have local LLMs set up, or your priority is always maximum output quality regardless of cost.

software-development LLM-ops developer-tools cost-optimization AI-coding-assistant
No License Stale 6m No Package No Dependents
Maintenance 2 / 25
Adoption 7 / 25
Maturity 8 / 25
Community 18 / 25

How are scores calculated?

Stars

41

Forks

12

Language

TypeScript

License

Last pushed

Jul 25, 2025

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/mcp/Heratiki/locallama-mcp"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.