houtini-ai/houtini-lm

Local or Cloud LLM support via MCP for your AI assistant with Houtini-LM - uses OpenAPI API for LM Studio, Cloud API and Ollama Compatibility. Save tokens by offloading some grunt work for your API - our tool description helps claude decide what work to assign and why.

48
/ 100
Emerging

This tool helps software developers save money on their AI assistant bills by intelligently routing coding tasks. It takes coding instructions from Claude Code and sends simpler, routine tasks like generating boilerplate, writing commit messages, or reviewing code to a local, cheaper large language model (LLM) server. This allows Claude to focus on complex, high-value tasks, while the developer's local LLM handles the grunt work, reducing token usage and costs.

Use this if you are a developer using Claude Code for software development and want to significantly reduce your monthly token expenditure by offloading common, repetitive coding tasks to a local or cheaper cloud LLM.

Not ideal if you primarily use AI assistants for short, quick questions that don't involve reading and analyzing large code files, as the overhead for very small tasks might not yield significant savings.

software-development AI-coding-assistant developer-tools code-review cost-optimization
No Package No Dependents
Maintenance 13 / 25
Adoption 5 / 25
Maturity 15 / 25
Community 15 / 25

How are scores calculated?

Stars

13

Forks

5

Language

JavaScript

License

MIT

Last pushed

Mar 14, 2026

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/mcp/houtini-ai/houtini-lm"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.