node-llama-cpp and llama-swap

These are complements: node-llama-cpp provides the local LLM inference engine for Node.js applications, while llama-swap manages dynamic model switching across compatible servers, allowing you to use them together to run and swap between multiple models locally.

node-llama-cpp
73
Verified
llama-swap
62
Established
Maintenance 13/25
Adoption 15/25
Maturity 25/25
Community 20/25
Maintenance 17/25
Adoption 10/25
Maturity 16/25
Community 19/25
Stars: 1,942
Forks: 176
Downloads:
Commits (30d): 2
Language: TypeScript
License: MIT
Stars: 2,775
Forks: 205
Downloads:
Commits (30d): 20
Language: Go
License: MIT
No risk flags
No Package No Dependents

About node-llama-cpp

withcatai/node-llama-cpp

Run AI models locally on your machine with node.js bindings for llama.cpp. Enforce a JSON schema on the model output on the generation level

This project helps JavaScript and TypeScript developers integrate advanced AI capabilities directly into their applications by running large language models (LLMs) on their own machines. Developers input a language model and prompts, and the tool outputs structured text, function calls, or embeddings, enabling features like smart chatbots, data summarization, or advanced search within their applications. It's designed for developers building AI-powered features without relying on external cloud services.

AI-powered applications local LLM deployment JavaScript development TypeScript development on-device AI

About llama-swap

mostlygeek/llama-swap

Reliable model swapping for any local OpenAI/Anthropic compatible server - llama.cpp, vllm, etc

This tool helps AI application developers manage multiple local generative AI models on their machines efficiently. It acts as a smart traffic controller, taking in requests for various AI tasks (like text generation, image creation, or speech processing) and automatically routing them to the correct local AI model server. Developers building AI applications will find this useful for testing and deploying different models without manual intervention.

AI application development local AI deployment model management generative AI AI server orchestration

Scores updated daily from GitHub, PyPI, and npm data. How scores work