tensorchord/modelz-llm

OpenAI compatible API for LLMs and embeddings (LLaMA, Vicuna, ChatGLM and many others)

50
/ 100
Established

This is an inference server for developers who want to run popular open-source large language models (LLMs) like LLaMA and ChatGLM. It takes your code (using OpenAI's Python SDK or LangChain) and lets you query these LLMs as if they were OpenAI's API, giving you generated text or embeddings back. The primary users are machine learning engineers, data scientists, and AI developers.

276 stars. No commits in the last 6 months. Available on PyPI.

Use this if you need to integrate and self-host various open-source LLMs into your applications using a familiar OpenAI-compatible API.

Not ideal if you are a non-technical end-user looking for a pre-built application that leverages LLMs without any coding.

AI-application-development machine-learning-engineering large-language-models natural-language-processing model-deployment
Stale 6m
Maintenance 0 / 25
Adoption 10 / 25
Maturity 25 / 25
Community 15 / 25

How are scores calculated?

Stars

276

Forks

27

Language

Python

License

Apache-2.0

Last pushed

Oct 11, 2023

Commits (30d)

0

Dependencies

4

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/tensorchord/modelz-llm"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.