tensorchord/modelz-llm
OpenAI compatible API for LLMs and embeddings (LLaMA, Vicuna, ChatGLM and many others)
This is an inference server for developers who want to run popular open-source large language models (LLMs) like LLaMA and ChatGLM. It takes your code (using OpenAI's Python SDK or LangChain) and lets you query these LLMs as if they were OpenAI's API, giving you generated text or embeddings back. The primary users are machine learning engineers, data scientists, and AI developers.
276 stars. No commits in the last 6 months. Available on PyPI.
Use this if you need to integrate and self-host various open-source LLMs into your applications using a familiar OpenAI-compatible API.
Not ideal if you are a non-technical end-user looking for a pre-built application that leverages LLMs without any coding.
Stars
276
Forks
27
Language
Python
License
Apache-2.0
Category
Last pushed
Oct 11, 2023
Commits (30d)
0
Dependencies
4
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/tensorchord/modelz-llm"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related models
NVIDIA/sphinx-llm
LLM extensions for Sphinx Documentation
synacktraa/tool-parse
Making LLM Tool-Calling Simpler.
gusye1234/llm-as-function
Embed your LLM into a python function
caua1503/llm-tool-fusion
llm-tool-fusion é uma biblioteca Python que unifica e simplifica o uso de ferramentas com LLMs....
murphyhoucn/llm-dev
LLM Dev