c0sogi/llama-api
An OpenAI-like LLaMA inference API
This project helps developers integrate large language models (LLMs) like LLaMA into their applications by providing a local server that mimics the OpenAI API. You feed it model definitions and it provides an API endpoint that can generate text or embeddings. It's designed for developers building applications that need to use powerful AI models without relying solely on external services.
113 stars. No commits in the last 6 months.
Use this if you are a software developer who wants to run LLaMA-based language models locally or on your own infrastructure with an OpenAI-compatible API.
Not ideal if you are an end-user looking for a no-code solution or if you prefer to use managed cloud services for LLM inference.
Stars
113
Forks
9
Language
Python
License
MIT
Category
Last pushed
Sep 17, 2023
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/c0sogi/llama-api"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
ludwig-ai/ludwig
Low-code framework for building custom LLMs, neural networks, and other AI models
withcatai/node-llama-cpp
Run AI models locally on your machine with node.js bindings for llama.cpp. Enforce a JSON schema...
mudler/LocalAI
:robot: The free, Open Source alternative to OpenAI, Claude and others. Self-hosted and...
zhudotexe/kani
kani (カニ) is a highly hackable microframework for tool-calling language models. (NLP-OSS @ EMNLP 2023)
SciSharp/LLamaSharp
A C#/.NET library to run LLM (🦙LLaMA/LLaVA) on your local device efficiently.