c0sogi/llama-api

An OpenAI-like LLaMA inference API

36
/ 100
Emerging

This project helps developers integrate large language models (LLMs) like LLaMA into their applications by providing a local server that mimics the OpenAI API. You feed it model definitions and it provides an API endpoint that can generate text or embeddings. It's designed for developers building applications that need to use powerful AI models without relying solely on external services.

113 stars. No commits in the last 6 months.

Use this if you are a software developer who wants to run LLaMA-based language models locally or on your own infrastructure with an OpenAI-compatible API.

Not ideal if you are an end-user looking for a no-code solution or if you prefer to use managed cloud services for LLM inference.

AI-application-development local-LLM-deployment API-development machine-learning-engineering
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 9 / 25
Maturity 16 / 25
Community 11 / 25

How are scores calculated?

Stars

113

Forks

9

Language

Python

License

MIT

Last pushed

Sep 17, 2023

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/c0sogi/llama-api"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.