fbaldassarri/llama-cpp-container

Docker image to deploy a llama-cpp container with conda-ready environments

27
/ 100
Experimental

This helps data scientists and AI researchers deploy large language models (LLMs) on their own Linux machines with NVIDIA GPUs. It provides a ready-to-use Docker container for running `llama-cpp` models, allowing you to quickly get LLMs up and running without complex setup. You provide the GPU-enabled Linux environment, and it gives you a containerized LLM inference server.

No commits in the last 6 months.

Use this if you are a data scientist or AI researcher who needs to quickly deploy and experiment with `llama-cpp` based large language models on your local GPU-equipped Linux workstation.

Not ideal if you don't have a Linux machine with an NVIDIA GPU, or if you need to run LLMs in a cloud environment or on different hardware.

AI-deployment large-language-models GPU-computing machine-learning-operations local-AI-inference
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 6 / 25
Maturity 16 / 25
Community 5 / 25

How are scores calculated?

Stars

17

Forks

1

Language

Dockerfile

License

MIT

Last pushed

Apr 27, 2023

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/fbaldassarri/llama-cpp-container"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.