worker-vllm and runpod-worker-oobabooga

These are ecosystem siblings — both are RunPod worker templates that serve different LLM inference backends (vLLM vs. Oobabooga), allowing users to choose which serving framework best fits their model and performance requirements.

worker-vllm
61
Established
Maintenance 10/25
Adoption 10/25
Maturity 16/25
Community 25/25
Maintenance 13/25
Adoption 3/25
Maturity 16/25
Community 14/25
Stars: 406
Forks: 290
Downloads:
Commits (30d): 0
Language: Python
License: MIT
Stars: 3
Forks: 3
Downloads:
Commits (30d): 0
Language: Python
License: GPL-3.0
No Package No Dependents
No Package No Dependents

About worker-vllm

runpod-workers/worker-vllm

The RunPod worker template for serving our large language model endpoints. Powered by vLLM.

This project helps developers deploy and manage large language models (LLMs) as highly performant, serverless API endpoints. It takes a chosen LLM (like Llama-3.1-8B-Instruct or OpenChat-3.5) and serves it through an API that's compatible with OpenAI's format. The primary users are developers who need to integrate custom LLM capabilities into their applications with speed and efficiency.

AI-application-development MLOps API-development backend-development large-language-model-deployment

About runpod-worker-oobabooga

ashleykleynhans/runpod-worker-oobabooga

RunPod Serverless Worker for Oobabooga Text Generation API for LLMs

Scores updated daily from GitHub, PyPI, and npm data. How scores work