worker-vllm and runpod-worker-oobabooga
These are ecosystem siblings — both are RunPod worker templates that serve different LLM inference backends (vLLM vs. Oobabooga), allowing users to choose which serving framework best fits their model and performance requirements.
About worker-vllm
runpod-workers/worker-vllm
The RunPod worker template for serving our large language model endpoints. Powered by vLLM.
This project helps developers deploy and manage large language models (LLMs) as highly performant, serverless API endpoints. It takes a chosen LLM (like Llama-3.1-8B-Instruct or OpenChat-3.5) and serves it through an API that's compatible with OpenAI's format. The primary users are developers who need to integrate custom LLM capabilities into their applications with speed and efficiency.
About runpod-worker-oobabooga
ashleykleynhans/runpod-worker-oobabooga
RunPod Serverless Worker for Oobabooga Text Generation API for LLMs
Scores updated daily from GitHub, PyPI, and npm data. How scores work