Cyclenerd/google-cloud-litellm-proxy
🚅 LiteLLM Proxy for Google Cloud Generative AI
This project helps developers integrate various Google Cloud Vertex AI Large Language Models (LLMs) like Gemini, Claude, Llama 3, and Mistral AI Large into their applications. It takes model requests formatted for the OpenAI API and routes them to the specified LLM on Google Cloud, outputting the LLM's response. A software developer or MLOps engineer looking to simplify LLM integrations would use this.
Use this if you want to use Google Cloud's Vertex AI LLMs but prefer to interact with them using the familiar OpenAI API format.
Not ideal if you prefer to use Google Cloud's native SDKs for LLM interactions or do not use Google Cloud for your AI workloads.
Stars
61
Forks
8
Language
Dockerfile
License
Apache-2.0
Category
Last pushed
Mar 01, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/Cyclenerd/google-cloud-litellm-proxy"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
BerriAI/litellm
Python SDK, Proxy Server (AI Gateway) to call 100+ LLM APIs in OpenAI (or native) format, with...
vava-nessa/free-coding-models
Find, benchmark and install in CLI 158 FREE coding LLM models across 20 providers in real time
envoyproxy/ai-gateway
Manages Unified Access to Generative AI Services built on Envoy Gateway
theopenco/llmgateway
Route, manage, and analyze your LLM requests across multiple providers with a unified API interface.
Portkey-AI/gateway
A blazing fast AI Gateway with integrated guardrails. Route to 200+ LLMs, 50+ AI Guardrails with...