moeru-ai/demodel
🚀🛸 Easily boost the speed of pulling your models and datasets from various of inference runtimes. (e.g. 🤗 HuggingFace, 🐫 Ollama, vLLM, and more!)
When working with large AI models and datasets, this tool helps overcome slow download speeds and difficult sharing challenges. It lets you quickly pull models from platforms like HuggingFace or Ollama, even if your internet connection is poor or you want to share data with colleagues on a local network. Data scientists, machine learning engineers, and AI researchers who frequently download and deploy models will find this useful.
No commits in the last 6 months.
Use this if you need to speed up the process of accessing and sharing AI models and datasets across different machines or with teammates locally.
Not ideal if your workflow primarily involves training models from scratch or if you don't frequently work with pre-trained models from online repositories.
Stars
10
Forks
2
Language
Go
License
MIT
Category
Last pushed
Jun 25, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/moeru-ai/demodel"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
vllm-project/vllm
A high-throughput and memory-efficient inference and serving engine for LLMs
sgl-project/sglang
SGLang is a high-performance serving framework for large language models and multimodal models.
alibaba/MNN
MNN: A blazing-fast, lightweight inference engine battle-tested by Alibaba, powering...
xorbitsai/inference
Swap GPT for any LLM by changing a single line of code. Xinference lets you run open-source,...
tensorzero/tensorzero
TensorZero is an open-source stack for industrial-grade LLM applications. It unifies an LLM...