ai-decentralized/BloomBee

Decentralized LLMs fine-tuning and inference with offloading

55
/ 100
Established

This project helps AI developers and researchers overcome the challenge of running large language models (LLMs) when they lack a single, powerful GPU. It takes large LLMs like LLaMA and splits them across multiple machines, allowing smaller GPUs to contribute. The output is a functional, distributed LLM for inference or fine-tuning, accessible to anyone building AI applications or performing model research.

111 stars.

Use this if you need to run or fine-tune very large language models but only have access to several machines with modest GPU resources, not one extremely powerful GPU.

Not ideal if you already have access to a single, high-end GPU or a dedicated cluster powerful enough to run your LLMs efficiently without decentralization.

distributed-AI LLM-deployment AI-infrastructure model-fine-tuning resource-optimization
No Package No Dependents
Maintenance 10 / 25
Adoption 9 / 25
Maturity 16 / 25
Community 20 / 25

How are scores calculated?

Stars

111

Forks

23

Language

Python

License

Apache-2.0

Last pushed

Mar 12, 2026

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/ai-decentralized/BloomBee"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.