Mixture of Experts LLMs LLM Tools

Techniques, implementations, and optimizations for Mixture-of-Experts (MoE) architectures in language models, including compression, routing, expert specialization, and training methods. Does NOT include general model compression, non-MoE attention mechanisms, or domain-specific applications using MoE as a black box.

There are 8 mixture of experts llms tools tracked. 1 score above 70 (verified tier). The highest-rated is InternLM/xtuner at 76/100 with 5,096 stars. 1 of the top 10 are actively maintained.

Get all 8 projects as JSON

curl "https://pt-edge.onrender.com/api/v1/datasets/quality?domain=llm-tools&subcategory=mixture-of-experts-llms&limit=20"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.

# Tool Score Tier
1 InternLM/xtuner

A Next-Generation Training Engine Built for Ultra-Large MoE Models

76
Verified
2 AmanPriyanshu/GPT-OSS-MoE-ExpertFingerprinting

ExpertFingerprinting: Behavioral Pattern Analysis and Specialization Mapping...

41
Emerging
3 arm-education/Advanced-AI-Mixture-of-Experts

Hands-on course materials for ML engineers to implement and optimize Mixture...

38
Emerging
4 SuperBruceJia/Awesome-Mixture-of-Experts

Awesome Mixture of Experts (MoE): A Curated List of Mixture of Experts (MoE)...

38
Emerging
5 sumitdotml/moe-emergence

a project highlighting the emergent expert specialization in Mixture of...

26
Experimental
6 iahuang/cosmoe

Enabling inference of large mixture-of-experts (MoE) models on Apple Silicon...

21
Experimental
7 Renuu007/Mixture-of-Idiots

🤪🧠💥 Mixture of Idiots (MoI): A Python project exploring 'Mixture of Models'...

10
Experimental
8 reshalfahsi/gpt2moe-instruct

Instruction Fine-tuning of the GPT2MoE Model: GPT-2 with Mixture-of-Experts

10
Experimental