ilyalasy/moe-routing

Analysis of token routing for different implementations of Mixture of Experts

13
/ 100
Experimental

This tool helps researchers and AI practitioners understand how different Mixture of Experts (MoE) Large Language Models (LLMs) distribute input tokens to their specialized 'expert' subnetworks. You provide a RedPajama dataset, and it produces data and visualizations showing how tokens are routed. This is primarily for those researching or working with the architecture and efficiency of MoE LLMs.

No commits in the last 6 months.

Use this if you are developing or studying Mixture of Experts LLMs and need to analyze the token routing patterns to optimize performance or understand architectural behavior.

Not ideal if you are looking for a tool to train, fine-tune, or simply use an LLM for text generation or other end-user applications.

Large Language Models MoE Architecture AI Research Deep Learning Optimization Model Analysis
No License Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 5 / 25
Maturity 8 / 25
Community 0 / 25

How are scores calculated?

Stars

10

Forks

Language

Jupyter Notebook

License

Last pushed

Mar 22, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/ilyalasy/moe-routing"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.