OpenSparseLLMs/LLaMA-MoE-v2

🚀 LLaMA-MoE v2: Exploring Sparsity of LLaMA from Perspective of Mixture-of-Experts with Post-Training

40
/ 100
Emerging

This project offers a collection of pre-trained, smaller, and more efficient language models built from LLaMA3, designed for advanced natural language processing. It takes existing large language models and outputs versions optimized for specific tasks, making them more affordable to run. Data scientists, machine learning engineers, and AI researchers would use this to deploy specialized AI.

No commits in the last 6 months.

Use this if you need to deploy a high-performing language model for specific tasks, but are concerned about the computational cost and resource demands of larger models.

Not ideal if you need an out-of-the-box, general-purpose large language model that doesn't require specialized fine-tuning or model construction.

natural-language-processing machine-learning-deployment AI-optimization model-fine-tuning computational-efficiency
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 9 / 25
Maturity 16 / 25
Community 15 / 25

How are scores calculated?

Stars

93

Forks

13

Language

Python

License

Apache-2.0

Last pushed

Dec 03, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/OpenSparseLLMs/LLaMA-MoE-v2"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.