AGI-Arena/MARS

The official implementation of MARS: Unleashing the Power of Variance Reduction for Training Large Models

54
/ 100
Established

When training large-scale deep learning models, particularly large language models like GPT-2, this project helps optimize the training process. It takes your model architecture and training data as input and outputs a trained model that converges faster and achieves better performance (lower validation loss) compared to traditional methods. Machine learning engineers and researchers who are pretraining or fine-tuning large models would use this.

716 stars. Actively maintained with 2 commits in the last 30 days.

Use this if you are a machine learning engineer or researcher looking to significantly improve the efficiency and final performance of your large model training, especially for natural language processing tasks.

Not ideal if you are working with smaller models or simpler machine learning tasks where traditional optimizers already perform adequately, as the overhead might not be justified.

large-language-models deep-learning-optimization model-pretraining natural-language-processing vision-models
No Package No Dependents
Maintenance 13 / 25
Adoption 10 / 25
Maturity 16 / 25
Community 15 / 25

How are scores calculated?

Stars

716

Forks

49

Language

Python

License

Apache-2.0

Last pushed

Mar 04, 2026

Commits (30d)

2

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/AGI-Arena/MARS"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.