GithubX-F/DynaMO-RL

Dynamic Rollout Allocation and Advantage Modulation for Policy Optimization (DynaMO) - Official Implementation

34
/ 100
Emerging

This project helps large language model (LLM) developers fine-tune their models for complex reasoning tasks, especially in mathematics. It takes your LLM and a set of reasoning problems, then dynamically optimizes the training process. The output is a more accurate and robust LLM capable of solving mathematical challenges with higher success rates.

Use this if you are developing or training large language models for tasks requiring verifiable reasoning, particularly in mathematical domains, and want to improve their performance and training efficiency.

Not ideal if you are working with LLMs for creative writing, summarization, or other non-reasoning-focused tasks, or if you are not directly involved in LLM training and optimization.

LLM training mathematical reasoning AI model optimization reinforcement learning for LLMs language model fine-tuning
No Package No Dependents
Maintenance 10 / 25
Adoption 9 / 25
Maturity 11 / 25
Community 4 / 25

How are scores calculated?

Stars

86

Forks

2

Language

Python

License

Apache-2.0

Last pushed

Mar 10, 2026

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/GithubX-F/DynaMO-RL"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.