QwenLM/ParScale

Parallel Scaling Law for Language Model — Beyond Parameter and Inference Time Scaling

32
/ 100
Emerging

ParScale introduces a new way to make large language models (LLMs) more capable without drastically increasing their size or slowing them down. By running multiple variations of an input through the model in parallel, it generates richer, more accurate outputs, particularly for complex tasks like coding or math. This is designed for AI researchers and engineers who build and deploy LLMs.

476 stars. No commits in the last 6 months.

Use this if you need to enhance the performance and reasoning abilities of your LLMs, especially for tasks requiring deep understanding, while being mindful of computational resources like memory and inference time.

Not ideal if your primary goal is to simply scale model parameters or inference speed using traditional methods, as this introduces a new parallel computation paradigm.

large-language-models model-optimization AI-research computational-efficiency reasoning-tasks
No License Stale 6m No Package No Dependents
Maintenance 2 / 25
Adoption 10 / 25
Maturity 7 / 25
Community 13 / 25

How are scores calculated?

Stars

476

Forks

24

Language

Python

License

Last pushed

May 17, 2025

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/QwenLM/ParScale"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.