hao-ai-lab/DistCA

Efficient Long-context Language Model Training by Core Attention Disaggregation

36
/ 100
Emerging

This system helps AI researchers and deep learning engineers train large language models (LLMs) more efficiently, especially when dealing with very long input texts. It takes your LLM training setup and data, and produces a faster, more scalable training process, allowing you to build more capable models without excessive hardware or time. It is designed for those pushing the boundaries of what LLMs can understand.

Use this if you are training large language models with extremely long input contexts and are struggling with slow training times, workload imbalances across GPUs, or high communication overhead.

Not ideal if you are working with shorter context lengths or do not require highly distributed training across many GPUs, as the overhead of this system may not provide significant benefits.

large-language-model-training deep-learning-research distributed-training high-performance-computing natural-language-processing
No License No Package No Dependents
Maintenance 10 / 25
Adoption 9 / 25
Maturity 7 / 25
Community 10 / 25

How are scores calculated?

Stars

93

Forks

7

Language

Python

License

Last pushed

Mar 05, 2026

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/hao-ai-lab/DistCA"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.