fla-org/flame

🔥 A minimal training framework for scaling FLA models

52
/ 100
Established

This project provides a training framework for creating highly efficient large language models, specifically those using Flash Linear Attention (FLA). It takes raw text datasets, like the FineWeb-Edu corpus, and outputs a trained language model ready for use in various applications. It's designed for machine learning researchers and engineers focused on developing custom, performant language models.

355 stars.

Use this if you are building and training your own large language models with a focus on high efficiency and scalability, especially when working with massive text datasets.

Not ideal if you're looking to simply fine-tune existing, pre-trained models or if you don't need to train models from scratch on large-scale datasets.

large-language-models model-training natural-language-processing machine-learning-engineering deep-learning
No Package No Dependents
Maintenance 6 / 25
Adoption 10 / 25
Maturity 16 / 25
Community 20 / 25

How are scores calculated?

Stars

355

Forks

58

Language

Python

License

MIT

Last pushed

Nov 15, 2025

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/fla-org/flame"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.