bigscience-workshop/bigscience
Central place for the engineering/scaling WG: documentation, SLURM scripts and logs, compute environment and data.
This project helps machine learning researchers and engineers manage and document their large language model training experiments. It provides a central repository for experiment specifications, scripts, logs, and detailed chronicles of findings. Researchers can track the progress of ongoing trainings and analyze results, contributing to a collective understanding of model behavior.
1,010 stars. No commits in the last 6 months.
Use this if you are a machine learning researcher or engineer actively involved in training large language models and need a structured way to document, track, and share your experimental setups and results.
Not ideal if you are looking for a pre-trained model to use directly, or if your focus is on deploying existing models rather than researching and training new large language models.
Stars
1,010
Forks
101
Language
Shell
License
—
Category
Last pushed
Jul 29, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/mlops/bigscience-workshop/bigscience"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Related tools
1minds3t/omnipkg
Universal Python Runtime Orchestrator - One environment. Infinite Pythons and packages. <1ms...
maharshi95/submititnow
A toolkit to create, launch and monitor SLURM jobs over existing python scripts.
LthreeC/pyruns
A lightweight Python CLI for batch experiments: auto-generate tasks from configs, run in...
synpse-hq/slurm-cluster
A quick and automated way to setup your Slurm mini cluster