feifeibear/Odysseus-Transformer

Odysseus: Playground of LLM Sequence Parallelism

27
/ 100
Experimental

This project is a playground for developers who are pushing the boundaries of large language models (LLMs) with extremely long input sequences. It explores and compares different methods for 'sequence parallelism,' which is a way to distribute the computational load of processing long texts across multiple GPUs. Developers can experiment with different parallelization techniques to find the most efficient way to train or run these advanced models.

No commits in the last 6 months.

Use this if you are a machine learning engineer or researcher developing or fine-tuning large language models and need to optimize their performance when working with very long input texts, especially on multi-GPU setups.

Not ideal if you are a general LLM user or a developer looking for a high-level API to simply use an existing LLM, as this project focuses on low-level parallelization strategies.

LLM development model training GPU optimization distributed deep learning long context window
No License Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 9 / 25
Maturity 8 / 25
Community 10 / 25

How are scores calculated?

Stars

79

Forks

6

Language

Python

License

Last pushed

Jun 17, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/feifeibear/Odysseus-Transformer"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.