jshuadvd/LongRoPE

Implementation of the LongRoPE: Extending LLM Context Window Beyond 2 Million Tokens Paper

31
/ 100
Emerging

This project extends the context window of large language models (LLMs) like LLaMA2 and Mistral, allowing them to process and understand much longer texts. It takes a pre-trained LLM and, through a progressive extension strategy, enables it to handle inputs up to 2 million tokens while maintaining accuracy. This is designed for AI practitioners and researchers who need LLMs to analyze or generate content from extremely long documents or conversations.

151 stars. No commits in the last 6 months.

Use this if you need your LLM to effectively process and reason over very long documents, extensive dialogues, or large sets of information, far beyond typical context limits.

Not ideal if your application only deals with short texts or if you are not working with large language models.

large-language-models natural-language-processing long-document-analysis AI-research text-generation
No License Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 10 / 25
Maturity 8 / 25
Community 13 / 25

How are scores calculated?

Stars

151

Forks

14

Language

Python

License

Last pushed

Jul 20, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/jshuadvd/LongRoPE"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.