hrlics/CoPE
CoPE: Clipped RoPE as A Scalable Free Lunch for Long Context LLMs
This project offers a simple enhancement for Large Language Models (LLMs) that makes them much better at understanding and processing very long texts. It takes an existing LLM model and modifies its internal workings to improve its ability to handle lengthy documents or conversations. This is for AI researchers and developers who are building or fine-tuning LLMs and want to improve their performance on long-context tasks without significant computational overhead.
Use this if you are developing or fine-tuning Large Language Models and need them to perform reliably and accurately with extremely long inputs, such as entire books, extensive reports, or long dialogue histories.
Not ideal if you are an end-user simply looking to apply an existing LLM, or if your tasks only involve short-to-medium length texts.
Stars
10
Forks
—
Language
Python
License
—
Category
Last pushed
Feb 06, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/hrlics/CoPE"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
zhuhanqing/APOLLO
APOLLO: SGD-like Memory, AdamW-level Performance; MLSys'25 Oustanding Paper Honorable Mention
zhenye234/xcodec
AAAI 2025: Codec Does Matter: Exploring the Semantic Shortcoming of Codec for Audio Language Model
HITESHLPATEL/Mamba-Papers
Awesome Mamba Papers: A Curated Collection of Research Papers , Tutorials & Blogs
Y-Research-SBU/CSRv2
Official Repository for CSRv2 - ICLR 2026
psychofict/llm-effective-context-length
Investigating Why the Effective Context Length of LLMs Falls Short (Based on STRING, ICLR 2025)