VITA-Group/TAPE
[ICML'25] "Rethinking Addressing in Language Models via Contextualized Equivariant Positional Encoding" by Jiajun Zhu, Peihao Wang, Ruisi Cai, Jason D. Lee, Pan Li, Zhangyang Wang
This project offers a new way to help large language models (LLMs) understand the order of information in long texts, such as documents, articles, or books. By improving how these models address and process the position of words, it allows them to handle much longer inputs more effectively. This is particularly useful for researchers and practitioners who work with advanced AI models and need them to accurately process very lengthy textual data.
No commits in the last 6 months.
Use this if you are an AI researcher or machine learning engineer training or fine-tuning large language models and struggling with their ability to accurately understand context in extremely long documents.
Not ideal if you are looking for an off-the-shelf application to summarize or analyze documents, as this is a foundational enhancement for language models, not an end-user tool.
Stars
14
Forks
2
Language
Python
License
MIT
Category
Last pushed
Jun 06, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/VITA-Group/TAPE"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
ZHZisZZ/dllm
dLLM: Simple Diffusion Language Modeling
pengzhangzhi/Open-dLLM
Open diffusion language model for code generation — releasing pretraining, evaluation,...
EnnengYang/Awesome-Model-Merging-Methods-Theories-Applications
Model Merging in LLMs, MLLMs, and Beyond: Methods, Theories, Applications and Opportunities. ACM...
THUDM/LongWriter
[ICLR 2025] LongWriter: Unleashing 10,000+ Word Generation from Long Context LLMs
AIoT-MLSys-Lab/SVD-LLM
[ICLR 2025🔥] SVD-LLM & [NAACL 2025🔥] SVD-LLM V2