JIA-Lab-research/Q-LLM

This is the official repo of "QuickLLaMA: Query-aware Inference Acceleration for Large Language Models"

25
/ 100
Experimental

This project helps you get accurate answers from Large Language Models (LLMs) when working with very long documents, like entire books or extensive research papers. It takes a long document and your specific question, then uses an LLM to quickly find and summarize the most relevant information to provide a precise answer. This is ideal for researchers, analysts, or anyone who needs to extract detailed insights from massive texts without manually sifting through them.

No commits in the last 6 months.

Use this if you frequently need to query and get specific answers from extremely long text documents using an LLM, without sacrificing accuracy or waiting a long time.

Not ideal if your primary use case involves short, conversational interactions with an LLM or if you are not working with lengthy documents that exceed typical LLM context windows.

document-analysis information-retrieval research-assist text-summarization knowledge-extraction
No License Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 8 / 25
Maturity 8 / 25
Community 9 / 25

How are scores calculated?

Stars

55

Forks

4

Language

Python

License

Last pushed

Jul 16, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/JIA-Lab-research/Q-LLM"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.