holarissun/Prompt-OIRL

code for paper Query-Dependent Prompt Evaluation and Optimization with Offline Inverse Reinforcement Learning

37
/ 100
Emerging

This project helps AI developers and researchers improve how Large Language Models (LLMs) perform on specific tasks, especially for complex reasoning like arithmetic. It takes existing demonstration data of how different prompts perform with LLMs and learns an offline reward model to evaluate new prompts without needing to query the LLM directly. The output is an optimized prompt that is tailored to specific user queries, leading to better and more consistent LLM responses.

No commits in the last 6 months.

Use this if you are a machine learning engineer or researcher looking to efficiently optimize prompts for Large Language Models to get better, query-specific results without high computational costs.

Not ideal if you are an end-user simply looking for a better way to phrase your queries to an LLM without delving into model training or prompt engineering methodologies.

Large Language Models Prompt Engineering AI Optimization Machine Learning Research Natural Language Processing
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 8 / 25
Maturity 16 / 25
Community 13 / 25

How are scores calculated?

Stars

43

Forks

6

Language

Python

License

MIT

Last pushed

Mar 20, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/prompt-engineering/holarissun/Prompt-OIRL"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.