Praful932/llmsearch
Find better generation parameters for your LLM
This tool helps machine learning practitioners fine-tune their large language models (LLMs) to perform better on specific tasks. You provide your LLM, a dataset, and a way to measure performance, and it automatically finds the best generation settings. This is for data scientists and ML engineers who want to get the best possible output from their LLMs.
No commits in the last 6 months.
Use this if you are an ML practitioner struggling to get optimal output from your LLM for generative tasks and need to systematically find the best generation parameters.
Not ideal if your task involves short, fixed-length responses or discrete outputs, as generation parameters have limited impact in these cases.
Stars
27
Forks
1
Language
Python
License
Apache-2.0
Category
Last pushed
Jun 09, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/Praful932/llmsearch"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
eth-sri/matharena
Evaluation of LLMs on latest math competitions
tatsu-lab/alpaca_eval
An automatic evaluator for instruction-following language models. Human-validated, high-quality,...
HPAI-BSC/TuRTLe
TuRTLe: A Unified Evaluation of LLMs for RTL Generation 🐢 (MLCAD 2025)
nlp-uoregon/mlmm-evaluation
Multilingual Large Language Models Evaluation Benchmark
haesleinhuepf/human-eval-bia
Benchmarking Large Language Models for Bio-Image Analysis Code Generation