amazon-science/llm-code-preference

Training and Benchmarking LLMs for Code Preference.

29
/ 100
Experimental

This project helps AI researchers evaluate and train models that can judge the quality of code. It takes in pairs of code snippets and determines which one is better based on criteria like correctness, efficiency, and security. Researchers working on improving AI models for software development would use this to refine their code-generating systems.

No commits in the last 6 months.

Use this if you are an AI researcher developing or evaluating large language models that generate code and need to assess their outputs based on various quality criteria.

Not ideal if you are a software developer looking for a tool to automatically fix or improve your application code, as this is a research tool for model training and evaluation.

AI research LLM evaluation code quality model training software engineering AI
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 7 / 25
Maturity 16 / 25
Community 6 / 25

How are scores calculated?

Stars

38

Forks

2

Language

Python

License

Last pushed

Nov 15, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/amazon-science/llm-code-preference"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.