general-preference/general-preference-model

[ICML 2025] Beyond Bradley-Terry Models: A General Preference Model for Language Model Alignment (https://arxiv.org/abs/2410.02197)

36
/ 100
Emerging

This project helps AI developers and researchers refine the quality of large language models (LLMs). It takes preference data, where humans have ranked or compared different LLM responses, and uses it to train a General Preference Model (GPM). The GPM then provides a more accurate way to evaluate and align LLMs, moving beyond simpler methods like Bradley-Terry models.

No commits in the last 6 months.

Use this if you are an AI developer or researcher looking to improve how you train and evaluate large language models based on human feedback.

Not ideal if you are a business user looking for a no-code solution to apply existing LLMs, as this is a tool for building and refining the underlying models.

LLM-alignment AI-research natural-language-processing model-training human-feedback-learning
Stale 6m No Package No Dependents
Maintenance 2 / 25
Adoption 7 / 25
Maturity 16 / 25
Community 11 / 25

How are scores calculated?

Stars

39

Forks

5

Language

Python

License

Apache-2.0

Last pushed

Sep 08, 2025

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/general-preference/general-preference-model"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.