xrsrke/instructGOOSE

Implementation of Reinforcement Learning from Human Feedback (RLHF)

41
/ 100
Emerging

This tool helps machine learning engineers fine-tune large language models to better follow human instructions. By taking a pre-trained language model and feedback data from human preferences, it allows you to train the model to produce responses that are more aligned with desired human outputs. It's designed for ML practitioners who want to customize existing LLMs for specific tasks.

174 stars. No commits in the last 6 months.

Use this if you are a machine learning engineer looking to implement Reinforcement Learning from Human Feedback (RLHF) to align your language models with human preferences.

Not ideal if you are a non-developer or do not have experience with machine learning frameworks like PyTorch and Hugging Face Transformers.

large-language-models model-fine-tuning natural-language-processing AI-alignment reinforcement-learning
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 10 / 25
Maturity 16 / 25
Community 15 / 25

How are scores calculated?

Stars

174

Forks

21

Language

Jupyter Notebook

License

MIT

Last pushed

Apr 07, 2023

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/xrsrke/instructGOOSE"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.