ksm26/Reinforcement-Learning-from-Human-Feedback
Embark on the "Reinforcement Learning from Human Feedback" course and align Large Language Models (LLMs) with human values.
This course helps AI developers and researchers align Large Language Models (LLMs) with human values and preferences. It teaches how to take an LLM and human feedback on different outputs, then train the model to produce responses that humans prefer. The primary users are individuals responsible for developing and refining AI models to ensure ethical and relevant outputs.
No commits in the last 6 months.
Use this if you need to train a Large Language Model (LLM) to better reflect human preferences and values, moving beyond basic fine-tuning.
Not ideal if you are not working with Large Language Models or if you are looking for a pre-trained, ready-to-use solution rather than a training methodology.
Stars
12
Forks
9
Language
Jupyter Notebook
License
—
Category
Last pushed
Jan 31, 2024
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/ksm26/Reinforcement-Learning-from-Human-Feedback"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
agentscope-ai/Trinity-RFT
Trinity-RFT is a general-purpose, flexible and scalable framework designed for reinforcement...
OpenRLHF/OpenRLHF
An Easy-to-use, Scalable and High-performance Agentic RL Framework based on Ray (PPO & DAPO &...
zjunlp/EasyEdit
[ACL 2024] An Easy-to-use Knowledge Editing Framework for LLMs.
huggingface/alignment-handbook
Robust recipes to align language models with human and AI preferences
hyunwoongko/nanoRLHF
nanoRLHF: from-scratch journey into how LLMs and RLHF really work.