FareedKhan-dev/Improve-Weak-LLM-Using-SPIN-Technique

After RLHF and SFT show promising results, a new technique named SPIN is invented for 2024

12
/ 100
Experimental

This project helps improve a weaker Large Language Model (LLM) using a technique called SPIN, without needing more human-annotated data. You input an existing LLM that has been fine-tuned with supervised learning, and the output is a more capable LLM that generates responses almost indistinguishable from human-written text. This is for researchers and developers working on enhancing LLMs, particularly those facing constraints on human annotation resources.

No commits in the last 6 months.

Use this if you have a base LLM that has undergone supervised fine-tuning and want to significantly improve its performance and human-like response generation without requiring extensive new human annotations.

Not ideal if you are looking for an initial training method for an LLM from scratch or if you have ample human-annotated data available for traditional fine-tuning methods like RLHF.

LLM-fine-tuning natural-language-generation AI-model-enhancement machine-learning-research
No License Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 4 / 25
Maturity 8 / 25
Community 0 / 25

How are scores calculated?

Stars

8

Forks

Language

License

Last pushed

Jan 17, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/FareedKhan-dev/Improve-Weak-LLM-Using-SPIN-Technique"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.