zhilizju/Awesome-instruction-tuning
A curated list of awesome instruction tuning datasets, models, papers and repositories.
This project helps AI researchers and practitioners find the best resources for 'instruction tuning,' a technique to make large language models better at following specific commands. It compiles a list of datasets and models, including those adapted from traditional natural language processing (NLP) tasks and those generated by other large language models. The end-user is typically an AI developer or researcher focused on building or improving custom large language models for specific applications.
347 stars. No commits in the last 6 months.
Use this if you are a researcher or developer looking for high-quality, open-source instruction tuning datasets, models, or academic papers to train or fine-tune large language models.
Not ideal if you are looking for a ready-to-use, off-the-shelf language model for general tasks, or if you are not involved in the technical development of AI models.
Stars
347
Forks
14
Language
Python
License
Apache-2.0
Category
Last pushed
Jun 12, 2023
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/zhilizju/Awesome-instruction-tuning"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
DaoD/INTERS
This is the repository for our paper "INTERS: Unlocking the Power of Large Language Models in...
declare-lab/instruct-eval
This repository contains code to quantitatively evaluate instruction-tuned models such as Alpaca...
Haiyang-W/TokenFormer
[ICLR2025 Spotlightš„] Official Implementation of TokenFormer: Rethinking Transformer Scaling...
hkust-nlp/deita
Deita: Data-Efficient Instruction Tuning for Alignment [ICLR2024]
kehanlu/DeSTA2
Code and model for ICASSP 2025 Paper "Developing Instruction-Following Speech Language Model...