ShiZhengyan/InstructionModelling

[NeurIPS 2024 Main Track] Code for the paper titled "Instruction Tuning With Loss Over Instructions"

32
/ 100
Emerging

This project offers a new method to fine-tune large language models (LLMs) to better understand and follow instructions. It takes existing LLMs and instruction-output training datasets, applying a unique loss function that considers the instruction part of the data, not just the desired output. The result is a more accurate and responsive LLM, particularly useful for researchers and practitioners working on improving LLM performance on diverse natural language processing and generation tasks.

No commits in the last 6 months.

Use this if you are a researcher or AI engineer working to improve the instruction-following capabilities of language models, especially when training with datasets that have long instructions and short desired outputs, or when working with limited training data.

Not ideal if you are a non-technical end-user simply looking for a ready-to-use application, as this project requires deep technical knowledge of LLM training and infrastructure.

language-model-fine-tuning natural-language-processing AI-model-training LLM-performance-optimization machine-learning-research
No License Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 7 / 25
Maturity 8 / 25
Community 17 / 25

How are scores calculated?

Stars

38

Forks

8

Language

Python

License

Last pushed

May 24, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/ShiZhengyan/InstructionModelling"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.