liziniu/GEM

Code for Paper (Preserving Diversity in Supervised Fine-tuning of Large Language Models)

28
/ 100
Experimental

This project offers an alternative method to train large language models (LLMs) more effectively. By replacing the standard training loss, it helps ensure that the fine-tuned LLMs can generate a wider variety of responses and avoid becoming too specialized or "overfit" to their training data. It's for researchers or engineers who fine-tune LLMs for various applications.

No commits in the last 6 months.

Use this if you are training or fine-tuning large language models and want to improve their ability to generate diverse and high-quality outputs without overfitting.

Not ideal if you are looking for a plug-and-play solution for using LLMs without involvement in their training or fine-tuning process.

large-language-models model-training natural-language-generation ai-research deep-learning-engineering
No License Stale 6m No Package No Dependents
Maintenance 2 / 25
Adoption 8 / 25
Maturity 8 / 25
Community 10 / 25

How are scores calculated?

Stars

52

Forks

5

Language

Python

License

Last pushed

May 12, 2025

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/liziniu/GEM"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.