MNoorFawi/curlora

The code repository for the CURLoRA research paper. Stable LLM continual fine-tuning and catastrophic forgetting mitigation.

34
/ 100
Emerging

CURLoRA helps you continually update large language models (LLMs) with new information without making them "forget" what they already know. It takes an existing LLM and training data for new tasks, producing a fine-tuned model that remembers its prior knowledge while learning new capabilities. This is for machine learning engineers and researchers who need to adapt LLMs efficiently and reliably over time.

No commits in the last 6 months.

Use this if you need to fine-tune an LLM on new datasets sequentially and want to prevent the model from losing previously learned knowledge (catastrophic forgetting), especially with limited data.

Not ideal if you are looking for a plug-and-play solution for quantized models or if you are not comfortable modifying Python code to integrate this technique.

LLM fine-tuning continual learning catastrophic forgetting mitigation natural language processing model adaptation
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 8 / 25
Maturity 16 / 25
Community 10 / 25

How are scores calculated?

Stars

53

Forks

5

Language

Jupyter Notebook

License

MIT

Category

llm-fine-tuning

Last pushed

Aug 28, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/MNoorFawi/curlora"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.