Llm Fine Tuning Transformer Models
There are 212 llm fine tuning models tracked. 4 score above 50 (established tier). The highest-rated is OptimalScale/LMFlow at 59/100 with 8,489 stars. 1 of the top 10 are actively maintained.
Get all 212 projects as JSON
curl "https://pt-edge.onrender.com/api/v1/datasets/quality?domain=transformers&subcategory=llm-fine-tuning&limit=20"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
| # | Model | Score | Tier |
|---|---|---|---|
| 1 |
OptimalScale/LMFlow
An Extensible Toolkit for Finetuning and Inference of Large Foundation... |
|
Established |
| 2 |
adithya-s-k/AI-Engineering.academy
Mastering Applied AI, One Concept at a Time |
|
Established |
| 3 |
jax-ml/jax-llm-examples
Minimal yet performant LLM examples in pure JAX |
|
Established |
| 4 |
young-geng/scalax
A simple library for scaling up JAX programs |
|
Established |
| 5 |
riyanshibohra/TuneKit
Upload your data → Get a fine-tuned SLM. Free. |
|
Emerging |
| 6 |
JIA-Lab-research/LongLoRA
Code and documents of LongLoRA and LongAlpaca (ICLR 2024 Oral) |
|
Emerging |
| 7 |
georgian-io/LLM-Finetuning-Toolkit
Toolkit for fine-tuning, ablating and unit-testing open-source LLMs. |
|
Emerging |
| 8 |
kyegomez/Finetuning-Suite
Finetune any model on HF in less than 30 seconds |
|
Emerging |
| 9 |
MaximeRobeyns/bayesian_lora
Bayesian Low-Rank Adaptation for Large Language Models |
|
Emerging |
| 10 |
NVlabs/EoRA
[ICLRW'26] EoRA: Fine-tuning-free Compensation for Compressed LLM with... |
|
Emerging |
| 11 |
ZinYY/TreeLoRA
A pytorch implementation of the paper "TreeLoRA: Efficient Continual... |
|
Emerging |
| 12 |
SakanaAI/text-to-lora
Hypernetworks that adapt LLMs for specific benchmark tasks using only... |
|
Emerging |
| 13 |
rohan-paul/LLM-FineTuning-Large-Language-Models
LLM (Large Language Model) FineTuning |
|
Emerging |
| 14 |
SensAI-PT/LLaMa2lang
Convenience scripts to finetune (chat-)LLaMa3 and other models for any language |
|
Emerging |
| 15 |
A-baoYang/alpaca-7b-chinese
Finetune LLaMA-7B with Chinese instruction datasets |
|
Emerging |
| 16 |
VectorInstitute/vectorlm
LLM finetuning in resource-constrained environments. |
|
Emerging |
| 17 |
bigscience-workshop/xmtf
Crosslingual Generalization through Multitask Finetuning |
|
Emerging |
| 18 |
NVlabs/DoRA
[ICML2024 (Oral)] Official PyTorch implementation of DoRA: Weight-Decomposed... |
|
Emerging |
| 19 |
liuqidong07/MOELoRA-peft
[SIGIR'24] The official implementation code of MOELoRA. |
|
Emerging |
| 20 |
punica-ai/punica
Serving multiple LoRA finetuned LLM as one |
|
Emerging |
| 21 |
sandy1990418/Finetune-Qwen2.5-VL
Fine-tuning Qwen2.5-VL for vision-language tasks | Optimized for Vision... |
|
Emerging |
| 22 |
architkaila/Fine-Tuning-LLMs-for-Medical-Entity-Extraction
Exploring the potential of fine-tuning Large Language Models (LLMs) like... |
|
Emerging |
| 23 |
molbal/llm-text-completion-finetune
Guide on text completion large language model fine-tuning, including example... |
|
Emerging |
| 24 |
rasbt/blog-finetuning-llama-adapters
Supplementary material for "Understanding Parameter-Efficient Finetuning of... |
|
Emerging |
| 25 |
metriccoders/one-line-llm-tuner
This repository is the source code for fine tuning any LLM in just one line 🔥 |
|
Emerging |
| 26 |
AlexandrosChrtn/llama-fine-tune-guide
Fine-tune the newly released Llama-3.2 lightweight models. |
|
Emerging |
| 27 |
rasbt/dora-from-scratch
LoRA and DoRA from Scratch Implementations |
|
Emerging |
| 28 |
neuralwork/instruct-finetune-mistral
Fine-tune Mistral 7B to generate fashion style suggestions |
|
Emerging |
| 29 |
anchen1011/FireAct
FireAct: Toward Language Agent Fine-tuning |
|
Emerging |
| 30 |
EricLBuehler/xlora
X-LoRA: Mixture of LoRA Experts |
|
Emerging |
| 31 |
TrelisResearch/install-guides
Various installation guides for Large Language Models |
|
Emerging |
| 32 |
di37/finetuning-quantize-evaluate
Fine-Tune, Quantize, Evaluate: The Complete Guide — LLMs, VLMs, and Embedding Models |
|
Emerging |
| 33 |
ymoslem/Adaptive-MT-LLM-Fine-tuning
Fine-tuning Open-Source LLMs for Adaptive Machine Translation |
|
Emerging |
| 34 |
GiovanniGatti/socratic-llm
Training pipeline for fine tuning Phi-3-mini-instruct to follow the Socratic method |
|
Emerging |
| 35 |
readytensor/rt-llm-eng-cert-week3
Week 3 of LLM Engineering Certification: Learn to fine-tune large language... |
|
Emerging |
| 36 |
aws-samples/fine-tuning-llm-with-domain-knowledge
This repo walks you through how to use transfer learning to fine tune a LLM... |
|
Emerging |
| 37 |
zjohn77/lightning-mlflow-hf
Use QLoRA to tune LLM in PyTorch-Lightning w/ Huggingface + MLflow |
|
Emerging |
| 38 |
promptslab/LLMtuner
FineTune LLMs in few lines of code (Text2Text, Text2Speech, Speech2Text) |
|
Emerging |
| 39 |
openmedlab/PULSE
PULSE: Pretrained and Unified Language Service Engine |
|
Emerging |
| 40 |
ksm26/Finetuning-Large-Language-Models
Unlock the potential of finetuning Large Language Models (LLMs). Learn from... |
|
Emerging |
| 41 |
poloclub/Fine-tuning-LLMs
Finetune Llama 2 on Colab for free on your own data: step-by-step tutorial |
|
Emerging |
| 42 |
NgJaBach/dark-kit
Collect and share guidance + code snippets for running LM-related tasks. |
|
Emerging |
| 43 |
SculptAI/GIMKit
Guided Infilling Modeling Toolkit |
|
Emerging |
| 44 |
Yog-Sotho/LLM-fine-tuner
Powerful no-code LLM fine-tuner: upload data → train → deploy in minutes.... |
|
Emerging |
| 45 |
researchim-ai/models-at-home
training models at home |
|
Emerging |
| 46 |
ngoanpv/llama2_vietnamese
A fine-tuned Large Language Model (LLM) for the Vietnamese language based on... |
|
Emerging |
| 47 |
eliahuhorwitz/Spectral-DeTuning
Official PyTorch Implementation for the "Recovering the Pre-Fine-Tuning... |
|
Emerging |
| 48 |
MNoorFawi/curlora
The code repository for the CURLoRA research paper. Stable LLM continual... |
|
Emerging |
| 49 |
CristianCristanchoT/chivito
Implementación de un LLM basado en Llama finetuneado en español empleando... |
|
Emerging |
| 50 |
rasbt/gradient-accumulation-blog
Finetuning BLOOM on a single GPU using gradient-accumulation |
|
Emerging |
| 51 |
Pengxin-Guo/FedSA-LoRA
Selective Aggregation for Low-Rank Adaptation in Federated Learning [ICLR 2025] |
|
Emerging |
| 52 |
GURPREETKAURJETHRA/Llama-3-ORPO-Fine-Tuning
Llama 3 ORPO Fine Tuning on A100 in Colab Pro. |
|
Emerging |
| 53 |
XavierSpycy/hands-on-lora
Explore practical fine-tuning of LLMs with Hands-on Lora. Dive into examples... |
|
Emerging |
| 54 |
DoubleVII/lithft
Pretrain, finetune any LLMs from huggingface on your own data. |
|
Emerging |
| 55 |
jianzhnie/LLMToolkit
LLMToolkit is a toolkit for NLP(Natural Language Processing) and LLM(Large... |
|
Emerging |
| 56 |
ramalamadingdong/onnx-rubikpi
ONNX LLM runtime on RUBIK-Pi with Gemma 1B and Llama 3.2 1B |
|
Emerging |
| 57 |
juzhengz/LoRI
[COLM 2025] LoRI: Reducing Cross-Task Interference in Multi-Task Low-Rank Adaptation |
|
Emerging |
| 58 |
mddunlap924/PyTorch-LLM
Fine-tuning an LLM using a Generic Workflow and Best Practices with PyTorch |
|
Emerging |
| 59 |
BFCmath/FinetuneAI_Learning
How to effectively finetune CV/LLM models (without local gpu) |
|
Emerging |
| 60 |
samadon1/LLM-From-Scratch
Medical Language Model fine-tuned using pretraining, instruction tuning, and... |
|
Emerging |
| 61 |
naity/finetune-esm
Scalable Protein Language Model Finetuning with Distributed Learning and... |
|
Emerging |
| 62 |
j-webtek/Local-LLM_FineTune
Finetune Your Local LLM |
|
Emerging |
| 63 |
yangjianxin1/LongQLoRA
LongQLoRA: Extent Context Length of LLMs Efficiently |
|
Emerging |
| 64 |
serp-ai/LLaMA-8bit-LoRA
Repository for Chat LLaMA - training a LoRA for the LLaMA (1 or 2) models on... |
|
Emerging |
| 65 |
graphcore-research/jax-scalify
JAX Scalify: end-to-end scaled arithmetics |
|
Emerging |
| 66 |
Followb1ind1y/Medical-LLM-Fine-tuning
Fine-tunes LLaMA-3-8B on PubMedQA with QLoRA, optimized via DeepSpeed and... |
|
Emerging |
| 67 |
ambideXtrous9/GRPO-and-SFT-Finetune-Qwen3-using-Unsloth-Reasoning-and-Non-Reasoning-Dataset
GRPO and SFT Finetune Qwen3 using Unsloth : Reasoning and Non-Reasoning Dataset |
|
Emerging |
| 68 |
sukanyabag/Finetuning-Qwen2-7B-VQA-on-Radiology-Scans
This repository is doing the finetuning of the Qwen2 7B VLM for performing... |
|
Emerging |
| 69 |
DianaDorobantu/legal-llm
Develop a Romanian legal domain Large Language Model (LLM) using pre-trained... |
|
Emerging |
| 70 |
francoislanc/midistral
LLM finetuned for generating symbolic music |
|
Experimental |
| 71 |
Atomic-man007/falcon-7b-lora-fine-tuning
falcon-7b-lora-fine-tuning |
|
Experimental |
| 72 |
mehdihosseinimoghadam/AVA-Llama-3
Fine-Tuned Llama 3 Persian Large Language Model LLM / Persian Llama 3 |
|
Experimental |
| 73 |
PardhuSreeRushiVarma20060119/OpenLoRA
"OpenLoRa" is designed to streamline and elevate the fine-tuning of large... |
|
Experimental |
| 74 |
Abhi0323/Fine-Tuning-LLaMA-2-with-QLORA-and-PEFT
This project enhances the LLaMA-2 model using Quantized Low-Rank Adaptation... |
|
Experimental |
| 75 |
roy-sub/LLM-FineTuning
Fine-Tuned Language Models Exploration using LoRA and Hugging Face's... |
|
Experimental |
| 76 |
YanSte/NLP-LLM-Fine-tuning-Llame-2-QLoRA-2024
Natural Language Processing (NLP) and Large Language Models (LLM) with... |
|
Experimental |
| 77 |
YuanheZ/LoRA-One
LoRA-One: One-Step Full Gradient Could Suffice for Fine-Tuning Large ... |
|
Experimental |
| 78 |
TobyYang7/Llava_Qwen2
Visual Instruction Tuning for Qwen2 Base Model |
|
Experimental |
| 79 |
MusfiqDehan/Llama2-Finetuned-for-Translation
Fine-Tuned Llama-2 For Machine Translation |
|
Experimental |
| 80 |
Marker-Inc-Korea/KO-Platypus
[KO-Platy🥮] Korean-Open-platypus를 활용하여 llama-2-ko를 fine-tuning한 KO-platypus model |
|
Experimental |
| 81 |
jkanalakis/finetuning-llama-model-for-text-generation-using-unsloth
Fine-tuning Llama 3.2 3B Instruct model for text generation using Unsloth AI |
|
Experimental |
| 82 |
rambodazimi/KD-LoRA
KD-LoRA: A Hybrid Approach to Efficient Fine-Tuning with LoRA and Knowledge... |
|
Experimental |
| 83 |
GURPREETKAURJETHRA/LLMs-Inference-and-Fine-Tuning
Estimate Memory Consumption of LLMs Inference and Fine Tuning |
|
Experimental |
| 84 |
aniquetahir/JORA
JORA: JAX Tensor-Parallel LoRA Library (ACL 2024) |
|
Experimental |
| 85 |
adithya-s-k/Indic-llm
A open-source framework designed to adapt pre-trained Language Models... |
|
Experimental |
| 86 |
Rs-py/HowToFineTuneLlama3.1
Quick tutorial showing how to fine-tune Llama3.1 with nothing but free tools... |
|
Experimental |
| 87 |
LimDoHyeon/EEG-LLM
Fine-tuned LLM for electroencephalography(EEG) data classification |
|
Experimental |
| 88 |
ambideXtrous9/Finetune-Qwen3-using-Unsloth
Finetune Qwen3 using Unsloth : Reasoning and Non-Reasoning Dataset |
|
Experimental |
| 89 |
HenryNdubuaku/super-lazy-autograd
Hand-derived memory-efficient VJPs for tuning LLMs on laptops. |
|
Experimental |
| 90 |
daniau23/LoRAfrica
LoRAfrica: Scaling LLM Fine Tuning for African History |
|
Experimental |
| 91 |
strickvl/isafpr_finetune
Finetuning an LLM for structured data extraction from press releases |
|
Experimental |
| 92 |
krishnaplwl/Homework_Solver_LLM
A fine-tuned LLM to solve homework questions ranging from maths to science... |
|
Experimental |
| 93 |
inuwamobarak/Meta-Llama-3-8B
Experiments with the Meta-Llama-3-8B |
|
Experimental |
| 94 |
heyisula/infosage-13b
LLM pretraining pipeline using the FineWeb-Edu Dataset |
|
Experimental |
| 95 |
nikisetti01/MTL-LORA-for-PubMedQA-and-Riddle
🚀 Fine-tuning LLaMA 1B for a medical chatbot using LoRA and a custom... |
|
Experimental |
| 96 |
sovit-123/lm_sft
Various LMs/LLMs below 3B parameters (for now) trained using SFT (Supervised... |
|
Experimental |
| 97 |
mattialoszach/LoRA-Agentic-Output-Format
Fine-tuning LLMs for structured agent-style outputs (e.g. JSON), built for... |
|
Experimental |
| 98 |
Emart29/phi4-finance-finetuning
Fine-tuning Microsoft Phi-4 Mini 3.8B on SEC 10-K financial Q&A using QLoRA... |
|
Experimental |
| 99 |
mbeps/llama3.1_fine-tuning_mult-it
Fine-tuning various Llama 3.1 family of models on the Mult-It dataset |
|
Experimental |
| 100 |
YYZhang2025/Pali-Gemma
Implement Multi-Modality-LLM and fine tuning the model using LoRA. Only... |
|
Experimental |
| 101 |
mbeps/magistral_mult-it_fine-tuning
Parameter Efficient Fine-Tuning of Magistral Small model on the Mult-It... |
|
Experimental |
| 102 |
SergiuDeveloper/yoro-finetuning
YORO (You-Only-Reason-Once) - a novel LLM architecture that runs the main... |
|
Experimental |
| 103 |
fkuhne/doctune
A fine-tuning pipeline for SLMs |
|
Experimental |
| 104 |
paulocoutinhox/mini-llm
Simple and lightweight tool to fine-tune GPT models (like GPT-2 and GPT-Neo)... |
|
Experimental |
| 105 |
arifme071/llm-finetuning-engineering-domain
Fine-tuned BERT (94.2% accuracy) + LoRA Mistral-7B on railroad AI domain... |
|
Experimental |
| 106 |
anmolg1997/Domain-Adaptive-LLM
Domain-specialized LLM fine-tuning — medical, legal, finance, code domains... |
|
Experimental |
| 107 |
stperrakis/ULM-fit
This repository contains an implementation of the ULMfit (Universal Language... |
|
Experimental |
| 108 |
TLILIFIRAS/Efficient-Fine-Tuning-of-Vision-Language-Models-with-LoRA-Quantization
This project demonstrates parameter-efficient fine-tuning of large... |
|
Experimental |
| 109 |
EN10/BabyLlama
Train and run a small Llama 2 model from scratch on the TinyStories dataset. |
|
Experimental |
| 110 |
Abdur-azure/xlmtec
xlmtec is a powerful, modular, and interactive command-line tool for... |
|
Experimental |
| 111 |
Arlchoose-code/Indonesian-LLM-Finetune
Fine-tune your Indonesian LLM with LoRA — instruction tuning kit designed to... |
|
Experimental |
| 112 |
Abeshith/FineTuning_LanguageModels
🎯 Fine-tune large language models and use them for text-related tasks. This... |
|
Experimental |
| 113 |
garystafford/duke-fine-tuning-llama
DUKE (Document Understanding and Knowledge Extraction) along with... |
|
Experimental |
| 114 |
PriyaDas258/llm-biomedical-finetuning-lab
Fine-tune TinyLlama, Phi-2, and Mistral on PubMedQA using LoRA/QLoRA —... |
|
Experimental |
| 115 |
mbeps/qwen3_fine-tune_mult-it
Parameter Efficient Fine-Tuning of various Qwen3 family of models on the... |
|
Experimental |
| 116 |
renaldiangsar/Medical-LLM-Fine-Tuning
Fine-tuning Large Language Models (LLMs) for medical reasoning to enhances... |
|
Experimental |
| 117 |
khadimhussain0/kllm
Fine-tune state-of-the-art LLMs with LoRA/QLoRA on consumer hardware. |
|
Experimental |
| 118 |
louisc-s/QLoRA-Fine-tuning-for-Film-Character-Styled-Responses-from-LLM
Code for fine-tuning Llama2 LLM with custom text dataset to produce film... |
|
Experimental |
| 119 |
DNLab2024/BGP_LLaMA
BGP-LLaMA: Fine-tuning Open-Source LLM on BGP Routing Knowledge and Analysis |
|
Experimental |
| 120 |
jmaczan/c-137
🦙 Llama 2 7B fine-tuned to revive Rick |
|
Experimental |
| 121 |
YanSte/NLP-LLM-Fine-tuning-DeepSpeed
Natural Language Processing (NLP) and Large Language Models (LLM) with... |
|
Experimental |
| 122 |
Tommaso-Sgroi/VojoLe-LM
DL24-25 project. The goal is Fine-Tuning a LLM on Italian Dialect. |
|
Experimental |
| 123 |
Abu-Sameer-66/ChemLLM-Tox-OLMo
Fine-tuning OLMo-7B with QLoRA & DeepChem for Molecular Toxicity Prediction... |
|
Experimental |
| 124 |
r-kovalch/omnigec-models
Reproducible QLoRA recipes and configs that fine‑tune Aya‑Expanse‑8B and... |
|
Experimental |
| 125 |
ph-ausseil/llm-training-dataset-builder
Streamlines the creation of dataset to train a Large Language Model with... |
|
Experimental |
| 126 |
AIdventures/flora
Fine-tuning LLMs with LoRA |
|
Experimental |
| 127 |
nv-legate/multimesh-jax
PjRt plugin and Python APIs for MPMD workflows in Jax |
|
Experimental |
| 128 |
HEMANGANI/Fine-Tuning-LLM-for-QA
Fine-Tuning Large Language Models for Question Answering |
|
Experimental |
| 129 |
arunpshankar/VAI-FineTuning-LLMs
"Clean and comprehensive examples for fine-tuning LLMs supported by Vertex... |
|
Experimental |
| 130 |
Eric-he-cn/Qwen3-QLoRA-News
This project enables the model to directly generate structured summaries... |
|
Experimental |
| 131 |
SauravMaheshkar/nanollm
JAX LLM playground |
|
Experimental |
| 132 |
zufeshan12/fine-tuning-and-reinforcement-learning-on-llms
supervised fine tuning and RLAIF on DeepSeek-math-7b-base using LoRA... |
|
Experimental |
| 133 |
neoheartbeats/neoheartbeats-kernel
An architecture for LLMs' continual-learning and long-term memories |
|
Experimental |
| 134 |
chaithanyasai18/LLMs-finetuning
This repository consists of python scripts for LLM finetuning (SFT, LoRA,... |
|
Experimental |
| 135 |
priyam-hub/LLM-Fine-Tuning-Pipeline
A comprehensive pipeline for Different Fine-Tuning Methods for Large... |
|
Experimental |
| 136 |
Pavansomisetty21/Visual-Question-Answering-Pixtral_Vision_Finetuning_Unsloth
In this we finetune Pixtral-12B-2409 model using unsloth for visual Question... |
|
Experimental |
| 137 |
RenaudGaudron/llm-quantisation-performance-study
Code and data accompanying the article "The impact of quantising a small... |
|
Experimental |
| 138 |
jwliao1209/Taiwan-LLaMa-Instruction-Tuning
2023 NTU CSIE ADL Homework 3 |
|
Experimental |
| 139 |
OutllierRejects/Intellihack_OutlierRejects_Task3
LLM Fine-tuning Challenge Enhancing Qwen 2.5 3B for AI Research QA |
|
Experimental |
| 140 |
c4dt/pitfalls_in_fine_tuning_llms
Jupyter notebooks for the LLM fine-tuning pitfalls hands-on workshop |
|
Experimental |
| 141 |
manufactai/finetuning-cookbook
A collection of practical examples and tutorials for fine-tuning large... |
|
Experimental |
| 142 |
giankev/Ancient-to-Modern-Italian-Automatic-Translation
Finetuning and evaluating LLMs on Ancient-to-Modern Italian translation task. |
|
Experimental |
| 143 |
atasoglu/turkish-llava-notebooks
A useful collection of notebooks for quantization, fine-tuning, and... |
|
Experimental |
| 144 |
erraji-jo/LLM-Finutune-based-on-customData
The project aims to showcase the process of fine-tuning LLMs on... |
|
Experimental |
| 145 |
jo-valer/machine-translation-ladin-fascian
Repository of our paper Nesciun Lengaz Lascià Endò: Machine Translation for... |
|
Experimental |
| 146 |
monadicarts/mistral-7b-trainer
Mistral 7b v0.3 LLM Model Trainer |
|
Experimental |
| 147 |
garyfanhku/Galore-pytorch
GaLore: Memory-Efficient LLM Training by Gradient Low-Rank Projection |
|
Experimental |
| 148 |
sjsayedkader/FineTuning-paris2024-olympics
End-to-end LLM fine-tuning: Paris 2024 Olympics Q&A using Databricks, AWS... |
|
Experimental |
| 149 |
NgJaBach/Language-Models-Utilities
Collect and share guidance + code snippets for running LM-related tasks. |
|
Experimental |
| 150 |
rvats20/LLM-Classification-Finetuning
Welcome to the LLM Classification Finetuning repository! This project... |
|
Experimental |
| 151 |
mehrdadalmasi2020/microsoft_MiniLM_L12_H384_uncased
A library that leverages the pre-trained microsoft_MiniLM-L12-H384-uncased... |
|
Experimental |
| 152 |
shizheng-rlfresh/llm-opt
Fine-tuning LLMs with LoRA and Hessian-free optimizers |
|
Experimental |
| 153 |
PrathamLearnsToCode/Fine-tuning-FLAN-T5-with-LoRA-WandB
Fine tune an LLM for summarization task using Low rank adaptation |
|
Experimental |
| 154 |
slv-ai/Fine-Tune-LLMs-with-DPO
Fine-tuning Microsoft’s Phi-2 Machine Learning Model with DPO |
|
Experimental |
| 155 |
sanskaryo/LLM-Finetuning-Projects
This repository contains various projects focused on fine-tuning Large... |
|
Experimental |
| 156 |
clement-cvll/AIMO-Math-Finetuning
Fine tuning of a model for AIMO 2 math competition on Kaggle |
|
Experimental |
| 157 |
sahilfaizal01/Kaggle-Contest---Fine-tuning-Llama-3.1-LLM-
We used the Llama-3.1 8B (LLM) model to verify math problem solutions via... |
|
Experimental |
| 158 |
mfaizan-ai/NewsQA
News QA generation and fine tuning an LLM for QA generation (under development) |
|
Experimental |
| 159 |
Rishabh9559/medical-llama-3.2-3B-model
This is all about fine-tuning the Llama3.2-3B model on your medical textbook. |
|
Experimental |
| 160 |
dineshsoudagar/llm-lab-from-scratch-to-fine-tuning
Comprehensive resources and scripts for training and fine-tuning Large... |
|
Experimental |
| 161 |
sparkup/medical-llm-finetuning-alignment
Medical LLM fine-tuning and preference alignment using SFT and DPO, with... |
|
Experimental |
| 162 |
spatialft/spatialft.github.io
LoRA fine-tuning of LFM2.5-1.2B to improve spatial reasoning on StepGame —... |
|
Experimental |
| 163 |
igna-s/QLoRA-Experiments
A collection of SFT and distillation pipelines to train specialized medical... |
|
Experimental |
| 164 |
Gholamrezadar/finetuning_llm_on_letter_counting
Fine-tuning Gemma-3 4B on the letter-counting dataset |
|
Experimental |
| 165 |
YounesBensafia/Algeria-2-0-FineTuning-workshop
This repository contains resources and examples used in my workshop for... |
|
Experimental |
| 166 |
Pects1949/LLM-Fine-tuning-Toolkit
A comprehensive toolkit for fine-tuning and deploying Large Language Models... |
|
Experimental |
| 167 |
Witurpred64/LLM-FineTuning-Toolkit
A comprehensive toolkit for fine-tuning Large Language Models (LLMs) with... |
|
Experimental |
| 168 |
di37/full-fine-tuning-nvidia-question-and-answering
Flan-t5-base model was fine-tuned on Nvidia Question and Answer Pair Dataset... |
|
Experimental |
| 169 |
Isha1600/LLM-Finetuning
Fine-tuning Large Language Models (LLMs) using custom datasets for improved... |
|
Experimental |
| 170 |
codershiyar/llama-google-colab-tutorial
Step-by-step tutorial on loading and using Llama 3.1 8B Instruct in Google... |
|
Experimental |
| 171 |
aakarsh31/qlora-llm-finetuning
QLoRA fine-tuning of Llama 3.2 3B on MedQA with full LoRA rank ablation... |
|
Experimental |
| 172 |
ahmad-albasha/Frankenstein-LLM-Model-fine-tuning-code
Fine-tuning Mistral-7B-v0.1 on Mary Shelley's Frankenstein using LoRA/QLoRA... |
|
Experimental |
| 173 |
jinda-liu/R-LoRA
This repository contains the source code and related resources for R-LoRA. |
|
Experimental |
| 174 |
Gyldenn/storywriter
Fine-tuning Mistral 7B with LoRA (QLoRA 4-bit) to generate Shakespearean... |
|
Experimental |
| 175 |
gazelle93/llm-fine-tuning-sft-lora-qlora
Practical examples for fine-tuning large language models (LLMs) with SFT,... |
|
Experimental |
| 176 |
alinourian/Fine-tuning-Mistral-7b-QA
Fine tuning Mistral-7b with PEFT(Parameter Efficient Fine-Tuning) and... |
|
Experimental |
| 177 |
alvi75/MultiTask-QLoRA-NFAnalysis
Official implementation of "Parameter-Efficient Multi-Task Fine-Tuning in... |
|
Experimental |
| 178 |
Akarsh1/Exploring-Unsloth-Library-for-Fine-Tuning
This is a sample notebook that can be used for exploring the fine-tuning of... |
|
Experimental |
| 179 |
ayushtiwari134/llm_fine_tuning
This model is fine-tuned to respond like Michael Gary Scott, Regional... |
|
Experimental |
| 180 |
FlorinAndrei/llm-social-media-cheap
LLMs fine-tuned with social media comments on cheap hardware |
|
Experimental |
| 181 |
MSWagner/qwen-lora-grpo-letter-counting
Fine-tuning Qwen2.5-3B-Instruct model with LoRa (Low-Rank Adaptation) and... |
|
Experimental |
| 182 |
Yousefbadr0/GPT-Neo_Medical_Fine-Tuning_using_LoRA
Fine-tuning GPT-Neo-125M using LoRA on a medical QA dataset, achieving... |
|
Experimental |
| 183 |
AparnaRoy76/LLM-finetuning
A comprehensive toolkit for fine-tuning Large Language Models (LLMs) using... |
|
Experimental |
| 184 |
nglguarino/code-completion
Fine-tuned 3 LLMs (Phi-2, Gemma, Llama2) on 100K+ instruction CodeInstruct... |
|
Experimental |
| 185 |
Sahar-Sheikhi/CRM-Data-Automation-Llama-3.2-Finetuned-
A memory-efficient fine-tuning pipeline using Llama-3.2-3B and QLoRA to... |
|
Experimental |
| 186 |
adityanaranje/FineTune-LLM
Fine-tuned a pretrained language model using Unsloth to specialize domain... |
|
Experimental |
| 187 |
Utshav-paudel/Finetuning-Mistral7B-on-google-colab
Finetuning Mistral 7B on google colab |
|
Experimental |
| 188 |
AbdulSametTurkmenoglu/unsloth_llama_news
Llama 2 7B - Turkish News Dataset Fine-Tuning |
|
Experimental |
| 189 |
Atomheart-Father/LoRA-SFT-vs-LoRA-DPO-A-Comparative-Study-of-Small-Factual-Updates-in-LLMs
This paper studies small factual updates: updates that preserve the subject... |
|
Experimental |
| 190 |
AvinashBolleddula/Domain-Adaptive-LLM-Fine-Tuning-for-Enterprise-Policy-QA
Production-grade pipeline for domain-adaptive fine-tuning of a small LLM... |
|
Experimental |
| 191 |
mltrev23/flan-t5-fine-tune
Flan-t5 model fine tune LoRA and Langchain |
|
Experimental |
| 192 |
luochang212/sft-note
三种方法实现监督微调 (SFT):LLaMA Factory, trl 和 unsloth |
|
Experimental |
| 193 |
chatterjeesaurabh/Dialogue-Summarization-with-Large-Language-Model
Explored In-Context prompt learning, Full Fine-Tuning, Parameter-Efficient... |
|
Experimental |
| 194 |
jistiak/finetune-gpt-deepspeed
Sample codes and guidelines on how to finetune any opensource GPT models... |
|
Experimental |
| 195 |
Muneeb1030/FineTune-Tiny-Llama
Fine-tuning the Tiny Llama model to mimic my professor's writing style using... |
|
Experimental |
| 196 |
serkanars/llm-fine-tuning-with-lora
LoRA yaklaşımıyla Mistral-7b-v0.1 modelini spesifik bir task için fine-tune etme |
|
Experimental |
| 197 |
tensor-fusion/sophia-jax
JAX implementation of 'Sophia: A Scalable Stochastic Second-order Optimizer... |
|
Experimental |
| 198 |
Shreyash-Gaur/TensorFlow_Python_Code_Generation
Fine-tuning CodeT5 for Python code generation on the MBPP dataset. Features... |
|
Experimental |
| 199 |
pracheeeeez/Fine_tuning_Llama2
This project focuses on fine-tuning the powerful Llama2 language model and... |
|
Experimental |
| 200 |
Holy-Morphism/VLM
Fine-Tuning a Generative VLM for Image Describing |
|
Experimental |
| 201 |
leodeveloper/phi3-vision-multimodel
Microsoft Phi-3 Vision-the first Multimodal model By Microsoft- Demo With Huggingface |
|
Experimental |
| 202 |
ako1983/Llama2-finetuned-mindsdb
Llama2 7-b-hf Fine-tuned on MindsDB Docs |
|
Experimental |
| 203 |
aloobun/llama2-7b-openhermes-15k
A 4-bit qlora refinement of llama-v2-guanaco, fine tuned on the 15k rows of... |
|
Experimental |
| 204 |
thatomaelane/Building-a-Domain-Expert-Model
This project aims to fine-tune the Meta Llama 2 7B foundation model to... |
|
Experimental |
| 205 |
mayur-kun/finetuning-llama2-7b-chat
This repository demonstrates fine-tuning an Large Language Model (LLM) on... |
|
Experimental |
| 206 |
NavodPeiris/Vulnerability-Analyst-Qwen2.5-1.5B-Instruct
Fine-tune Qwen2.5-1.5B-Instruct model for code vulnerability analysis |
|
Experimental |
| 207 |
Siddhesh19991/Llama-3-8B-Fine-tune
This project demonstrates how to Fine-Tune Llama-3-8B model on medical data... |
|
Experimental |
| 208 |
Thiraput01/QwenMed
Qwen3 fine-tuned on medical datasets with reasoning data |
|
Experimental |
| 209 |
AbdulHadi806/LLM_fune_tuning_Hackathon
In the recent competition, we were challenged to finetune a model that can... |
|
Experimental |
| 210 |
khaouitiabdelhakim/llm_fine_tuning
Fine-tuning essentially involves taking a pre-trained LLM, already equipped... |
|
Experimental |
| 211 |
ahmadalsharef994/Langchain_LlamaCPP_Mistral_7B_Fine_Tuning_Example
A comprehensive example of fine-tuning Mistral 7B models with Langchain and... |
|
Experimental |
| 212 |
Pragateeshwaran/LoRA-From-Scratch
This project implements a Low-Rank Adaptation (LoRA) technique from scratch... |
|
Experimental |