Ebimsv/LLM-Lab
Pretraining and Finetuning Language Model
This project helps machine learning engineers and researchers pretrain new causal language models (LLMs) from scratch or further train existing ones. You provide raw text data or a Hugging Face dataset, and it outputs a trained language model that can generate text. It's designed for individuals working on developing custom text generation capabilities for specialized domains.
Use this if you need to build or adapt a large language model specifically for a unique text dataset, such as internal company documents, scientific papers, or domain-specific literature.
Not ideal if you're a casual user looking to simply fine-tune an existing, general-purpose LLM on a small dataset without deep customization or infrastructure setup.
Stars
21
Forks
1
Language
Jupyter Notebook
License
—
Category
Last pushed
Jan 03, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/Ebimsv/LLM-Lab"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
gustavecortal/gpt-j-fine-tuning-example
Fine-tuning 6-Billion GPT-J (& other models) with LoRA and 8-bit compression
msmrexe/pytorch-lora-from-scratch
A from-scratch PyTorch implementation of Low-Rank Adaptation (LoRA) to efficiently fine-tune...
linhaowei1/Fine-tuning-Scaling-Law
🌹[ICML 2024] Selecting Large Language Model to Fine-tune via Rectified Scaling Law
aamanlamba/phi3-tune-payments
Bidirectional fine-tuning of Microsoft's Phi-3-Mini model for payment transaction processing...
HamzahDrawsheh/fine-tuning-and-instruction-tuning-of-large-language-models
This project demonstrates the use of Large Language Models (LLMs) for Natural Language...