waltonfuture/InstructionGPT-4
InstructionGPT-4
This project helps AI researchers and practitioners refine large language models that understand both images and text. It takes a large, general dataset of image-text instructions and processes it to identify the most impactful, high-quality examples. The output is a smaller, highly effective dataset that can be used to fine-tune models like MiniGPT-4, leading to better performance with less data.
No commits in the last 6 months.
Use this if you are a machine learning engineer or AI researcher looking to efficiently improve the performance of multimodal large language models by curating high-quality training data.
Not ideal if you are looking for a pre-trained, ready-to-use chatbot or image analysis tool, as this project focuses on the data preparation step for model training.
Stars
42
Forks
3
Language
Python
License
MIT
Category
Last pushed
Dec 29, 2023
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/waltonfuture/InstructionGPT-4"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
AI-Hypercomputer/maxtext
A simple, performant and scalable Jax LLM!
rasbt/reasoning-from-scratch
Implement a reasoning LLM in PyTorch from scratch, step by step
mindspore-lab/mindnlp
MindSpore + 🤗Huggingface: Run any Transformers/Diffusers model on MindSpore with seamless...
mosaicml/llm-foundry
LLM training code for Databricks foundation models
rickiepark/llm-from-scratch
<밑바닥부터 만들면서 공부하는 LLM>(길벗, 2025)의 코드 저장소