FuxiaoLiu/LRV-Instruction

[ICLR'24] Mitigating Hallucination in Large Multi-Modal Models via Robust Instruction Tuning

37
/ 100
Emerging

This project helps large language models (LLMs) that can understand images (multi-modal models) become more reliable. It takes existing image-text datasets, enhances them with detailed instructions and answers, and uses this to train models. The output is an improved model that can more accurately interpret visual information and provide correct responses, which is useful for AI researchers or developers working with these advanced models.

297 stars. No commits in the last 6 months.

Use this if you are a researcher or developer aiming to reduce factual errors or 'hallucinations' in your multi-modal AI models when they interpret images and text.

Not ideal if you are an end-user looking for a ready-to-use application, as this project provides tools and methods for improving underlying AI models rather than a direct consumer product.

AI Research Multi-modal AI Large Language Models Computer Vision Natural Language Processing
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 10 / 25
Maturity 16 / 25
Community 11 / 25

How are scores calculated?

Stars

297

Forks

15

Language

Python

License

BSD-3-Clause

Last pushed

Mar 13, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/FuxiaoLiu/LRV-Instruction"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.