JIA-Lab-research/LLMGA
This project is the official implementation of 'LLMGA: Multimodal Large Language Model based Generation Assistant', ECCV2024 Oral
LLMGA helps creative professionals like designers, illustrators, and content creators generate and edit images through conversational interactions. You describe what you want in natural language, and it produces detailed images or modifies existing ones via text-to-image, inpainting, outpainting, or instruction-based editing. It's designed for anyone needing to create visual content efficiently without deep technical expertise.
398 stars. No commits in the last 6 months.
Use this if you need an AI assistant to generate images, create story illustrations, design posters, or produce picture books based on your conversational prompts.
Not ideal if you need a simple, single-purpose image generation tool without conversational capabilities or advanced editing features.
Stars
398
Forks
25
Language
Python
License
Apache-2.0
Category
Last pushed
Jun 02, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/JIA-Lab-research/LLMGA"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
jingyaogong/minimind-v
🚀 「大模型」1小时从0训练26M参数的视觉多模态VLM!🌏 Train a 26M-parameter VLM from scratch in just 1 hours!
roboflow/vision-ai-checkup
Take your LLM to the optometrist.
SkyworkAI/Skywork-R1V
Skywork-R1V is an advanced multimodal AI model series developed by Skywork AI, specializing in...
zai-org/GLM-TTS
GLM-TTS: Controllable & Emotion-Expressive Zero-shot TTS with Multi-Reward Reinforcement Learning
NExT-GPT/NExT-GPT
Code and models for ICML 2024 paper, NExT-GPT: Any-to-Any Multimodal Large Language Model