xjywhu/Awesome-Multimodal-LLM-for-Code
Multimodal Large Language Models for Code Generation under Multimodal Scenarios
This is a curated collection of research papers and benchmarks on tools that automatically generate code from various visual inputs. It covers methods for converting UI designs (like screenshots or mockups) into front-end code for web and mobile apps, or turning scientific plots and charts into the code that generated them. Software developers, UI/UX designers, and researchers in AI and software engineering would find this useful for understanding the state-of-the-art in multimodal code generation.
221 stars.
Use this if you are a developer, designer, or researcher interested in generating various types of code (UI, scientific plots, etc.) directly from visual or multi-modal descriptions and want to explore the latest academic advancements and benchmarks in this field.
Not ideal if you are looking for a ready-to-use software tool or library to implement multimodal code generation directly, as this repository primarily links to academic papers.
Stars
221
Forks
9
Language
—
License
—
Category
Last pushed
Mar 16, 2026
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/transformers/xjywhu/Awesome-Multimodal-LLM-for-Code"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
oripress/AlgoTune
AlgoTune is a NeurIPS 2025 benchmark made up of 154 math, physics, and computer science...
jie-jw-wu/human-eval-comm
HumanEvalComm: Evaluating Communication Skill of Code LLM and LLM Agent
juyongjiang/CodeUp
CodeUp: A Multilingual Code Generation Llama-X Model with Parameter-Efficient Instruction-Tuning
JHansiduYapa/Fine-Tuning-a-Small-Language-Model-for-Cypher-Query-Generation
This project fine-tunes Unsloth's Gemma-3 4B IT (4-bit) model to translate natural language into...
Gen-Verse/ReasonFlux
[NeurIPS 2025 Spotlight] LLM post-training suite — featuring ReasonFlux, ReasonFlux-PRM, and...