JochenYang/luma-mcp

Multi-Model Visual Understanding MCP Server, GLM-4.6V, DeepSeek-OCR (free), and Qwen3-VL-Flash. Provide visual processing capabilities for AI coding models that do not support image understanding.多模型视觉理解MCP服务器,GLM-4.6V、DeepSeek-OCR(免费)和Qwen3-VL-Flash等。为不支持图片理解的 AI 编码模型提供视觉处理能力。

50
/ 100
Established

This tool helps developers integrate advanced image understanding capabilities into their existing AI coding assistants that don't natively support visual input. You provide an image (a local file, URL, or data URI) and a question about it, and the tool returns a detailed analysis. This is ideal for software developers, QA testers, or technical writers who use AI assistants for tasks like debugging code from screenshots, analyzing UI layouts, or extracting text from documentation.

Available on npm.

Use this if your current AI coding assistant lacks the ability to 'see' and interpret images, and you need it to understand screenshots of code, user interfaces, error messages, or dense text documents.

Not ideal if your AI assistant already has strong native visual understanding capabilities, or if you primarily need a standalone image analysis tool for non-development tasks.

AI-assisted coding software-development debugging UI-analysis technical-documentation
Maintenance 10 / 25
Adoption 8 / 25
Maturity 22 / 25
Community 10 / 25

How are scores calculated?

Stars

48

Forks

5

Language

TypeScript

License

MIT

Last pushed

Mar 06, 2026

Commits (30d)

0

Dependencies

4

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/mcp/JochenYang/luma-mcp"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.