zai-org/VisualGLM-6B

Chinese and English multimodal conversational language model | 多模态中英双语对话语言模型

46
/ 100
Emerging

This project offers a versatile tool for understanding and interacting with images and text in both Chinese and English. You can provide an image and ask questions about its content, and the model will generate descriptive answers. It's designed for anyone who needs to quickly extract information from images or engage in multi-modal conversations.

4,169 stars. No commits in the last 6 months.

Use this if you need to ask questions about an image's content or engage in text-based conversations that incorporate visual information, especially in Chinese or English.

Not ideal if you require absolute factual accuracy for critical applications, as it's still an early version with known limitations like occasional factual errors or hallucination.

image-analysis content-description multilingual-communication visual-question-answering
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 10 / 25
Maturity 16 / 25
Community 20 / 25

How are scores calculated?

Stars

4,169

Forks

425

Language

Python

License

Apache-2.0

Last pushed

Aug 23, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/zai-org/VisualGLM-6B"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.