tosiyuki/LLaVA-JP

LLaVA-JP is a Japanese VLM trained by LLaVA method

42
/ 100
Emerging

This project provides pre-trained models and code to help you create systems that can understand and respond to questions about images in Japanese. You provide an image and a question in Japanese, and the system generates a natural language answer describing what's in the image or responding to the query. This is useful for developers and AI researchers building multimodal applications for Japanese speakers.

No commits in the last 6 months.

Use this if you need to build or customize a Vision-Language Model (VLM) specifically for understanding Japanese text in conjunction with images, especially if you are working with lightweight Large Language Models.

Not ideal if you are looking for a ready-to-use application and don't have experience with model training or fine-tuning, as this provides code and models for development, not an end-user tool.

AI development Japanese language processing image understanding multimodal AI natural language processing
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 8 / 25
Maturity 16 / 25
Community 18 / 25

How are scores calculated?

Stars

64

Forks

13

Language

Python

License

Apache-2.0

Last pushed

Jul 03, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/tosiyuki/LLaVA-JP"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.