minjiyoon/MMGL

Multimodal Graph Learning: how to encode multiple multimodal neighbors with their relations into LLMs

25
/ 100
Experimental

This project helps content creators and researchers generate more accurate and contextually rich summaries. It takes diverse information sources like text and images that are connected in complex ways, and produces concise text summaries. Anyone who needs to summarize large amounts of multimodal content, such as a content strategist or a data analyst, would find this useful.

No commits in the last 6 months.

Use this if you need to summarize information where text and images are intricately linked, and simple one-to-one pairings aren't enough to capture the full context.

Not ideal if your data consists only of simple text-only or image-only inputs, or if the relationships between different types of information are straightforward.

content-creation information-summarization multimodal-data knowledge-graph-analysis research-assistants
No License Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 8 / 25
Maturity 8 / 25
Community 9 / 25

How are scores calculated?

Stars

67

Forks

5

Language

Python

License

Last pushed

Jul 02, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/minjiyoon/MMGL"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.