claws-lab/multimodal-robustness

Code and resources for EMNLP 2022 paper on 'Robustness of Fusion-based Multimodal Classifiers to Cross-Modal Content Dilutions'

20
/ 100
Experimental

This project helps evaluate how robust multimodal AI classifiers are when presented with 'diluted' information, where crucial details might be missing from either images or text. It takes in images and text, extracts key information from both, and then generates modified, 'diluted' versions to test the AI's ability to maintain correct classifications. This tool is for AI researchers and practitioners working on systems that combine information from different sources, like images and text, to make decisions or classify data.

No commits in the last 6 months.

Use this if you need to understand how well your multimodal classification AI performs when one of its input channels (like images or text) contains less relevant information or is 'diluted'.

Not ideal if you are looking for a general-purpose, off-the-shelf multimodal classifier for direct deployment in a production environment without needing to test its robustness to diluted content.

AI robustness multimodal classification natural language processing computer vision AI model evaluation
No License Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 5 / 25
Maturity 8 / 25
Community 7 / 25

How are scores calculated?

Stars

10

Forks

1

Language

Python

License

Last pushed

Mar 11, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/nlp/claws-lab/multimodal-robustness"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.