eric-ai-lab/MSSBench

[ICLR 2025] Official codebase for the ICLR 2025 paper "Multimodal Situational Safety"

31
/ 100
Emerging

This project helps evaluate multimodal AI models to determine if they can identify and avoid unsafe situations based on images and instructions. It takes in multimodal AI models, a dataset of images and associated safe/unsafe descriptions or instructions, and outputs a safety score or evaluation for the models. It is designed for AI safety researchers and developers working on responsible AI.

No commits in the last 6 months.

Use this if you are developing or evaluating AI systems that interact with users through images and text and need to ensure they operate safely.

Not ideal if you are looking for a pre-built safety filter or an AI model to directly apply to a production system.

AI safety responsible AI multimodal AI AI evaluation human-AI interaction
Stale 6m No Package No Dependents
Maintenance 2 / 25
Adoption 7 / 25
Maturity 16 / 25
Community 6 / 25

How are scores calculated?

Stars

30

Forks

2

Language

Python

License

MIT

Last pushed

Jun 23, 2025

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/agents/eric-ai-lab/MSSBench"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.