MAC-AutoML/SocialOmni

Benchmarking Audio-Visual Social Interactivity in Omni Models

20
/ 100
Experimental

This project helps evaluate how well AI models can participate in natural, multi-person conversations, especially in video calls or real-world interactions. It takes audio-visual data of people talking and measures if a model understands who is speaking, when to interject naturally, and how to respond appropriately. This is for researchers and developers who are building or improving AI models that engage in complex social dialogue.

Use this if you are developing or benchmarking large language models that need to understand and participate in dynamic, multi-speaker audio-visual conversations with natural timing and content.

Not ideal if you are looking for a tool to analyze existing human conversations or for simple, static question-and-answer evaluations of AI models.

AI-model-evaluation social-robotics conversational-AI multimodal-interaction dialogue-systems
No License No Package No Dependents
Maintenance 13 / 25
Adoption 6 / 25
Maturity 1 / 25
Community 0 / 25

How are scores calculated?

Stars

17

Forks

Language

Python

License

Last pushed

Mar 18, 2026

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/llm-tools/MAC-AutoML/SocialOmni"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.