AceCHQ/MMIQ
This repo contains evaluation code for MM-IQ benchmark.
This evaluation code helps researchers assess how well multimodal AI models understand complex visual and textual information and perform abstract reasoning. It takes the responses from an AI model to a set of curated test items and produces a performance score, allowing AI researchers to quantify the model's cognitive capabilities. This is for AI researchers and cognitive scientists developing or evaluating advanced multimodal AI.
No commits in the last 6 months.
Use this if you are developing new multimodal AI models and need a standardized way to measure their core reasoning abilities, beyond simple recognition or classification tasks.
Not ideal if you are looking for a tool to apply existing AI models to specific business problems or for general-purpose AI model development without a focus on fundamental cognitive evaluation.
Stars
10
Forks
1
Language
Jupyter Notebook
License
Apache-2.0
Category
Last pushed
May 18, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/AceCHQ/MMIQ"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
open-mmlab/mmpretrain
OpenMMLab Pre-training Toolbox and Benchmark
facebookresearch/mmf
A modular framework for vision & language multimodal research from Facebook AI Research (FAIR)
HuaizhengZhang/Awsome-Deep-Learning-for-Video-Analysis
Papers, code and datasets about deep learning and multi-modal learning for video analysis
KaiyangZhou/pytorch-vsumm-reinforce
Unsupervised video summarization with deep reinforcement learning (AAAI'18)
adambielski/siamese-triplet
Siamese and triplet networks with online pair/triplet mining in PyTorch