GlitchBench/Benchmark

Code and Data for GlitchBench

13
/ 100
Experimental

This project helps evaluate how well large AI models can spot unusual or 'glitched' situations in video game videos. It takes video game footage as input and assesses if an AI model can identify and explain unexpected events or strange visual compositions. Anyone working with or developing large AI models for visual understanding, especially in complex or unusual scenarios, would use this.

No commits in the last 6 months.

Use this if you need to benchmark the common-sense reasoning and visual recognition abilities of large multimodal AI models, particularly their capacity to detect anomalies in video game content.

Not ideal if you are looking for a tool to develop new AI models or directly apply AI for real-time glitch detection in games, as this is purely for academic benchmarking.

AI-model-evaluation computer-vision-benchmarking multimodal-AI-testing video-game-analysis AI-model-auditing
No License Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 5 / 25
Maturity 8 / 25
Community 0 / 25

How are scores calculated?

Stars

13

Forks

Language

Python

License

Last pushed

Feb 27, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/transformers/GlitchBench/Benchmark"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.