OpenGVLab/perception_test_iccv2023
Champion Solutions repository for Perception Test challenges in ICCV2023 workshop.
This project helps computer vision researchers and AI developers identify specific actions or sounds within video recordings. By inputting video and audio data, it precisely outputs the start and end times of events like a dog barking or a person waving. Researchers working on video analysis, surveillance, or content moderation would find this useful.
No commits in the last 6 months.
Use this if you need to accurately pinpoint the exact moments when specific actions or sounds occur in video content, especially for research or developing new AI models.
Not ideal if you are looking for a ready-to-use application with a graphical interface, as this project provides core models and code for integration by developers.
Stars
14
Forks
—
Language
Python
License
MIT
Category
Last pushed
Oct 18, 2023
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/computer-vision/OpenGVLab/perception_test_iccv2023"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
roboflow/supervision
We write your reusable computer vision tools. 💜
insight-platform/Savant
Python Computer Vision & Video Analytics Framework With Batteries Included
PeculiarVentures/GammaCV
GammaCV is a WebGL accelerated Computer Vision library for browser
ROCm/MIVisionX
MIVisionX toolkit is a set of comprehensive computer vision and machine intelligence libraries,...
cleanlab/cleanvision
Automatically find issues in image datasets and practice data-centric computer vision.