thuiar/MIntRec
MIntRec: A New Dataset for Multimodal Intent Recognition (ACM MM 2022)
This project offers a unique collection of real-world conversational data, including text, video, and audio, drawn from a TV series. It helps researchers and AI practitioners analyze and recognize human intentions in complex, multimodal interactions. The output is a classification of an individual's intent into one of 20 categories, such as 'Complain,' 'Advise,' or 'Thank,' based on their combined verbal and nonverbal cues. This is primarily for AI researchers and data scientists focused on understanding human communication.
129 stars. No commits in the last 6 months.
Use this if you are an AI researcher or data scientist needing a rich, real-world dataset and a framework to develop and benchmark models for understanding human intent from spoken language, body language, and tone of voice.
Not ideal if you are looking for a plug-and-play solution for intent recognition in a deployed application, as this is a research dataset and benchmark framework, not an end-user product.
Stars
129
Forks
15
Language
Python
License
MIT
Category
Last pushed
May 02, 2025
Commits (30d)
0
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/thuiar/MIntRec"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Higher-rated alternatives
open-mmlab/mmpretrain
OpenMMLab Pre-training Toolbox and Benchmark
facebookresearch/mmf
A modular framework for vision & language multimodal research from Facebook AI Research (FAIR)
adambielski/siamese-triplet
Siamese and triplet networks with online pair/triplet mining in PyTorch
HuaizhengZhang/Awsome-Deep-Learning-for-Video-Analysis
Papers, code and datasets about deep learning and multi-modal learning for video analysis
KaiyangZhou/pytorch-vsumm-reinforce
Unsupervised video summarization with deep reinforcement learning (AAAI'18)