wenliangdai/Modality-Transferable-MER

Modality-Transferable-MER, multimodal emotion recognition model with zero-shot and few-shot abilities.

30
/ 100
Emerging

This project helps researchers and developers understand emotions from various sources like spoken language, facial expressions, and text. It takes audio, visual, and textual data as input and determines the emotional content, even for emotions it hasn't seen much of before. Social science researchers, human-computer interaction designers, and AI developers working on emotion-aware systems would find this useful.

No commits in the last 6 months.

Use this if you need to analyze emotions across different data types (like voice, video, and text) and are dealing with limited examples for certain emotions or entirely new emotion categories.

Not ideal if you only work with a single data type, have abundant labeled data for all emotion categories, or require real-time, low-latency emotion detection in a production environment without further optimization.

emotion-analysis sentiment-recognition multimodal-data natural-language-processing affective-computing
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 8 / 25
Maturity 16 / 25
Community 6 / 25

How are scores calculated?

Stars

66

Forks

3

Language

Python

License

CC-BY-4.0

Last pushed

Apr 23, 2021

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/nlp/wenliangdai/Modality-Transferable-MER"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.