HiBorn4/TensorFusion_Network_for_Multimodal_sentiment_analysis

This repository implements the Tensor Fusion Network (TFN) for multimodal sentiment analysis using the CMU-MOSI dataset. TFN integrates language, visual, and acoustic modalities to predict sentiment intensity, enhancing sentiment prediction accuracy by modeling unimodal, bimodal, and trimodal interactions.

28
/ 100
Experimental

This project helps researchers and data scientists analyze sentiment from video opinions by integrating spoken language, facial expressions, and acoustic cues. It takes video data with these three modalities and outputs a sentiment intensity score or classification (e.g., positive, negative). This is ideal for those studying emotional responses or public opinion expressed in video content.

No commits in the last 6 months.

Use this if you need highly accurate sentiment analysis for video data, especially when subtle emotional nuances across speech, visuals, and sound are important.

Not ideal if your data is purely text-based or if you don't have access to detailed visual and acoustic features from video content.

social-media-analytics market-research customer-feedback video-content-analysis computational-linguistics
No License Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 5 / 25
Maturity 8 / 25
Community 15 / 25

How are scores calculated?

Stars

13

Forks

5

Language

Jupyter Notebook

License

Last pushed

May 21, 2024

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/HiBorn4/TensorFusion_Network_for_Multimodal_sentiment_analysis"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.