fbpx

Multi- Modal Sentiment Analysis

By Orisys Academy on 20th January 2024

Problem Statement

Traditional sentiment analysis models often focus on analyzing text alone,
leaving out valuable information from images and audio data. There is a need for
a more comprehensive sentiment analysis system that can capture user
sentiment across multiple modalities

Abstract

The Multi-Modal Sentiment Analysis project aims to develop a sentiment analysis
model that integrates text, images, and audio data. By combining information
from different media, the model will provide a more holistic understanding of
user sentiment, allowing for more accurate and nuanced sentiment analysis.

Outcome

● Improved accuracy in sentiment analysis by considering multiple
modalities.
● Enhanced understanding of user sentiment across diverse media types.
● Increased applicability in real-world scenarios, such as social media
monitoring and customer feedback analysis.

Reference

In this information age, opinion mining which is also known as sentiment analysis turns up to be the most important task in the field of natural language processing. Previous literature in area of sentiment analysis which mostly focused on single modality that is on textual data. Almost all the latest advancement in the sentiment analysis are using textual dataset and resources only. With the invent of internet which increases the use of social media, people are using vlogs, videos, pictures, audios, emojis and microblogs to represent their opinions on different web platforms. In this new media age, every day 720k hours of videos are uploaded on alone Youtube only. We have number of such platforms like YouTube. In the classical methods other modalities’ expressiveness is overlooked and thus these methods fail to generate accurate results. Numerous commercial applications used the aggregation of sentiments and opinions of individuals by anticipating large population. 

1.Mohammad Soleymani, David Garcia, Brendan Jou, Bjorn Schuller, Shih–Fu Chang and Maja Pantic, “A survey of multimodal sentiment analysis”, Image and Vision Computing, vol. 65, pp. 314, September 2017.

2.Fakhreddine Karray, Alemzadeh Milad, Jamil Abou Saleh and Arab Mo Nours, “Human–Computer Interaction: Overview on State of the Art”, (PDF). International Journal on Smart Sensing and Intelligent Systems, vol. 1, pp. 137-159, 2008.

3.Soujanya Poria, Erik Cambria, Rajiv Bajpai and Amir Hussain, “A review of affective computing: From unimodal analysis to multimodal fusion”, Information Fusion, vol. 37, pp. 98-125, 2017.

5.L. S. Chen, T. S. Huang, T. Miyasato and R. Nakatsu, “Multimodal human emotion/expression recognition”, Proceedings of the Third IEEE International Conference on Automatic Face and Gesture Recognition, pp. 366-371, 1998.

4.L. C. De Silva, T. Miyasato and R. Nakatsu, “Facial emotion recognition using multi–modal information”, Proceedings of ICICS, vol. 1, pp. 397-401, 1997.