TY - JOUR ID - 139376 TI - Multimodal Sentiment Analysis of Social Media Posts Using Deep Neural Networks JO - International Journal of Web Research JA - IJWR LA - en SN - 2645-4335 AU - Naseri Karimvand, Aria AU - Nemati, Shahla AU - Salehi Ghegeni, Reza AU - Basiri, Mohammad Ehsan AD - Department of Computer Eng., Shahrekord University, Shahrekord, Iran AD - Department of Computer Engineering, Faculty of Engineering Shahrekord University Shahrekord, Iran AD - Computer Engineering Dept., Shahrekord University, Shahrekord, Iran Y1 - 2021 PY - 2021 VL - 4 IS - 1 SP - 1 EP - 9 KW - Social Networks KW - Persian language KW - Sentiment analysis KW - Deep Learning KW - Instagram posts DO - 10.22133/ijwr.2021.289091.1096 N2 - With the fast growth of social media, they have become the most important platform for posting multimodal content generated by users. Much of the data on social networks such as Instagram and Telegram is multimodal data. With the aim of analyzing such multimodal data in social networks, multimodal sentiment analysis has become one of the most significant subjects for researchers in the field of emotion recognition and data mining. Although multimodal sentiment analysis of social media data for English language has been addressed in several researches recently, few studies addressed the problem for the Persian language which is the official language of more than 120 million of people around the word. In this study, a multimodal deep learning model is proposed to address this problem. The proposed method utilizes a bi-directional long short-term memory (bi-LSTM) for processing text posts and a VGG16 convolutional network for analyzing images. A new dataset of Instagram and Telegram posts, MPerSocial, containing 1000 pairs of images and Persian comments is introduced in the current study and used for evaluating the proposed method. The results of experiments show that using the fusion of textual and image modalities improves sentiment polarity detection accuracy by 20% and 8% compared with the scenario in which image and text modalities in isolation. Also, the performance of the proposed model is better than three similar deep and four traditional machine learning models. All codes and dataset used in the current study are publicly available at GitHub.   UR - https://ijwr.usc.ac.ir/article_139376.html L1 - https://ijwr.usc.ac.ir/article_139376_a37f13533b926cde65f4492417ca486f.pdf ER -